diff --git "a/065.jsonl" "b/065.jsonl"
new file mode 100644--- /dev/null
+++ "b/065.jsonl"
@@ -0,0 +1,1146 @@
+{"seq_id": "3101271927", "text": "\"\"\"tests vulture_manager.py\"\"\"\nfrom pathlib import Path\n\nimport pytest\n\nfrom vulture.core import Item\n\nfrom pytest_vulture.conf.reader import IniReader\nfrom pytest_vulture.vulture.manager import VultureManager\n\n\n# pylint: disable=protected-access,too-many-arguments\n\n\n\n\n\n@pytest.mark.parametrize(\n \"results,file,answer\",\n [\n ([], \"test.py\", None),\n ([Item(\"test\", \"function\", Path(\"src/test.py\"), 8, 8, \"unused function 'test'\", 50)], \"not_test.py\", None),\n ([Item(\"test\", \"function\", Path(\"src/test.py\"), 8, 8, \"unused function 'test'\", 50)], \"test.py\", None),\n ([Item(\"test\", \"function\", Path(\"src/test.py\"), 8, 8, \"unused function 'test'\", 50)], \"src/test.py\",\n \"line 8 : unused function 'test'\"),\n ([Item(\"test\", \"function\", Path(\"src/test.py\"), 9, 9, \"unused function 'toto'\", 50)], \"src/test.py\",\n \"line 9 : unused function 'toto'\"),\n ]\n)\ndef test_get_file_errors(tmp_path, results, file, answer):\n \"\"\"Tests the getting file\"\"\"\n manager = VultureManager(tmp_path / \"test2\", IniReader(tmp_path / \"test\"))\n manager._results = results\n\n assert manager.get_file_errors(file) == answer\n", "repo_name": "Gatewatcher/pytest-vulture", "sub_path": "test/tests/unit/test_vulture_manager.py", "file_name": "test_vulture_manager.py", "file_ext": "py", "file_size_in_byte": 1159, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 4, "dataset": "github-code", "pt": "24", "api": [{"api_name": "pytest_vulture.vulture.manager.VultureManager", "line_number": 32, "usage_type": "call"}, {"api_name": "pytest_vulture.conf.reader.IniReader", "line_number": 32, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 18, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 18, "usage_type": "attribute"}, {"api_name": "vulture.core.Item", "line_number": 22, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 22, "usage_type": "call"}, {"api_name": "vulture.core.Item", "line_number": 23, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 23, "usage_type": "call"}, {"api_name": "vulture.core.Item", "line_number": 24, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 24, "usage_type": "call"}, {"api_name": "vulture.core.Item", "line_number": 26, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 26, "usage_type": "call"}]}
+{"seq_id": "35535981815", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nimport os\nfrom Mul_models import models_select\nimport torch\nfrom PIL import Image\nfrom os.path import join\nfrom torchvision import transforms\nimport torchvision\nimport torch.nn.functional as F\n\ndef imgshow(img):\n img=img/2+0.5\n npimg=img.numpy()\n plt.imshow(np.transpose(npimg, (1, 2, 0)))\n plt.show()\n\nclass Test:\n def __init__(self,model,weight_path=None,image_path='./samples',transforms=None):\n self.weight_path = weight_path\n self.image_path=image_path\n self.transforms=transforms\n self.model=model\n Net = models_select(class_num=2)\n self.net = Net.net(self.model)\n self.net.load_state_dict(torch.load(self.weight_path))\n self.net.eval()\n #print(self.net)\n def result(self):\n for image in os.listdir(self.image_path):\n img=Image.open(join(self.image_path,image))\n img=self.transforms(img)\n #imgshow(img)\n img = img.unsqueeze(0)\n #print(img.size())\n output=self.net(img)\n print(output)\n _, predicted = torch.max(output, 1)\n print(image,predicted)\n\nif __name__=='__main__':\n if not os.path.isdir('./samples'):\n os.mkdir('./samples')\n #transform = transforms.Compose([transforms.Resize((224, 224)), transforms.ToTensor(),\n # transforms.Normalize((0.485, 0.456, 0.406), (0.229, 0.224, 0.225))])\n transform = transforms.Compose([transforms.Resize((224, 224)), transforms.ToTensor()])\n T=Test(model='ResNet50',weight_path='./Weights/best_ResNet50_1_99.pth',transforms=transform)\n T.result()\n", "repo_name": "JXQI/crack_Identify", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 1689, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 5, "dataset": "github-code", "pt": "24", "api": [{"api_name": "matplotlib.pyplot.imshow", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "numpy.transpose", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "torchvision.transforms", "line_number": 22, "usage_type": "name"}, {"api_name": "Mul_models.models_select", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 26, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 30, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 31, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 31, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 43, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 46, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 46, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 46, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 46, "usage_type": "call"}]}
+{"seq_id": "29740344441", "text": "from scripts.report_scripts.data_mixin import *\nimport pandas as pd\n#Подключение гугл таблиц\nfrom scripts.filemanager.googlesheets_upload_lib import Googlesheets \nfrom scripts.filemanager.googlesheets_upload_lib import ListConventer\n\n\"\"\"\nБазовый отчет по Расходам в Яндекс Директ\n\"\"\"\n\nclass DataPrepare: \n def __init__ (self, client):\n self.client = client\n self.work()\n \n def work (self):\n \n \"\"\"\n Подкласс получения данных\n \"\"\"\n class GetData (GetMixinGenerateBigData):\n client = self.client \n dataset_name = 'yandex_direct_campaign_report' \n deep = 12\n gt = GetData()\n self.data = gt.data\n \n\n\nclass PayDirectReport:\n \n def __init__ (self, client):\n self.client = client\n dp = DataPrepare (client)\n self.data = dp.data\n self.go_report()\n def go_report(self):\n \n data = self.data\n client = self.client\n\n #Базовая подготовка данных\n\n def campaign_func (row):\n split = row['CampaignName'].split('.')\n if len(split) > 1:\n return split[1]\n return split[0]\n\n def type_func (row):\n return row['CampaignName'].split('.')[0]\n\n data['Campaign'] = data.apply(campaign_func, axis=1)\n data['Type'] = data.apply(type_func, axis=1)\n data['SumCTR'] = data.apply (lambda x: x['Ctr'] * x['Clicks'], \n axis=1)\n data['SumCRate'] = data.apply (lambda x: x['ConversionRate'] * x['Clicks'], \n axis=1)\n \n #Базовый отчет\n data_longtime = data.groupby('startOfMonth').sum().\\\n sort_values('startOfMonth', ascending=True).reset_index()\n\n def expand_func(data): \n\n data['Ctr'] = \\\n data.apply(lambda x: int(x['SumCTR'] / x['Clicks'] * 100) / 100 \\\n if x['SumCTR'] != 0\\\n else 0, axis=1)\n data['ConversionRate'] = \\\n data.apply(lambda x: int(x['SumCRate'] / x['Clicks'] * 100) / 100 \\\n if x['SumCRate'] !=0\\\n else 0, axis=1)\n\n return data\n\n data_longtime = expand_func(data_longtime)\n data_longtime = data_longtime[['startOfMonth', 'Impressions', 'Clicks', 'Ctr',\n 'Campaign_cost', 'ConversionRate']] \n \n #Отчет по типам трафика\n data_type = data.groupby(['startOfMonth', 'Type'])\\\n .sum().sort_values('startOfMonth', ascending=True).reset_index()\n\n\n data_type = data_type\\\n [ (data_type['Type']=='Поиск') | \\\n (data_type['Type']=='РСЯ') | \\\n (data_type['Type']=='Баннер на поиске')]\n\n data_type = expand_func(data_type)\n data_type = data_type[['startOfMonth', 'Type', 'Impressions', 'Clicks', 'Ctr',\n 'Campaign_cost', 'ConversionRate']] \n \n #Отчет по кампаниям\n data_campaign = data.groupby(['startOfMonth', 'Campaign'])\\\n .sum().sort_values('startOfMonth', ascending=True).reset_index()\n\n data_campaign = expand_func(data_campaign)\n data_campaign = data_campaign[['startOfMonth', 'Campaign', \n 'Impressions', 'Clicks', 'Ctr',\n 'Campaign_cost', 'ConversionRate']] \n \n \n \n lc = ListConventer(data_longtime,\n 'width_headers')\n lc.conventer()\n\n maxlonglist = 14\n maxwidthlist = 6\n width = [''] * maxwidthlist\n maxlist = [width] * maxlonglist\n\n Googlesheets('data pay direct!a1:f14', maxlist, client) \n Googlesheets('data pay direct!a1:f14', lc.datsheets, client) \n\n\n\n lc = ListConventer(data_type,\n 'width_headers')\n lc.conventer()\n\n maxlonglist = 60\n maxwidthlist = 6\n width = [''] * maxwidthlist\n maxlist = [width] * maxlonglist\n\n Googlesheets('data pay direct!h1:z60', maxlist, client) \n Googlesheets('data pay direct!h1:z60', lc.datsheets, client) \n\n\n lc = ListConventer(data_campaign,\n 'width_headers')\n lc.conventer()\n\n maxlonglist = 700\n maxwidthlist = 7\n width = [''] * maxwidthlist\n maxlist = [width] * maxlonglist\n\n Googlesheets('data pay direct!a18:g800', maxlist, client) \n Googlesheets('data pay direct!a18:g800', lc.datsheets, client)\n", "repo_name": "ustsl/IAB_FW_CLEAN", "sub_path": "reports/report list 1/pay_direct_report.py", "file_name": "pay_direct_report.py", "file_ext": "py", "file_size_in_byte": 4808, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "scripts.filemanager.googlesheets_upload_lib.ListConventer", "line_number": 106, "usage_type": "call"}, {"api_name": "scripts.filemanager.googlesheets_upload_lib.Googlesheets", "line_number": 115, "usage_type": "call"}, {"api_name": "scripts.filemanager.googlesheets_upload_lib.Googlesheets", "line_number": 116, "usage_type": "call"}, {"api_name": "scripts.filemanager.googlesheets_upload_lib.ListConventer", "line_number": 120, "usage_type": "call"}, {"api_name": "scripts.filemanager.googlesheets_upload_lib.Googlesheets", "line_number": 129, "usage_type": "call"}, {"api_name": "scripts.filemanager.googlesheets_upload_lib.Googlesheets", "line_number": 130, "usage_type": "call"}, {"api_name": "scripts.filemanager.googlesheets_upload_lib.ListConventer", "line_number": 133, "usage_type": "call"}, {"api_name": "scripts.filemanager.googlesheets_upload_lib.Googlesheets", "line_number": 142, "usage_type": "call"}, {"api_name": "scripts.filemanager.googlesheets_upload_lib.Googlesheets", "line_number": 143, "usage_type": "call"}]}
+{"seq_id": "37384237854", "text": "''' Help the user achieve a high score in a real game of threes by using a move searcher.\n\nThis assistant takes manual input from the user, allowing it to be used with any game. '''\n\nfrom __future__ import print_function\nimport os\nimport numpy as np\nimport re\nimport sys\n\nfrom base_assistant import run_assistant, movenames\nfrom threes import do_move, get_lines\n\nPY3 = sys.version_info[0] >= 3\n\nif not PY3:\n range = xrange\n input = raw_input\n\ndef to_ind(val):\n try:\n return {0:0, 1:1, 2:2, 3:3, 6:4, 12:5, 24:6, 48:7, 96:8, 192:9, 384:10, 768:11, 1536:12, 3072:13, 6144:14}[val]\n except KeyError as e:\n raise Exception(\"Invalid value %s\" % val)\n\nclass ManualAssistant:\n def __init__(self):\n self.last_board = None\n self.last_move = None\n\n def _ask_tileset(self):\n tileset = input(\"Upcoming tile(s)? \")\n tileset = {'blue': '1', 'red': '2', 'white': '3+'}.get(tileset, tileset)\n if tileset in ('3+', '6+'):\n return tileset # will be fixed up\n tileset = re.split(r'[\\s,]', tileset)\n return {to_ind(int(v)) for v in tileset}\n\n def _fixup_tileset(self, tileset, board):\n if tileset not in ('3+', '6+'):\n return tileset\n\n maxval = board.max()\n out = set(range(4, maxval-3+1))\n if tileset == '3+':\n out |= {3}\n else:\n out |= {4} # make sure the tileset isn't empty\n return out\n\n def _parse_delta(self, ind, val=None, move=None):\n if self.last_board is None:\n raise Exception(\"Can't specify a delta: last board is unknown\")\n\n ind = int(ind)\n if val is None:\n if len(self.last_tiles) > 1:\n raise Exception(\"Can't omit tile value: multiple possible previous tiles\")\n val = list(self.last_tiles)[0]\n else:\n val = to_ind(int(val))\n if val not in self.last_tiles:\n raise Exception(\"New tile wasn't in previous tile set\")\n\n if move is None:\n move = self.last_move\n\n move = movenames.index(move)\n newboard = self.last_board.copy()\n changed = do_move(newboard, move)\n line = get_lines(newboard, move)[ind-1]\n if line[-1] != 0:\n raise Exception(\"Incorrect changed row/col\")\n line[-1] = val\n return newboard\n\n def _parse_board(self, bits):\n out = np.array([to_ind(int(x)) if x else 0 for x in bits], dtype=int)\n return out.reshape((4,4))\n\n def _ask_board(self):\n if self.last_board is None:\n print(\"Current board?\")\n else:\n print(\"Current board or difference from last board?\")\n\n bits = []\n while 1:\n line = re.split(r'[\\s,]+', input())\n bits += line\n if 1 <= len(bits) < 4:\n return self._parse_delta(*bits)\n elif len(bits) == 16:\n return self._parse_board(bits)\n elif len(bits) > 16:\n raise Exception(\"More than 16 numbers specified!\")\n\n def gen_board(self):\n while 1:\n while 1:\n try:\n board = self._ask_board()\n break\n except Exception as e:\n print(\"Didn't understand your input:\", e)\n\n while 1:\n try:\n tileset = self._ask_tileset()\n break\n except Exception as e:\n print(\"Didn't understand your input:\", e)\n\n tileset = self._fixup_tileset(tileset, board)\n yield board, tileset, False\n self.last_board = board\n self.last_tiles = tileset\n\n def make_move(self, move):\n print(\"*** Suggested move:\", move)\n print()\n self.last_move = move\n\ndef parse_args(argv):\n import argparse\n parser = argparse.ArgumentParser(description=\"Suggest moves for Threes!\")\n\n args = parser.parse_args(argv)\n return args\n\ndef main(argv):\n from itertools import count\n args = parse_args(argv)\n\n print('Welcome to the Threes! assistant. See README.md for help on input formats.')\n assistant = ManualAssistant()\n run_assistant(assistant.gen_board(), assistant.make_move, False)\n\nif __name__ == '__main__':\n import sys\n exit(main(sys.argv[1:]))\n", "repo_name": "nneonneo/threes-ai", "sub_path": "manual_assistant.py", "file_name": "manual_assistant.py", "file_ext": "py", "file_size_in_byte": 4339, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 50, "dataset": "github-code", "pt": "24", "api": [{"api_name": "sys.version_info", "line_number": 14, "usage_type": "attribute"}, {"api_name": "re.split", "line_number": 36, "usage_type": "call"}, {"api_name": "base_assistant.movenames.index", "line_number": 68, "usage_type": "call"}, {"api_name": "base_assistant.movenames", "line_number": 68, "usage_type": "name"}, {"api_name": "threes.do_move", "line_number": 70, "usage_type": "call"}, {"api_name": "threes.get_lines", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 78, "usage_type": "call"}, {"api_name": "re.split", "line_number": 89, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 126, "usage_type": "call"}, {"api_name": "base_assistant.run_assistant", "line_number": 137, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 141, "usage_type": "attribute"}]}
+{"seq_id": "31294166628", "text": "import os\n\nimport numpy as np\nfrom analysis.simulation import Simulator\nfrom data import dataOperation\nfrom data.dataOperation import combineTopicChat\nfrom uti.utility import write_to_json_file, read_csv_todf, write_to_csv_file, create_fix_random_matrix\n\n\nclass Synthesizer:\n def __init__(self, text_source, output_tu_json, output_synthesized_text_s, sequence_max):\n self.sequence_max = sequence_max\n self.output_tu_json = output_tu_json\n self.text_source = text_source\n self.output_synthesized_text_s = output_synthesized_text_s\n def synthesize(self, A):\n if os.path.exists(self.output_tu_json):\n os.remove(self.output_tu_json)\n if os.path.exists(self.output_synthesized_text_s):\n os.remove(self.output_synthesized_text_s)\n chatSequence = []\n sequence_ID = 1\n csv_columns = ['sequenceID','text']\n all_messages = [csv_columns]\n df = read_csv_todf(self.text_source)\n index = 1\n for i in range(10000):\n simulator1 = Simulator(A, D, 0)\n simulator2 = Simulator(A, D, 2)\n simulator1.simulation()\n if len(simulator1.time_sequence) < 7 or len(simulator1.time_sequence)>12:\n continue\n simulator2.simulation()\n if len(simulator2.time_sequence) < 7 or len(simulator2.time_sequence)>12:\n continue\n result, l1, l2 = combineTopicChat(simulator1, simulator2)\n group_sequence = result[\"group_sequence\"]\n\n sequence1 = df[df.sequenceID ==index]\n index+=1\n indexlist = sequence1.index\n count = len(indexlist)\n while l1 > count:\n sequence1 = df[df.sequenceID == index]\n indexlist = sequence1.index\n count = len(indexlist)\n index+=1\n sequence2 = df[df.sequenceID == index]\n indexlist = sequence2.index\n count = len(indexlist)\n while l2 > count:\n sequence2 = df[df.sequenceID == index]\n indexlist = sequence2.index\n count = len(indexlist)\n index += 1\n\n textlist1 = sequence1.text.to_list()\n textlist2 = sequence2.text.to_list()\n p1=p2=0\n for group_id in group_sequence:\n if group_id == 0:\n all_messages.append([sequence_ID, textlist1[p1]])\n p1+=1\n else :\n all_messages.append([sequence_ID, textlist2[p2]])\n p2+=1\n chatSequence.append(result)\n sequence_ID+=1\n if sequence_ID == self.sequence_max:\n break\n\n write_to_csv_file(self.output_synthesized_text_s, all_messages)\n write_to_json_file(self.output_tu_json, chatSequence)\n\n def synthesize_random(self, A):\n if os.path.exists(self.output_tu_json):\n os.remove(self.output_tu_json)\n if os.path.exists(self.output_synthesized_text_s):\n os.remove(self.output_synthesized_text_s)\n chatSequence = []\n sequence_ID = 1\n csv_columns = ['sequenceID','text']\n all_messages = [csv_columns]\n df = read_csv_todf(self.text_source)\n index = 1\n for i in range(10000):\n simulator1 = Simulator(A, D, 0)\n simulator1.simulation()\n if len(simulator1.time_sequence) < 7 or len(simulator1.time_sequence)> 12:\n continue\n\n l1 = len(simulator1.time_sequence)\n group_sequence = simulator1.group_sequence\n\n sequence1 = df[df.sequenceID ==index]\n index+=1\n indexlist = sequence1.index\n count = len(indexlist)\n while l1 > count:\n sequence1 = df[df.sequenceID == index]\n indexlist = sequence1.index\n count = len(indexlist)\n index+=1\n\n\n textlist1 = sequence1.text.to_list()\n\n p1=0\n for group_id in group_sequence:\n all_messages.append([sequence_ID, textlist1[p1]])\n p1+=1\n\n chatSequence.append({\"time_sequence\":simulator1.time_sequence, \"user_sequence\":simulator1.user_sequence})\n sequence_ID+=1\n if sequence_ID == self.sequence_max:\n break\n\n write_to_csv_file(self.output_synthesized_text_s, all_messages)\n write_to_json_file(self.output_tu_json, chatSequence)\n\n\n\nif __name__ == '__main__':\n D = 4\n A = create_fix_random_matrix(D)\n np.fill_diagonal(A, 0)\n '''\n A[0, 1] = 0.9\n A[0, 2] = 0.01\n A[0, 3] = 0.01\n A[1, 0] = 0.9\n A[1, 2] = 0.01\n A[1, 3] = 0.01\n A[2, 0] = 0.01\n A[2, 1] = 0.01\n A[2, 3] = 0.9\n A[3, 0] = 0.01\n A[3, 1] = 0.01\n A[3, 2] = 0.9\n '''\n folder = \"/home/chauncey/PycharmProjects/Parsing_Telegram_Chat_History/data/synthesizer_data/blended_skill_talk/\"\n source_file = os.path.join(folder, \"test.csv\" )\n output_tu_json =os.path.join(folder, \"/tu_sequence.json\" )\n output_synthesized_text_s = os.path.join(folder, \"/text_sequence.csv\")\n synthesizer =Synthesizer(source_file,output_tu_json,output_synthesized_text_s,201)\n synthesizer.synthesize(A)", "repo_name": "ChaunceyCXC/Infectivity_Network", "sub_path": "preprocess/Synthesizer.py", "file_name": "Synthesizer.py", "file_ext": "py", "file_size_in_byte": 5367, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "os.path.exists", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 20, "usage_type": "call"}, {"api_name": "uti.utility.read_csv_todf", "line_number": 25, "usage_type": "call"}, {"api_name": "analysis.simulation.Simulator", "line_number": 28, "usage_type": "call"}, {"api_name": "analysis.simulation.Simulator", "line_number": 29, "usage_type": "call"}, {"api_name": "data.dataOperation.combineTopicChat", "line_number": 36, "usage_type": "call"}, {"api_name": "uti.utility.write_to_csv_file", "line_number": 72, "usage_type": "call"}, {"api_name": "uti.utility.write_to_json_file", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 79, "usage_type": "call"}, {"api_name": "uti.utility.read_csv_todf", "line_number": 84, "usage_type": "call"}, {"api_name": "analysis.simulation.Simulator", "line_number": 87, "usage_type": "call"}, {"api_name": "uti.utility.write_to_csv_file", "line_number": 118, "usage_type": "call"}, {"api_name": "uti.utility.write_to_json_file", "line_number": 119, "usage_type": "call"}, {"api_name": "uti.utility.create_fix_random_matrix", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.fill_diagonal", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 142, "usage_type": "call"}, {"api_name": "os.path", "line_number": 142, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "attribute"}]}
+{"seq_id": "39091185249", "text": "from flask import render_template, flash, request, url_for, redirect, abort, session, Markup\nfrom flask_login import login_user, current_user, logout_user, login_required\nfrom flask_mail import Message\nfrom application import app, bcrypt, mail, login_manager\n\nfrom application.classes.course import Course \nfrom application.classes.user import User\nfrom application.classes.assignment import Assignment\nfrom application.forms.forms import AssignmentForm\n\nimport os \nimport json \nimport re\nfrom datetime import datetime \nfrom bson import ObjectId\n\n## Routesin this file\n# /add_assignment\n# /update_assignment\n\n@app.route(\"/add_assignment\", methods=[\"GET\", \"POST\"], defaults={'close': False})\n@app.route(\"/add_assignment/\", methods=[\"GET\", \"POST\"])\n@login_required\ndef add_assignment(close):\n form = AssignmentForm()\n\n courses = sorted(current_user.courses, key = lambda x: x.period)\n choices = [(str(course.id), f\"{str(course.name)} - {course.period}\") for course in courses]\n form.course.choices = choices\n\n if form.validate_on_submit():\n a = Assignment(str(ObjectId()), form.name.data, form.a_type.data, form.course.data, \n datetime.combine(form.due_date.data, form.due_time.data), form.notes.data)\n current_user.add_assignment(a)\n\n if close:\n return \"\"\n\n flash('Assignment added successfuly', 'success')\n return redirect(url_for(\"dashboard\"))\n\n return render_template('add_assignment.html', form=form, update=False, close=close)\n\n@app.route(\"/update_assignment/\", methods=[\"GET\", \"POST\"], defaults={'close': False})\n@app.route(\"/update_assignment//\", methods=[\"GET\", \"POST\"])\n@login_required\ndef update_assignment(assignment_id, close):\n form = AssignmentForm()\n\n assignment = current_user.get_assignment_by_id(assignment_id)\n if not assignment:\n abort(404)\n\n form = AssignmentForm(name=assignment.name, course_id=assignment.course_id, due_date=assignment.due_date.date(),\n due_time=assignment.due_date.time(), notes=assignment.notes, a_type=assignment.a_type)\n\n form.submit.label.text = \"Update Assignment\"\n\n courses = sorted(current_user.courses, key = lambda x: x.period)\n choices = [(str(course.id), f\"{str(course.name)} - {course.period}\") for course in courses]\n form.course.choices = choices\n\n if form.validate_on_submit():\n current_user.update_assignment(assignment_id, name=form.name.data, a_type=form.a_type.data,\n course_id=form.course.data, notes=form.notes.data,\n due_date=datetime.combine(form.due_date.data, form.due_time.data))\n flash('Assignment updated successfuly', 'success')\n return redirect(url_for(\"dashboard\"))\n\n form.course.data = assignment.course_id\n\n return render_template('add_assignment.html', form=form, update=True, assignment=assignment, close=close)\n\n@app.route(\"/complete_assignment/\", methods=[\"POST\"])\ndef complete_assignment(assignment_id):\n assignment = current_user.get_assignment_by_id(assignment_id)\n if not assignment:\n return {\"Status\": \"Failure\"}\n\n current_user.delete_assignment(assignment_id)\n current_user.increment_assignment_count()\n\n return {\"Status\": \"Success\"}\n\n@app.route(\"/delete_assignment/\", methods=[\"POST\"])\ndef delete_assignment(assignment_id):\n assignment = current_user.get_assignment_by_id(assignment_id)\n if not assignment:\n return {\"Status\": \"Failure\"}\n\n current_user.delete_assignment(assignment_id)\n\n return {\"Status\": \"Success\"}\n\n@app.route(\"/get_assignments//\", methods=[\"GET\"])\ndef get_assignments(course, a_type):\n assignments = [] if not current_user.assignments else current_user.assignments\n assignments = sorted(assignments, key=lambda x: x.due_date)\n assignments = [(a, current_user.get_course_by_id(a.course_id), a.due_date.strftime(\"%a, %m/%d/%y %I:%M %p\")) for a in assignments]\n \n if course != \"all\":\n assignments = [a for a in assignments if a[0].course_id == course]\n if a_type != \"all\":\n assignments = [a for a in assignments if a[0].a_type == a_type]\n\n return render_template('assignments_update.html', assignments=assignments)", "repo_name": "ronnachum11/locker", "sub_path": "application/routes/assignment_routes.py", "file_name": "assignment_routes.py", "file_ext": "py", "file_size_in_byte": 4404, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "24", "api": [{"api_name": "application.forms.forms.AssignmentForm", "line_number": 25, "usage_type": "call"}, {"api_name": "flask_login.current_user.courses", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 27, "usage_type": "name"}, {"api_name": "application.classes.assignment.Assignment", "line_number": 32, "usage_type": "call"}, {"api_name": "bson.ObjectId", "line_number": 32, "usage_type": "call"}, {"api_name": "datetime.datetime.combine", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 33, "usage_type": "name"}, {"api_name": "flask_login.current_user.add_assignment", "line_number": 34, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 39, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 42, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 21, "usage_type": "call"}, {"api_name": "application.app", "line_number": 21, "usage_type": "name"}, {"api_name": "application.app.route", "line_number": 22, "usage_type": "call"}, {"api_name": "application.app", "line_number": 22, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 23, "usage_type": "name"}, {"api_name": "application.forms.forms.AssignmentForm", "line_number": 48, "usage_type": "call"}, {"api_name": "flask_login.current_user.get_assignment_by_id", "line_number": 50, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 50, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 52, "usage_type": "call"}, {"api_name": "application.forms.forms.AssignmentForm", "line_number": 54, "usage_type": "call"}, {"api_name": "flask_login.current_user.courses", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 59, "usage_type": "name"}, {"api_name": "flask_login.current_user.update_assignment", "line_number": 64, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 64, "usage_type": "name"}, {"api_name": "datetime.datetime.combine", "line_number": 66, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 66, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 72, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 44, "usage_type": "call"}, {"api_name": "application.app", "line_number": 44, "usage_type": "name"}, {"api_name": "application.app.route", "line_number": 45, "usage_type": "call"}, {"api_name": "application.app", "line_number": 45, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 46, "usage_type": "name"}, {"api_name": "flask_login.current_user.get_assignment_by_id", "line_number": 76, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 76, "usage_type": "name"}, {"api_name": "flask_login.current_user.delete_assignment", "line_number": 80, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 80, "usage_type": "name"}, {"api_name": "flask_login.current_user.increment_assignment_count", "line_number": 81, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 81, "usage_type": "name"}, {"api_name": "application.app.route", "line_number": 74, "usage_type": "call"}, {"api_name": "application.app", "line_number": 74, "usage_type": "name"}, {"api_name": "flask_login.current_user.get_assignment_by_id", "line_number": 87, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 87, "usage_type": "name"}, {"api_name": "flask_login.current_user.delete_assignment", "line_number": 91, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 91, "usage_type": "name"}, {"api_name": "application.app.route", "line_number": 85, "usage_type": "call"}, {"api_name": "application.app", "line_number": 85, "usage_type": "name"}, {"api_name": "flask_login.current_user.assignments", "line_number": 97, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 97, "usage_type": "name"}, {"api_name": "flask_login.current_user.get_course_by_id", "line_number": 99, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 99, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 106, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 95, "usage_type": "call"}, {"api_name": "application.app", "line_number": 95, "usage_type": "name"}]}
+{"seq_id": "73507352383", "text": "import argparse \n\nimport torch \nimport torch.nn as nn\nimport torch.optim as optim\n\nfrom sleep_classification.trainer import Trainer\nfrom sleep_classification.data_loader import load_dataloader_for_featureNet\n\nfrom sleep_classification.models import DeepSleepNet\n\ndef define_argparser():\n p = argparse.ArgumentParser() \n \n p.add_argument('--model_fn', required=True)\n p.add_argument('--log_dir', default=\"/tensorboard_logs\")\n p.add_argument('--gpu_id', type= int,default=0 if torch.cuda.is_available() else -1)\n\n p.add_argument('--train_ratio', type=float,default=0.9)\n p.add_argument('--data_dir', type=str,default='G:/내 드라이브/EEG_classification/output')\n p.add_argument('--n_fold',type=int,default=20)\n p.add_argument('--fold_idx', type=int,required=True)\n \n p.add_argument('--batch_size',type=int,default=512)\n p.add_argument('--n_epochs',type=int,default=200)\n p.add_argument('--verbose',type=int,default=2)\n\n p.add_argument('--use_dropout',type=bool, default=True)\n p.add_argument('--use_rnn',type=bool, default=True)\n \n p.add_argument('--max_grad', type=float, default=-1)\n\n config = p.parse_args()\n\n return config\n\ndef main(config):\n device = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n\n train_loader, valid_loader, test_loader = load_dataloader_for_featureNet(config)\n\n print(\"Train:\", len(train_loader.dataset))\n print(\"Valid:\", len(valid_loader.dataset))\n print(\"Test:\", len(test_loader.dataset))\n\n model = DeepSleepNet(input_dim=1,n_classes=5,is_train=True,use_dropout=config.use_dropout,use_rnn=config.use_rnn).to(device)\n optimizer = optim.Adam(model.parameters())\n crit = nn.CrossEntropyLoss()\n\n trainer = Trainer(config)\n\n trainer.tb_logger.writer.add_graph(model=model,input_to_model=torch.randn(128,1,3000).to(device),verbose=True)\n\n if config.verbose >= 2:\n print(model)\n print(optimizer)\n print(crit)\n\n trainer.train(model, crit, optimizer, train_loader, valid_loader)\n trainer.test(test_loader)\n trainer.tb_logger.close()\n\n\n\nif __name__ == '__main__':\n config = define_argparser()\n main(config)\n", "repo_name": "Ldoun/EEG-AI", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 2171, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 17, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 38, "usage_type": "attribute"}, {"api_name": "sleep_classification.data_loader.load_dataloader_for_featureNet", "line_number": 40, "usage_type": "call"}, {"api_name": "sleep_classification.models.DeepSleepNet", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 47, "usage_type": "name"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 48, "usage_type": "name"}, {"api_name": "sleep_classification.trainer.Trainer", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 52, "usage_type": "call"}]}
+{"seq_id": "74944157501", "text": "from django import forms\nfrom .models import Url\nfrom django.forms.widgets import TextInput\n\n\nclass UrlForm(forms.ModelForm):\n\n long_url = forms.URLField(\n max_length=200, \n label=\"URL\",\n # help_text=\"Please enter the URL of the page.\", \n initial=\"http://\",\n widget=TextInput\n )\n\n class Meta:\n model = Url\n fields= ['long_url']\n", "repo_name": "theduckfliesagain/url-shortener", "sub_path": "shortener/urls/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 387, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.forms.ModelForm", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 6, "usage_type": "name"}, {"api_name": "django.forms.URLField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 8, "usage_type": "name"}, {"api_name": "django.forms.widgets.TextInput", "line_number": 13, "usage_type": "name"}, {"api_name": "models.Url", "line_number": 17, "usage_type": "name"}]}
+{"seq_id": "15667700901", "text": "import typing\r\n\r\nfrom kombu.exceptions import OperationalError\r\nfrom redis.exceptions import ConnectionError as RedisConnectionError\r\n\r\nfrom project.server.extensions import celery, redis_client\r\n\r\n\r\ndef celery_status() -> typing.Optional[dict]:\r\n \"\"\" Try to get status of celery. Returns None is celery is not active\"\"\"\r\n try:\r\n i = celery.control.inspect()\r\n stats = i.stats()\r\n registered_tasks = i.registered()\r\n active_tasks = i.active()\r\n scheduled_tasks = i.scheduled()\r\n result = {\r\n 'stats': stats,\r\n 'registered_tasks': registered_tasks,\r\n 'active_tasks': active_tasks,\r\n 'scheduled_tasks': scheduled_tasks\r\n }\r\n return result\r\n except OperationalError:\r\n return None\r\n\r\n\r\ndef redis_status() -> typing.Optional[bool]:\r\n \"\"\" Try to ping redis. Return None on error and True on success. \"\"\"\r\n try:\r\n redis = redis_client.redis\r\n return redis.ping()\r\n except RedisConnectionError:\r\n return None\r\n", "repo_name": "M0r13n/Smartphoniker-shop", "sub_path": "project/server/common/stats.py", "file_name": "stats.py", "file_ext": "py", "file_size_in_byte": 1049, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3, "dataset": "github-code", "pt": "24", "api": [{"api_name": "project.server.extensions.celery.control.inspect", "line_number": 12, "usage_type": "call"}, {"api_name": "project.server.extensions.celery.control", "line_number": 12, "usage_type": "attribute"}, {"api_name": "project.server.extensions.celery", "line_number": 12, "usage_type": "name"}, {"api_name": "kombu.exceptions.OperationalError", "line_number": 24, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 9, "usage_type": "attribute"}, {"api_name": "redis.exceptions", "line_number": 31, "usage_type": "name"}, {"api_name": "project.server.extensions.redis_client.redis", "line_number": 31, "usage_type": "attribute"}, {"api_name": "project.server.extensions.redis_client", "line_number": 31, "usage_type": "name"}, {"api_name": "redis.exceptions.ping", "line_number": 32, "usage_type": "call"}, {"api_name": "redis.exceptions", "line_number": 32, "usage_type": "name"}, {"api_name": "redis.exceptions.ConnectionError", "line_number": 33, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 28, "usage_type": "attribute"}]}
+{"seq_id": "9894409543", "text": "import asyncio\nimport sqlite3\nfrom fetch_data import get_info\n\nconn = sqlite3.connect('../../mobile.db')\nresult = asyncio.get_event_loop().run_until_complete(get_info())\nc = conn.cursor()\n# Create table - STATIONS\nc.execute('''CREATE TABLE STATIONS\n ([generated_id] INTEGER PRIMARY KEY ,[Station_id] TEXT, [Station_Name] TEXT, [Latitude] FLOAT, [Longnitude] FLOAT)''')\nfor r in result:\n c.execute(\"INSERT INTO STATIONS VALUES (NULL, :Station_id, :Station_Name, :Latitude, :Longnitude)\", r)\n\nconn.commit()\n", "repo_name": "Kadir94/djangoAPI", "sub_path": "djangoapp/mobiletasks/connectiondb.py", "file_name": "connectiondb.py", "file_ext": "py", "file_size_in_byte": 520, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "sqlite3.connect", "line_number": 5, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 6, "usage_type": "call"}, {"api_name": "fetch_data.get_info", "line_number": 6, "usage_type": "call"}]}
+{"seq_id": "11776707362", "text": "# pyre-strict\n\nimport setuptools\n\nwith open('README.md', 'r') as fh:\n long_description = fh.read()\n\nsetuptools.setup(\n name='paytext',\n version='0.0.2',\n author='Yngve Høiseth',\n author_email='yngve@hoiseth.net',\n description='Generalize texts from payment card transactions',\n long_description=long_description,\n long_description_content_type='text/markdown',\n url='https://github.com/yhoiseth/paytext',\n packages=setuptools.find_packages(),\n classifiers=(\n 'Programming Language :: Python :: 3',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n ),\n)\n", "repo_name": "yhoiseth/paytext", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 644, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "setuptools.setup", "line_number": 8, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 17, "usage_type": "call"}]}
+{"seq_id": "8509312399", "text": "from django.test import TestCase\n\nfrom ui.voyager import get_open_library_item_title\n\n\nclass OpenLibraryTitleTest(TestCase):\n def test_correct_title(self):\n \"this is from issue 487 where open libray link points to correct title\"\n title = \"Life on the Mississippi\"\n open_library_link = \"http://openlibrary.org/books/OL6710196M/Life_on_the_Mississippi\"\n open_library_title = get_open_library_item_title(open_library_link)\n self.assertEqual(title[0:10], open_library_title[0:10])\n\n def test_incorrect_title(self):\n \"from issue 420\"\n title = \"Frank Lloyd Wright's Hanna House : the clients' report Paul R. and Jean S. Hanna\"\n open_library_link = \"http://openlibrary.org/books/OL24933180M/The_Baptist_position_as_to_the_Bible\"\n open_library_title = get_open_library_item_title(open_library_link)\n self.assertNotEqual(title[0:10], open_library_title[0:10])\n", "repo_name": "gwu-libraries/launchpad", "sub_path": "lp/ui/tests/open_library_title_test.py", "file_name": "open_library_title_test.py", "file_ext": "py", "file_size_in_byte": 927, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 15, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.test.TestCase", "line_number": 6, "usage_type": "name"}, {"api_name": "ui.voyager.get_open_library_item_title", "line_number": 11, "usage_type": "call"}, {"api_name": "ui.voyager.get_open_library_item_title", "line_number": 18, "usage_type": "call"}]}
+{"seq_id": "41129438773", "text": "import scrapy\nfrom scrapy.linkextractors import LinkExtractor\nfrom scrapy.spiders import CrawlSpider, Rule\nfrom scrapy.item import Item, Field\n\n\nfrom items import awayTeamRushItem\nimport logging\nimport re\n\nclass EspnspiderSpider(scrapy.Spider):\n name = \"espnSpider\"\n allowed_domains = [\"espn.com\"]\n start_urls = (\n 'http://www.espn.com/nfl/boxscore?gameId=400874586',\n )\n\n rules = (\n Rule(LinkExtractor(), callback='parse_item', follow=False),\n )\n\n global awayItem\n awayItem = awayTeamRushItem()\n\n def parse(self, response):\n rushers = response.xpath('//*[@id=\"gamepackage-rushing\"]/div/div[1]/div/div/table/tbody/*/td[1]/a/span[1]/text()').extract()\n\n carries = response.xpath('//*[@id=\"gamepackage-rushing\"]/div/div[1]/div/div/table/tbody/*/td[2]/text()').extract()\n\n yards = response.xpath('//*[@id=\"gamepackage-rushing\"]/div/div[1]/div/div/table/tbody/*/td[3]/text()').extract()\n\n averages = response.xpath('//*[@id=\"gamepackage-rushing\"]/div/div[1]/div/div/table/tbody/*/td[4]/text()').extract()\n\n touchdowns = response.xpath('//*[@id=\"gamepackage-rushing\"]/div/div[1]/div/div/table/tbody/*/td[5]/text()').extract()\n\n longs = response.xpath('//*[@id=\"gamepackage-rushing\"]/div/div[1]/div/div/table/tbody/*/td[6]/text()').extract()\n\n awayItemIndex = 0\n\n for rusher in rushers:\n awayItem = awayTeamRushItem()\n\n awayItem['car'] = str(carries[awayItemIndex])\n awayItem['yds'] = str(yards[awayItemIndex])\n awayItem['avg'] = str(averages[awayItemIndex])\n awayItem['td'] = str(touchdowns[awayItemIndex])\n awayItem['longest'] = str(longs[awayItemIndex])\n awayItem['rusher'] = str(rusher)\n\n awayItemIndex+=1\n\n yield awayItem\n", "repo_name": "gannonk08/scrapy-demo", "sub_path": "scrapy_demo/scrapy_demo/spiders/espnSpider.py", "file_name": "espnSpider.py", "file_ext": "py", "file_size_in_byte": 1817, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "scrapy.Spider", "line_number": 11, "usage_type": "attribute"}, {"api_name": "scrapy.spiders.Rule", "line_number": 19, "usage_type": "call"}, {"api_name": "scrapy.linkextractors.LinkExtractor", "line_number": 19, "usage_type": "call"}, {"api_name": "items.awayTeamRushItem", "line_number": 23, "usage_type": "call"}, {"api_name": "items.awayTeamRushItem", "line_number": 41, "usage_type": "call"}]}
+{"seq_id": "35886338081", "text": "import torch\nfrom tqdm import tqdm\nimport json, os, sys\nfrom rouge import Rouge\nfrom transformers import AutoTokenizer\nfrom nltk import sent_tokenize\nimport argparse\n\ndef get_args(): \n parser = argparse.ArgumentParser()\n parser.add_argument('--ckpt', default=125000, type=int)\n args = parser.parse_args()\n return args\n\nargs = get_args()\nckpt = args.ckpt\n\nsys.path.append('src') \n\ndevice = 'cuda'\nckpt_list = [_.split('-')[-1] for _ in os.listdir('/workspace/ckpt/kobart_ckpt') if _.startswith('checkpoint')]\nckpt_list.sort()\n\ntest_dataset = json.load(open('data/article/valid_dataset.json', 'r', encoding='utf-8'))\ntokenizer = AutoTokenizer.from_pretrained('models/kobart')\nrouge_scorer = Rouge()\ntarget_list = [target['abs'] for target in test_dataset]\n\ndef tokenize_list(data_list):\n return [' '.join(tokenizer.tokenize(data)) for data in data_list]\n\n\ntokenized_target_list = tokenize_list(target_list)\nbest_rouge = 0\n\n# for ckpt in ckpt_list:\nresult_list = []\nmodel_path = f'/workspace/ckpt/kobart_ckpt/checkpoint-{ckpt}'\nmodel = torch.load(os.path.join(model_path, 'multitask_ext_abs_summary_model.pt')).model.to(device)\nprint(f\"Current doing... {model_path.split('/')[-1]}\")\n\nfor data in tqdm(test_dataset):\n doc = ' '.join(data['sentences'])\n input_ids = tokenizer(doc, return_tensors=\"pt\").input_ids.to(device)\n\n output = model.generate(input_ids, num_beams=5, eos_token_id=1, repetition_penalty=1.2, no_repeat_ngram_size=1, early_stopping=True,\n max_length=150)\n result = tokenizer.decode(output[0], skip_special_toknes=True).replace('', '')\n result = sent_tokenize(result)[0]\n result_list.append(result)\n\ntokenized_result_list = tokenize_list(result_list)\nscores = rouge_scorer.get_scores(tokenized_result_list, tokenized_target_list, avg=True)\nrouge_score = scores['rouge-l']['f']\nif rouge_score > best_rouge:\n best_rouge = rouge_score\n best_score = scores\n best_ckpt = ckpt\nprint(f\"Best CKPT: {best_ckpt}, Best score: {best_score}\")\n\nwith open(f'results/infer/{ckpt}_results.txt', 'w', encoding='utf-8') as f:\n f.write('\\n'.join(result_list))\n\nwith open(f'results/rouge/{ckpt}_rouge.json', 'w') as f:\n json.dump(best_score, f, indent='\\t')", "repo_name": "CommoMo/Ext-Abs-Summ", "sub_path": "get_rouge.py", "file_name": "get_rouge.py", "file_ext": "py", "file_size_in_byte": 2233, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 18, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 21, "usage_type": "call"}, {"api_name": "json.load", "line_number": 24, "usage_type": "call"}, {"api_name": "transformers.AutoTokenizer.from_pretrained", "line_number": 25, "usage_type": "call"}, {"api_name": "transformers.AutoTokenizer", "line_number": 25, "usage_type": "name"}, {"api_name": "rouge.Rouge", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 42, "usage_type": "call"}, {"api_name": "nltk.sent_tokenize", "line_number": 49, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 65, "usage_type": "call"}]}
+{"seq_id": "4180674774", "text": "# -*- coding:utf-8 -*-\r\n\"\"\"\r\n 爬虫案例流程\r\n 1,明确需求,需要爬取那些信息?\r\n 2,分析信息来自与哪里?\r\n 开发者工具,抓包分析,数据包来源与哪里?\r\n 视频 m3u8,分片段模式\r\n 分析网页源代码进行分析\r\n 想要视频内容 -------》 分片段ts文件 ---------》 m3u8文件里面 ----》 网页源代码\r\n headers ----> cookies host referer ua\r\n 3,\r\n 代码具体实现:\r\n 1,发送请求,网站url发起请求\r\n 2,获取数据,获取服务器相应html数据,并通过re模块正则表达式匹配\r\n 3,解析数据,提取我们想要的,m3u8 Ts,url请求链接\r\n\r\n 4,发送请求 m3u8url 获取 Ts数据\r\n 5,获取数据\r\n 6,解析数据\r\n 7,保存数据\r\n\"\"\"\r\nimport requests\r\nimport re\r\nimport json\r\nfrom pathlib import Path\r\nimport time\r\n# 伪装浏览器 headers\r\nheaders = {\r\n \"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/86.0.4240.198 Safari/537.36\"\r\n}\r\nvedio_dir = Path('./vedio')\r\nvedio_dir.mkdir(exist_ok=True)\r\n\r\ndef afun_init(url):\r\n # 发送请求,获取数据\r\n response = requests.get(url=url, headers=headers)\r\n # 提取视频标题\r\n title = re.findall('\"title\":\"(.*?)\"', response.text)[1]\r\n print(title)\r\n # 获取m3u8链接,获取Ts数据\r\n u3u8_str = re.findall('window.pageInfo = window.videoInfo = (.*?)};', response.text)[0]\r\n u3u8_str += '}'\r\n # 字符串类型数据\r\n # 根据键值对取m3u8url\r\n u3u8_json = json.loads(u3u8_str)\r\n # 二次转换\r\n # backupUrl 视频m3u8链接 对应的视频画质\r\n m3u8_url = \\\r\n json.loads(u3u8_json['currentVideoInfo']['ksPlayJson'])['adaptationSet'][0]['representation'][0]['backupUrl'][0]\r\n print(m3u8_url)\r\n # 获取m3u8链接完毕,获取Ts链接全部文件\r\n m3u8_data = requests.get(url=m3u8_url, headers=headers).text\r\n # 解析数据,获取Ts文件\r\n # 拆分分割符号\r\n m3u8_data = re.sub('#E.*', '', m3u8_data).split()\r\n return m3u8_data, title\r\n\r\nif __name__ == '__main__':\r\n print(\"A站视频下载器|单线程版\")\r\n while True:\r\n print(\"注: url值为 0 退出系统\")\r\n url = input(\"请输入视频url: \")\r\n if url == \"0\":\r\n break\r\n start_time = time.time()\r\n m3u8_data = afun_init(url)[0]\r\n title = afun_init(url)[1]\r\n # for 循环遍历 ts 内容\r\n print(m3u8_data)\r\n for ts in m3u8_data:\r\n ts_url = 'https://ali-safety-video.acfun.cn/mediacloud/acfun/acfun_video/' + ts\r\n # 获取视频二进制数据\r\n ts_content = requests.get(url=ts_url, headers=headers, timeout=5).content\r\n # 若视频标题不符合规范将会报错\r\n with open('./vedio/' + m3u8_data[0][0:5] + '.mp4', 'ab') as f:\r\n f.write(ts_content)\r\n print(ts_url)\r\n end_time = time.time()\r\n print(\"视频下载时间\", end_time - start_time)\r\n print(\"下载完毕\\n\")\r\n", "repo_name": "gaogaotwo/Project-Note-Code", "sub_path": "A站.py", "file_name": "A站.py", "file_ext": "py", "file_size_in_byte": 3129, "program_lang": "python", "lang": "zh", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "pathlib.Path", "line_number": 31, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 36, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 38, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 41, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 45, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 49, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 52, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 55, "usage_type": "call"}, {"api_name": "time.time", "line_number": 65, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 73, "usage_type": "call"}, {"api_name": "time.time", "line_number": 78, "usage_type": "call"}]}
+{"seq_id": "34303176417", "text": "\n\nimport torch\nimport torch.nn.functional as F\n\nfrom diffusion import DiffusionTrainer\nfrom bert import DiffusionBERT\nfrom transformers import BertTokenizer\n\nT = 100\ntokenizer = BertTokenizer.from_pretrained(\"bert-base-uncased\")\nmodel = DiffusionBERT(vocab_size=tokenizer.vocab_size)\nmaxlen = 128\n\ntest = DiffusionTrainer(T=T,\n tokenizer=tokenizer,\n model=model,\n maxlen=maxlen)\n\nprint(\"mask_token_ids \", tokenizer.mask_token_id)\nprint(\"pad_token_ids \", tokenizer.pad_token_id)\nprint(\"number \", tokenizer.convert_ids_to_tokens([10000]))\n\nx_t = F.one_hot(torch.tensor(tokenizer.mask_token_id), num_classes=30522).repeat(128, 1).unsqueeze(0).cuda()\nattention_mask = torch.ones((1, 128)).cuda()\n\nprint(x_t.shape)\nprint(attention_mask.shape)\n\nresult = test.predict_text(x_t=x_t, attention_mask=attention_mask)\n\nprint(result)\n", "repo_name": "YHL04/diffusionbert", "sub_path": "__test__.py", "file_name": "__test__.py", "file_ext": "py", "file_size_in_byte": 890, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "transformers.BertTokenizer.from_pretrained", "line_number": 11, "usage_type": "call"}, {"api_name": "transformers.BertTokenizer", "line_number": 11, "usage_type": "name"}, {"api_name": "bert.DiffusionBERT", "line_number": 12, "usage_type": "call"}, {"api_name": "diffusion.DiffusionTrainer", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn.functional.one_hot", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 25, "usage_type": "call"}]}
+{"seq_id": "39993290159", "text": "from django.urls import path, include\n\nfrom .views import * \n\napp_name = 'blog'\nurlpatterns = [\n path('',IndexView.as_view(),name='index'),\n path('/', BlogSingleView.as_view(), name='blog_single'),\n path('contact/', ContactView.as_view(), name='contact'),\n path('about/', AboutView.as_view(), name='about'),\n path('subscribe/', subscribe, name='subscribe'),\n path('user/', include('blog.user_post_interactions.urls'))\n]", "repo_name": "ferizoozoo/MediumLikeBlog", "sub_path": "blog/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 445, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "20", "api": [{"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 12, "usage_type": "call"}]}
+{"seq_id": "11429339586", "text": "# ============================================================================\n# FILE: type.py\n# AUTHOR: Shougo Matsushita \n# License: MIT license\n# ============================================================================\n\nfrom pynvim import Nvim\nimport typing\n\nfrom defx.base.column import Base, Highlights\nfrom defx.context import Context\nfrom defx.util import Candidate, len_bytes\nfrom defx.view import View\n\n\nclass Column(Base):\n\n def __init__(self, vim: Nvim) -> None:\n super().__init__(vim)\n\n self.name = 'type'\n types = [\n {\n 'name': 'text', 'globs': ['*.txt', '*.md', 'README'],\n 'icon': '[T]', 'highlight': 'Constant'\n },\n {\n 'name': 'image', 'globs': ['*.jpg'],\n 'icon': '[I]', 'highlight': 'Type'\n },\n {\n 'name': 'archive', 'globs': ['*.zip'],\n 'icon': '[A]', 'highlight': 'Special'\n },\n {\n 'name': 'executable', 'globs': ['*.exe'],\n 'icon': '[X]', 'highlight': 'Statement'\n },\n ]\n self.vars = {\n 'types': types,\n }\n self.has_get_with_highlights = True\n\n self._length: int = 0\n\n def on_init(self, view: View, context: Context) -> None:\n self._length = max([self.vim.call('strwidth', x['icon'])\n for x in self.vars['types']])\n\n def get_with_highlights(\n self, context: Context, candidate: Candidate\n ) -> typing.Tuple[str, Highlights]:\n for t in self.vars['types']:\n for glob in t['globs']:\n if not candidate['action__path'].match(glob):\n continue\n return (str(t['icon']), [\n (f\"{self.highlight_name}_{t['name']}\",\n self.start, len_bytes(t['icon']))\n ])\n\n return (' ' * self._length, [])\n\n def length(self, context: Context) -> int:\n return self._length\n\n def syntaxes(self) -> typing.List[str]:\n return [self.syntax_name + '_' + x['name'] for x\n in self.vars['types']]\n\n def highlight_commands(self) -> typing.List[str]:\n commands: typing.List[str] = []\n for t in self.vars['types']:\n commands.append(\n 'highlight default link {}_{} {}'.format(\n self.highlight_name, t['name'], t['highlight']))\n return commands\n", "repo_name": "Shougo/defx.nvim", "sub_path": "rplugin/python3/defx/column/type.py", "file_name": "type.py", "file_ext": "py", "file_size_in_byte": 2514, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1167, "dataset": "github-code", "pt": "24", "api": [{"api_name": "defx.base.column.Base", "line_number": 16, "usage_type": "name"}, {"api_name": "pynvim.Nvim", "line_number": 18, "usage_type": "name"}, {"api_name": "defx.view.View", "line_number": 47, "usage_type": "name"}, {"api_name": "defx.context.Context", "line_number": 47, "usage_type": "name"}, {"api_name": "defx.context.Context", "line_number": 52, "usage_type": "name"}, {"api_name": "defx.util.Candidate", "line_number": 52, "usage_type": "name"}, {"api_name": "defx.util.len_bytes", "line_number": 60, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 53, "usage_type": "attribute"}, {"api_name": "defx.base.column.Highlights", "line_number": 53, "usage_type": "name"}, {"api_name": "defx.context.Context", "line_number": 65, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 68, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 73, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 72, "usage_type": "attribute"}]}
+{"seq_id": "30780282300", "text": "# -*- coding: utf-8 -*-\n\"\"\"User views.\"\"\"\n\nfrom flask import (\n Blueprint,\n current_app,\n flash,\n jsonify,\n redirect,\n render_template,\n request,\n url_for,\n)\nfrom flask_login import current_user, login_required\n\nfrom interview_simulator.extensions import db\nfrom interview_simulator.user.models import UserFile\n\nfrom .forms import UploadForm\nfrom .services import chat_gpt, gpt_questions, transcribe_audio_with_whisper\n\nblueprint = Blueprint(\"user\", __name__, url_prefix=\"/users\", static_folder=\"../static\")\n\n\n@blueprint.route(\"/upload\", methods=[\"GET\", \"POST\"])\n@login_required\ndef upload():\n \"\"\"\n Handles the uploading of the Resume and Job Description files.\n\n Returns:\n - A string representing the HTML page displaying the form for uploading the files.\n \"\"\"\n form = UploadForm()\n if form.validate_on_submit():\n resume_text = form.resume_text.data\n job_description = form.job_description.data\n\n # Save the uploaded resume and job description to the database\n user_file = UserFile(\n file_name=\"Resume\", file_content=resume_text, user=current_user\n )\n db.session.add(user_file)\n\n user_file = UserFile(\n file_name=\"Job Description\", file_content=job_description, user=current_user\n )\n db.session.add(user_file)\n\n db.session.commit()\n\n flash(\"Resume and Job Description uploaded successfully!\", \"success\")\n return redirect(url_for(\"user.home_logged_in\"))\n return render_template(\"users/upload.html\", form=form)\n\n\n@blueprint.route(\"/check_uploads\")\n@login_required\ndef check_uploads():\n \"\"\"\n Checks if the user has uploaded a resume and job description and returns a JSON response.\n\n Returns:\n - A string representing the JSON response indicating if the user has uploaded the files.\n \"\"\"\n latest_resume = (\n UserFile.query.filter_by(user_id=current_user.id, file_name=\"Resume\")\n .order_by(UserFile.upload_date.desc())\n .first()\n )\n latest_job_description = (\n UserFile.query.filter_by(user_id=current_user.id, file_name=\"Job Description\")\n .order_by(UserFile.upload_date.desc())\n .first()\n )\n\n if latest_resume and latest_job_description:\n return jsonify(\n {\n \"uploaded\": True,\n \"resume\": latest_resume.file_content,\n \"job_description\": latest_job_description.file_content,\n }\n )\n else:\n return jsonify({\"uploaded\": False, \"resume\": None, \"job_description\": None})\n\n\n@blueprint.route(\"/start_game\", methods=[\"POST\"])\n@login_required\ndef start_game():\n \"\"\"\n Starts the game by calling the gpt_questions() function.\n\n Returns:\n - A string representing the JSON response containing the interview questions.\n \"\"\"\n resume = request.json.get(\"resume\")\n job_description = request.json.get(\"job_description\")\n questions = gpt_questions(resume, job_description)\n return jsonify(questions)\n\n\n@blueprint.route(\"/transcribe\", methods=[\"POST\"])\n@login_required\ndef transcribe():\n \"\"\"\n Transcribes an audio file using the Whisper ASR API and returns a JSON response.\n\n Returns:\n - A string representing the JSON response containing the transcribed text,\n the ChatGPT API response and the question asked.\n \"\"\"\n # Get the audio file from the request\n audio_file = request.files.get(\"audio\")\n question = request.form.get(\"question\")\n\n # log the audio file\n current_app.logger.info(f\"Audio file: {audio_file}\")\n\n if audio_file:\n # Extract the audio data from the file\n audio_data = audio_file.read()\n\n # Extract the transcribed text from the API response\n transcription = transcribe_audio_with_whisper(audio_data)\n\n # Call the ChatGPT API\n response = chat_gpt(question, transcription)\n\n # Return the transcription as a JSON response\n return jsonify(\n {\"transcription\": transcription, \"response\": response, \"question\": question}\n )\n else:\n # Return an error response if no audio file was provided\n return jsonify({\"error\": \"No audio file provided\"}), 400\n\n\n@blueprint.route(\"/home_logged_in\", methods=[\"GET\", \"POST\"])\n@login_required\ndef home_logged_in():\n \"\"\"\n Handles the home page for logged-in users.\n\n This function renders the home_logged_in.html template, which displays the form for inputting a message to ChatGPT.\n\n Returns:\n - A string representing the HTML page displaying the form for inputting a message to ChatGPT.\n \"\"\"\n return render_template(\"users/home_logged_in.html\")\n", "repo_name": "theuerc/interview_simulator", "sub_path": "interview_simulator/user/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4658, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 5, "dataset": "github-code", "pt": "24", "api": [{"api_name": "flask.Blueprint", "line_number": 22, "usage_type": "call"}, {"api_name": "forms.UploadForm", "line_number": 34, "usage_type": "call"}, {"api_name": "interview_simulator.user.models.UserFile", "line_number": 40, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 41, "usage_type": "name"}, {"api_name": "interview_simulator.extensions.db.session.add", "line_number": 43, "usage_type": "call"}, {"api_name": "interview_simulator.extensions.db.session", "line_number": 43, "usage_type": "attribute"}, {"api_name": "interview_simulator.extensions.db", "line_number": 43, "usage_type": "name"}, {"api_name": "interview_simulator.user.models.UserFile", "line_number": 45, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 46, "usage_type": "name"}, {"api_name": "interview_simulator.extensions.db.session.add", "line_number": 48, "usage_type": "call"}, {"api_name": "interview_simulator.extensions.db.session", "line_number": 48, "usage_type": "attribute"}, {"api_name": "interview_simulator.extensions.db", "line_number": 48, "usage_type": "name"}, {"api_name": "interview_simulator.extensions.db.session.commit", "line_number": 50, "usage_type": "call"}, {"api_name": "interview_simulator.extensions.db.session", "line_number": 50, "usage_type": "attribute"}, {"api_name": "interview_simulator.extensions.db", "line_number": 50, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 53, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 53, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 54, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 26, "usage_type": "name"}, {"api_name": "interview_simulator.user.models.UserFile.query.filter_by", "line_number": 67, "usage_type": "call"}, {"api_name": "interview_simulator.user.models.UserFile.query", "line_number": 67, "usage_type": "attribute"}, {"api_name": "interview_simulator.user.models.UserFile", "line_number": 67, "usage_type": "name"}, {"api_name": "flask_login.current_user.id", "line_number": 67, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 67, "usage_type": "name"}, {"api_name": "interview_simulator.user.models.UserFile.upload_date.desc", "line_number": 68, "usage_type": "call"}, {"api_name": "interview_simulator.user.models.UserFile.upload_date", "line_number": 68, "usage_type": "attribute"}, {"api_name": "interview_simulator.user.models.UserFile", "line_number": 68, "usage_type": "name"}, {"api_name": "interview_simulator.user.models.UserFile.query.filter_by", "line_number": 72, "usage_type": "call"}, {"api_name": "interview_simulator.user.models.UserFile.query", "line_number": 72, "usage_type": "attribute"}, {"api_name": "interview_simulator.user.models.UserFile", "line_number": 72, "usage_type": "name"}, {"api_name": "flask_login.current_user.id", "line_number": 72, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 72, "usage_type": "name"}, {"api_name": "interview_simulator.user.models.UserFile.upload_date.desc", "line_number": 73, "usage_type": "call"}, {"api_name": "interview_simulator.user.models.UserFile.upload_date", "line_number": 73, "usage_type": "attribute"}, {"api_name": "interview_simulator.user.models.UserFile", "line_number": 73, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 78, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 86, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 98, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 98, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 98, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 99, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 99, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 99, "usage_type": "name"}, {"api_name": "services.gpt_questions", "line_number": 100, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 101, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 90, "usage_type": "name"}, {"api_name": "flask.request.files.get", "line_number": 115, "usage_type": "call"}, {"api_name": "flask.request.files", "line_number": 115, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 115, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 116, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 116, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 116, "usage_type": "name"}, {"api_name": "flask.current_app.logger.info", "line_number": 119, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 119, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 119, "usage_type": "name"}, {"api_name": "services.transcribe_audio_with_whisper", "line_number": 126, "usage_type": "call"}, {"api_name": "services.chat_gpt", "line_number": 129, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 132, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 137, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 105, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 151, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 141, "usage_type": "name"}]}
+{"seq_id": "35252557940", "text": "__author__ = \"Thibault Dayris\"\n__copyright__ = \"Copyright 2023, Thibault Dayris\"\n__email__ = \"thibault.dayris@gustaveroussy.fr\"\n__license__ = \"MIT\"\n\n\nfrom tempfile import TemporaryDirectory\nfrom snakemake.shell import shell\n\n\nlog = snakemake.log_fmt_shell(stdout=True, stderr=True, append=True)\nextra = snakemake.params.get(\"extra\", \"\")\n\n# pyroe uses the flank-length value to name its output files\n# in the result directory. We need this value to acquired output\n# files and let snakemake-wrapper choose its output file names.\nread_length = snakemake.params.get(\"read_length\", 101)\nflank_trim_length = snakemake.params.get(\"flank_trim_length\", 5)\nflank_length = read_length - flank_trim_length\n\nspliced = snakemake.input.get(\"spliced\", \"\")\nif spliced:\n spliced = \"--extra-spliced \" + spliced\n\n\nunspliced = snakemake.input.get(\"unspliced\", \"\")\nif unspliced:\n unspliced = \"--extra-unspliced \" + unspliced\n\n\nwith TemporaryDirectory() as tempdir:\n shell(\n \"pyroe make-spliced+intronic \"\n \"{extra} {spliced} \"\n \"{unspliced} \"\n \"{snakemake.input.fasta} \"\n \"{snakemake.input.gtf} \"\n \"{read_length} \"\n \"{tempdir} \"\n \"{log}\"\n )\n\n if snakemake.output.get(\"fasta\", False):\n shell(\n \"mv --verbose \"\n \"{tempdir}/splici_fl{flank_length}.fa \"\n \"{snakemake.output.fasta} {log}\"\n )\n\n if snakemake.output.get(\"gene_id_to_name\", False):\n shell(\n \"mv --verbose \"\n \"{tempdir}/gene_id_to_name.tsv \"\n \"{snakemake.output.gene_id_to_name} {log}\"\n )\n\n if snakemake.output.get(\"t2g\", False):\n shell(\n \"mv --verbose \"\n \"{tempdir}/splici_fl{flank_length}_t2g_3col.tsv \"\n \"{snakemake.output.t2g} {log} \"\n )\n", "repo_name": "snakemake/snakemake-wrappers", "sub_path": "bio/pyroe/makesplicedintronic/wrapper.py", "file_name": "wrapper.py", "file_ext": "py", "file_size_in_byte": 1796, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 182, "dataset": "github-code", "pt": "24", "api": [{"api_name": "snakemake.shell.log_fmt_shell", "line_number": 11, "usage_type": "call"}, {"api_name": "snakemake.shell", "line_number": 11, "usage_type": "name"}, {"api_name": "snakemake.shell.params.get", "line_number": 12, "usage_type": "call"}, {"api_name": "snakemake.shell.params", "line_number": 12, "usage_type": "attribute"}, {"api_name": "snakemake.shell", "line_number": 12, "usage_type": "name"}, {"api_name": "snakemake.shell.params.get", "line_number": 17, "usage_type": "call"}, {"api_name": "snakemake.shell.params", "line_number": 17, "usage_type": "attribute"}, {"api_name": "snakemake.shell", "line_number": 17, "usage_type": "name"}, {"api_name": "snakemake.shell.params.get", "line_number": 18, "usage_type": "call"}, {"api_name": "snakemake.shell.params", "line_number": 18, "usage_type": "attribute"}, {"api_name": "snakemake.shell", "line_number": 18, "usage_type": "name"}, {"api_name": "snakemake.shell.input.get", "line_number": 21, "usage_type": "call"}, {"api_name": "snakemake.shell.input", "line_number": 21, "usage_type": "attribute"}, {"api_name": "snakemake.shell", "line_number": 21, "usage_type": "name"}, {"api_name": "snakemake.shell.input.get", "line_number": 26, "usage_type": "call"}, {"api_name": "snakemake.shell.input", "line_number": 26, "usage_type": "attribute"}, {"api_name": "snakemake.shell", "line_number": 26, "usage_type": "name"}, {"api_name": "tempfile.TemporaryDirectory", "line_number": 31, "usage_type": "call"}, {"api_name": "snakemake.shell.shell", "line_number": 32, "usage_type": "call"}, {"api_name": "snakemake.shell.output.get", "line_number": 43, "usage_type": "call"}, {"api_name": "snakemake.shell.output", "line_number": 43, "usage_type": "attribute"}, {"api_name": "snakemake.shell", "line_number": 43, "usage_type": "name"}, {"api_name": "snakemake.shell.shell", "line_number": 44, "usage_type": "call"}, {"api_name": "snakemake.shell.output.get", "line_number": 50, "usage_type": "call"}, {"api_name": "snakemake.shell.output", "line_number": 50, "usage_type": "attribute"}, {"api_name": "snakemake.shell", "line_number": 50, "usage_type": "name"}, {"api_name": "snakemake.shell.shell", "line_number": 51, "usage_type": "call"}, {"api_name": "snakemake.shell.output.get", "line_number": 57, "usage_type": "call"}, {"api_name": "snakemake.shell.output", "line_number": 57, "usage_type": "attribute"}, {"api_name": "snakemake.shell", "line_number": 57, "usage_type": "name"}, {"api_name": "snakemake.shell.shell", "line_number": 58, "usage_type": "call"}]}
+{"seq_id": "29737401719", "text": "import os\nfrom setuptools import setup\n\nwith open('requirements.txt') as f:\n required = f.read().splitlines()\n\nsetup(\n name='django-keen',\n version='0.1.3',\n author='Jannis Gebauer',\n author_email='ja.geb@pricemesh.io',\n packages=['dkeen',],\n url='http://pypi.python.org/pypi/django-keen/',\n license='LICENSE.txt',\n description='Simple wrapper for django around the official keen.io client',\n long_description=open('README.md').read(),\n install_requires=required\n)", "repo_name": "jayfk/django-keen", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 497, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "setuptools.setup", "line_number": 7, "usage_type": "call"}]}
+{"seq_id": "22541081394", "text": "# import dependencies\nimport cv2 # connect webcam, process images\nimport mediapipe as mp # holistic API\nimport time # to calculate FPS\n# import tensorflow as tf, keras # to create model for training\n# from keras.models import load_model # load pretrained model\nimport numpy as np # processing\n\nimport imageio\n\n\n\nif __name__ == '__main__':\n width, height = 1280, 720\n # SETUP MEDIAPIPE\n print('Setting up................')\n mp_drawing = mp.solutions.drawing_utils # help draw the detections\n mp_holistic = mp.solutions.holistic # a Holistic class object\n\n # GET REALTIME WEBCAM FEED\n print('Getting webcam feed.................')\n ## define a video capture object, 0 is the webcam\n ## by default, each frame has size (480x640) (480 x 640)\n start, end = 0, 0 # helper variables to calculate FPS\n demo_path = '.\\\\assets\\\\demo\\\\mantalking.mp4'\n cap = cv2.VideoCapture(0)\n cap.set(3, width)\n cap.set(4, height)\n print('Initiate Holistic Model') \n # Initiate holistic model\n # dataset = []\n demo_frames = []\n with mp_holistic.Holistic( \\\n # model_complexity=2,\n min_detection_confidence=0.5, \\\n min_tracking_confidence=0.5) as holistic:\n print('Opening webcam feed........... Press q to stop')\n while cap.isOpened():\n\n start = time.time()\n # Capture the video frame\n # by frame\n success, frame = cap.read()\n if not success:\n print('Cannot receive frame from camera')\n break\n\n # flip the image vertically for later selfie view display\n # recolor feed from BGR to RGB so that the model will have good performance\n frame = cv2.cvtColor(cv2.flip(frame, 1), cv2.COLOR_BGR2RGB)\n\n # to improve performance, mark the image as not writeable to\n # pass by reference instead of making a copy\n frame.flags.writeable = False\n \n # make detection\n results = holistic.process(frame) # store all different kinds of landmarks...\n\n # enable drawing landmark annotation on the frame\n frame.flags.writeable = True \n cv2.imshow('ko che', cv2.cvtColor(frame, cv2.COLOR_RGB2BGR))\n frame = np.zeros(frame.shape) \n # recolor feed from RGB to BGR so it can be displayed\n # frame = cv2.cvtColor(frame, cv2.COLOR_RGB2BGR)\n \n # extract mouth features\n # mouth_landmarks = []\n # if results.face_landmarks:\n # mouth_landmarks.append(np.array([results.face_landmarks.landmark[61].x * width, results.face_landmarks.landmark[61].y * height]))\n # mouth_landmarks.append(np.array([results.face_landmarks.landmark[291].x * width, results.face_landmarks.landmark[291].y * height]))\n # mouth_landmarks.append(np.array([results.face_landmarks.landmark[0].x * width, results.face_landmarks.landmark[0].y * height]))\n # mouth_landmarks.append(np.array([results.face_landmarks.landmark[17].x * width, results.face_landmarks.landmark[17].y * height]))\n\n # mouth_landmarks.append(np.array([results.face_landmarks.landmark[14].x * width, results.face_landmarks.landmark[14].y * height]))\n # # mouth_landmarks.append(np.array([results.face_landmarks.landmark[87].x * width, results.face_landmarks.landmark[87].y * height]))\n # # mouth_landmarks.append(np.array([results.face_landmarks.landmark[312].x * width, results.face_landmarks.landmark[312].y * height]))\n # # mouth_landmarks.append(np.array([results.face_landmarks.landmark[317].x * width, results.face_landmarks.landmark[317].y * height]))\n # vector_1 = mouth_landmarks[4] - mouth_landmarks[1]\n # vector_2 = mouth_landmarks[4] - mouth_landmarks[0]\n # angle = np.arccos(np.dot(vector_1 / np.linalg.norm(vector_1), \\\n # vector_2 / np.linalg.norm(vector_2)))*57.2958\n # for x, y in mouth_landmarks:\n # frame = cv2.circle(frame, (int(x), int(y)), radius=5, color=(0, 0, 255), thickness=5)\n # # draw facemesh \n # mp_drawing.draw_landmarks(frame, \\\n # results.face_landmarks,\\\n # mp_holistic.FACEMESH_TESSELATION,\\\n # # stylizing\n # mp_drawing.DrawingSpec(color=(245,117,66), thickness=1, circle_radius=1), \n # mp_drawing.DrawingSpec(color=(245,66,230), thickness=1, circle_radius=1))\n # cv2.putText(frame, str(f'angle: {round(angle, 3)} degree'), (10, 680), cv2.FONT_HERSHEY_COMPLEX, 3, (153,43,37), 3)\n # # draw pose landmarks\n # mp_drawing.draw_landmarks(frame, results.pose_landmarks, mp_holistic.POSE_CONNECTIONS)\n # draw left hand landmarks\n mp_drawing.draw_landmarks(frame, results.left_hand_landmarks, mp_holistic.HAND_CONNECTIONS)\n # if results.left_hand_landmarks: print(results.left_hand_landmarks)\n # draw right hand landmarks\n mp_drawing.draw_landmarks(frame, results.right_hand_landmarks, mp_holistic.HAND_CONNECTIONS)\n \n # calculate how long this code takes to process a frame on a CPU\n end = time.time() \n fps = 1/(end - start)\n # display FPS on the frame\n # cv2.putText(frame, str(f'FPS: {int(fps)}'), (10, 70), cv2.FONT_HERSHEY_COMPLEX, 3, (255, 255, 255), 3)\n \n # demo_frames.append(frame)\n # Display the resulting frame\n cv2.imshow('Webcam Feed', frame)\n\n # the 'q' button is set as the\n # quitting button you may use any\n # desired button of your choice\n if cv2.waitKey(1) & 0xFF == ord('q'):\n break\n\n # After the loop release the cap object\n cap.release()\n # Destroy all the windows\n cv2.destroyAllWindows()\n # imageio.mimsave('./proof/mouth_angle.gif', demo_frames, fps=10)\n print('Saving mouth dataset..........')\n # print(dataset)\n", "repo_name": "uyenbhku/CS231_ImageProcessingProject", "sub_path": "dev_files/baseline.py", "file_name": "baseline.py", "file_ext": "py", "file_size_in_byte": 6330, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "mediapipe.solutions", "line_number": 17, "usage_type": "attribute"}, {"api_name": "mediapipe.solutions", "line_number": 18, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 26, "usage_type": "call"}, {"api_name": "time.time", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 50, "usage_type": "call"}, {"api_name": "cv2.flip", "line_number": 50, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 50, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 61, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 61, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 61, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 62, "usage_type": "call"}, {"api_name": "time.time", "line_number": 101, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 108, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 113, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 119, "usage_type": "call"}]}
+{"seq_id": "19494866258", "text": "# ---\n# jupyter:\n# jupytext:\n# formats: ipynb,py:light\n# text_representation:\n# extension: .py\n# format_name: light\n# format_version: '1.4'\n# jupytext_version: 1.2.4\n# kernelspec:\n# display_name: Python 3\n# language: python\n# name: python3\n# ---\n\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport pymc3 as pm\n# #!pip3 install jupytext\n\ndf = pd.read_csv('./poverty.csv')\n\ndf.head()\n\nX = df['PovPct']\ny = df['Brth15to17']\n\nplt.scatter(X, y)\n\nbasic_model = pm.Model()\nwith basic_model:\n alpha = pm.Normal('alpha', mu=0, sigma=10)\n beta = pm.Normal('beta', mu=0, sigma=10)\n sigma = pm.HalfNormal('sigma', sigma=1)\n \n mu = alpha + beta*X\n \n y_obs = pm.Normal('Y_obs', mu=mu, sigma=sigma, observed=y) # observed apparently makes this the likelihood\n # presumably for glm one replaces normal with the appropriate residual.\n\nmap_estimate = pm.find_MAP(model=basic_model)\nmap_estimate\n\nwith basic_model:\n trace = pm.sample(5000)\n\npm.traceplot(trace)\n\n# +\n# https://docs.pymc.io/notebooks/GLM-hierarchical.html\n", "repo_name": "mwpb/bayesian-regression", "sub_path": "pym3_testing.py", "file_name": "pym3_testing.py", "file_ext": "py", "file_size_in_byte": 1104, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "pandas.read_csv", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "pymc3.Model", "line_number": 31, "usage_type": "call"}, {"api_name": "pymc3.Normal", "line_number": 33, "usage_type": "call"}, {"api_name": "pymc3.Normal", "line_number": 34, "usage_type": "call"}, {"api_name": "pymc3.HalfNormal", "line_number": 35, "usage_type": "call"}, {"api_name": "pymc3.Normal", "line_number": 39, "usage_type": "call"}, {"api_name": "pymc3.find_MAP", "line_number": 42, "usage_type": "call"}, {"api_name": "pymc3.sample", "line_number": 46, "usage_type": "call"}, {"api_name": "pymc3.traceplot", "line_number": 48, "usage_type": "call"}]}
+{"seq_id": "74268902448", "text": "import numpy as np\nimport torch\nfrom torch import functional\nfrom maze1_env import Maze\nfrom DQN import DeepQNetwork\n\nclass Config(object):\n TARGET_REPLACE_ITER = 100 # target update frequency\n BATCH_SIZE = 32\n LR = 0.01 # learning rate\n EPSILON = 0.9 # greedy policy\n GAMMA = 0.9 # reward discount\n MEMORY_SIZE = 500\n MAX_EPISODE = 500\n\nopt = Config()\nenv = Maze()\n\ndef update_dqn(RL):\n step = 0\n for episode in range(opt.MAX_EPISODE):\n # initial observation\n observation = env.reset()\n\n while True:\n # fresh env\n env.render()\n # RL choose action based on observation\n action = RL.choose_action(observation)\n # RL take action and get next observation and reward\n observation_, reward, done = env.step(action)\n RL.store_transition(observation, action, reward, observation_)\n if (step > 200) and (step % 5 == 0):\n RL.learn()\n # swap observation\n observation = observation_\n # break while loop when end of this episode\n if done:\n print('--------episode:{0}-------'.format(episode))\n break\n step += 1\n\n # end of game\n print('game over')\n env.destroy()\n\ndef train(**kwargs):\n for k_, v_ in kwargs.items():\n setattr(opt, k_, v_) \n\n RL = DeepQNetwork(env.n_actions, env.n_features, opt)\n env.after(100, update_dqn(RL))\n env.mainloop()\n \nif __name__ == '__main__': \n import fire\n fire.Fire()\n", "repo_name": "yuanyilikl/Reinforcement-Learning", "sub_path": "DQN_run.py", "file_name": "DQN_run.py", "file_ext": "py", "file_size_in_byte": 1607, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "20", "api": [{"api_name": "maze1_env.Maze", "line_number": 17, "usage_type": "call"}, {"api_name": "DQN.DeepQNetwork", "line_number": 51, "usage_type": "call"}, {"api_name": "fire.Fire", "line_number": 57, "usage_type": "call"}]}
+{"seq_id": "74416463422", "text": "import abc\nimport gin\nimport tensorflow as tf\n\n\ndef vtrace(values, rewards, done_terminated, done_abandoned, discount_factor,\n target_action_log_probs, behaviour_action_log_probs, lambda_=1.0,\n max_importance_weight=1., name='vtrace'):\n r\"\"\"Calculates V-trace value targets and advantages.\n\n Args:\n values: A float32 tensor of shape [T+1, B] with the value function estimates\n wrt. the target policy, i.e., for the time steps, i, i+1, ..., i+T.\n rewards: A float32 tensor of shape [T, B] containing rewards generated by\n following the behaviour policy after time steps i, i+1, ..., i+T-1.\n done_terminated: A boolean tensor of shape [T, B] signifying if the agent\n terminated after the actions in steps i, i+1, ..., i+T-1. This is\n equivalent to going into a terminal state with infinite rewards of 0.\n done_abandoned: A boolean tensor of shape [T, B] signifying if the agent did\n not further act after the actions in steps i, i+1, ..., i+T-1. This is not\n the same as termination and can be used if the maximum episode length is\n reached. This will set the advantage of that state to zero and the target\n value to the input value (which generally results in a zero gradient).\n discount_factor: Float with the discount factor to be used.\n target_action_log_probs: A float32 tensor of shape [T, B] with\n log-probabilities of taking the action by the current policy.\n behaviour_action_log_probs: A float32 tensor of shape [T, B] with\n log-probabilities of taking the action by the behavioural policy.\n lambda_: Float that determines the mix between 1-step (lambda_=0) and n-step\n (lambda_=1) bootstrapping\n max_importance_weight: Bigger importance weights are clipped.\n name: String with the name scope that all operations will be created in.\n\n Returns:\n A float32 tensor of shape [T, B] with value targets that can be used to\n train a baseline (V(x_t) - vs_t)^2.\n A float32 tensor of shape [T, B] of advantages.\n \"\"\"\n with tf.name_scope(name):\n # Compute importance sampling weights.\n log_rhos = target_action_log_probs - behaviour_action_log_probs\n log_rhos = tf.minimum(log_rhos, tf.math.log(max_importance_weight))\n rhos = tf.exp(log_rhos)\n\n # We compute the temporal differences with special handling of episodes\n # which ended. We consider two cases:\n # - Termination: In this case, the agent took a decision that led to proper\n # termination of the episode. In this case, the future value of the\n # policy is enforced to be zero. This is done by setting the next step\n # bootstrapping value to zero and not to the next value function (which\n # is the value of the state after the reset).\n not_terminated_mask = tf.cast(~done_terminated, tf.float32)\n next_step_bootstrap = not_terminated_mask * values[1:]\n\n # - Abandonment: The current episode was abandoned, e.g., due to a maximum\n # epsiode length. If the policy would have continued, it would have\n # continued to obtain rewards. We handle this by setting the temporal\n # difference and thus the advantage to zero.\n not_abandoned_mask = tf.cast(~done_abandoned, tf.float32)\n deltas = rewards + discount_factor * next_step_bootstrap - values[:-1]\n deltas *= not_abandoned_mask\n\n # For both cases, we do not propagate future temporal differences as they\n # relate to different episodes.\n propagate_future = not_terminated_mask * not_abandoned_mask\n\n # We accumulate temporal differences by iterating backwards in time and\n # computing advantages as we go using dynamic programming.\n accumulator = tf.zeros_like(values[0])\n targets = []\n advantages = []\n for i in range(int(rewards.shape[0]) - 1, -1, -1):\n future = propagate_future[i] * discount_factor * lambda_ * accumulator\n # For advantages we don't use importance weights because this is\n # the advantage exactly for the action which was taken.\n advantages.append(deltas[i] + future)\n # On the other hand, the accumulator corresponds to the value for the\n # current state so both terms are multiplied by rho.\n accumulator = rhos[i] * (deltas[i] + future)\n targets.append(values[i] + accumulator)\n\n # We need to return targets and values with stopped gradients, as we do not\n # want to differentiate through the generalized advantage estimator.\n targets = tf.convert_to_tensor(targets[::-1], dtype=tf.float32)\n advantages = tf.convert_to_tensor(advantages[::-1], dtype=tf.float32)\n return tf.stop_gradient(targets), tf.stop_gradient(advantages)\n\n\n\n\ndef gae(values, rewards, done_terminated, done_abandoned, discount_factor,\n target_action_log_probs=None, behaviour_action_log_probs=None,\n lambda_=1.0, name='gae'):\n \"\"\"Generalized Advantages Estimator.\n\n Args:\n See V-trace above.\n\n Returns:\n A float32 tensor of shape [T, B] with value targets that can be used to\n train a baseline (V(x_t) - vs_t)^2.\n A float32 tensor of shape [T, B] of advantages.\n \"\"\"\n return vtrace(values, rewards, done_terminated, done_abandoned,\n discount_factor,\n tf.zeros_like(rewards), tf.zeros_like(rewards),\n lambda_, 1., name)\n\n\nclass AdvantageEstimator(tf.Module, metaclass=abc.ABCMeta):\n \"\"\"Abstract base class for advantage estimators.\"\"\"\n\n @abc.abstractmethod\n def __call__(self, values, rewards, done_terminated, done_abandoned,\n discount_factor, target_action_log_probs,\n behaviour_action_log_probs):\n r\"\"\"Computes advantages and value function targets.\n\n Args:\n values: A float32 tensor of shape [T+1, B] with the value function\n estimates wrt. the target policy, i.e., for the time steps, i, i+1,\n ..., i+T.\n rewards: A float32 tensor of shape [T, B] containing rewards generated by\n following the behaviour policy after time steps i, i+1, ..., i+T-1.\n done_terminated: A boolean tensor of shape [T, B] signifying if the agent\n terminated after the actions in steps i, i+1, ..., i+T-1. This is\n equivalent to going into a terminal state with infinite rewards of 0.\n done_abandoned: A boolean tensor of shape [T, B] signifying if the agent\n did not further act after the actions in steps i, i+1, ..., i+T-1. This\n is not the same as termination and can be used if the maximum episode\n length is reached. This will set the advantage of that state to zero and\n the target value to the input value (which generally results in a zero\n gradient).\n discount_factor: Float with the discount factor to be used.\n target_action_log_probs: A float32 tensor of shape [T, B] with\n log-probabilities of taking the action by the current policy\n behaviour_action_log_probs: A float32 tensor of shape [T, B] with\n log-probabilities of taking the action by the behavioural policy\n\n\n Returns:\n A float32 tensor of shape [T, B] with value targets that can be used to\n train a baseline (V(x_t) - vs_t)^2.\n A float32 tensor of shape [T, B] of advantages.\n \"\"\"\n raise NotImplementedError('`__call__()` is not implemented!')\n\n\n@gin.configurable\nclass GAE(AdvantageEstimator):\n\n def __init__(self, lambda_, name='gam'):\n super().__init__()\n self.lambda_ = lambda_\n\n def __call__(self, *args, **kwargs):\n return gae(*args, **kwargs, lambda_=self.lambda_, name=self.name)\n\n\n@gin.configurable\nclass VTrace(AdvantageEstimator):\n\n def __init__(self, lambda_, max_importance_weight=1., name='vtrace'):\n super().__init__(name)\n self.lambda_ = lambda_\n self.max_importance_weight = max_importance_weight\n\n def __call__(self, *args, **kwargs):\n return vtrace(*args, **kwargs,\n max_importance_weight=self.max_importance_weight,\n lambda_=self.lambda_, name=self.name)\n\n\n@gin.configurable\nclass NStep(AdvantageEstimator):\n \"\"\"N-step returns.\"\"\"\n\n def __init__(self, n, name='nstep2'):\n super().__init__(name)\n self.n = n\n\n def __call__(self, values, rewards, done_terminated, done_abandoned,\n discount_factor, target_action_log_probs,\n behaviour_action_log_probs):\n with tf.name_scope(self.name):\n # We compute the n-step returns in min(n, unroll_length) steps.\n unroll_length = int(rewards.shape[0])\n eff_n = self.n if self.n < unroll_length else unroll_length\n\n # We pad the dimension with n-1 additional values with abandon=True so\n # that we don't have to handle the last n-1 steps differently.\n values_pad = tf.zeros((eff_n - 1, values.shape[1]), dtype=tf.float32)\n done_terminated_pad = tf.zeros((eff_n - 1, values.shape[1]),\n dtype=tf.bool)\n done_abandoned_pad = tf.ones((eff_n - 1, values.shape[1]), dtype=tf.bool)\n rewards_pad = tf.zeros((eff_n - 1, values.shape[1]), dtype=tf.float32)\n\n nvalues = tf.concat([values, values_pad], axis=0)\n ndone_terminated = tf.concat([done_terminated, done_terminated_pad],\n axis=0)\n ndone_abandoned = tf.concat([done_abandoned, done_abandoned_pad], axis=0)\n nrewards = tf.concat([rewards, rewards_pad], axis=0)\n\n future_value = nvalues[eff_n:]\n\n window_size = rewards.shape[0]\n\n for i in range(eff_n):\n # Extract relevant sub tensors.\n start = eff_n - i - 1\n end = start + window_size\n rel_n_values = nvalues[start:end]\n rel_rewards = nrewards[start:end]\n rel_done_terminated = ndone_terminated[start:end]\n rel_done_abandoned = ndone_abandoned[start:end]\n\n # We compute the targets with special handling of episodes\n # which ended. We consider two cases:\n # - Termination: In this case, the agent took a decision that led to\n # proper termination of the episode. In this case, the future value\n # of the policy is enforced to be zero. This is done by setting the\n # next step bootstrapping value to zero and not to the next value\n # function (which is the value of the state after the reset).\n not_terminated_mask = tf.cast(~rel_done_terminated, tf.float32)\n next_step_bootstrap = not_terminated_mask * future_value\n\n # - Abandonment: The current episode was abandoned, e.g., due to a\n # maximum episode length (or padding). If the policy would have\n # continued, it would have continued to obtain rewards. We handle\n # this by setting the value to the current value.\n not_abandoned_mask = tf.cast(~rel_done_abandoned, tf.float32)\n abandoned_mask = tf.cast(rel_done_abandoned, tf.float32)\n\n one_step_bootstrap = rel_rewards + discount_factor * next_step_bootstrap\n\n future_value = (not_abandoned_mask*one_step_bootstrap +\n abandoned_mask*rel_n_values)\n\n advantages = future_value - values[:-1]\n return tf.stop_gradient(future_value), tf.stop_gradient(advantages)\n", "repo_name": "google-research/seed_rl", "sub_path": "agents/policy_gradient/modules/advantages.py", "file_name": "advantages.py", "file_ext": "py", "file_size_in_byte": 11097, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 783, "dataset": "github-code", "pt": "24", "api": [{"api_name": "tensorflow.name_scope", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.minimum", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.math.log", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tensorflow.exp", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow.zeros_like", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.convert_to_tensor", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 84, "usage_type": "attribute"}, {"api_name": "tensorflow.convert_to_tensor", "line_number": 85, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 85, "usage_type": "attribute"}, {"api_name": "tensorflow.stop_gradient", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.zeros_like", "line_number": 106, "usage_type": "call"}, {"api_name": "tensorflow.Module", "line_number": 110, "usage_type": "attribute"}, {"api_name": "abc.ABCMeta", "line_number": 110, "usage_type": "attribute"}, {"api_name": "abc.abstractmethod", "line_number": 113, "usage_type": "attribute"}, {"api_name": "gin.configurable", "line_number": 149, "usage_type": "attribute"}, {"api_name": "gin.configurable", "line_number": 160, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 185, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 192, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 192, "usage_type": "attribute"}, {"api_name": "tensorflow.zeros", "line_number": 193, "usage_type": "call"}, {"api_name": "tensorflow.bool", "line_number": 194, "usage_type": "attribute"}, {"api_name": "tensorflow.ones", "line_number": 195, "usage_type": "call"}, {"api_name": "tensorflow.bool", "line_number": 195, "usage_type": "attribute"}, {"api_name": "tensorflow.zeros", "line_number": 196, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 196, "usage_type": "attribute"}, {"api_name": "tensorflow.concat", "line_number": 198, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 199, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 201, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 202, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 224, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 224, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 231, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 231, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 232, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 232, "usage_type": "attribute"}, {"api_name": "tensorflow.stop_gradient", "line_number": 240, "usage_type": "call"}, {"api_name": "gin.configurable", "line_number": 174, "usage_type": "attribute"}]}
+{"seq_id": "74516767730", "text": "import torch\r\nimport os\r\nfrom glob import glob\r\n\r\nmodel_folder = './models'\r\nfolder = os.listdir(model_folder)\r\n\r\nfor i in range(len(folder)):\r\n filenames = glob(os.path.join(model_folder, folder[i], '*.pth'))\r\n for j in range(len(filenames)):\r\n model = torch.load(filenames[j], map_location='cpu')\r\n backbone = 0\r\n neck = 0\r\n head = 0\r\n all = 0\r\n for key in list(model['state_dict'].keys()):\r\n if 'backbone' in key:\r\n # if key.startswith('img_backbone'):\r\n backbone += model['state_dict'][key].nelement()\r\n elif 'neck' in key:\r\n neck += model['state_dict'][key].nelement()\r\n elif 'head' in key:\r\n head += model['state_dict'][key].nelement()\r\n\r\n all += model['state_dict'][key].nelement()\r\n print(filenames[j])\r\n print(f\"Backbone param: {backbone / 1e6}M\")\r\n print(f\"Neck param: {neck / 1e6}M\")\r\n print(f\"Head param: {head / 1e6}M\")\r\n print(f\"Total param: {all / 1e6}M\")\r\n\r\n# smaller 63374123\r\n# v4 69140395\r\n", "repo_name": "Daniel-xsy/BEV-Attack", "sub_path": "mmdet_adv/tools/analysis_tools/get_params.py", "file_name": "get_params.py", "file_ext": "py", "file_size_in_byte": 1092, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 10, "dataset": "github-code", "pt": "20", "api": [{"api_name": "os.listdir", "line_number": 6, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 11, "usage_type": "call"}]}
+{"seq_id": "17096658626", "text": "import requests\n\nservice_url = 'https://example.com'\n\nexpected_status_code = 200\n\ndef check_service_status(url,expected_code):\n try:\n response = requests.get(url)\n if response.status_code == expected_code:\n return True\n else:\n return False\n except requests.exceptions.RequestException:\n return False\nif check_service_status(service_url,expected_status_code):\n print(\"The service at {} is up and running.\".format(service_url))\nelse:\n print(\"The service at {} is up and running.\".format(service_url))", "repo_name": "Shreyashbhise/PythonForDevOps", "sub_path": "practiceexampl3.py", "file_name": "practiceexampl3.py", "file_ext": "py", "file_size_in_byte": 561, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "20", "api": [{"api_name": "requests.get", "line_number": 9, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 14, "usage_type": "attribute"}]}
+{"seq_id": "38994124664", "text": "import os\nimport discord\nimport random\nfrom replit import db\n\n\nclient = discord.Client()\n\n######################### UPDATE em r_palavras_acrescentadas #####################################\ndef uptade_encouragements(encouraging_message):\n if \"encouragements\" in db.keys():\n encouragements = db[\"encouragements\"]\n encouragements.append(encouraging_message)\n db[\"encouragements\"] = encouragements\n else:\n db[\"encouragements\"] = [encouraging_message]\n\ndef delete_encouragements(index):\n encouragements = db[\"encouragements\"]\n if len (encouragements) > index:\n del encouragements [index]\n db[\"encouragements\"] = encouragements\n################################################################################################\n######################## UPDATE em palavras_acrescentadas #####################################\ndef uptade_encouragements2(encouraging_message2):\n if \"encouragements2\" in db.keys():\n encouragements2 = db[\"encouragements2\"]\n encouragements2.append(encouraging_message2)\n db[\"encouragements2\"] = encouragements2\n else:\n db[\"encouragements2\"] = [encouraging_message2]\n\ndef delete_encouragements2(index2):\n encouragements2 = db[\"encouragements2\"]\n if len (encouragements2) > index2:\n del encouragements2 [index2]\n db[\"encouragements2\"] = encouragements2\n################################################################################################3\n\n@client.event\nasync def on_ready():\n #print(\"We have logged in as {0.user}\".format(client))\n print(f\"Connected succesfully as {client.user}\")\n\npalavras_acrescentadas = [\"jogo\",\"mano\"]\nr_palavras_acrescentadas = [\"bacana\",\"ha ha ha\"]\n\ndiferente = [\"quero uivar\",\"Tô com fome\", \"que canseira\", \"relaxa carinha\",\"acho que tô com pulga\"]\nr_diferente = [\"quero uivar\",\"Tô com fome\", \"que canseira\", \"relaxa carinha\",\"acho que tô com pulga\"]\n\npalavrax = [\"oi \", \"ola \"]\nr_palavrax = [\"olá amigo como vc está hoje?\",\" tô perdido aqui\",\"au au au\",\"eu sou o cão coragem\"]\n\ncumprimentosx = [\"bem\",\"bom\",\"legal\"]\nr_cumprimentosx = [\"vc disse bem uhuu que bacana\",\"vc disse bom, que legal carinha\", \"vc escreveu legal e eu concordo com vc!!\"]\n\ngx = [\"bom dia\",\"boa noite\", \"boa tarde\"]\nr_gx = [\"para vc também\", \"bom para todos\", \"bom mesmo uhu\"]\n\nif \"responding\" not in db.keys():\n db[\"responding\"] = True\n\n@client.event\nasync def on_message(message):\n if message.author == client.user:\n return\n############## acrescenta palavra na lista r_palavras_acrescentadas #######\n \n if db[\"responding\"]:\n options = r_palavras_acrescentadas\n if \"encouragements\" in db.keys():\n options.extend(db[\"encouragements\"]) #adiciona lista em outra lista\n #options = options + db[\"encouragements\"]\n \n if any(word in message.content for word in palavras_acrescentadas):\n await message.channel.send(random.choice(options))\n \n if message.content.startswith(\"$new \"):\n encouraging_message = message.content.split(\"$new \",1)[1]\n uptade_encouragements(encouraging_message)\n await message.channel.send(\"New encouraging message added.\")\n\n if message.content.startswith(\"$del\"):\n encouragements = []\n if \"encouragements\" in db.keys():\n index = int(message.content.split(\"$del\",1)[1])\n delete_encouragements(index)\n encouragements = db[\"encouragements\"]\n await message.channel.send(encouragements)\n############################################################################\n if message.content.startswith(\"$list\"):\n encouragements = []\n if \"encouragements\" in db.keys():\n encouragements = db[\"encouragements\"]\n await message.channel.send(encouragements)\n\n if message.content.startswith(\"$responding\"):\n value = message.content.split(\"$responding \",1)[1]\n\n if value.lower() == \"true\":\n db[\"responding\"] = True\n await message.channel.send(\"cao coragem ligado!\")\n else:\n db[\"responding\"] = False\n await message.channel.send(\"cao coragem desligado\")\n\n #############################################################################\n ############## acrescenta palavra na lista palavras_acrescentadas #######\n options2 = palavras_acrescentadas\n if \"encouragements2\" in db.keys():\n options2.extend(db[\"encouragements2\"]) #adiciona lista em outra lista\n #options = options + db[\"encouragements\"]\n \n if any(word in message.content for word in r_palavras_acrescentadas):\n await message.channel.send(random.choice(options2))\n \n if message.content.startswith(\"$new_palavras \"):\n encouraging_message2 = message.content.split(\"$new_palavras \",1)[1]\n uptade_encouragements2(encouraging_message2)\n await message.channel.send(\"New encouraging message added.\")\n\n if message.content.startswith(\"$del_palavras\"):\n encouragements2 = []\n if \"encouragements2\" in db.keys():\n index2 = int(message.content.split(\"$del_palavras\",1)[1])\n delete_encouragements2(index2)\n encouragements2 = db[\"encouragements2\"]\n await message.channel.send(encouragements2)\n ############################################################################# \n\n if message.content == \"duliano\":\n await message.channel.send(\"escreva no particular para o seu amigo\")\n dm = await message.author.create_dm() # Creates a dm channel with the user\n await dm.send(\"o que você quer falar com duliano?\") # Sends the user the message\n\n if any (palavra in message.content for palavra in diferente):\n await message.channel.send(random.choice(r_diferente))\n\n if any (palavra in message.content + (\" \") for palavra in palavrax):\n await message.channel.send(random.choice(r_palavrax))\n\n if any (palavra in message.content != \"bom\" for palavra in gx):\n await message.channel.send(random.choice(r_gx))\n\n if any (palavra in message.content != \"bom dia\" for palavra in cumprimentosx):\n await message.channel.send(random.choice(r_cumprimentosx))\n\n '''else:\n await message.author.send(\"Ainda não aprendi essa palavra \\nMe ensina?\")'''\n\nclient.run(os.getenv(\"REC\"))\nmy_secret = os.environ['REC']", "repo_name": "cardosodbc/bot_discord_cao_coragem", "sub_path": "_discord.py", "file_name": "_discord.py", "file_ext": "py", "file_size_in_byte": 6354, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "discord.Client", "line_number": 7, "usage_type": "call"}, {"api_name": "replit.db.keys", "line_number": 11, "usage_type": "call"}, {"api_name": "replit.db", "line_number": 11, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 12, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 14, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 16, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 19, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 22, "usage_type": "name"}, {"api_name": "replit.db.keys", "line_number": 26, "usage_type": "call"}, {"api_name": "replit.db", "line_number": 26, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 27, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 29, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 31, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 34, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 37, "usage_type": "name"}, {"api_name": "replit.db.keys", "line_number": 60, "usage_type": "call"}, {"api_name": "replit.db", "line_number": 60, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 61, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 69, "usage_type": "name"}, {"api_name": "replit.db.keys", "line_number": 71, "usage_type": "call"}, {"api_name": "replit.db", "line_number": 71, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 72, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 76, "usage_type": "call"}, {"api_name": "replit.db.keys", "line_number": 85, "usage_type": "call"}, {"api_name": "replit.db", "line_number": 85, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 88, "usage_type": "name"}, {"api_name": "replit.db.keys", "line_number": 93, "usage_type": "call"}, {"api_name": "replit.db", "line_number": 93, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 94, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 101, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 104, "usage_type": "name"}, {"api_name": "replit.db.keys", "line_number": 110, "usage_type": "call"}, {"api_name": "replit.db", "line_number": 110, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 111, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 115, "usage_type": "call"}, {"api_name": "replit.db.keys", "line_number": 124, "usage_type": "call"}, {"api_name": "replit.db", "line_number": 124, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 127, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 137, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 140, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 143, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 146, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 151, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 152, "usage_type": "attribute"}]}
+{"seq_id": "667133840", "text": "import os\nimport logging\nimport tempfile\nfrom pathlib import Path\n\nimport torch.multiprocessing as mp\n\nfrom mindsdb.__about__ import __version__ as mindsdb_version\nfrom mindsdb.interfaces.database.database import DatabaseWrapper\nfrom mindsdb.interfaces.storage.db import session, Predictor\nfrom mindsdb.interfaces.storage.fs import FsSotre\nfrom mindsdb.utilities.config import Config\nfrom mindsdb.utilities.fs import create_process_mark, delete_process_mark\n\n\nctx = mp.get_context('spawn')\n\n\ndef create_learn_mark():\n if os.name == 'posix':\n p = Path(tempfile.gettempdir()).joinpath('mindsdb/learn_processes/')\n p.mkdir(parents=True, exist_ok=True)\n p.joinpath(f'{os.getpid()}').touch()\n\n\ndef delete_learn_mark():\n if os.name == 'posix':\n p = Path(tempfile.gettempdir()).joinpath('mindsdb/learn_processes/').joinpath(f'{os.getpid()}')\n if p.exists():\n p.unlink()\n\n\ndef run_learn(name, db_name, from_data, to_predict, kwargs, datasource_id, company_id):\n import mindsdb_native\n import mindsdb_datasources\n import mindsdb\n import torch\n import gc\n\n if 'join_learn_process' in kwargs:\n del kwargs['join_learn_process']\n\n create_process_mark('learn')\n\n config = Config()\n fs_store = FsSotre()\n mdb = mindsdb_native.Predictor(name=name, run_env={'trigger': 'mindsdb'})\n\n predictor_record = Predictor.query.filter_by(company_id=company_id, name=db_name).first()\n predictor_record.datasource_id = datasource_id\n predictor_record.to_predict = to_predict\n predictor_record.native_version = mindsdb_native.__version__\n predictor_record.mindsdb_version = mindsdb_version\n predictor_record.learn_args = {\n 'to_predict': to_predict,\n 'kwargs': kwargs\n }\n predictor_record.data = {\n 'name': db_name,\n 'status': 'training'\n }\n session.commit()\n\n to_predict = to_predict if isinstance(to_predict, list) else [to_predict]\n data_source = getattr(mindsdb_datasources, from_data['class'])(*from_data['args'], **from_data['kwargs'])\n try:\n mdb.learn(\n from_data=data_source,\n to_predict=to_predict,\n **kwargs\n )\n\n except Exception as e:\n log = logging.getLogger('mindsdb.main')\n log.error(f'Predictor learn error: {e}')\n predictor_record.data = {\n 'name': db_name,\n 'status': 'error'\n }\n session.commit()\n delete_process_mark('learn')\n\n fs_store.put(name, f'predictor_{company_id}_{predictor_record.id}', config['paths']['predictors'])\n\n model_data = mindsdb_native.F.get_model_data(name)\n\n try:\n torch.cuda.empty_cache()\n except Exception as e:\n pass\n gc.collect()\n\n predictor_record = Predictor.query.filter_by(company_id=company_id, name=db_name).first()\n predictor_record.data = model_data\n session.commit()\n\n model_data['name'] = db_name\n DatabaseWrapper(company_id).register_predictors([model_data])\n delete_process_mark('learn')\n\n\nclass LearnProcess(ctx.Process):\n daemon = True\n\n def __init__(self, *args):\n super(LearnProcess, self).__init__(args=args)\n\n def run(self):\n '''\n running at subprocess due to\n ValueError: signal only works in main thread\n\n this is work for celery worker here?\n '''\n run_learn(*self._args)\n", "repo_name": "sbsreekanth/mindsdb", "sub_path": "mindsdb/interfaces/model/learn_process.py", "file_name": "learn_process.py", "file_ext": "py", "file_size_in_byte": 3382, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "github-code", "pt": "20", "api": [{"api_name": "torch.multiprocessing.get_context", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.multiprocessing", "line_number": 16, "usage_type": "name"}, {"api_name": "os.name", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 21, "usage_type": "call"}, {"api_name": "tempfile.gettempdir", "line_number": 21, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 23, "usage_type": "call"}, {"api_name": "os.name", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 28, "usage_type": "call"}, {"api_name": "tempfile.gettempdir", "line_number": 28, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 28, "usage_type": "call"}, {"api_name": "mindsdb.utilities.fs.create_process_mark", "line_number": 43, "usage_type": "call"}, {"api_name": "mindsdb.utilities.config.Config", "line_number": 45, "usage_type": "call"}, {"api_name": "mindsdb.interfaces.storage.fs.FsSotre", "line_number": 46, "usage_type": "call"}, {"api_name": "mindsdb_native.Predictor", "line_number": 47, "usage_type": "call"}, {"api_name": "mindsdb.interfaces.storage.db.Predictor.query.filter_by", "line_number": 49, "usage_type": "call"}, {"api_name": "mindsdb.interfaces.storage.db.Predictor.query", "line_number": 49, "usage_type": "attribute"}, {"api_name": "mindsdb.interfaces.storage.db.Predictor", "line_number": 49, "usage_type": "name"}, {"api_name": "mindsdb_native.__version__", "line_number": 52, "usage_type": "attribute"}, {"api_name": "mindsdb.__about__.__version__", "line_number": 53, "usage_type": "name"}, {"api_name": "mindsdb.interfaces.storage.db.session.commit", "line_number": 62, "usage_type": "call"}, {"api_name": "mindsdb.interfaces.storage.db.session", "line_number": 62, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 74, "usage_type": "call"}, {"api_name": "mindsdb.interfaces.storage.db.session.commit", "line_number": 80, "usage_type": "call"}, {"api_name": "mindsdb.interfaces.storage.db.session", "line_number": 80, "usage_type": "name"}, {"api_name": "mindsdb.utilities.fs.delete_process_mark", "line_number": 81, "usage_type": "call"}, {"api_name": "mindsdb_native.F.get_model_data", "line_number": 85, "usage_type": "call"}, {"api_name": "mindsdb_native.F", "line_number": 85, "usage_type": "attribute"}, {"api_name": "torch.cuda.empty_cache", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 88, "usage_type": "attribute"}, {"api_name": "gc.collect", "line_number": 91, "usage_type": "call"}, {"api_name": "mindsdb.interfaces.storage.db.Predictor.query.filter_by", "line_number": 93, "usage_type": "call"}, {"api_name": "mindsdb.interfaces.storage.db.Predictor.query", "line_number": 93, "usage_type": "attribute"}, {"api_name": "mindsdb.interfaces.storage.db.Predictor", "line_number": 93, "usage_type": "name"}, {"api_name": "mindsdb.interfaces.storage.db.session.commit", "line_number": 95, "usage_type": "call"}, {"api_name": "mindsdb.interfaces.storage.db.session", "line_number": 95, "usage_type": "name"}, {"api_name": "mindsdb.interfaces.database.database.DatabaseWrapper", "line_number": 98, "usage_type": "call"}, {"api_name": "mindsdb.utilities.fs.delete_process_mark", "line_number": 99, "usage_type": "call"}]}
+{"seq_id": "30495040041", "text": "\"\"\"\nTesting sales_inserter plus the DB Commands (Which is imported into sales_inserter\n\"\"\"\n\nfrom SQL import sales_inserter as si\nfrom datetime import date\n\nii = si.InsertItem()\n# print('Asserting if _id_lookup correctly returns an id.')\n# assert ii._id_lookup('patients', 'patient_name', 'Angeles Pollard') == 1332\n# ii.db.connect()\n# ii.quick_sale(\"1500\", None, 'VSP', [1, 5, 6])\n# ii.db.commit_close()\n# ii.insert_sale('Newton Powers', purchase_items=[('Eye Exam', 8500), ('Refraction', 4900)]) # Works!\n\n# ii.db.delete('sale', ('patient', 1332))\n\n# ii.db.update(['products', ('id', 1), ('id', 29)])\n\n#\n# sales_items = ii.db.view('sale_item')\n# print(sales_items)\n\n# patient = ii.db.view('patients', ('patient_name', 'Angeles Pollard'))\n# print(patient)\n\n# print(ii.db.view('sale', ('purchase_time::date', '2019-02-23')))\n#\n# ii.db._connect()\n# for i in range(2100):\n# ii.db.update_avg_dollar(i, slow=False) # Works!\n#\n# ii.db._commit_close()\n\nif __name__ == '__main__':\n products = ii.db.view('products')\n for item in products:\n print(f\"({item[0]}) {item[1]} - ${(item[2] / 100)}\")\n\n sales = ii.db.view('sale')\n print(sales)\n\n sale_time = date(2014, 2, 20)\n patient_id = 102\n sale_id = ii.db.view(f\"sale WHERE purchase_time = '{sale_time}' AND patient = {patient_id}\", field=\"id\")\n print(sale_id)\n\n", "repo_name": "KarlKeisel/Eyewear-Saleswebsite", "sub_path": "tests/test_sales_inserter.py", "file_name": "test_sales_inserter.py", "file_ext": "py", "file_size_in_byte": 1341, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "SQL.sales_inserter.InsertItem", "line_number": 8, "usage_type": "call"}, {"api_name": "SQL.sales_inserter", "line_number": 8, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 43, "usage_type": "call"}]}
+{"seq_id": "24169378647", "text": "import win32api, win32gui, win32con\r\nimport win32com.client as wclt\r\nimport win32clipboard as clipbd\r\nimport time, re, datetime\r\nfrom tkinter import Label, Button, Tk, Entry, Frame, StringVar, IntVar, DoubleVar, Text, Spinbox, Checkbutton, BooleanVar, messagebox, HORIZONTAL\r\nfrom tkinter.constants import END\r\nimport threading as thd\r\nfrom pythoncom import CoInitialize\r\n\r\nclass GUI_MAIN_APP(Frame):\r\n def __init__(self, master = None) -> None:\r\n super().__init__(master)\r\n self.master = master\r\n self.pack()\r\n self.create_widget()\r\n\r\n def w_size(self, content: str) -> int:\r\n return len(content) * 12\r\n\r\n def create_widget(self) -> None:\r\n self.place(x = 0, y = 0, width=450, height=700)\r\n note = '''说明: 在输入框种输入要传输的内容选择文件复选框后,\r\n 会在倒计时之后直接发送剪贴板中的内容, 默认的发送日期为当天日期,\r\n 在日期输入框中输入日期格式例如2022-01-01, \r\n 时间格式例如13:45:23,7:23:05.'''\r\n self.note_label = Label(self, text = note, justify = 'left').place(x = 10, y = 0, width = 400, height = 65)\r\n\r\n l1_ct = '窗口名称:'\r\n Label(self, text=l1_ct).place(x = 10, y = 70, width = self.w_size(l1_ct))\r\n \r\n wl_ct = '文件传输助手'\r\n self.wlct = StringVar()\r\n self.wlct.set(wl_ct)\r\n self.window_name_entry = Entry(self, textvariable = self.wlct)\r\n self.window_name_entry.place(x = self.w_size(l1_ct) + 10, y = 70, width = 196)\r\n \r\n l2_ct = '要发送的内容: '\r\n Label(self, text = l2_ct).place(x = 10, y = 100, width = self.w_size(l2_ct) - 10)\r\n self.content_text = Text(self)\r\n self.content_text.place(x = 10, y = 130, width = 400, height = 300)\r\n\r\n l3_ct = '循环次数:'\r\n self.lts_ct = IntVar()\r\n Label(self, text = l3_ct).place(x = 10, y = 435, width = self.w_size(l3_ct))\r\n self.loop_times_spinbox = Spinbox(self, from_ = 1, to = 100000, textvariable = self.lts_ct)\r\n self.loop_times_spinbox.place(x = self.w_size(l3_ct) + 10, y = 435)\r\n\r\n l4_ct = '循环时间间隔(秒):'\r\n Label(self, text = l4_ct).place(x = 10, y = 465, width = self.w_size(l4_ct) - 15)\r\n\r\n self.lti_ct = DoubleVar()\r\n self.lti_ct.set(1)\r\n self.loop_time_interval_entry = Entry(self, textvariable = self.lti_ct)\r\n self.loop_time_interval_entry.place(x = self.w_size(l4_ct) + 10, y = 465, width = 60)\r\n\r\n l5_ct = '日期:'\r\n Label(self, text = l5_ct).place(x = 10, y = 495, width = self.w_size(l5_ct))\r\n\r\n dt = time.strftime('%Y-%m-%d', time.localtime())\r\n self.dt_ct = StringVar()\r\n self.dt_ct.set(dt)\r\n self.send_date_entry = Entry(self, textvariable = self.dt_ct)\r\n self.send_date_entry.place(x = self.w_size(l5_ct) + 10, y = 495, width = self.w_size(dt))\r\n\r\n l6_ct = '时间:'\r\n Label(self, text = l6_ct).place(x = self.w_size(dt) + self.w_size(l5_ct) + 20,\r\n y = 495, width = self.w_size(l6_ct))\r\n\r\n tt = time.strftime('%H:%M:%S', time.localtime())\r\n self.tt_ct = StringVar()\r\n self.tt_ct.set(tt)\r\n self.send_time_entry = Entry(self, textvariable = self.tt_ct)\r\n self.send_time_entry.place(x = self.w_size(dt) + self.w_size(l5_ct) + self.w_size(l6_ct) + 30,\r\n y = 495, width = self.w_size(tt))\r\n\r\n l7_ct = '是否为文件(勾选后为True, 不勾选为False)'\r\n Label(self, text = l7_ct).place(x = 10, y = 525, width = round(self.w_size(l7_ct) * 0.72))\r\n\r\n self.fcbv = BooleanVar()\r\n self.file_check_box = Checkbutton(self, command = self.file_check_box_action)\r\n self.file_check_box.place(x = round(self.w_size(l7_ct) * 0.72) + 10, y = 525)\r\n\r\n self.indicator_content = StringVar()\r\n Label(self, textvariable = self.indicator_content).place(x = 150, y = 580, width = 120)\r\n \r\n self.send_button = Button(self, text = '发送', command = self.send_button_action)\r\n self.send_button.place(x = 155, y = 600, width = 100, height = 50)\r\n\r\n def file_check_box_action(self) -> None:\r\n if not self.fcbv.get():\r\n self.fcbv.set(not self.fcbv.get())\r\n else:\r\n self.fcbv.set(not self.fcbv.get())\r\n\r\n def send_button_action(self) -> None:\r\n send_date = self.dt_ct.get()\r\n send_time = self.tt_ct.get()\r\n rgx = re.compile(r'\\d{4}\\-\\d{2}-\\d{2} \\d{1,2}:\\d{2}:\\d{2}')\r\n input_date_time = f'{send_date} {send_time}'\r\n if rgx.match(input_date_time):\r\n time_left = get_date_time_sub(send_time, send_date)\r\n if thd.active_count() < 2:\r\n loop_times = self.lts_ct.get()\r\n loop_time_interval = self.lti_ct.get()\r\n file_flag = self.fcbv.get()\r\n window_name = self.wlct.get()\r\n send_content = self.content_text.get(1.0, END)\r\n self.indicator_content.set(f'已经过:{0}%')\r\n msg = {'window_name': window_name, 'send_content': send_content,\r\n 'loop_times': loop_times, 'loop_time_interval': loop_time_interval,\r\n 'file_flag': file_flag, 'send_date': send_date, 'send_time': send_time,\r\n 'input_date_time': input_date_time}\r\n mt = thd.Thread(target = self.indicator_increse, args = (time_left, msg))\r\n mt.start()\r\n else:\r\n messagebox.showerror('时间日期错误', '请重新输入时间或日期再次尝试')\r\n\r\n def indicator_increse(self, time_left: int, msg: dict) -> None:\r\n for i in range(time_left + 1):\r\n self.indicator_content.set(f'已经过:{i / time_left * 100:.2f}%')\r\n time.sleep(1)\r\n self.update()\r\n messagebox.showinfo('发送提醒', '正在执行发送任务...')\r\n loop_execute(msg.get('window_name'), msg.get('send_content'), msg.get('loop_time_interval'),\r\n msg.get('loop_times'), msg.get('file_flag'))\r\n messagebox.showinfo('任务提醒', '发送任务已完成。')\r\n\r\ndef get_window_handler(wname: str) -> int:\r\n win_ha = win32gui.FindWindow('ChatWnd', wname)\r\n win32gui.BringWindowToTop(win_ha)\r\n CoInitialize()\r\n sl = wclt.Dispatch('WScript.Shell')\r\n sl.SendKeys('%')\r\n win32gui.SetForegroundWindow(win_ha)\r\n return win_ha\r\n\r\ndef message_send(win_ha: int) -> None:\r\n win32api.keybd_event(17, 0, 0, 0)\r\n time.sleep(0.1)\r\n win32gui.SendMessage(win_ha, win32con.WM_KEYDOWN, 86, 0)\r\n time.sleep(0.1)\r\n win32gui.SendMessage(win_ha, win32con.WM_KEYDOWN, win32con.VK_RETURN, 0)\r\n win32api.keybd_event(17, 0, win32con.KEYEVENTF_KEYUP, 0)\r\n\r\ndef content_copy_to_clipboard(text):\r\n clipbd.OpenClipboard()\r\n clipbd.EmptyClipboard()\r\n clipbd.SetClipboardText(text)\r\n clipbd.CloseClipboard()\r\n\r\ndef loop_execute(window_name: str,\r\n sending_text: str,\r\n time_interval: float,\r\n loop_times: int,\r\n file_flg: bool) -> None:\r\n \r\n for _ in range(1, loop_times + 1):\r\n time.sleep(time_interval)\r\n if not file_flg:\r\n content_copy_to_clipboard(sending_text)\r\n w_ha = get_window_handler(window_name)\r\n message_send(w_ha)\r\n\r\ndef get_file_flg() -> bool:\r\n cont = input('If file send else type here: ').strip()\r\n if cont == 'True':\r\n return True\r\n else:\r\n return False\r\n\r\ndef get_date_time_sub(send_time: str, send_date: str = datetime.datetime.today().date().strftime('%H:%M:%S')) -> int:\r\n ti = time.localtime()\r\n fdt = f'{send_date} {send_time}'\r\n td = time.strptime(fdt, '%Y-%m-%d %H:%M:%S')\r\n tx = datetime.datetime(ti.tm_year, ti.tm_mon, ti.tm_mday, ti.tm_hour, ti.tm_min, ti.tm_sec)\r\n ty = datetime.datetime(td.tm_year, td.tm_mon, td.tm_mday, td.tm_hour, td.tm_min, td.tm_sec)\r\n return (ty - tx).days * 24 * 3600 + (ty - tx).seconds\r\n\r\nif __name__ == '__main__':\r\n root = Tk()\r\n root.title('微信轰炸工具 Ver GM 1.0.1')\r\n root.geometry('420x700+200+50')\r\n app = GUI_MAIN_APP(root)\r\n app.mainloop()\r\n", "repo_name": "ItiharaYuuko/wechat_loop", "sub_path": "wechat_loop.pyw", "file_name": "wechat_loop.pyw", "file_ext": "pyw", "file_size_in_byte": 8314, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "tkinter.Frame", "line_number": 10, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 26, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 29, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 32, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 34, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 38, "usage_type": "call"}, {"api_name": "tkinter.Text", "line_number": 39, "usage_type": "call"}, {"api_name": "tkinter.IntVar", "line_number": 43, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 44, "usage_type": "call"}, {"api_name": "tkinter.Spinbox", "line_number": 45, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 49, "usage_type": "call"}, {"api_name": "tkinter.DoubleVar", "line_number": 51, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 53, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 57, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 59, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 59, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 60, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 62, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 66, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 69, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 69, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 70, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 72, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 77, "usage_type": "call"}, {"api_name": "tkinter.BooleanVar", "line_number": 79, "usage_type": "call"}, {"api_name": "tkinter.Checkbutton", "line_number": 80, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 83, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 84, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 86, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 98, "usage_type": "call"}, {"api_name": "threading.active_count", "line_number": 102, "usage_type": "call"}, {"api_name": "tkinter.constants.END", "line_number": 107, "usage_type": "argument"}, {"api_name": "threading.Thread", "line_number": 113, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 116, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 116, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 121, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 123, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 123, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 126, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 126, "usage_type": "name"}, {"api_name": "win32gui.FindWindow", "line_number": 129, "usage_type": "call"}, {"api_name": "win32gui.BringWindowToTop", "line_number": 130, "usage_type": "call"}, {"api_name": "pythoncom.CoInitialize", "line_number": 131, "usage_type": "call"}, {"api_name": "win32com.client.Dispatch", "line_number": 132, "usage_type": "call"}, {"api_name": "win32com.client", "line_number": 132, "usage_type": "name"}, {"api_name": "win32gui.SetForegroundWindow", "line_number": 134, "usage_type": "call"}, {"api_name": "win32api.keybd_event", "line_number": 138, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 139, "usage_type": "call"}, {"api_name": "win32gui.SendMessage", "line_number": 140, "usage_type": "call"}, {"api_name": "win32con.WM_KEYDOWN", "line_number": 140, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 141, "usage_type": "call"}, {"api_name": "win32gui.SendMessage", "line_number": 142, "usage_type": "call"}, {"api_name": "win32con.WM_KEYDOWN", "line_number": 142, "usage_type": "attribute"}, {"api_name": "win32con.VK_RETURN", "line_number": 142, "usage_type": "attribute"}, {"api_name": "win32api.keybd_event", "line_number": 143, "usage_type": "call"}, {"api_name": "win32con.KEYEVENTF_KEYUP", "line_number": 143, "usage_type": "attribute"}, {"api_name": "win32clipboard.OpenClipboard", "line_number": 146, "usage_type": "call"}, {"api_name": "win32clipboard.EmptyClipboard", "line_number": 147, "usage_type": "call"}, {"api_name": "win32clipboard.SetClipboardText", "line_number": 148, "usage_type": "call"}, {"api_name": "win32clipboard.CloseClipboard", "line_number": 149, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 158, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 171, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 171, "usage_type": "attribute"}, {"api_name": "time.localtime", "line_number": 172, "usage_type": "call"}, {"api_name": "time.strptime", "line_number": 174, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 175, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 176, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 180, "usage_type": "call"}]}
+{"seq_id": "3112870991", "text": "import sys\nimport pygame as p\nfrom config import *\nfrom utils import *\nfrom Sorting import *\n\n\ndef draw_sorted(window, algo, count):\n\ttry:\n\t\t# Get current list and the numbers compared this turn\n\t\tsorted_list, i = next(count)\n\t\tn_num = len(sorted_list)\n\n\t\tdraw_bg(window, algo)\n\n\t\tfor j, v in enumerate(sorted_list):\n\t\t\trect = p.Rect(*create_rect(j, v, n_num))\n\t\t\tif j <= i:\n\t\t\t\tp.draw.rect(window, PASTEL_GREEN, rect)\n\t\t\telse:\n\t\t\t\tp.draw.rect(window, WHITE, rect)\n\n\t\tp.display.flip()\n\n\texcept StopIteration:\n\t\treturn True\t\n\n\ndef draw_sorting(window, curr_list, v1, v2):\n\tn_num = len(curr_list)\n\tfor i, v in enumerate(curr_list):\n\t\trect = p.Rect(*create_rect(i, v, n_num))\n\t\tif v == v1:\n\t\t\tp.draw.rect(window, PASTEL_PINK, rect)\n\t\telif v == v2:\n\t\t\tp.draw.rect(window, PASTEL_BLUE, rect)\n\t\telse:\n\t\t\tp.draw.rect(window, WHITE, rect)\n\n\ndef draw_bg(window, algo):\n\twindow.fill(BACKGROUND)\n\tcomicsans = p.font.SysFont('Helvetica', 20)\n\tname_surface = comicsans.render(algo, False, WHITE)\n\twindow.blit(name_surface, (20, 20))\n\ndef draw(window, algo, l, n):\n\ttry:\n\t\t# Get current list and the numbers compared this turn\n\t\tcurr_list, v1, v2 = next(l)\n\t\tdraw_bg(window, algo)\n\t\tdraw_sorting(window, curr_list, v1, v2)\n\n\t\tp.display.flip()\n\n\texcept StopIteration:\n\t\treturn True\n\n\ndef main():\n\t# Initialize pygame\n\tp.init()\n\tp.font.init()\n\n\t# Create pygame window\n\twindow = p.display.set_mode((WINDOW_WIDTH, WINDOW_HEIGHT))\n\tclock = p.time.Clock()\n\n\tsorting_algorithms = {\n\t\t'Bubble Sort': bubble_sort,\n\t\t'Selection Sort': selection_sort,\n\t\t'Insertion Sort': insertion_sort,\n\t}\n\n\t# Initialize parameters\n\tn = 30\n\tascending = False\n\tlst = generate_list(n)\n\tcount = counter(sorted(lst, reverse = not ascending))\n\n\talgo = 'Selection Sort'\n\tl = sorting_algorithms[algo](lst, ascending = ascending)\n\n\t# Initialize flags\n\trunning = True\n\treset = False\n\tfinished = False\n\n\twhile running:\n\t\tclock.tick(FPS)\n\n\t\tif reset:\n\t\t\tlst = generate_list(n)\n\t\t\tl = bubble_sort(lst)\n\n\t\tfor e in p.event.get():\n\t\t\tif e.type == p.QUIT:\n\t\t\t\trunning = False\n\t\t\t\tp.quit()\n\t\t\t\tsys.exit()\n\n\t\tif not finished and draw(window, algo, l, n):\n\t\t\tfinished = True\n\n\t\tif finished and draw_sorted(window, algo, count):\n\t\t\tp.time.wait(2000)\n\t\t\trunning = False\n\t\t\tp.quit()\n\nif __name__ == '__main__':\n\tmain()", "repo_name": "kaiwinut/pygame-sorting-visualizer", "sub_path": "src/sortviz/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2257, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "pygame.Rect", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 21, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 34, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 36, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 38, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 43, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 54, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 62, "usage_type": "call"}, {"api_name": "pygame.font.init", "line_number": 63, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 63, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 66, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 66, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 67, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 67, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 96, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 96, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 97, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 99, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 100, "usage_type": "call"}, {"api_name": "pygame.time.wait", "line_number": 106, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 106, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 108, "usage_type": "call"}]}
+{"seq_id": "21911083259", "text": "# coding: utf-8\n\nimport six\n\nfrom huaweicloudsdkcore.utils.http_utils import sanitize_for_serialization\n\n\nclass AgreeTenantAuthorizationV2Req:\n\n \"\"\"\n Attributes:\n openapi_types (dict): The key is attribute name\n and the value is attribute type.\n attribute_map (dict): The key is attribute name\n and the value is json key in definition.\n \"\"\"\n sensitive_list = []\n\n openapi_types = {\n 'auth_detail_list': 'list[TenantAgreeAuthDetailV2]',\n 'auth_effective_time': 'int',\n 'auth_expire_time': 'int',\n 'group_id': 'str',\n 'agency_id': 'str'\n }\n\n attribute_map = {\n 'auth_detail_list': 'auth_detail_list',\n 'auth_effective_time': 'auth_effective_time',\n 'auth_expire_time': 'auth_expire_time',\n 'group_id': 'group_id',\n 'agency_id': 'agency_id'\n }\n\n def __init__(self, auth_detail_list=None, auth_effective_time=None, auth_expire_time=None, group_id=None, agency_id=None):\n \"\"\"AgreeTenantAuthorizationV2Req\n\n The model defined in huaweicloud sdk\n\n :param auth_detail_list: 授权详情列表\n :type auth_detail_list: list[:class:`huaweicloudsdkosm.v2.TenantAgreeAuthDetailV2`]\n :param auth_effective_time: 授权生效时间\n :type auth_effective_time: int\n :param auth_expire_time: 授权到期时间\n :type auth_expire_time: int\n :param group_id: 组id\n :type group_id: str\n :param agency_id: 委托id\n :type agency_id: str\n \"\"\"\n \n \n\n self._auth_detail_list = None\n self._auth_effective_time = None\n self._auth_expire_time = None\n self._group_id = None\n self._agency_id = None\n self.discriminator = None\n\n if auth_detail_list is not None:\n self.auth_detail_list = auth_detail_list\n if auth_effective_time is not None:\n self.auth_effective_time = auth_effective_time\n if auth_expire_time is not None:\n self.auth_expire_time = auth_expire_time\n if group_id is not None:\n self.group_id = group_id\n if agency_id is not None:\n self.agency_id = agency_id\n\n @property\n def auth_detail_list(self):\n \"\"\"Gets the auth_detail_list of this AgreeTenantAuthorizationV2Req.\n\n 授权详情列表\n\n :return: The auth_detail_list of this AgreeTenantAuthorizationV2Req.\n :rtype: list[:class:`huaweicloudsdkosm.v2.TenantAgreeAuthDetailV2`]\n \"\"\"\n return self._auth_detail_list\n\n @auth_detail_list.setter\n def auth_detail_list(self, auth_detail_list):\n \"\"\"Sets the auth_detail_list of this AgreeTenantAuthorizationV2Req.\n\n 授权详情���表\n\n :param auth_detail_list: The auth_detail_list of this AgreeTenantAuthorizationV2Req.\n :type auth_detail_list: list[:class:`huaweicloudsdkosm.v2.TenantAgreeAuthDetailV2`]\n \"\"\"\n self._auth_detail_list = auth_detail_list\n\n @property\n def auth_effective_time(self):\n \"\"\"Gets the auth_effective_time of this AgreeTenantAuthorizationV2Req.\n\n 授权生效时间\n\n :return: The auth_effective_time of this AgreeTenantAuthorizationV2Req.\n :rtype: int\n \"\"\"\n return self._auth_effective_time\n\n @auth_effective_time.setter\n def auth_effective_time(self, auth_effective_time):\n \"\"\"Sets the auth_effective_time of this AgreeTenantAuthorizationV2Req.\n\n 授权生效时间\n\n :param auth_effective_time: The auth_effective_time of this AgreeTenantAuthorizationV2Req.\n :type auth_effective_time: int\n \"\"\"\n self._auth_effective_time = auth_effective_time\n\n @property\n def auth_expire_time(self):\n \"\"\"Gets the auth_expire_time of this AgreeTenantAuthorizationV2Req.\n\n 授权到期时间\n\n :return: The auth_expire_time of this AgreeTenantAuthorizationV2Req.\n :rtype: int\n \"\"\"\n return self._auth_expire_time\n\n @auth_expire_time.setter\n def auth_expire_time(self, auth_expire_time):\n \"\"\"Sets the auth_expire_time of this AgreeTenantAuthorizationV2Req.\n\n 授权到期时间\n\n :param auth_expire_time: The auth_expire_time of this AgreeTenantAuthorizationV2Req.\n :type auth_expire_time: int\n \"\"\"\n self._auth_expire_time = auth_expire_time\n\n @property\n def group_id(self):\n \"\"\"Gets the group_id of this AgreeTenantAuthorizationV2Req.\n\n 组id\n\n :return: The group_id of this AgreeTenantAuthorizationV2Req.\n :rtype: str\n \"\"\"\n return self._group_id\n\n @group_id.setter\n def group_id(self, group_id):\n \"\"\"Sets the group_id of this AgreeTenantAuthorizationV2Req.\n\n 组id\n\n :param group_id: The group_id of this AgreeTenantAuthorizationV2Req.\n :type group_id: str\n \"\"\"\n self._group_id = group_id\n\n @property\n def agency_id(self):\n \"\"\"Gets the agency_id of this AgreeTenantAuthorizationV2Req.\n\n 委托id\n\n :return: The agency_id of this AgreeTenantAuthorizationV2Req.\n :rtype: str\n \"\"\"\n return self._agency_id\n\n @agency_id.setter\n def agency_id(self, agency_id):\n \"\"\"Sets the agency_id of this AgreeTenantAuthorizationV2Req.\n\n 委托id\n\n :param agency_id: The agency_id of this AgreeTenantAuthorizationV2Req.\n :type agency_id: str\n \"\"\"\n self._agency_id = agency_id\n\n def to_dict(self):\n \"\"\"Returns the model properties as a dict\"\"\"\n result = {}\n\n for attr, _ in six.iteritems(self.openapi_types):\n value = getattr(self, attr)\n if isinstance(value, list):\n result[attr] = list(map(\n lambda x: x.to_dict() if hasattr(x, \"to_dict\") else x,\n value\n ))\n elif hasattr(value, \"to_dict\"):\n result[attr] = value.to_dict()\n elif isinstance(value, dict):\n result[attr] = dict(map(\n lambda item: (item[0], item[1].to_dict())\n if hasattr(item[1], \"to_dict\") else item,\n value.items()\n ))\n else:\n if attr in self.sensitive_list:\n result[attr] = \"****\"\n else:\n result[attr] = value\n\n return result\n\n def to_str(self):\n \"\"\"Returns the string representation of the model\"\"\"\n import simplejson as json\n if six.PY2:\n import sys\n reload(sys)\n sys.setdefaultencoding(\"utf-8\")\n return json.dumps(sanitize_for_serialization(self), ensure_ascii=False)\n\n def __repr__(self):\n \"\"\"For `print`\"\"\"\n return self.to_str()\n\n def __eq__(self, other):\n \"\"\"Returns true if both objects are equal\"\"\"\n if not isinstance(other, AgreeTenantAuthorizationV2Req):\n return False\n\n return self.__dict__ == other.__dict__\n\n def __ne__(self, other):\n \"\"\"Returns true if both objects are not equal\"\"\"\n return not self == other\n", "repo_name": "huaweicloud/huaweicloud-sdk-python-v3", "sub_path": "huaweicloud-sdk-osm/huaweicloudsdkosm/v2/model/agree_tenant_authorization_v2_req.py", "file_name": "agree_tenant_authorization_v2_req.py", "file_ext": "py", "file_size_in_byte": 7199, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 104, "dataset": "github-code", "pt": "20", "api": [{"api_name": "six.iteritems", "line_number": 186, "usage_type": "call"}, {"api_name": "six.PY2", "line_number": 212, "usage_type": "attribute"}, {"api_name": "sys.setdefaultencoding", "line_number": 215, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 216, "usage_type": "call"}, {"api_name": "huaweicloudsdkcore.utils.http_utils.sanitize_for_serialization", "line_number": 216, "usage_type": "call"}]}
+{"seq_id": "33150573155", "text": "import random\r\nimport sys\r\nimport json\r\nfrom musicInfo import *\r\n\r\nclass NGramModel(object):\r\n\r\n def __init__(self):\r\n \"\"\"\r\n Requires: nothing\r\n Modifies: self (this instance of the NGramModel object)\r\n Effects: This is the NGramModel constructor. It sets up an empty\r\n dictionary as a member variable. It is called from the\r\n constructors of the NGramModel child classes. This\r\n function is done for you.\r\n \"\"\"\r\n self.nGramCounts = {}\r\n\r\n def __str__(self):\r\n \"\"\"\r\n Requires: nothing\r\n Modifies: nothing\r\n Effects: Returns the string to print when you call print on an\r\n NGramModel object. This string will be formatted in JSON\r\n and display the currently trained dataset.\r\n This function is done for you.\r\n \"\"\"\r\n return self.__class__.__name__ + ':\\n' +\\\r\n json.dumps(\r\n self.nGramCounts,\r\n sort_keys=True,\r\n indent=4,\r\n separators=(',', ': ')\r\n )\r\n\r\n def prepData(self, text):\r\n \"\"\"\r\n Requires: text is a list of lists of strings\r\n Modifies: nothing\r\n Effects: returns a copy of text where each inner list starts with\r\n the symbols '^::^' and '^:::^', and ends with the symbol\r\n '$:::$'. For example, if an inner list in text were\r\n ['hello', 'goodbye'], that list would become\r\n ['^::^', '^:::^', 'hello', 'goodbye', '$:::$'] in the\r\n returned copy.\r\n \"\"\"\r\n textCopy = []\r\n for line in text:\r\n textCopy.append(['^::^', '^:::^'] + line + ['$:::$'])\r\n return textCopy\r\n\r\n def trainModel(self, text):\r\n \"\"\"\r\n Requires: text is a list of lists of strings\r\n Modifies: self.nGramCounts\r\n Effects: this function populates the self.nGramCounts dictionary.\r\n It does not need to be modified here because you will\r\n override it in the NGramModel child classes according\r\n to the spec.\r\n \"\"\"\r\n pass\r\n\r\n def trainingDataHasNGram(self, sentence):\r\n \"\"\"\r\n Requires: sentence is a list of strings, and trainingDataHasNGram\r\n has returned True for this particular language model\r\n Modifies: nothing\r\n Effects: returns a bool indicating whether or not this n-gram model\r\n can be used to choose the next token for the current\r\n sentence. This function does not need to be modified because\r\n you will override it in NGramModel child classes according\r\n to the spec.\r\n \"\"\"\r\n pass\r\n\r\n def getCandidateDictionary(self, sentence):\r\n \"\"\"\r\n Requires: sentence is a list of strings\r\n Modifies: nothing\r\n Effects: returns the dictionary of candidate next words to be added\r\n to the current sentence. This function does not need to be\r\n modified because you will override it in the NGramModel child\r\n classes according to the spec.\r\n \"\"\"\r\n pass\r\n\r\n def weightedChoice(self, candidates):\r\n \"\"\"\r\n Requires: candidates is a dictionary; the keys of candidates are items\r\n you want to choose from and the values are integers\r\n Modifies: nothing\r\n Effects: returns a candidate item (a key in the candidates dictionary)\r\n based on the algorithm described in the spec.\r\n \"\"\"\r\n #create a list of the keys in candidates\r\n words = []\r\n for key in candidates:\r\n words.append(key)\r\n #create a list of the values in candidates\r\n values = []\r\n for key in candidates:\r\n values.append(candidates[key])\r\n #create the list of cumulative values\r\n cumulative = []\r\n count = 0\r\n for number in values:\r\n count += number\r\n cumulative.append(count)\r\n #get a random number between [1, last number in cumulative]\r\n x = random.randrange(0, cumulative[-1])\r\n #as soon as the cumulative value is higher than random number, return key\r\n for i in range(len(words)):\r\n if cumulative[i] > x:\r\n return words[i]\r\n\r\n def getNextToken(self, sentence):\r\n \"\"\"\r\n Requires: sentence is a list of strings, and this model can be used to\r\n choose the next token for the current sentence\r\n Modifies: nothing\r\n Effects: returns the next token to be added to sentence by calling\r\n the getCandidateDictionary and weightedChoice functions.\r\n For more information on how to put all these functions\r\n together, see the spec.\r\n \"\"\"\r\n return self.weightedChoice(self.getCandidateDictionary(sentence))\r\n\r\n def getNextNote(self, musicalSentence, possiblePitches):\r\n \"\"\"\r\n Requires: musicalSentence is a list of PySynth tuples,\r\n possiblePitches is a list of possible pitches for this\r\n line of music (in other words, a key signature), and this\r\n model can be used to choose the next note for the current\r\n musical sentence\r\n Modifies: nothing\r\n Effects: returns the next note to be added to the \"musical sentence\".\r\n For details on how to do this and how this will differ\r\n from getNextToken, see the spec.\r\n \"\"\"\r\n #makes dict consisting of possible notes\r\n allCandidates = self.getCandidateDictionary(musicalSentence)\r\n #makes new dict consisting of possible notes that are also in the key signature\r\n constrainedCandidates = {}\r\n for note in allCandidates:\r\n if note[0][: -1] in possiblePitches or note == '$:::$':\r\n constrainedCandidates[note] = allCandidates[note]\r\n if constrainedCandidates != {}:\r\n return self.weightedChoice(constrainedCandidates)\r\n else:\r\n return (random.choice(possiblePitches) + '4', random.choice(NOTE_DURATIONS))\r\n\r\n###############################################################################\r\n# Main\r\n###############################################################################\r\n\r\nif __name__ == '__main__':\r\n # Add your tests here\r\n text = [ ['the', 'quick', 'brown', 'fox'], ['the', 'lazy', 'dog'] ]\r\n choices = { 'the': 2, 'quick': 1, 'brown': 1 }\r\n mod = NGramModel()\r\n print(mod)\r\n dict = {'the' : 2, 'birthday' : 5, 'python' : 1, 'jessica' : 10, 'aaron' : 9, 'brad' : 48, 'seven' : 6}\r\n sentence = ['happy', 'birthday']\r\n mod.weightedChoice(dict)\r\n", "repo_name": "aawill/Creative-AI-Music-Generator", "sub_path": "models/nGramModel.py", "file_name": "nGramModel.py", "file_ext": "py", "file_size_in_byte": 6912, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "json.dumps", "line_number": 29, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 110, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 150, "usage_type": "call"}]}
+{"seq_id": "25062275689", "text": "import db\nimport cv2\nimport face\nimport frame\nimport yaml\nimport typedef\nimport utils\nimport numpy as np\nimport copy\nimport pickle\n\n\ndef main(config):\n ss = config['source_scale']\n frame_drawer = frame.drawer.Drawer()\n capturer = cv2.VideoCapture(config['source'])\n face_detector = face.detectors.get(**config['face_detector'])\n face_validators = face.validators.get_list(config['face_validators'])\n frame_filters = frame.filters.get_list(config['frame_filters'])\n face_buffer = face.buffer.FaceBuffer(config['face_buffer_size'])\n face_encoder = face.encoders.get(**config['face_encoder'])\n face_recognizer = face.recognizers.get(face_encoder, **config['face_recognizer'])\n storage = db.initialize(**config['database'])\n face_trackers = []\n\n try:\n with open('storage.pkl', 'rb') as db_file:\n storage = pickle.load(db_file)\n print(\"Database was loaded from file.\")\n print(f\"Number of persons in DB: {len(storage.get_face_ids())}\")\n except:\n print(\"No databese to load.\")\n\n while True:\n read_ok, image = capturer.read()\n if not read_ok:\n cv2.imshow(\"Frame\", typedef.NO_VIDEO_FRAME)\n key = cv2.waitKey(1) & 0xFF\n if key == ord(\"q\"):\n capturer.release()\n break\n continue\n image_copy = image.copy()\n\n image = cv2.resize(image, None, fx=ss, fy=ss, interpolation=cv2.INTER_CUBIC)\n image = frame.filters.apply(frame_filters, image)\n face_boxes = face_detector(image)\n face_boxes = face.validators.apply(face_validators, image, face_boxes)\n\n face_ids, face_boxes = face.trackers.apply(face_trackers, image, face_boxes)\n face_trackers = face.trackers.drop_wasted(face_trackers)\n _face_boxes = copy.deepcopy(face_boxes)\n face_boxes = face.validators.apply(face_validators, image, face_boxes)\n _face_ids = []\n for face_id, _face_box in zip(face_ids, _face_boxes):\n for face_box in face_boxes:\n if _face_box == face_box:\n _face_ids.append(face_id)\n\n for face_id, face_box in zip(_face_ids, face_boxes):\n\n face_image = utils.crop(image, *face_box)\n face_image = cv2.resize(face_image, tuple(config['face_shape']),\n interpolation=cv2.INTER_AREA)\n\n if face_id == typedef.UNKNOWN_FACE_ID:\n face_id = utils.generate_tmp_face_id()\n tracker = face.trackers.get(**config['face_tracker'])\n tracker.init(image, face_box, face_id)\n face_trackers.append(tracker)\n\n if utils.is_tmp_id(face_id):\n face_buffer.update(face_id, face_image)\n if face_buffer.is_full(face_id):\n mean_face = face_buffer.get_mean_face(face_id)\n recognized_ok, rec_face_id = face_recognizer(mean_face, storage)\n\n tracked_face_id = face_id\n if not recognized_ok:\n encoded_mean_face = face_encoder(mean_face)\n face_id = storage.generate_face_id()\n storage.add(face_id, encoded_mean_face)\n else:\n face_id = rec_face_id\n face.trackers.update_face_ids(face_trackers, [tracked_face_id], [face_id])\n\n face_box = tuple(np.int64(np.array(face_box) * (1 / ss)))\n frame_drawer.draw_box(image_copy, face_box)\n frame_drawer.draw_face_id(image_copy, face_box, face_id)\n\n cv2.imshow(\"Frame\", image_copy)\n key = cv2.waitKey(1) & 0xFF\n if key == ord(\"q\"):\n capturer.release()\n break\n cv2.destroyAllWindows()\n with open('storage.pkl', 'wb') as db_file:\n pickle.dump(storage, db_file)\n\n\nif __name__ == '__main__':\n with open('config.yml', 'r') as file:\n main(yaml.safe_load(file))\n", "repo_name": "shaxov/spy-eye", "sub_path": "run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 3989, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "frame.drawer.Drawer", "line_number": 15, "usage_type": "call"}, {"api_name": "frame.drawer", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 16, "usage_type": "call"}, {"api_name": "face.detectors.get", "line_number": 17, "usage_type": "call"}, {"api_name": "face.detectors", "line_number": 17, "usage_type": "attribute"}, {"api_name": "face.validators.get_list", "line_number": 18, "usage_type": "call"}, {"api_name": "face.validators", "line_number": 18, "usage_type": "attribute"}, {"api_name": "frame.filters.get_list", "line_number": 19, "usage_type": "call"}, {"api_name": "frame.filters", "line_number": 19, "usage_type": "attribute"}, {"api_name": "face.buffer.FaceBuffer", "line_number": 20, "usage_type": "call"}, {"api_name": "face.buffer", "line_number": 20, "usage_type": "attribute"}, {"api_name": "face.encoders.get", "line_number": 21, "usage_type": "call"}, {"api_name": "face.encoders", "line_number": 21, "usage_type": "attribute"}, {"api_name": "face.recognizers.get", "line_number": 22, "usage_type": "call"}, {"api_name": "face.recognizers", "line_number": 22, "usage_type": "attribute"}, {"api_name": "db.initialize", "line_number": 23, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 37, "usage_type": "call"}, {"api_name": "typedef.NO_VIDEO_FRAME", "line_number": 37, "usage_type": "attribute"}, {"api_name": "cv2.waitKey", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 45, "usage_type": "call"}, {"api_name": "cv2.INTER_CUBIC", "line_number": 45, "usage_type": "attribute"}, {"api_name": "frame.filters.apply", "line_number": 46, "usage_type": "call"}, {"api_name": "frame.filters", "line_number": 46, "usage_type": "attribute"}, {"api_name": "face.validators.apply", "line_number": 48, "usage_type": "call"}, {"api_name": "face.validators", "line_number": 48, "usage_type": "attribute"}, {"api_name": "face.trackers.apply", "line_number": 50, "usage_type": "call"}, {"api_name": "face.trackers", "line_number": 50, "usage_type": "attribute"}, {"api_name": "face.trackers.drop_wasted", "line_number": 51, "usage_type": "call"}, {"api_name": "face.trackers", "line_number": 51, "usage_type": "attribute"}, {"api_name": "copy.deepcopy", "line_number": 52, "usage_type": "call"}, {"api_name": "face.validators.apply", "line_number": 53, "usage_type": "call"}, {"api_name": "face.validators", "line_number": 53, "usage_type": "attribute"}, {"api_name": "utils.crop", "line_number": 62, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 63, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 64, "usage_type": "attribute"}, {"api_name": "typedef.UNKNOWN_FACE_ID", "line_number": 66, "usage_type": "attribute"}, {"api_name": "utils.generate_tmp_face_id", "line_number": 67, "usage_type": "call"}, {"api_name": "face.trackers.get", "line_number": 68, "usage_type": "call"}, {"api_name": "face.trackers", "line_number": 68, "usage_type": "attribute"}, {"api_name": "utils.is_tmp_id", "line_number": 72, "usage_type": "call"}, {"api_name": "face.trackers.update_face_ids", "line_number": 85, "usage_type": "call"}, {"api_name": "face.trackers", "line_number": 85, "usage_type": "attribute"}, {"api_name": "numpy.int64", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 87, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 91, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 92, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 96, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 98, "usage_type": "call"}, {"api_name": "yaml.safe_load", "line_number": 103, "usage_type": "call"}]}
+{"seq_id": "34440016677", "text": "from abc import ABC, abstractmethod\nfrom typing import Optional\n\nfrom torchdata.dataloader2 import DataLoader2, DistributedReadingService, MultiProcessingReadingService, SequentialReadingService \nfrom torch.utils.data import DataLoader\n\nimport pytorch_lightning as pl\nfrom pytorch_lightning.utilities.exceptions import MisconfigurationException\n\nimport logging\npl_logger = logging.getLogger('pytorch_lightning')\n\nclass BaseTDM(pl.LightningDataModule, ABC):\n\tdef __init__(self, \n\t\t\ttrain_urls:Optional[list]=None,\n\t\t\ttest_urls:Optional[list]=None,\n\t\t\tvalid_urls:Optional[list]=None,\n\t\t\tpredict_urls:Optional[list]=None,\n\t\t\tbatch_size:Optional[int]=1,\n\t\t\tnum_workers:Optional[int]=0,\n\t\t\tpersistent_workers:Optional[bool]=True,\n\t\t\tshuffle:Optional[bool]=True,\n\t\t):\n\t\tsuper().__init__()\n\n\t\tself.train_data_dir = train_urls\n\t\tself.test_data_dir = test_urls\n\t\tself.valid_data_dir = valid_urls\n\t\tself.predict_data_dir = predict_urls\n\n\t\tself.shuffle = shuffle\n\t\tself.batch_size = batch_size\n\t\tself.num_workers = num_workers\n\t\tself.persistent_workers = persistent_workers\n\n\t@abstractmethod\n\tdef to_sampels(self, data):\n\t\tpass\n\t\n\t@abstractmethod\n\tdef create_pipeline(self, data_dir):\n\t\tpass\n\n\t@abstractmethod\n\tdef collate_fn(self, data):\n\t\tpass\n\n\tdef setup(self, stage:Optional[str] = None):\n\t\tif self.train_data_dir and len(self.train_data_dir)>0:\n\t\t\tself.train = self.create_pipeline(self.train_data_dir)\n\n\t\tif self.test_data_dir and len(self.test_data_dir)>0:\n\t\t\tself.test = self.create_pipeline(self.test_data_dir)\n\n\t\tif self.valid_data_dir and len(self.valid_data_dir)>0:\n\t\t\tself.valid = self.create_pipeline(self.valid_data_dir)\n\n\t\tif self.predict_data_dir and len(self.predict_data_dir)>0:\n\t\t\tself.predict = self.create_pipeline(self.predict_data_dir)\n\n\tdef _dataloader2(self, dataset):\n\t\tservice = [\n\t\t\tDistributedReadingService(),\n\t\t\tMultiProcessingReadingService(num_workers=self.num_workers),\n\t\t]\n\t\treading_service = SequentialReadingService(*service)\n\t\treturn DataLoader2(dataset, reading_service=reading_service)\n\n\tdef _dataloader(self, dataset):\n\t\treturn DataLoader(dataset, num_workers=self.num_workers, batch_size=self.batch_size, collate_fn=self.collate_fn)\n\n\tdef train_dataloader(self):\n\t\tif not self.train_data_dir:\n\t\t\traise MisconfigurationException('train_urls not set.')\n\t\treturn self._dataloader(self.train)\n\n\tdef val_dataloader(self):\n\t\tif not self.valid_data_dir:\n\t\t\traise MisconfigurationException('valid_urls not set.')\n\t\treturn self._dataloader(self.valid)\n\n\tdef test_dataloader(self):\n\t\tif not self.test_data_dir:\n\t\t\traise MisconfigurationException('test_urls not set.')\n\t\treturn self._dataloader(self.test)\n\n\tdef predict_dataloader(self):\n\t\tif not self.predict_data_dir:\n\t\t\traise MisconfigurationException('predict_urls not set.')\n\t\treturn self._dataloader(self.predict)", "repo_name": "knoriy/video2caption", "sub_path": "src/datamodule/base.py", "file_name": "base.py", "file_ext": "py", "file_size_in_byte": 2790, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "pytorch_lightning.LightningDataModule", "line_number": 13, "usage_type": "attribute"}, {"api_name": "abc.ABC", "line_number": 13, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 15, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 16, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 18, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 19, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 20, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 21, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 22, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 36, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 40, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 44, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 48, "usage_type": "name"}, {"api_name": "torchdata.dataloader2.DistributedReadingService", "line_number": 63, "usage_type": "call"}, {"api_name": "torchdata.dataloader2.MultiProcessingReadingService", "line_number": 64, "usage_type": "call"}, {"api_name": "torchdata.dataloader2.SequentialReadingService", "line_number": 66, "usage_type": "call"}, {"api_name": "torchdata.dataloader2.DataLoader2", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 70, "usage_type": "call"}, {"api_name": "pytorch_lightning.utilities.exceptions.MisconfigurationException", "line_number": 74, "usage_type": "call"}, {"api_name": "pytorch_lightning.utilities.exceptions.MisconfigurationException", "line_number": 79, "usage_type": "call"}, {"api_name": "pytorch_lightning.utilities.exceptions.MisconfigurationException", "line_number": 84, "usage_type": "call"}, {"api_name": "pytorch_lightning.utilities.exceptions.MisconfigurationException", "line_number": 89, "usage_type": "call"}]}
+{"seq_id": "37326198417", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport os\nimport random\nimport re\nfrom PIL import Image\nfrom pylab import *\nimport sys\nfrom tensorflow.keras.models import Model\nfrom tensorflow.keras.regularizers import l2\nfrom tensorflow.keras.layers import *\nfrom tensorflow.keras.applications.vgg16 import *\nfrom tensorflow.keras.models import *\nimport tensorflow.keras.backend as K\nimport tensorflow as tf\nfrom tensorflow.keras.optimizers import Adam\nfrom tensorflow.keras.layers import Convolution2D, ZeroPadding2D, MaxPooling2D, Cropping2D, Conv2D\nfrom tensorflow.keras.layers import Input, Add, Dropout, Permute, add\nfrom tensorflow.compat.v1.layers import conv2d_transpose\nfrom tensorflow.keras.preprocessing.image import ImageDataGenerator\nfrom tensorflow.python.keras.callbacks import TensorBoard, ModelCheckpoint, EarlyStopping, ReduceLROnPlateau\n\nlabel_codes=[(255,0,0), (0,255,0), (255,255,0),(0,0,0)]\nlabel_names=['comp_1','comp2_2', 'both','background']\nDATA_PATH='./dataset/'\ncode2id = {v:k for k,v in enumerate(label_codes)}\nid2code = {k:v for k,v in enumerate(label_codes)}\nname2id = {v:k for k,v in enumerate(label_names)}\nid2name = {k:v for k,v in enumerate(label_names)}\n\ndata_gen_args = dict(rescale=1./255)\nmask_gen_args = dict()\n\ntrain_frames_datagen = ImageDataGenerator(**data_gen_args)\ntrain_masks_datagen = ImageDataGenerator(**mask_gen_args)\nval_frames_datagen = ImageDataGenerator(**data_gen_args)\nval_masks_datagen = ImageDataGenerator(**mask_gen_args)\ndef _read_to_tensor(fname, output_height=256, output_width=256, normalize_data=False):\n '''Function to read images from given image file path, and provide resized images as tensors\n Inputs:\n fname - image file path\n output_height - required output image height\n output_width - required output image width\n normalize_data - if True, normalize data to be centered around 0 (mean 0, range 0 to 1)\n Output: Processed image tenso\n '''\n\n # Read the image as a tensor\n img_strings = tf.io.read_file(fname)\n imgs_decoded = tf.image.decode_jpeg(img_strings)\n\n # Resize the image\n output = tf.image.resize(imgs_decoded, [output_height, output_width])\n\n # Normalize if required\n if normalize_data:\n output = (output - 128) / 128\n return output\n\n#reading frames and masks\ndef read_images(img_dir):\n '''Function to get all image directories, read images and masks in separate tensors\n Inputs:\n img_dir - file directory\n Outputs\n frame_tensors, masks_tensors, frame files list, mask files list\n '''\n\n # Get the file names list from provided directory\n file_list = [f for f in os.listdir(img_dir) if os.path.isfile(os.path.join(img_dir, f))]\n\n # Separate frame and mask files lists, exclude unnecessary files\n frames_list = [file for file in file_list if ('_labeled' not in file) and ('txt' not in file) and ('m' not in file)]\n frames_list.sort()\n masks_list = [file for file in file_list if ('_labeled' in file) and ('txt' not in file) and ('m' not in file)]\n masks_list.sort()\n print('{} frame files found in the provided directory.'.format(len(frames_list)))\n print('{} mask files found in the provided directory.'.format(len(masks_list)))\n\n # Create file paths from file names\n frames_paths = [os.path.join(img_dir, fname) for fname in frames_list]\n masks_paths = [os.path.join(img_dir, fname) for fname in masks_list]\n\n # Create dataset of tensors\n frame_data = tf.data.Dataset.from_tensor_slices(frames_paths)\n masks_data = tf.data.Dataset.from_tensor_slices(masks_paths)\n\n # Read images into the tensor dataset\n frame_tensors = frame_data.map(_read_to_tensor)\n masks_tensors = masks_data.map(_read_to_tensor)\n\n print('Completed importing {} frame images from the provided directory.'.format(len(frames_list)))\n print('Completed importing {} mask images from the provided directory.'.format(len(masks_list)))\n\n return frame_tensors, masks_tensors, frames_list, masks_list\n\n\ndef parse_code(l):\n '''Function to parse lines in a text file, returns separated elements (label codes and names in this case)\n '''\n if len(l.strip().split(\"\\t\")) == 2:\n a, b = l.strip().split(\"\\t\")\n return tuple(int(i) for i in a.split(' ')), b\n else:\n a, b, c = l.strip().split(\"\\t\")\n return tuple(int(i) for i in a.split(' ')), c\n\ndef generate_image_folder_structure(DATA_PATH,frames, masks, frames_list, masks_list):\n '''Function to save images in the appropriate folder directories\n Inputs:\n frames - frame tensor dataset\n masks - mask tensor dataset\n frames_list - frame file paths\n masks_list - mask file paths\n '''\n # Create iterators for frames and masks\n frame_batches = tf.compat.v1.data.make_one_shot_iterator(\n frames) # outside of TF Eager, we would use make_one_shot_iterator\n mask_batches = tf.compat.v1.data.make_one_shot_iterator(masks)\n\n # Iterate over the train images while saving the frames and masks in appropriate folders\n dir_name = 'train'\n for file in zip(frames_list[:-round(0.2 * len(frames_list))], masks_list[:-round(0.2 * len(masks_list))]):\n # Convert tensors to numpy arrays\n frame = frame_batches.next().numpy().astype(np.uint8)\n mask = mask_batches.next().numpy().astype(np.uint8)\n\n # Convert numpy arrays to images\n frame = Image.fromarray(frame)\n mask = Image.fromarray(mask)\n\n # Save frames and masks to correct directories\n frame.save(DATA_PATH + '{}_frames/{}'.format(dir_name, dir_name) + '/' + file[0])\n mask.save(DATA_PATH + '{}_masks/{}'.format(dir_name, dir_name) + '/' + file[1])\n\n # Iterate over the val images while saving the frames and masks in appropriate folders\n dir_name = 'val'\n for file in zip(frames_list[-round(0.2 * len(frames_list)):], masks_list[-round(0.2 * len(masks_list)):]):\n # Convert tensors to numpy arrays\n frame = frame_batches.next().numpy().astype(np.uint8)\n mask = mask_batches.next().numpy().astype(np.uint8)\n\n # Convert numpy arrays to images\n frame = Image.fromarray(frame)\n mask = Image.fromarray(mask)\n\n # Save frames and masks to correct directories\n frame.save(DATA_PATH + '{}_frames/{}'.format(dir_name, dir_name) + '/' + file[0])\n mask.save(DATA_PATH + '{}_masks/{}'.format(dir_name, dir_name) + '/' + file[1])\n\n print(\"Saved {} frames to directory {}\".format(len(frames_list), DATA_PATH))\n print(\"Saved {} masks to directory {}\".format(len(masks_list), DATA_PATH))\n\ndef rgb_to_onehot(rgb_image, colormap = id2code):\n '''Function to one hot encode RGB mask labels\n Inputs:\n rgb_image - image matrix (eg. 256 x 256 x 3 dimension numpy ndarray)\n colormap - dictionary of color to label id\n Output: One hot encoded image of dimensions (height x width x num_classes) where num_classes = len(colormap)\n '''\n num_classes = len(colormap)\n shape = rgb_image.shape[:2]+(num_classes,)\n encoded_image = np.zeros( shape, dtype=np.int8 )\n for i, cls in enumerate(colormap):\n encoded_image[:,:,i] = np.all(rgb_image.reshape( (-1,3) ) == colormap[i], axis=1).reshape(shape[:2])\n return encoded_image\n\n\ndef onehot_to_rgb(onehot, colormap = id2code):\n '''Function to decode encoded mask labels\n Inputs:\n onehot - one hot encoded image matrix (height x width x num_classes)\n colormap - dictionary of color to label id\n Output: Decoded RGB image (height x width x 3)\n '''\n single_layer = np.argmax(onehot, axis=-1)\n output = np.zeros( onehot.shape[:2]+(3,) )\n for k in colormap.keys():\n output[single_layer==k] = colormap[k]\n return np.uint8(output)\n\n\ndef TrainAugmentGenerator(seed=1, batch_size=5 ):\n '''Train Image data generator\n Inputs:\n seed - seed provided to the flow_from_directory function to ensure aligned data flow\n batch_size - number of images to import at a time\n Output: Decoded RGB image (height x width x 3)\n '''\n train_image_generator = train_frames_datagen.flow_from_directory(\n DATA_PATH + 'train_frames/',\n batch_size=batch_size, seed=seed)\n\n train_mask_generator = train_masks_datagen.flow_from_directory(\n DATA_PATH + 'train_masks/',\n batch_size=batch_size, seed=seed)\n\n while True:\n X1i = train_image_generator.next()\n X2i = train_mask_generator.next()\n\n # One hot encoding RGB images\n mask_encoded = [rgb_to_onehot(X2i[0][x, :, :, :], id2code) for x in range(X2i[0].shape[0])]\n\n yield X1i[0], np.asarray(mask_encoded)\n\n\ndef ValAugmentGenerator(seed=1, batch_size=5):\n '''Validation Image data generator\n Inputs:\n seed - seed provided to the flow_from_directory function to ensure aligned data flow\n batch_size - number of images to import at a time\n Output: Decoded RGB image (height x width x 3)\n '''\n val_image_generator = val_frames_datagen.flow_from_directory(\n DATA_PATH + 'val_frames/',\n batch_size=batch_size, seed=seed)\n\n val_mask_generator = val_masks_datagen.flow_from_directory(\n DATA_PATH + 'val_masks/',\n batch_size=batch_size, seed=seed)\n\n while True:\n X1i = val_image_generator.next()\n X2i = val_mask_generator.next()\n\n # One hot encoding RGB images\n mask_encoded = [rgb_to_onehot(X2i[0][x, :, :, :], id2code) for x in range(X2i[0].shape[0])]\n yield X1i[0], np.asarray(mask_encoded)\n\n\ndef read_test_images(img_dir):\n '''Function to get all image directories, read images and masks in separate tensors\n Inputs:\n img_dir - file directory\n Outputs\n frame_tensors, masks_tensors, frame files list, mask files list\n '''\n\n # Get the file names list from provided directory\n file_list = [f for f in os.listdir(img_dir) if os.path.isfile(os.path.join(img_dir, f))]\n\n # Separate frame and mask files lists, exclude unnecessary files\n frames_list = [file for file in file_list if ('_labeled' not in file) and ('txt' not in file) and ('m' not in file)]\n frames_list.sort()\n print('{} frame files found in the provided directory.'.format(len(frames_list)))\n\n # Create file paths from file names\n frames_paths = [os.path.join(img_dir, fname) for fname in frames_list]\n\n # Create dataset of tensors\n frame_data = tf.data.Dataset.from_tensor_slices(frames_paths)\n\n # Read images into the tensor dataset\n frame_tensors = frame_data.map(_read_to_tensor)\n\n print('Completed importing {} frame images from the provided directory.'.format(len(frames_list)))\n\n return frame_tensors, frames_list\n\ndef tversky_loss(y_true, y_pred):\n alpha = 0.5\n beta = 0.5\n\n ones = K.ones(K.shape(y_true))\n p0 = y_pred # proba that voxels are class i\n p1 = ones - y_pred # proba that voxels are not class i\n g0 = y_true\n g1 = ones - y_true\n\n num = K.sum(p0 * g0, (0, 1, 2, 3))\n den = num + alpha * K.sum(p0 * g1, (0, 1, 2, 3)) + beta * K.sum(p1 * g0, (0, 1, 2, 3))\n\n T = K.sum(num / den) # when summing over classes, T has dynamic range [0 Ncl]\n\n Ncl = K.cast(K.shape(y_true)[-1], 'float32')\n return Ncl - T\n\n\ndef dice_coef(y_true, y_pred):\n smooth=1\n y_true_f = K.flatten(y_true)\n y_pred_f = K.flatten(y_pred)\n intersection = K.sum(y_true_f * y_pred_f)\n return (2. * intersection + smooth) / (K.sum(y_true_f*y_true_f) + K.sum(y_pred_f*y_pred_f) + smooth)\n\n\ndef dice_coef_loss(y_true, y_pred):\n return 1.-dice_coef(y_true, y_pred)", "repo_name": "aidanamv/Unet-Segmenetation", "sub_path": "data_processing.py", "file_name": "data_processing.py", "file_ext": "py", "file_size_in_byte": 11708, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "tensorflow.keras.preprocessing.image.ImageDataGenerator", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.image.ImageDataGenerator", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.image.ImageDataGenerator", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.image.ImageDataGenerator", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.io.read_file", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.io", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tensorflow.image.decode_jpeg", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 50, "usage_type": "attribute"}, {"api_name": "tensorflow.image.resize", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "tensorflow.data.Dataset.from_tensor_slices", "line_number": 85, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 85, "usage_type": "attribute"}, {"api_name": "tensorflow.data.Dataset.from_tensor_slices", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 86, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1.data.make_one_shot_iterator", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 117, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1.data.make_one_shot_iterator", "line_number": 119, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 119, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 125, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 126, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 129, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 129, "usage_type": "name"}, {"api_name": "PIL.Image.fromarray", "line_number": 130, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 130, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 140, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 141, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 144, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 144, "usage_type": "name"}, {"api_name": "PIL.Image.fromarray", "line_number": 145, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 145, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 163, "usage_type": "attribute"}, {"api_name": "numpy.all", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 229, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path", "line_number": 241, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 249, "usage_type": "call"}, {"api_name": "os.path", "line_number": 249, "usage_type": "attribute"}, {"api_name": "tensorflow.data.Dataset.from_tensor_slices", "line_number": 252, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 252, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.backend.ones", "line_number": 265, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 265, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.shape", "line_number": 265, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend.sum", "line_number": 271, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 271, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.sum", "line_number": 272, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 272, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.sum", "line_number": 274, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 274, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.cast", "line_number": 276, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 276, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.shape", "line_number": 276, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend.flatten", "line_number": 282, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 282, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.flatten", "line_number": 283, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 283, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.sum", "line_number": 284, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 284, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.sum", "line_number": 285, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 285, "usage_type": "name"}]}
+{"seq_id": "21884309689", "text": "# coding: utf-8\n\nimport six\n\nfrom huaweicloudsdkcore.utils.http_utils import sanitize_for_serialization\n\n\nclass ShowDataRequest:\n\n \"\"\"\n Attributes:\n openapi_types (dict): The key is attribute name\n and the value is attribute type.\n attribute_map (dict): The key is attribute name\n and the value is json key in definition.\n \"\"\"\n sensitive_list = []\n\n openapi_types = {\n 'x_need_content': 'bool',\n 'eihealth_project_id': 'str',\n 'path': 'str'\n }\n\n attribute_map = {\n 'x_need_content': 'X-Need-Content',\n 'eihealth_project_id': 'eihealth_project_id',\n 'path': 'path'\n }\n\n def __init__(self, x_need_content=None, eihealth_project_id=None, path=None):\n \"\"\"ShowDataRequest\n\n The model defined in huaweicloud sdk\n\n :param x_need_content: 返回文件内容\n :type x_need_content: bool\n :param eihealth_project_id: 医疗智能体平台项目ID,您可以在EIHealth平台单击所需的项目名称,进入项目设置页面查看。\n :type eihealth_project_id: str\n :param path: 对象全路径(项目名称:|路径)\n :type path: str\n \"\"\"\n \n \n\n self._x_need_content = None\n self._eihealth_project_id = None\n self._path = None\n self.discriminator = None\n\n if x_need_content is not None:\n self.x_need_content = x_need_content\n self.eihealth_project_id = eihealth_project_id\n self.path = path\n\n @property\n def x_need_content(self):\n \"\"\"Gets the x_need_content of this ShowDataRequest.\n\n 返回文件内容\n\n :return: The x_need_content of this ShowDataRequest.\n :rtype: bool\n \"\"\"\n return self._x_need_content\n\n @x_need_content.setter\n def x_need_content(self, x_need_content):\n \"\"\"Sets the x_need_content of this ShowDataRequest.\n\n 返回文件内容\n\n :param x_need_content: The x_need_content of this ShowDataRequest.\n :type x_need_content: bool\n \"\"\"\n self._x_need_content = x_need_content\n\n @property\n def eihealth_project_id(self):\n \"\"\"Gets the eihealth_project_id of this ShowDataRequest.\n\n 医疗智能体平台项目ID,您可以在EIHealth平台单击所需的项目名称,进入项目设置页面查看。\n\n :return: The eihealth_project_id of this ShowDataRequest.\n :rtype: str\n \"\"\"\n return self._eihealth_project_id\n\n @eihealth_project_id.setter\n def eihealth_project_id(self, eihealth_project_id):\n \"\"\"Sets the eihealth_project_id of this ShowDataRequest.\n\n 医疗智能体平台项目ID,您可以在EIHealth平台单击所需的项目名称,进入项目设置页面查看。\n\n :param eihealth_project_id: The eihealth_project_id of this ShowDataRequest.\n :type eihealth_project_id: str\n \"\"\"\n self._eihealth_project_id = eihealth_project_id\n\n @property\n def path(self):\n \"\"\"Gets the path of this ShowDataRequest.\n\n 对象全路径(项目名称:|路径)\n\n :return: The path of this ShowDataRequest.\n :rtype: str\n \"\"\"\n return self._path\n\n @path.setter\n def path(self, path):\n \"\"\"Sets the path of this ShowDataRequest.\n\n 对象全路径(项目名称:|路径)\n\n :param path: The path of this ShowDataRequest.\n :type path: str\n \"\"\"\n self._path = path\n\n def to_dict(self):\n \"\"\"Returns the model properties as a dict\"\"\"\n result = {}\n\n for attr, _ in six.iteritems(self.openapi_types):\n value = getattr(self, attr)\n if isinstance(value, list):\n result[attr] = list(map(\n lambda x: x.to_dict() if hasattr(x, \"to_dict\") else x,\n value\n ))\n elif hasattr(value, \"to_dict\"):\n result[attr] = value.to_dict()\n elif isinstance(value, dict):\n result[attr] = dict(map(\n lambda item: (item[0], item[1].to_dict())\n if hasattr(item[1], \"to_dict\") else item,\n value.items()\n ))\n else:\n if attr in self.sensitive_list:\n result[attr] = \"****\"\n else:\n result[attr] = value\n\n return result\n\n def to_str(self):\n \"\"\"Returns the string representation of the model\"\"\"\n import simplejson as json\n if six.PY2:\n import sys\n reload(sys)\n sys.setdefaultencoding(\"utf-8\")\n return json.dumps(sanitize_for_serialization(self), ensure_ascii=False)\n\n def __repr__(self):\n \"\"\"For `print`\"\"\"\n return self.to_str()\n\n def __eq__(self, other):\n \"\"\"Returns true if both objects are equal\"\"\"\n if not isinstance(other, ShowDataRequest):\n return False\n\n return self.__dict__ == other.__dict__\n\n def __ne__(self, other):\n \"\"\"Returns true if both objects are not equal\"\"\"\n return not self == other\n", "repo_name": "huaweicloud/huaweicloud-sdk-python-v3", "sub_path": "huaweicloud-sdk-eihealth/huaweicloudsdkeihealth/v1/model/show_data_request.py", "file_name": "show_data_request.py", "file_ext": "py", "file_size_in_byte": 5186, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 104, "dataset": "github-code", "pt": "20", "api": [{"api_name": "six.iteritems", "line_number": 126, "usage_type": "call"}, {"api_name": "six.PY2", "line_number": 152, "usage_type": "attribute"}, {"api_name": "sys.setdefaultencoding", "line_number": 155, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 156, "usage_type": "call"}, {"api_name": "huaweicloudsdkcore.utils.http_utils.sanitize_for_serialization", "line_number": 156, "usage_type": "call"}]}
+{"seq_id": "70807824383", "text": "import binascii\nimport collections\nimport itertools\nimport logging\nimport re\nimport socket\nimport struct\n\nfrom typing import List, Optional, Tuple\n\nlogger = logging.getLogger(__name__)\n\n\ndef target_to_ipv4(target: str) -> Optional[List]:\n \"\"\"Attempt to return a single IPv4 host list from a target string.\"\"\"\n\n try:\n socket.inet_pton(socket.AF_INET, target)\n return [target]\n except socket.error:\n return None\n\n\ndef target_to_ipv6(target: str) -> Optional[List]:\n \"\"\"Attempt to return a single IPv6 host list from a target string.\"\"\"\n\n try:\n socket.inet_pton(socket.AF_INET6, target)\n return [target]\n except socket.error:\n return None\n\n\ndef ipv4_range_to_list(start_packed, end_packed) -> Optional[List]:\n \"\"\"Return a list of IPv4 entries from start_packed to end_packed.\"\"\"\n\n new_list = list()\n start = struct.unpack('!L', start_packed)[0]\n end = struct.unpack('!L', end_packed)[0]\n\n for value in range(start, end + 1):\n new_ip = socket.inet_ntoa(struct.pack('!L', value))\n new_list.append(new_ip)\n\n return new_list\n\n\ndef target_to_ipv4_short(target: str) -> Optional[List]:\n \"\"\"Attempt to return a IPv4 short range list from a target string.\"\"\"\n\n splitted = target.split('-')\n if len(splitted) != 2:\n return None\n\n try:\n start_packed = socket.inet_pton(socket.AF_INET, splitted[0])\n end_value = int(splitted[1])\n except (socket.error, ValueError):\n return None\n\n # For subnet with mask lower than /24, ip addresses ending in .0 are\n # allowed.\n # The next code checks for a range starting with a A.B.C.0.\n # For the octet equal to 0, bytes() returns an empty binary b'',\n # which must be handle in a special way.\n _start_value = bytes(start_packed[3])\n if _start_value:\n start_value = int(binascii.hexlify(_start_value), 16)\n elif _start_value == b'':\n start_value = 0\n else:\n return None\n\n if end_value < 0 or end_value > 255 or end_value < start_value:\n return None\n\n end_packed = start_packed[0:3] + struct.pack('B', end_value)\n\n return ipv4_range_to_list(start_packed, end_packed)\n\n\ndef target_to_ipv4_cidr(target: str) -> Optional[List]:\n \"\"\"Attempt to return a IPv4 CIDR list from a target string.\"\"\"\n\n splitted = target.split('/')\n if len(splitted) != 2:\n return None\n\n try:\n start_packed = socket.inet_pton(socket.AF_INET, splitted[0])\n block = int(splitted[1])\n except (socket.error, ValueError):\n return None\n\n if block <= 0 or block > 30:\n return None\n\n start_value = int(binascii.hexlify(start_packed), 16) >> (32 - block)\n start_value = (start_value << (32 - block)) + 1\n\n end_value = (start_value | (0xFFFFFFFF >> block)) - 1\n\n start_packed = struct.pack('!I', start_value)\n end_packed = struct.pack('!I', end_value)\n\n return ipv4_range_to_list(start_packed, end_packed)\n\n\ndef target_to_ipv6_cidr(target: str) -> Optional[List]:\n \"\"\"Attempt to return a IPv6 CIDR list from a target string.\"\"\"\n\n splitted = target.split('/')\n if len(splitted) != 2:\n return None\n\n try:\n start_packed = socket.inet_pton(socket.AF_INET6, splitted[0])\n block = int(splitted[1])\n except (socket.error, ValueError):\n return None\n\n if block <= 0 or block > 126:\n return None\n\n start_value = int(binascii.hexlify(start_packed), 16) >> (128 - block)\n start_value = (start_value << (128 - block)) + 1\n\n end_value = (start_value | (int('ff' * 16, 16) >> block)) - 1\n\n high = start_value >> 64\n low = start_value & ((1 << 64) - 1)\n\n start_packed = struct.pack('!QQ', high, low)\n\n high = end_value >> 64\n low = end_value & ((1 << 64) - 1)\n\n end_packed = struct.pack('!QQ', high, low)\n\n return ipv6_range_to_list(start_packed, end_packed)\n\n\ndef target_to_ipv4_long(target: str) -> Optional[List]:\n \"\"\"Attempt to return a IPv4 long-range list from a target string.\"\"\"\n\n splitted = target.split('-')\n if len(splitted) != 2:\n return None\n\n try:\n start_packed = socket.inet_pton(socket.AF_INET, splitted[0])\n end_packed = socket.inet_pton(socket.AF_INET, splitted[1])\n except socket.error:\n return None\n\n if end_packed < start_packed:\n return None\n\n return ipv4_range_to_list(start_packed, end_packed)\n\n\ndef ipv6_range_to_list(start_packed, end_packed) -> List:\n \"\"\"Return a list of IPv6 entries from start_packed to end_packed.\"\"\"\n\n new_list = list()\n\n start = int(binascii.hexlify(start_packed), 16)\n end = int(binascii.hexlify(end_packed), 16)\n\n for value in range(start, end + 1):\n high = value >> 64\n low = value & ((1 << 64) - 1)\n new_ip = socket.inet_ntop(\n socket.AF_INET6, struct.pack('!2Q', high, low)\n )\n new_list.append(new_ip)\n\n return new_list\n\n\ndef target_to_ipv6_short(target: str) -> Optional[List]:\n \"\"\"Attempt to return a IPv6 short-range list from a target string.\"\"\"\n\n splitted = target.split('-')\n if len(splitted) != 2:\n return None\n\n try:\n start_packed = socket.inet_pton(socket.AF_INET6, splitted[0])\n end_value = int(splitted[1], 16)\n except (socket.error, ValueError):\n return None\n\n start_value = int(binascii.hexlify(start_packed[14:]), 16)\n if end_value < 0 or end_value > 0xFFFF or end_value < start_value:\n return None\n\n end_packed = start_packed[:14] + struct.pack('!H', end_value)\n\n return ipv6_range_to_list(start_packed, end_packed)\n\n\ndef target_to_ipv6_long(target: str) -> Optional[List]:\n \"\"\"Attempt to return a IPv6 long-range list from a target string.\"\"\"\n\n splitted = target.split('-')\n if len(splitted) != 2:\n return None\n\n try:\n start_packed = socket.inet_pton(socket.AF_INET6, splitted[0])\n end_packed = socket.inet_pton(socket.AF_INET6, splitted[1])\n except socket.error:\n return None\n\n if end_packed < start_packed:\n return None\n\n return ipv6_range_to_list(start_packed, end_packed)\n\n\ndef target_to_hostname(target: str) -> Optional[List]:\n \"\"\"Attempt to return a single hostname list from a target string.\"\"\"\n\n if len(target) == 0 or len(target) > 255:\n return None\n\n if not re.match(r'^[\\w.-]+$', target):\n return None\n\n return [target]\n\n\ndef target_to_list(target: str) -> Optional[List]:\n \"\"\"Attempt to return a list of single hosts from a target string.\"\"\"\n\n # Is it an IPv4 address ?\n new_list = target_to_ipv4(target)\n # Is it an IPv6 address ?\n if not new_list:\n new_list = target_to_ipv6(target)\n # Is it an IPv4 CIDR ?\n if not new_list:\n new_list = target_to_ipv4_cidr(target)\n # Is it an IPv6 CIDR ?\n if not new_list:\n new_list = target_to_ipv6_cidr(target)\n # Is it an IPv4 short-range ?\n if not new_list:\n new_list = target_to_ipv4_short(target)\n # Is it an IPv4 long-range ?\n if not new_list:\n new_list = target_to_ipv4_long(target)\n # Is it an IPv6 short-range ?\n if not new_list:\n new_list = target_to_ipv6_short(target)\n # Is it an IPv6 long-range ?\n if not new_list:\n new_list = target_to_ipv6_long(target)\n # Is it a hostname ?\n if not new_list:\n new_list = target_to_hostname(target)\n\n return new_list\n\n\ndef target_str_to_list(target_str: str) -> Optional[List]:\n \"\"\"Parses a targets string into a list of individual targets.\n Return a list of hosts, None if supplied target_str is None or\n empty, or an empty list in case of malformed target.\n \"\"\"\n new_list = list()\n\n if not target_str:\n return None\n\n target_str = target_str.strip(',')\n\n for target in target_str.split(','):\n target = target.strip()\n target_list = target_to_list(target)\n\n if target_list:\n new_list.extend(target_list)\n else:\n logger.info(\"%s: Invalid target value\", target)\n return []\n\n return list(collections.OrderedDict.fromkeys(new_list))\n\n\ndef resolve_hostname(hostname: str) -> Optional[str]:\n \"\"\"Returns IP of a hostname.\"\"\"\n\n assert hostname\n try:\n return socket.gethostbyname(hostname)\n except socket.gaierror:\n return None\n\n\ndef is_valid_address(address: str) -> bool:\n if not address:\n return False\n\n try:\n socket.inet_pton(socket.AF_INET, address)\n except OSError:\n # invalid IPv4 address\n try:\n socket.inet_pton(socket.AF_INET6, address)\n except OSError:\n # invalid IPv6 address\n return False\n\n return True\n\n\ndef get_hostname_by_address(address: str) -> str:\n \"\"\"Returns hostname of an address.\"\"\"\n\n if not is_valid_address(address):\n return ''\n\n try:\n hostname = socket.getfqdn(address)\n except (socket.gaierror, socket.herror):\n return ''\n\n if hostname == address:\n return ''\n\n return hostname\n\n\ndef port_range_expand(portrange: str) -> Optional[List]:\n \"\"\"\n Receive a port range and expands it in individual ports.\n\n @input Port range.\n e.g. \"4-8\"\n\n @return List of integers.\n e.g. [4, 5, 6, 7, 8]\n \"\"\"\n if not portrange or '-' not in portrange:\n return None\n\n try:\n port_range_min = int(portrange[: portrange.index('-')])\n port_range_max = int(portrange[portrange.index('-') + 1 :]) + 1\n except (IndexError, ValueError) as e:\n logger.info(\"Invalid port range format %s\", e)\n return None\n\n port_list = list()\n\n for single_port in range(\n port_range_min,\n port_range_max,\n ):\n port_list.append(single_port)\n\n return port_list\n\n\ndef port_str_arrange(ports: str) -> str:\n \"\"\"Gives a str in the format (always tcp listed first).\n T:U:\n \"\"\"\n b_tcp = ports.find(\"T\")\n b_udp = ports.find(\"U\")\n\n if (b_udp != -1 and b_tcp != -1) and b_udp < b_tcp:\n return ports[b_tcp:] + ports[b_udp:b_tcp]\n\n return ports\n\n\ndef ports_str_check_failed(port_str: str) -> bool:\n \"\"\"\n Check if the port string is well formed.\n Return True if fail, False other case.\n \"\"\"\n pattern = r'[^TU:0-9, \\-\\n]'\n if (\n re.search(pattern, port_str)\n or port_str.count('T') > 1\n or port_str.count('U') > 1\n or '-\\n' in port_str\n or '\\n-' in port_str\n or port_str[0] == '-'\n or port_str[len(port_str) - 1] == '-'\n or port_str.count(':') < (port_str.count('T') + port_str.count('U'))\n ):\n logger.error(\"Invalid port range format\")\n return True\n\n index = 0\n while index <= len(port_str) - 1:\n if port_str[index] == '-':\n try:\n int(port_str[index - 1])\n int(port_str[index + 1])\n except (TypeError, ValueError) as e:\n logger.error(\"Invalid port range format: %s\", e)\n return True\n index += 1\n\n return False\n\n\ndef ports_as_list(port_str: str) -> Tuple[Optional[List], Optional[List]]:\n \"\"\"\n Parses a ports string into two list of individual tcp and udp ports.\n\n @input string containing a port list\n e.g. T:1,2,3,5-8 U:22,80,600-1024\n\n @return two list of sorted integers, for tcp and udp ports respectively.\n \"\"\"\n if not port_str:\n logger.info(\"Invalid port value\")\n return [None, None]\n\n if ports_str_check_failed(port_str):\n logger.info(\"{0}: Port list malformed.\")\n return [None, None]\n\n tcp_list = list()\n udp_list = list()\n\n ports = port_str.replace(' ', '')\n ports = ports.replace('\\n', '')\n\n b_tcp = ports.find(\"T\")\n b_udp = ports.find(\"U\")\n\n if b_tcp != -1 and \"T:\" not in ports:\n return [None, None]\n if b_udp != -1 and \"U:\" not in ports:\n return [None, None]\n\n if len(ports) > 1 and ports[b_tcp - 1] == ',':\n ports = ports[: b_tcp - 1] + ports[b_tcp:]\n if len(ports) > 1 and ports[b_udp - 1] == ',':\n ports = ports[: b_udp - 1] + ports[b_udp:]\n\n ports = port_str_arrange(ports)\n\n tports = ''\n uports = ''\n # TCP ports listed first, then UDP ports\n if b_udp != -1 and b_tcp != -1:\n tports = ports[ports.index('T:') + 2 : ports.index('U:')]\n uports = ports[ports.index('U:') + 2 :]\n # Only UDP ports\n elif b_tcp == -1 and b_udp != -1:\n uports = ports[ports.index('U:') + 2 :]\n # Only TCP ports\n elif b_udp == -1 and b_tcp != -1:\n tports = ports[ports.index('T:') + 2 :]\n else:\n tports = ports\n\n if tports:\n for port in tports.split(','):\n port_range_expanded = port_range_expand(port)\n if '-' in port and port_range_expanded:\n tcp_list.extend(port_range_expanded)\n elif port != '' and '-' not in port:\n tcp_list.append(int(port))\n\n tcp_list.sort()\n\n if uports:\n for port in uports.split(','):\n port_range_expanded = port_range_expand(port)\n if '-' in port and port_range_expanded:\n udp_list.extend(port_range_expanded)\n elif port and '-' not in port:\n udp_list.append(int(port))\n udp_list.sort()\n\n if len(tcp_list) == 0 and len(udp_list) == 0:\n return [None, None]\n\n return (tcp_list, udp_list)\n\n\ndef get_tcp_port_list(port_str: str) -> Optional[List]:\n \"\"\"Return a list with tcp ports from a given port list in string format\"\"\"\n return ports_as_list(port_str)[0]\n\n\ndef get_udp_port_list(port_str: str) -> Optional[List]:\n \"\"\"Return a list with udp ports from a given port list in string format\"\"\"\n return ports_as_list(port_str)[1]\n\n\ndef port_list_compress(port_list: List) -> str:\n \"\"\"Compress a port list and return a string.\"\"\"\n\n if not port_list or len(port_list) == 0:\n logger.info(\"Invalid or empty port list.\")\n return ''\n\n port_list = sorted(set(port_list))\n compressed_list = []\n\n for _key, group in itertools.groupby(\n enumerate(port_list), lambda t: t[1] - t[0]\n ):\n group = list(group)\n\n if group[0][1] == group[-1][1]:\n compressed_list.append(str(group[0][1]))\n else:\n compressed_list.append(str(group[0][1]) + '-' + str(group[-1][1]))\n\n return ','.join(compressed_list)\n\n\ndef valid_port_list(port_list: str) -> bool:\n \"\"\"Validate a port list string.\n Parameters:\n port_list: string containing UDP and/or TCP\n port list as ranges or single comma\n separated ports \"\n Return True if it is a valid port list, False otherwise.\n \"\"\"\n\n # No port list provided\n if not port_list:\n return False\n\n # Remove white spaces\n port_list = port_list.replace(' ', '')\n\n # Special case is ignored.\n if port_list == 'U:,T:':\n return True\n\n # Invalid chars in the port list, like \\0 or \\n\n if ports_str_check_failed(port_list):\n return False\n\n tcp, udp = ports_as_list(port_list)\n # There is a port list but no tcp and no udp.\n if not tcp and not udp:\n return False\n\n if tcp:\n for port in tcp:\n if port < 1 or port > 65535:\n return False\n if udp:\n for port in udp:\n if port < 1 or port > 65535:\n return False\n\n return True\n", "repo_name": "greenbone/ospd-openvas", "sub_path": "ospd/network.py", "file_name": "network.py", "file_ext": "py", "file_size_in_byte": 15418, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 58, "dataset": "github-code", "pt": "24", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "socket.inet_pton", "line_number": 18, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 18, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 20, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 14, "usage_type": "name"}, {"api_name": "socket.inet_pton", "line_number": 28, "usage_type": "call"}, {"api_name": "socket.AF_INET6", "line_number": 28, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 30, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 24, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 24, "usage_type": "name"}, {"api_name": "struct.unpack", "line_number": 38, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 39, "usage_type": "call"}, {"api_name": "socket.inet_ntoa", "line_number": 42, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 42, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 34, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 34, "usage_type": "name"}, {"api_name": "socket.inet_pton", "line_number": 56, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 56, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 58, "usage_type": "attribute"}, {"api_name": "binascii.hexlify", "line_number": 68, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 77, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 48, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 48, "usage_type": "name"}, {"api_name": "socket.inet_pton", "line_number": 90, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 90, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 92, "usage_type": "attribute"}, {"api_name": "binascii.hexlify", "line_number": 98, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 103, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 104, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 82, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 82, "usage_type": "name"}, {"api_name": "socket.inet_pton", "line_number": 117, "usage_type": "call"}, {"api_name": "socket.AF_INET6", "line_number": 117, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 119, "usage_type": "attribute"}, {"api_name": "binascii.hexlify", "line_number": 125, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 133, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 138, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 109, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 109, "usage_type": "name"}, {"api_name": "socket.inet_pton", "line_number": 151, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 151, "usage_type": "attribute"}, {"api_name": "socket.inet_pton", "line_number": 152, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 152, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 153, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 143, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 143, "usage_type": "name"}, {"api_name": "binascii.hexlify", "line_number": 167, "usage_type": "call"}, {"api_name": "binascii.hexlify", "line_number": 168, "usage_type": "call"}, {"api_name": "socket.inet_ntop", "line_number": 173, "usage_type": "call"}, {"api_name": "socket.AF_INET6", "line_number": 174, "usage_type": "attribute"}, {"api_name": "struct.pack", "line_number": 174, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 162, "usage_type": "name"}, {"api_name": "socket.inet_pton", "line_number": 189, "usage_type": "call"}, {"api_name": "socket.AF_INET6", "line_number": 189, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 191, "usage_type": "attribute"}, {"api_name": "binascii.hexlify", "line_number": 194, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 198, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 181, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 181, "usage_type": "name"}, {"api_name": "socket.inet_pton", "line_number": 211, "usage_type": "call"}, {"api_name": "socket.AF_INET6", "line_number": 211, "usage_type": "attribute"}, {"api_name": "socket.inet_pton", "line_number": 212, "usage_type": "call"}, {"api_name": "socket.AF_INET6", "line_number": 212, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 213, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 203, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 203, "usage_type": "name"}, {"api_name": "re.match", "line_number": 228, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 222, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 222, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 234, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 234, "usage_type": "name"}, {"api_name": "collections.OrderedDict.fromkeys", "line_number": 289, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 289, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 267, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 267, "usage_type": "name"}, {"api_name": "socket.gethostbyname", "line_number": 297, "usage_type": "call"}, {"api_name": "socket.gaierror", "line_number": 298, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 292, "usage_type": "name"}, {"api_name": "socket.inet_pton", "line_number": 307, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 307, "usage_type": "attribute"}, {"api_name": "socket.inet_pton", "line_number": 311, "usage_type": "call"}, {"api_name": "socket.AF_INET6", "line_number": 311, "usage_type": "attribute"}, {"api_name": "socket.getfqdn", "line_number": 326, "usage_type": "call"}, {"api_name": "socket.gaierror", "line_number": 327, "usage_type": "attribute"}, {"api_name": "socket.herror", "line_number": 327, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 336, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 336, "usage_type": "name"}, {"api_name": "re.search", "line_number": 387, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 413, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 413, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 413, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 491, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 491, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 496, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 496, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 501, "usage_type": "name"}, {"api_name": "itertools.groupby", "line_number": 511, "usage_type": "call"}]}
+{"seq_id": "23008404203", "text": "from django.shortcuts import redirect, render_to_response, get_object_or_404\r\nfrom track.forms.user import LoginForm, RegisterForm, ModifyForm\r\nfrom django.contrib import auth\r\nfrom django.template import RequestContext\r\nfrom django.conf import settings\r\nfrom django.contrib.auth.decorators import login_required\r\nfrom track.models import EUser, MapUser\r\nfrom django.http import Http404\r\nfrom datetime import datetime\r\n\r\n\r\n@login_required\r\ndef unstar(request, id):\r\n user = get_object_or_404(EUser, id=id)\r\n if request.user.username:\r\n try:\r\n mapu = MapUser.objects.get(Auser_id=request.user.id, Buser_id=id)\r\n except MapUser.DoesNotExist:\r\n raise Http404\r\n else:\r\n mapu.delete()\r\n user.fans -= 1\r\n user.save()\r\n return redirect('/users/%d/detail' % user.id)\r\n else:\r\n raise Http404\r\n\r\n\r\n@login_required\r\ndef star(request, id):\r\n user = get_object_or_404(EUser, id=id)\r\n if request.user.username:\r\n try:\r\n MapUser.objects.get(Auser_id=request.user.id, Buser_id=id)\r\n except MapUser.DoesNotExist:\r\n mapu = MapUser(Auser_id=request.user.id, Buser_id=id, time=datetime.now())\r\n mapu.save()\r\n user.fans += 1\r\n user.save()\r\n return redirect('/users/%d/detail' % user.id)\r\n else:\r\n raise Http404\r\n\r\n\r\n@login_required\r\ndef detail(request, id):\r\n user = get_object_or_404(EUser, id=id)\r\n is_followed = 1\r\n try:\r\n MapUser.objects.get(Auser_id=request.user.id, Buser_id=id)\r\n except MapUser.DoesNotExist:\r\n is_followed = 0\r\n if user.first_name is '':\r\n user.first_name = '保密'\r\n if user.last_name is '':\r\n user.last_name = '保密'\r\n if user.birth is None:\r\n user.birth = '保密'\r\n if user.nickname is None:\r\n user.nickname = '保密'\r\n if user.id != request.user.id:\r\n self = 0\r\n else:\r\n self = 1\r\n return render_to_response('user/detail.html', locals(), context_instance=RequestContext(request))\r\n\r\n\r\ndef get_logout(request):\r\n auth.logout(request)\r\n return redirect('/users/login')\r\n\r\n\r\ndef get_login(request, **kwargs):\r\n auth.logout(request)\r\n return render_to_response('user/login.html', kwargs, context_instance=RequestContext(request))\r\n\r\n\r\ndef post_login(request):\r\n form = LoginForm(request.POST)\r\n if not form.is_valid():\r\n return get_login(request, errors=form.errors)\r\n\r\n user = form.get_user()\r\n auth.login(request, user)\r\n\r\n return redirect('/users/%d/detail' % user.id)\r\n\r\n\r\ndef get_register(request, **kwargs):\r\n auth.logout(request)\r\n return render_to_response('user/register.html', kwargs, context_instance=RequestContext(request))\r\n\r\n\r\ndef post_register(request):\r\n form = RegisterForm(request.POST)\r\n\r\n if not form.is_valid():\r\n return get_register(request, errors=form.errors)\r\n\r\n user = form.save()\r\n user.set_password(form.cleaned_data.get('password'))\r\n user.save()\r\n return redirect(settings.LOGIN_URL)\r\n\r\n\r\n@login_required\r\ndef get_modify(request, **kwargs):\r\n return render_to_response('user/modify.html', kwargs, context_instance=RequestContext(request))\r\n\r\n\r\n@login_required\r\ndef post_modify(request):\r\n form = ModifyForm(request.POST)\r\n\r\n if not form.is_valid():\r\n return get_modify(request, errors=form.errors)\r\n\r\n user = request.user\r\n user.first_name = form.cleaned_data.get('firstname')\r\n user.last_name = form.cleaned_data.get('lastname')\r\n user.nickname = form.cleaned_data.get('nickname')\r\n user.email = form.cleaned_data.get('email')\r\n user.sex = form.cleaned_data.get('sex')\r\n user.birth = form.cleaned_data.get('birth')\r\n user.save()\r\n return redirect('/users/%d/detail' % request.user.id)", "repo_name": "zxpgo/news-website", "sub_path": "track/views/user.py", "file_name": "user.py", "file_ext": "py", "file_size_in_byte": 3812, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "20", "api": [{"api_name": "django.shortcuts.get_object_or_404", "line_number": 14, "usage_type": "call"}, {"api_name": "track.models.EUser", "line_number": 14, "usage_type": "argument"}, {"api_name": "track.models.MapUser.objects.get", "line_number": 17, "usage_type": "call"}, {"api_name": "track.models.MapUser.objects", "line_number": 17, "usage_type": "attribute"}, {"api_name": "track.models.MapUser", "line_number": 17, "usage_type": "name"}, {"api_name": "track.models.MapUser.DoesNotExist", "line_number": 18, "usage_type": "attribute"}, {"api_name": "track.models.MapUser", "line_number": 18, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 19, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 24, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 26, "usage_type": "name"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 12, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 31, "usage_type": "call"}, {"api_name": "track.models.EUser", "line_number": 31, "usage_type": "argument"}, {"api_name": "track.models.MapUser.objects.get", "line_number": 34, "usage_type": "call"}, {"api_name": "track.models.MapUser.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "track.models.MapUser", "line_number": 34, "usage_type": "name"}, {"api_name": "track.models.MapUser.DoesNotExist", "line_number": 35, "usage_type": "attribute"}, {"api_name": "track.models.MapUser", "line_number": 35, "usage_type": "name"}, {"api_name": "track.models.MapUser", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 36, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 40, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 42, "usage_type": "name"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 29, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 47, "usage_type": "call"}, {"api_name": "track.models.EUser", "line_number": 47, "usage_type": "argument"}, {"api_name": "track.models.MapUser.objects.get", "line_number": 50, "usage_type": "call"}, {"api_name": "track.models.MapUser.objects", "line_number": 50, "usage_type": "attribute"}, {"api_name": "track.models.MapUser", "line_number": 50, "usage_type": "name"}, {"api_name": "track.models.MapUser.DoesNotExist", "line_number": 51, "usage_type": "attribute"}, {"api_name": "track.models.MapUser", "line_number": 51, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 65, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 65, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 45, "usage_type": "name"}, {"api_name": "django.contrib.auth.logout", "line_number": 69, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 69, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 70, "usage_type": "call"}, {"api_name": "django.contrib.auth.logout", "line_number": 74, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 74, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 75, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 75, "usage_type": "call"}, {"api_name": "track.forms.user.LoginForm", "line_number": 79, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 84, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 84, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 86, "usage_type": "call"}, {"api_name": "django.contrib.auth.logout", "line_number": 90, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 90, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 91, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 91, "usage_type": "call"}, {"api_name": "track.forms.user.RegisterForm", "line_number": 95, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 103, "usage_type": "call"}, {"api_name": "django.conf.settings.LOGIN_URL", "line_number": 103, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 103, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 108, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 108, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 106, "usage_type": "name"}, {"api_name": "track.forms.user.ModifyForm", "line_number": 113, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 126, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 111, "usage_type": "name"}]}
+{"seq_id": "42685496135", "text": "from selenium import webdriver\r\nfrom selenium.webdriver.chrome.options import Options\r\nimport urllib.request, urllib.parse, urllib.error\r\nfrom bs4 import BeautifulSoup\r\nimport re\r\nimport csv\r\n\r\nnumChannels = 0\r\noptions = Options()\r\noptions.headless = True\r\ndriver = webdriver.Chrome('C:/chromedriver_win32/chromedriver.exe', chrome_options=options)\r\n\r\ndef BuildURL(url):\r\n try:\r\n driver.get(url)\r\n res = driver.execute_script(\"return document.documentElement.outerHTML\")\r\n soup = BeautifulSoup(res, \"lxml\")\r\n\r\n Ad = soup.find(\"ytd-app\")\r\n Ad = Ad.find(\"div\",{'id':'content'})\r\n Ad = Ad.find(\"ytd-page-manager\")\r\n return Ad\r\n except:\r\n print(\"Check internet connection while building URL\")\r\n return None\r\n\r\n\r\ndef getRelated(url,Ad):\r\n channelList = []\r\n\r\n # url = \"https://www.youtube.com/channel/UCCTtSp0T63xo2wQ4z9x3ORQ/about\"\r\n # url = \"https://www.youtube.com/channel/UCuXE1qQ4pqFhflKeQgcqlrg/about\"\r\n # url = input()\r\n\r\n try:\r\n # BuildURL\r\n\r\n relatedChannels = Ad.find(\"ytd-browse\",{'class':'style-scope ytd-page-manager'})\r\n relatedChannels = relatedChannels.find(\"ytd-two-column-browse-results-renderer\")\r\n relatedChannels = relatedChannels.find(\"div\",{\"id\":\"secondary\"})\r\n relatedChannels = relatedChannels.find(\"ytd-browse-secondary-contents-renderer\")\r\n relatedChannels = relatedChannels.find(\"div\",{\"id\":\"contents\"})\r\n except:\r\n print(\"Check internet connection while related\")\r\n return None\r\n\r\n linkFlag = 1\r\n while True:\r\n try:\r\n currChannels = relatedChannels.contents[linkFlag].find(\"div\",{\"id\":\"items\"})\r\n currChannels = currChannels.findAll(\"ytd-mini-channel-renderer\")\r\n # print(\"LOL\")\r\n for i in currChannels:\r\n i = str(i) # i is ytd-min-channel-renderer tag\r\n i = re.findall('href=\"(\\S+)\"',i)\r\n # i collects the channel sublink from href of tag in ytd-min-channel-renderer tag\r\n channelLink = \"https://www.youtube.com\" + i[0] + \"/about\" # channelLink is the whole link to channel\r\n channelList = channelList + [channelLink]\r\n linkFlag = linkFlag + 1\r\n # print(linkFlag)\r\n except:\r\n break\r\n return channelList\r\n\r\n\r\ndef getDescription(url,Ad):\r\n # url = \"https://www.youtube.com/channel/UCCTtSp0T63xo2wQ4z9x3ORQ/about\" \r\n try:\r\n #BuildURL\r\n\r\n description = Ad.find(\"ytd-browse\",{'class':'style-scope ytd-page-manager'})\r\n description = description.find(\"ytd-two-column-browse-results-renderer\")\r\n description = description.find(\"ytd-section-list-renderer\")\r\n description = description.find(\"div\",{'id':'contents'})\r\n description = description.find(\"ytd-item-section-renderer\")\r\n description = description.find(\"div\",{'id':'contents'})\r\n description = description.find(\"ytd-channel-about-metadata-renderer\")\r\n description = description.find(\"div\",{'id':'left-column'})\r\n description = description.find(\"div\",{'id':'description-container'})\r\n description = description.find(\"yt-formatted-string\",{'id':'description'})\r\n return str(Ad)\r\n except:\r\n print(\"Check internet connection while description\")\r\n return None\r\n\r\ndef getCountry(url,Ad):\r\n try:\r\n country = Ad.find(\"ytd-browse\",{'class':'style-scope ytd-page-manager'})\r\n country = country.find(\"ytd-two-column-browse-results-renderer\")\r\n country = country.find(\"ytd-section-list-renderer\")\r\n country = country.find(\"div\",{'id':'contents'})\r\n country = country.find(\"ytd-item-section-renderer\")\r\n country = country.find(\"div\",{'id':'contents'})\r\n country = country.find(\"ytd-channel-about-metadata-renderer\")\r\n country = country.find(\"div\",{'id':'left-column'})\r\n country = country.find(\"div\",{'id':'details-container'})\r\n country = country.find(\"table\",{'class':'style-scope ytd-channel-about-metadata-renderer'})\r\n country = country.find(\"tbody\",{'class':'style-scope ytd-channel-about-metadata-renderer'})\r\n country = country.findAll(\"tr\",{'class':'style-scope ytd-channel-about-metadata-renderer'})\r\n country = country[1].findAll(\"td\",{'class':'style-scope ytd-channel-about-metadata-renderer'})\r\n country = country[0].find(\"yt-formatted-string\",{'class':'style-scope ytd-channel-about-metadata-renderer'}).contents\r\n return str(country[0])\r\n except:\r\n return \"\"\r\n\r\ndef getSubscribers(url,Ad):\r\n try:\r\n subs = Ad.findAll(\"ytd-browse\",{'class':'style-scope ytd-page-manager'})\r\n subs = subs[0].find(\"div\",{'id':'header'})\r\n subs = subs.find(\"ytd-c4-tabbed-header-renderer\")\r\n subs = subs.find(\"app-header-layout\")\r\n subs = subs.find(\"div\",{'id':'wrapper'})\r\n subs = subs.find(\"app-header\",{'id':\"header\"})\r\n subs = subs.find(\"div\",{'id':'contentContainer'})\r\n subs = subs.find(\"div\",{'id':'channel-container'})\r\n subs = subs.find(\"div\",{'id':'channel-header'})\r\n subs = subs.find(\"div\",{'id':'channel-header-container'})\r\n subs = subs.find(\"div\",{'id':'inner-header-container'})\r\n subs = subs.find(\"yt-formatted-string\",{'id':'subscriber-count'}).contents\r\n subs = subs[0].split(' ')\r\n subs = subs[0].split(',')\r\n ans = \"\"\r\n for i in subs:\r\n ans = ans + str(i)\r\n return int(ans)\r\n except:\r\n return 0\r\n\r\ndef getName(url,Ad):\r\n try:\r\n name = Ad.findAll(\"ytd-browse\",{'class':'style-scope ytd-page-manager'})\r\n name = name[0].find(\"div\",{'id':'header'})\r\n name = name.find(\"ytd-c4-tabbed-header-renderer\")\r\n name = name.find(\"app-header-layout\")\r\n name = name.find(\"div\",{'id':'wrapper'})\r\n name = name.find(\"app-header\",{'id':\"header\"})\r\n name = name.find(\"div\",{'id':'contentContainer'})\r\n name = name.find(\"div\",{'id':'channel-container'})\r\n name = name.find(\"div\",{'id':'channel-header'})\r\n name = name.find(\"div\",{'id':'channel-header-container'})\r\n name = name.find(\"div\",{'id':'inner-header-container'})\r\n name = name.find(\"h1\",{'id':'channel-title-container'})\r\n name = name.find(\"span\",{'id':'channel-title'}).contents\r\n return str(name[0])\r\n except:\r\n return \" \"\r\n\r\n# check for mail an numbr regular expressions\r\ndef getRow(url,Ad):\r\n row = []\r\n name = str(getName(url,Ad))\r\n link = str(url)\r\n subs = int(getSubscribers(url,Ad))\r\n description = str(getDescription(url,Ad))\r\n mail = \" \"\r\n number = \" \"\r\n try:\r\n mail = re.findall('\\S+@\\S+',description)\r\n mail = str(mail[0])\r\n except:\r\n mail = \" \"\r\n try:\r\n number = re.findall(\"[0-9]\\S+\",description)\r\n i = 0\r\n for i in number:\r\n if len(i) >= 10:\r\n break\r\n number = str(i)\r\n except:\r\n number = \" \"\r\n\r\n # forming current row\r\n row.append(name)\r\n row.append(link)\r\n row.append(subs)\r\n row.append(mail)\r\n row.append(number)\r\n return row\r\n\r\ndef isPresent(channels,url):\r\n for i in channels:\r\n if i[1] == url:\r\n return False\r\n return True\r\n\r\ndef printChannel(url,channels):\r\n Ad = BuildURL(url)\r\n if Ad == None:\r\n print(\"Ad\")\r\n return None\r\n if getCountry(url,Ad) != \"India\":\r\n print(\"Coun\")\r\n return None\r\n if len(channels) >= numChannels:\r\n print(\"Reached\")\r\n return None\r\n \r\n channels.append(getRow(url,Ad)) # adding url to list\r\n list = getRelated(url,Ad) # getting related channels\r\n if list == None:\r\n print(\"Related\")\r\n return None\r\n for j in list:\r\n if isPresent(channels,str(j)): # Checking repitition\r\n printChannel(str(j),channels)\r\n\r\ndef buildCSV(list):\r\n csvList = open(\"urlList.csv\",'a',encoding='utf-8')\r\n for i in list:\r\n writer = csv.writer(csvList,delimiter=' ',lineterminator='\\r')\r\n writer.writerow(i)\r\n csvList.close()\r\n\r\n\r\n\r\n\r\nnumChannels = int(input(\"Enter number of channels you want to get : \"))\r\nurl = str(input(\"Enter URL : \"))\r\n#url = \"https://www.youtube.com/channel/UCCTtSp0T63xo2wQ4z9x3ORQ/about\"\r\n# https://www.youtube.com/channel/UCFda_3iggsKi_scQFbTIJPw/about Numchannels = 5 - error why?\r\n# https://www.youtube.com/channel/UCBnnsrvmuQ7tFdTL511dzBQ/about Numchannels = 5 - no error\r\n\r\nchannels = []\r\nprintChannel(url,channels)\r\ni = int(input(\"Continue to build csv press 1 : \"))\r\nif i == 1:\r\n buildCSV(channels)\r\n\r\nfor k in channels:\r\n print(k)\r\n", "repo_name": "6sr/Python", "sub_path": "YoutubeScraper/FormYoutuberList.py", "file_name": "FormYoutuberList.py", "file_ext": "py", "file_size_in_byte": 8731, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "selenium.webdriver.chrome.options.Options", "line_number": 9, "usage_type": "call"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 11, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 11, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 17, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 55, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 158, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 163, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 210, "usage_type": "call"}]}
+{"seq_id": "12213466482", "text": "'''\nAuthor: fujiawei0724\nDate: 2022-06-07 11:01:56\nLastEditors: fujiawei0724\nLastEditTime: 2022-06-29 16:13:26\nDescription: mcts algorithm.\n'''\n\nimport sys\nsys.path.append('..')\nimport random\nimport cProfile\nimport _pickle as cPickle\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom collections import defaultdict\nfrom shapely.geometry import Polygon\n\nfrom subEnvironment import SubEnvironment, State\nfrom rl_behavior_planner.utils import *\n\nVEHICLE_INTENTION_SET = [VehicleIntention(lat_beh, lon_vel) \n for lat_beh in LateralBehavior \n for lon_vel in np.arange(-5.0, 5.0 + 1e-3, 1.0)]\n\n\n# A group of states in time order\nclass MacroState:\n moves_num = 11 * 3\n def __init__(self, states=None, lane_change_num=None, lane_info_with_speed=None, intention=None):\n self.states_ = states\n self.lane_change_num_ = lane_change_num\n self.lane_info_with_speed_ = lane_info_with_speed\n \n # Record behavior information\n self.intention_ = intention \n\n def reward(self):\n # TODO: add consideration of nonexistent lanes\n cost, is_collision, _, _, _ = MctsPolicyEvaluator.praise(self)\n if is_collision:\n return -1.0\n return 1.0 / cost\n \n def terminal(self):\n if self.states_[-1].terminal():\n return True\n return False\n\n # Generate random next state to construct the default policy\n def next_state(self, env, cur_intention):\n # Load data to environment and generate next state\n env.loadState(self.lane_info_with_speed_, self.states_[-1])\n next_state = env.simulateSingleStep(cur_intention)\n\n return next_state\n \n # Generate next macro state\n def next_macro_state(self, env):\n # Select intention randomly\n cur_intention = random.choice(VEHICLE_INTENTION_SET)\n \n # Calculate next state\n next_state = self.next_state(env, cur_intention)\n\n # Integrate ego macro state and the generated next state\n # TODO: check the logic about the copy of the current object (try to avoid the use of 'copy')\n next_macro_state = MacroState()\n next_macro_state.states_ = copy.deepcopy(self.states_)\n next_macro_state.states_.append(next_state)\n next_macro_state.lane_change_num_ = self.lane_change_num_\n next_macro_state.intention_ = cur_intention\n if cur_intention.lat_beh_ == LateralBehavior.LaneChangeLeft or cur_intention.lat_beh_ == LateralBehavior.LaneChangeRight:\n next_macro_state.lane_change_num_ += 1\n next_macro_state.lane_info_with_speed_ = self.lane_info_with_speed_\n\n return next_macro_state\n \n # Visualization lane situation and vehicles states sequence\n def visualization(self, ax):\n # Construct lane server\n center_lane = None\n left_lane = None\n right_lane = None\n left_lane_exist, right_lane_exist, center_left_distance, center_right_distance = self.lane_info_with_speed_[0], self.lane_info_with_speed_[1], self.lane_info_with_speed_[2], self.lane_info_with_speed_[3]\n \n # Initialize lane with the assumption that the lane has 500m to drive at least\n center_lane_start_point = PathPoint(0.0, 0.0)\n center_lane_end_point = PathPoint(500.0, 0.0)\n center_lane = Lane(center_lane_start_point, center_lane_end_point, LaneId.CenterLane)\n # center_lane_points_array = Visualization.transformPathPointsToArray(center_lane.path_points_)\n if left_lane_exist:\n left_lane_start_point = PathPoint(0.0, center_left_distance)\n left_lane_end_point = PathPoint(500.0, center_left_distance)\n left_lane = Lane(left_lane_start_point, left_lane_end_point, LaneId.LeftLane)\n # left_lane_points_array = Visualization.transformPathPointsToArray(left_lane.path_points_)\n if right_lane_exist:\n right_lane_start_point = PathPoint(0.0, -center_right_distance)\n right_lane_end_point = PathPoint(500.0, -center_right_distance)\n right_lane = Lane(right_lane_start_point, right_lane_end_point, LaneId.RightLane)\n # right_lane_points_array = Visualization.transformPathPointsToArray(right_lane.path_points_)\n\n # Construct lane server\n lanes = dict()\n lanes[center_lane.id_] = center_lane\n if left_lane_exist:\n lanes[left_lane.id_] = left_lane\n if right_lane_exist:\n lanes[right_lane.id_] = right_lane\n lane_server = LaneServer(lanes)\n\n # Visualization lanes\n if LaneId.CenterLane in lane_server.lanes_:\n center_lane = lane_server.lanes_[LaneId.CenterLane]\n center_lane_points_array = Visualization.transformPathPointsToArray(center_lane.path_points_)\n ax.plot(center_lane_points_array[:, 0], center_lane_points_array[:, 1], c='m', linewidth=1.0)\n ax.plot(center_lane.left_boundary_points_[:, 0], center_lane.left_boundary_points_[:, 1], c='black',\n ls='--', linewidth=1.0)\n ax.plot(center_lane.right_boundary_points_[:, 0], center_lane.right_boundary_points_[:, 1], c='black',\n ls='--', linewidth=1.0)\n if LaneId.LeftLane in lane_server.lanes_:\n left_lane = lane_server.lanes_[LaneId.LeftLane]\n left_lane_points_array = Visualization.transformPathPointsToArray(left_lane.path_points_)\n ax.plot(left_lane_points_array[:, 0], left_lane_points_array[:, 1], c='m', linewidth=1.0)\n ax.plot(left_lane.left_boundary_points_[:, 0], left_lane.left_boundary_points_[:, 1], c='black', ls='--',\n linewidth=1.0)\n ax.plot(left_lane.right_boundary_points_[:, 0], left_lane.right_boundary_points_[:, 1], c='black', ls='--',\n linewidth=1.0)\n if LaneId.RightLane in lane_server.lanes_:\n right_lane = lane_server.lanes_[LaneId.RightLane]\n right_lane_points_array = Visualization.transformPathPointsToArray(right_lane.path_points_)\n ax.plot(right_lane_points_array[:, 0], right_lane_points_array[:, 1], c='m', linewidth=1.0)\n ax.plot(right_lane.left_boundary_points_[:, 0], right_lane.left_boundary_points_[:, 1], c='black', ls='--',\n linewidth=1.0)\n ax.plot(right_lane.right_boundary_points_[:, 0], right_lane.right_boundary_points_[:, 1], c='black',\n ls='--', linewidth=1.0)\n \n\n # Transform to trajectories\n ego_veh_states = []\n sur_vehs_states = defaultdict(list)\n for state in self.states_:\n ego_veh_states.append(state.ego_vehicle_)\n for sur_veh_id, sur_veh in state.surround_vehicles_.items():\n sur_vehs_states[sur_veh_id].append(sur_veh)\n ego_traj = Trajectory(ego_veh_states)\n sur_trajs = dict()\n for sur_veh_id, sur_veh_states in sur_vehs_states.items():\n sur_trajs[sur_veh_id] = Trajectory(sur_veh_states)\n \n # Visualization trajectories\n traj_length = len(ego_traj.vehicle_states_)\n for i in range(0, traj_length):\n if i == 0:\n # For current position\n ego_vehicle_polygon = Polygon(ego_traj.vehicle_states_[i].rectangle_.vertex_)\n ax.plot(*ego_vehicle_polygon.exterior.xy, c='r')\n # ax.text(ego_vehicle.position_.x_, ego_vehicle.position_.y_, 'id: {}, v: {}'.format(ego_vehicle.id_, ego_vehicle.velocity_), size=10.0)\n # Traverse surround vehicle\n for sur_veh_id, sur_veh_tra in sur_trajs.items():\n sur_vehicle_polygon = Polygon(sur_veh_tra.vehicle_states_[i].rectangle_.vertex_)\n ax.plot(*sur_vehicle_polygon.exterior.xy, c='green')\n # ax.text(sur_veh_tra.vehicle_states_[i].position_.x_, sur_veh_tra.vehicle_states_[i].position_.y_, 'id: {}, v: {}'.format(sur_veh_id, sur_veh_tra.vehicle_states_[i].velocity_), size=10.0)\n\n else:\n # For predicted position\n # For current position\n ego_vehicle_polygon = Polygon(ego_traj.vehicle_states_[i].rectangle_.vertex_)\n ax.plot(*ego_vehicle_polygon.exterior.xy, c='r', ls='--')\n # ax.text(lane_keeping_ego_trajectory.vehicle_states_[i].position_.x_, lane_keeping_ego_trajectory.vehicle_states_[i].position_.y_, 'id: {}, v: {}, time stamp: {}'.format(ego_vehicle.id_, lane_keeping_ego_trajectory.vehicle_states_[i].velocity_, lane_keeping_ego_trajectory.vehicle_states_[i].time_stamp_), size=10.0)\n # Traverse surround vehicle\n for sur_veh_id, sur_veh_tra in sur_trajs.items():\n sur_vehicle_polygon = Polygon(sur_veh_tra.vehicle_states_[i].rectangle_.vertex_)\n ax.plot(*sur_vehicle_polygon.exterior.xy, c='green', ls='--')\n # ax.text(sur_veh_tra.vehicle_states_[i].position_.x_, sur_veh_tra.vehicle_states_[i].position_.y_, 'id: {}, v: {}, time stamp: {}'.format(sur_veh_id, sur_veh_tra.vehicle_states_[i].velocity_, sur_veh_tra.vehicle_states_[i].time_stamp_), size=10.0)\n\n\n \n \n\n\n\n \n\n# Node in the search tree\nclass Node:\n def __init__(self, macro_state, parent=None):\n self.visit_num_ = 1\n self.reward_ = 0.0\n self.macro_state_ = macro_state\n self.children_ = []\n self.parent_ = parent\n \n def add_child(self, child_macro_state):\n child = Node(child_macro_state, self)\n self.children_.append(child)\n \n def update(self, reward):\n self.reward_ += reward\n self.visit_num_ += 1\n \n def fully_expanded(self):\n # TODO: add domain knowledge here to limit the scale of the search tree\n if len(self.children_) == self.macro_state_.moves_num:\n return True\n return False\n \n def best_policy(self):\n best_score = -np.inf\n best_children = []\n for c in self.children_:\n score = c.reward_ / c.visit_num_\n if score == best_score:\n best_children.append(c)\n if score > best_score:\n best_children = [c]\n best_score = score\n if len(best_children) == 0:\n print('Fatal error!!!')\n return random.choice(best_children)\n \n# Generate reward for the states sequence\nclass MctsPolicyEvaluator(PolicyEvaluator):\n \n @classmethod\n def calculateMultiLaneChangeCost(cls, change_num):\n return change_num * 0.3\n\n @classmethod\n def praise(cls, macro_state):\n # Reconstruct data \n ego_states = []\n sur_states = defaultdict(list)\n for st in macro_state.states_:\n ego_states.append(st.ego_vehicle_)\n for sur_id, sur_veh in st.surround_vehicles_.items():\n sur_states[sur_id].append(sur_veh)\n ego_traj = Trajectory(ego_states)\n sur_trajs = dict()\n for s_id, s_states in sur_states.items():\n sur_trajs[s_id] = Trajectory(s_states)\n \n # Calculate cost\n safety_cost, is_collision = cls.calculateSafetyCost(ego_traj, sur_trajs, macro_state.lane_info_with_speed_[-1])\n lane_change_cost = cls.calculateMultiLaneChangeCost(macro_state.lane_info_with_speed_[-1])\n efficiency_cost = cls.calculateEfficiencyCost(ego_traj, macro_state.lane_info_with_speed_[-1])\n comfort_cost = cls.calculateComfortCost(ego_traj)\n\n # # DEBUG\n # print('Safety cost: {}'.format(safety_cost))\n # print('Lane change cost: {}'.format(lane_change_cost))\n # print('Efficiency cost: {}'.format(efficiency_cost))\n # print('Comfort cost: {}'.format(comfort_cost))\n # print('All cost: {}'.format(safety_cost + lane_change_cost + efficiency_cost + comfort_cost))\n # # END DEBUG\n\n return safety_cost + lane_change_cost + efficiency_cost + comfort_cost, is_collision, safety_cost, lane_change_cost, efficiency_cost\n \n# Training the tree policy\nclass TreePolicyTrainer:\n def __init__(self, round_limit, time_limit, scalar):\n self.round_limit_ = round_limit\n self.time_limit_ = time_limit\n self.scalar_ = scalar\n \n '''\n description: train the tree policy\n param {root} root node of the search tree\n return {root} the result of the training process\n ''' \n def train(self, root, env):\n for iter in range(self.round_limit_):\n print('-------------------Start No. {} training epoch-------------------'.format(iter))\n front = self.tree_policy(root, env)\n reward = self.default_policy(front.macro_state_, env)\n self.backup(front, reward)\n return root\n \n '''\n description: stretch a tree node\n param {node} start node, also the node manipulated \n return {node} the ternimal node in the branch of the start node\n ''' \n def tree_policy(self, node, env):\n while node.macro_state_.terminal() == False:\n if len(node.children_) == 0:\n return self.expand(node, env)\n elif random.uniform(0, 1) < 0.5:\n node = self.best_child(node)\n else:\n if node.fully_expanded() == False:\n return self.expand(node, env)\n else:\n node = self.best_child(node)\n return node\n \n def expand(self, node, env):\n tried_children = [c.macro_state_ for c in node.children_]\n next_macro_state = node.macro_state_.next_macro_state(env)\n while next_macro_state in tried_children and next_macro_state.terminal() == False:\n next_macro_state = node.macro_state_.next_macro_state(env)\n node.add_child(next_macro_state)\n return node.children_[-1]\n\n def best_child(self, node):\n best_score = -np.inf\n best_children = []\n for c in node.children_:\n exploit = c.reward_ / c.visit_num_\n explore = np.sqrt(2.0 * np.log(node.visit_num_) / float(c.visit_num_))\n score = exploit + self.scalar_ * explore\n if score == best_score:\n best_children.append(c)\n if score > best_score:\n best_children = [c]\n best_score = score\n if len(best_children) == 0:\n print('Fatal error!!!')\n return random.choice(best_children)\n\n def default_policy(self, macro_state, env):\n while macro_state.terminal() == False:\n macro_state = macro_state.next_macro_state(env)\n\n # # DEBUG\n # macro_state.states_[-1].ego_vehicle_.print()\n # # END DEBUG\n\n return macro_state.reward()\n\n def backup(self, node, reward):\n while node != None:\n node.visit_num_ += 1\n node.reward_ += reward\n node = node.parent_\n return \n \n \n\nif __name__ == '__main__':\n # Load environment data randomly\n random.seed(0)\n left_lane_exist = random.randint(0, 1)\n right_lane_exist = random.randint(0, 1)\n center_left_distance = random.uniform(3.0, 4.5)\n center_right_distance = random.uniform(3.0, 4.5)\n lane_info = [left_lane_exist, right_lane_exist, center_left_distance, center_right_distance]\n lane_speed_limit = random.uniform(10.0, 25.0)\n lane_info_with_speed = [left_lane_exist, right_lane_exist, center_left_distance, center_right_distance, lane_speed_limit]\n\n # Construct ego vehicle and surround vehicles randomly\n ego_vehicle = EgoInfoGenerator.generateOnce()\n surround_vehicles_generator = AgentGenerator(left_lane_exist, right_lane_exist, center_left_distance, center_right_distance)\n surround_vehicles = surround_vehicles_generator.generateAgents(random.randint(0, 10))\n random.seed()\n\n # Block the situation with a wrong initialization information\n while ego_vehicle.velocity_ > lane_speed_limit:\n print('Reset vehicle information')\n ego_vehicle = EgoInfoGenerator.generateOnce()\n\n # Construct environment\n env = SubEnvironment()\n\n # Construct trainer\n scalar = 1.0 / (2.0 * np.sqrt(2.0))\n mcts_trainer = TreePolicyTrainer(1000, None, scalar)\n\n # Initialize start node \n root = Node(MacroState([State(ego_vehicle, surround_vehicles, 0.0)], 0, lane_info_with_speed))\n mcts_trainer.train(root, env)\n\n # # Test running performance\n # cProfile.run('mcts_trainer.train(root, env)')\n\n # Output the result\n while len(root.children_) != 0:\n root = root.best_policy()\n root.macro_state_.intention_.print()\n root.macro_state_.intention_.print()\n\n # Visualization\n fig = plt.figure(0)\n ax = plt.axes()\n ax.axis('equal')\n root.macro_state_.visualization(ax)\n plt.show()\n\n \n ", "repo_name": "fujiawei0724/motion_planning_scripts", "sub_path": "mcts_planner/mcts.py", "file_name": "mcts.py", "file_ext": "py", "file_size_in_byte": 16835, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 4, "dataset": "github-code", "pt": "20", "api": [{"api_name": "sys.path.append", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 24, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 61, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 141, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 156, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 161, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 168, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 209, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 220, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 233, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 288, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 306, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 310, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 310, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 319, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 342, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 343, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 344, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 345, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 346, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 348, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 354, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 355, "usage_type": "call"}, {"api_name": "subEnvironment.SubEnvironment", "line_number": 363, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 366, "usage_type": "call"}, {"api_name": "subEnvironment.State", "line_number": 370, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 383, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 383, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 384, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 384, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 387, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 387, "usage_type": "name"}]}
+{"seq_id": "23484700980", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nimport getopt\nimport sys\n\nfrom PIL import Image\nimport numpy as np\nfrom transforms.DqKT import DqKT\n# from transforms_gpgpu.dqkt_gpgpu import DqktGPGPU\nfrom transforms.DAT import DAT\nfrom block_tools.BlockTools import BlockTools\nfrom qr_tools.MyQR62 import MyQR62\n\nfrom scipy import misc\nimport math\n\n\ndef binary2int(binary):\n # Devuelve el entero correspondiente a una lista de binarios\n n = len(binary)\n v = 0\n for i in range(n):\n v += (2**(n-i-1))*binary[i]\n return v\n\n\ndef get_dwt(chromosome):\n \"\"\"\n Devuelve la subbanda de la DWT a utilizar (0, 1, 2, 3) -> (LL, LH, HL, HH)\n \"\"\"\n return binary2int(chromosome[0:2])\n\n\ndef zigzag(n):\n indexorder = sorted(\n ((x, y) for x in range(n) for y in range(n)), key=lambda s: (s[0]+s[1], -s[1] if (s[0]+s[1]) % 2 else s[1]))\n return {index: n for n, index in enumerate(indexorder)}\n\n\ndef get_indice(m):\n zarray = zigzag(8)\n indice = []\n n = int(len(zarray) ** 0.5 + 0.5)\n for x in range(n):\n for y in range(n):\n if zarray[(x, y)] == m:\n indice.append(x)\n indice.append(y)\n return indice\n\n\ndef mypwlcm(dic, valores):\n '''\n dic: diccionario compuesto por la semila y el valor de p\n valores: lista de valores a tratar\n '''\n valores_finales = []\n cantidad_valores = len(valores)\n posiciones = orden(dic, cantidad_valores)\n # print 'Posiciones: ', posiciones\n posiciones_distintas = lista_valores_distintos(posiciones)\n # print 'Posiciones distintas: ', posiciones_distintas\n if len(posiciones_distintas) == len(posiciones):\n v = []\n for i in range(len(posiciones_distintas)):\n v.append(valores[i])\n return v\n if len(posiciones_distintas) == 1:\n return valores\n for i in range(len(posiciones_distintas)):\n valores_finales.append(valores[posiciones_distintas[i]])\n posiciones_faltantes = lista_valores_faltantes(\n posiciones_distintas, cantidad_valores)\n # print 'Faltantes: ', posiciones_faltantes\n if len(posiciones_faltantes) > 0:\n v = []\n for i in range(len(posiciones_faltantes)):\n v.append(valores[posiciones_faltantes[i]])\n valores_finales.extend(mypwlcm(dic, v))\n return valores_finales\n\n\ndef pwlcm(dic):\n # dic: diccionario compuesto por la semila y el valor de p\n if (dic['semilla'] >= 0) and (dic['semilla'] < dic['p']):\n x = dic['semilla'] / dic['p']\n elif (dic['semilla'] >= dic['p']) and (dic['semilla'] < 0.5):\n x = (dic['semilla'] - dic['p']) / (0.5 - dic['p'])\n elif (dic['semilla'] >= 0.5) and (dic['semilla'] < 1):\n dic['semilla'] = 1 - dic['semilla']\n x = pwlcm(dic)\n return x\n\n\ndef orden(dic, cant=40):\n lista = []\n for i in range(cant):\n if i != 0:\n dic_a = {}\n dic_a['semilla'] = temp\n dic_a['p'] = dic['p']\n temp = pwlcm(dic_a)\n else:\n temp = pwlcm(dic)\n lista.append(int(math.floor(math.fmod(temp * 10**14, cant))))\n return lista\n\n\ndef lista_valores_distintos(lista):\n lista_nueva = []\n for i in lista:\n if i not in lista_nueva:\n lista_nueva.append(i)\n return lista_nueva\n\n\ndef lista_valores_faltantes(posiciones_distintas, cantidad_valores):\n posiciones_faltantes = []\n for i in range(cantidad_valores):\n if i not in posiciones_distintas:\n posiciones_faltantes.append(i)\n return posiciones_faltantes\n\n\ndef mypwlcm_limit(dic, valores, limite):\n '''\n dic: diccionario compuesto por la semila y el valor de p\n valores: lista de valores a tratar\n '''\n valores_finales = []\n cantidad_valores = len(valores)\n posiciones = orden(dic, cantidad_valores)\n # print 'Posiciones: ', posiciones\n posiciones_distintas = lista_valores_distintos(posiciones)\n # print 'Posiciones distintas: ', posiciones_distintas\n if len(valores_finales) > limite:\n return valores_finales\n if len(posiciones_distintas) == len(posiciones):\n v = []\n for i in range(len(posiciones_distintas)):\n v.append(valores[i])\n return v\n if len(posiciones_distintas) == 1:\n return valores\n for i in range(len(posiciones_distintas)):\n valores_finales.append(valores[posiciones_distintas[i]])\n posiciones_faltantes = lista_valores_faltantes(\n posiciones_distintas, cantidad_valores)\n # print 'Faltantes: ', posiciones_faltantes\n if len(posiciones_faltantes) > 0:\n v = []\n for i in range(len(posiciones_faltantes)):\n v.append(valores[posiciones_faltantes[i]])\n valores_finales.extend(mypwlcm_limit(dic, v, limite))\n return valores_finales\n\n\ndef extract(watermarked_filename):\n delta = 128\n c = [1, 19]\n from image_tools.ImageTools import ImageTools\n dqkt = DqKT()\n myqr = MyQR62()\n dat = DAT()\n itools = ImageTools()\n\n watermarked_image = Image.open(watermarked_filename)\n watermarked_ycbcr_image = itools.rgb2ycbcr(watermarked_image)\n watermarked_array = watermarked_ycbcr_image[:, :, 0]\n bt_of_watermarked_image_without_noise = BlockTools(watermarked_array)\n\n extract = []\n\n len_of_watermark = 3844\n\n # Utilizar Bloques segun key\n dic = {'semilla': 0.00325687, 'p': 0.22415897}\n valores = []\n cantidad = bt_of_watermarked_image_without_noise.max_blocks()\n for i in range(cantidad):\n valores.append(i)\n v = mypwlcm_limit(dic, valores, len_of_watermark)\n\n for i in range(len_of_watermark):\n\n dqkt_block = dqkt.dqkt2(\n np.array(\n bt_of_watermarked_image_without_noise.get_block(v[i]+1),\n dtype=np.float32))\n\n negative = False\n if dqkt_block[get_indice(c[1])[0], get_indice(c[1])[1]] < 0:\n negative = True\n\n C1 = (2*delta*round(abs(dqkt_block[get_indice(c[1])[0], get_indice(c[1])[1]])/(2.0*delta)) + delta/2.0) - abs(dqkt_block[get_indice(c[1])[0], get_indice(c[1])[1]])\n C0 = (2*delta*round(abs(dqkt_block[get_indice(c[1])[0], get_indice(c[1])[1]])/(2.0*delta)) - delta/2.0) - abs(dqkt_block[get_indice(c[1])[0], get_indice(c[1])[1]])\n\n if negative:\n C1 *= -1\n C0 *= -1\n if C0 < C1:\n extract.append(0)\n else:\n extract.append(1)\n\n wh = int(math.sqrt(len_of_watermark))\n extract_image = Image.new(\"1\", (wh, wh), 255)\n array_extract_image = misc.fromimage(extract_image)\n\n for i in range(wh):\n for y in range(wh):\n if extract[wh*i+y] == 0:\n array_extract_image[i, y] = 0\n\n watermark_array_image = misc.toimage(array_extract_image)\n for i in range(10):\n watermark_array_image = dat.dat2(watermark_array_image)\n\n b = BlockTools(misc.fromimage(watermark_array_image), 2, 2)\n for m in range(1, b.max_blocks()+1):\n b.set_color(m)\n\n return misc.toimage(myqr.get_resconstructed(b.get()))\n\n\ndef main(args):\n input_filename = None\n watermark_filename = None\n output_filename = None\n\n # print('ARGV :', sys.argv[1:])\n\n options, remainder = getopt.getopt(\n sys.argv[1:], 'i:w:o:v',\n ['input', 'watermark', 'output='])\n # print('OPTIONS :', options)\n\n for opt, arg in options:\n if opt in ('-i', '--input'):\n input_filename = arg\n elif opt in ('-w', '--watermark'):\n watermark_filename = arg\n elif opt in ('-o', '--output'):\n output_filename = arg\n\n if input_filename:\n extract_watermark = extract(input_filename)\n extract_watermark.save(output_filename, quality=100)\n\n return 0\n\n\nif __name__ == '__main__':\n import sys\n sys.exit(main(sys.argv))\n", "repo_name": "eadomenech/benchmark", "sub_path": "watermarking/static/watermarking/methods_examples/my_extract.py", "file_name": "my_extract.py", "file_ext": "py", "file_size_in_byte": 7791, "program_lang": "python", "lang": "es", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "20", "api": [{"api_name": "math.floor", "line_number": 105, "usage_type": "call"}, {"api_name": "math.fmod", "line_number": 105, "usage_type": "call"}, {"api_name": "transforms.DqKT.DqKT", "line_number": 162, "usage_type": "call"}, {"api_name": "qr_tools.MyQR62.MyQR62", "line_number": 163, "usage_type": "call"}, {"api_name": "transforms.DAT.DAT", "line_number": 164, "usage_type": "call"}, {"api_name": "image_tools.ImageTools.ImageTools", "line_number": 165, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 167, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 167, "usage_type": "name"}, {"api_name": "block_tools.BlockTools.BlockTools", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 189, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 206, "usage_type": "call"}, {"api_name": "PIL.Image.new", "line_number": 207, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 207, "usage_type": "name"}, {"api_name": "scipy.misc.fromimage", "line_number": 208, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 208, "usage_type": "name"}, {"api_name": "scipy.misc.toimage", "line_number": 215, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 215, "usage_type": "name"}, {"api_name": "block_tools.BlockTools.BlockTools", "line_number": 219, "usage_type": "call"}, {"api_name": "scipy.misc.fromimage", "line_number": 219, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 219, "usage_type": "name"}, {"api_name": "scipy.misc.toimage", "line_number": 223, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 223, "usage_type": "name"}, {"api_name": "getopt.getopt", "line_number": 233, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 234, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 255, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 255, "usage_type": "attribute"}]}
+{"seq_id": "13136919447", "text": "from datetime import datetime\n\nnow = datetime.now() # time object\n\nprint(\"now =\", now)\n\nfile1 = open('myfile.txt', 'w')\nL = [\"This is Delhi \\n\", \"This is Paris \\n\", \"This is London \\n\"]\ns = \"Hello\\n\"\n \n# Writing a string to file\nfile1.write(s)\n \n# Writing multiple strings\n# at a time\nfile1.writelines(L)\n \n# Closing file\nfile1.close()\n\nnow2 = datetime.now() # time object\n\nprint(\"now2 =\", now2)\nprint(\"difence\", now2-now)", "repo_name": "ermsharo/EP_Modelagem", "sub_path": "logs.py", "file_name": "logs.py", "file_ext": "py", "file_size_in_byte": 425, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "datetime.datetime.now", "line_number": 3, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 3, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "name"}]}
+{"seq_id": "29845512816", "text": "\"\"\"add name\n\nRevision ID: 32974039eb89\nRevises: f35e1eca70f7\nCreate Date: 2022-12-18 19:51:20.619707\n\n\"\"\"\nimport sqlalchemy as sa\n\nfrom alembic import op\n\n# revision identifiers, used by Alembic.\nrevision = \"32974039eb89\"\ndown_revision = \"f35e1eca70f7\"\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n op.add_column(\"plugin_testapi_reviews\", sa.Column(\"name\", sa.Text(), nullable=True))\n # ### end Alembic commands ###\n\n\ndef downgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n op.drop_column(\"plugin_testapi_reviews\", \"name\")\n # ### end Alembic commands ###\n", "repo_name": "bitcartcc/sample-plugin", "sub_path": "src/backend/testapi/versions/32974039eb89_add_name.py", "file_name": "32974039eb89_add_name.py", "file_ext": "py", "file_size_in_byte": 672, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "20", "api": [{"api_name": "alembic.op.add_column", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op.drop_column", "line_number": 27, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 27, "usage_type": "name"}]}
+{"seq_id": "6086577225", "text": "from __future__ import absolute_import, unicode_literals\n\nimport os\nimport json\nfrom django.conf import settings\nimport redis\nfrom celery import Celery\nfrom zipfile import ZipFile\nfrom io import BytesIO\nfrom urllib.request import Request, urlopen\nfrom datetime import date\n\nos.environ.setdefault('DJANGO_SETTINGS_MODULE', 'backend.settings')\napp = Celery('backend')\napp.config_from_object('django.conf:settings', namespace='CELERY')\napp.conf.enable_utc = False\napp.autodiscover_tasks()\n\n\n@app.task(bind=True)\ndef wake_up(self):\n redis_instance = redis.StrictRedis(host=settings.REDIS_HOST,\n port=settings.REDIS_PORT,\n password=settings.REDIS_PASSWORD,\n db=0)\n today = str(date.today())\n todays_date = list(today.split('-'))\n todays_date.reverse()\n dt = \"\"\n for x in todays_date:\n dt += x[-2:]\n req = Request(f'https://www.bseindia.com/download/BhavCopy/Equity/EQ{dt}_CSV.zip', headers={'User-Agent': 'Mozilla/5.0'})\n with ZipFile(BytesIO(urlopen(req).read())) as my_zip_file:\n for contained_file in my_zip_file.namelist():\n for line in my_zip_file.open(contained_file).readlines():\n output = line.decode()\n lst_info = output.split(',')\n dct = {\n \"code\": lst_info[0],\n \"name\": lst_info[1],\n \"open\": lst_info[4],\n \"high\": lst_info[5],\n \"low\": lst_info[6],\n \"close\": lst_info[7]\n }\n json_dct = json.dumps(dct)\n redis_instance.set(lst_info[1], json_dct)\n\n\n\n\n\n\n\n\n\n\n# @app.task(bind=True)\n# def debug_task(self):\n# print('Request: {0!r}'.format(self.request))\n\n# @app.task(bind=True)\n# def send_import_summary(self):\n# print(\"Hello world\")", "repo_name": "9643kavinder/bse-stator-backend", "sub_path": "backend/celery.py", "file_name": "celery.py", "file_ext": "py", "file_size_in_byte": 1908, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "os.environ.setdefault", "line_number": 13, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 13, "usage_type": "attribute"}, {"api_name": "celery.Celery", "line_number": 14, "usage_type": "call"}, {"api_name": "redis.StrictRedis", "line_number": 22, "usage_type": "call"}, {"api_name": "django.conf.settings.REDIS_HOST", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 22, "usage_type": "name"}, {"api_name": "django.conf.settings.REDIS_PORT", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.settings.REDIS_PASSWORD", "line_number": 24, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 24, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 26, "usage_type": "name"}, {"api_name": "urllib.request.Request", "line_number": 32, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 33, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 33, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 33, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 46, "usage_type": "call"}]}
+{"seq_id": "15311865107", "text": "# Primeiro mini-projeto\nfrom tweepy.streaming import StreamListener \nfrom tweepy import OAuthHandler\nfrom tweepy import Stream\nfrom datetime import datetime\nimport json\nimport os\n\n# Consumer Key\nconsumer_key = os.environ['tweeter_api_key']\n\n# Consumer Secret\nconsumer_secret = os.environ['tweeter_api_secret']\n\n# Access Token\naccess_token = os.environ['tweeter_token']\n\n# Token Secret\naccess_token_secret = os.environ['tweeter_token_secret']\n\n# Criando a autenticação\nauth = OAuthHandler(consumer_key, consumer_secret)\n\nauth.set_access_token(access_token, access_token_secret)\n\n# Classe para capturar stream de dados do tweeter\nclass MyListener(StreamListener):\n def on_data(self, dados):\n tweet = json.loads(dados)\n created_at = tweet['created_at']\n id_str = tweet['id_str']\n text = tweet['text']\n obj = {'created_at':created_at, 'id_str':id_str, 'text':text}\n tweetind = col.insertone(obj).inserted_id\n print(obj)\n return True\n\n# Criando o objeto mylistener\nmy_listener = MyListener()\n\nmy_stream = Stream(auth, listener = MyListener)", "repo_name": "EndriwMichel/estudoPython", "sub_path": "PythonDsa/cap6/mini_projeto.py", "file_name": "mini_projeto.py", "file_ext": "py", "file_size_in_byte": 1096, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "24", "api": [{"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 19, "usage_type": "attribute"}, {"api_name": "tweepy.OAuthHandler", "line_number": 22, "usage_type": "call"}, {"api_name": "tweepy.streaming.StreamListener", "line_number": 27, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 29, "usage_type": "call"}, {"api_name": "tweepy.Stream", "line_number": 41, "usage_type": "call"}]}
+{"seq_id": "12475773932", "text": "# coding=UTF-8\r\nimport requests\r\n\r\n#将网页转化成列表,\r\nurl = 'http://www.****.com/'\r\nheader = {'User-Agent': '***********'}\r\nresponse = requests.get(url,headers=header).text\r\n#切割字符串,转列表\r\na = response.split('\\n')\r\nz = []\r\n#for 循环读取每一个字符串,删除其中的\\n\\t\\d和空格,\r\nfor i in a:\r\n responses = i.strip()\r\n #组成新的列表\r\n z.append(responses)\r\n#去除列表中的None和空字符串\r\nc = list(filter(None, z))\r\nprint(c)\r\n\r\n#这是对比两个网页的不同之处\r\na = ['', '']\r\nb = ['a', '']\r\naa = set(a)\r\nbb = set(b)\r\na_b = bb - aa\r\nprint(str(a_b))\r\n", "repo_name": "miaoyongsen/look-here-now-", "sub_path": "网页转列表-对比两个网页不同之处.py", "file_name": "网页转列表-对比两个网页不同之处.py", "file_ext": "py", "file_size_in_byte": 776, "program_lang": "python", "lang": "zh", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "requests.get", "line_number": 7, "usage_type": "call"}]}
+{"seq_id": "2884177946", "text": "import os\nfrom rest_framework.response import Response\nfrom rest_framework.decorators import api_view\nfrom api.settings import PROJECT_ROOT\nfrom api.util.SentimentData import SentimentData\nimport pandas as pd\nfrom api.serialziers import Dictionary, DictionarySerializer\nimport pickle\nfrom rest_framework import serializers\n\n\n@api_view(['GET'])\ndef get_sentiment_analysis(request):\n ticker = request.GET.get('ticker', '')\n print(request.data)\n with open(os.path.join(PROJECT_ROOT + '/FinBert.pkl'), 'rb') as f:\n finbert = pickle.load(f)\n hot_posts = SentimentData.get_raddit_data(ticker)\n hot_tweets = SentimentData.get_stock_twit_data(ticker)\n data = []\n analysis = 0\n\n for tweet in hot_tweets:\n if len(tweet['body']) == 0 or len(tweet['body']) > 200:\n continue\n result = finbert(tweet['body'])\n\n score = calculateScore(result[0][\"score\"], result[0]['label'])\n analysis = analysis + score\n data.append({\n \"analysis\": score,\n \"label\": result[0]['label'],\n \"paragraph\": tweet['body'],\n \"name\": tweet['user']['name'],\n \"avatar_url\": tweet['user']['avatar_url'],\n \"like_count\": tweet['user']['like_count'],\n \"username\": tweet['user']['username'],\n \"followers\": tweet['user']['followers'],\n \"type\": \"stock_tweet\"\n })\n\n for post in hot_posts:\n if len(post.title) == 0 or len(post.title) > 3000:\n continue\n result = finbert(post.title)\n score = calculateScore(result[0][\"score\"], result[0]['label'])\n analysis = analysis + score\n data.append({\n \"analysis\": score,\n \"label\": result[0]['label'],\n \"title\": post.title,\n \"paragraph\": post.selftext,\n \"url\": post.url,\n \"likes\": post.likes,\n # \"subreddit\": post.subreddit,\n \"ups\": post.ups,\n \"subreddit_subscribers\": post.subreddit_subscribers,\n \"downs\": post.downs,\n \"vote\": post.ups-post.downs,\n \"type\": \"reddit\",\n })\n # analysis = analysis/\n\n analysis = analysis / (len(data))\n # GenericSzl = getGenericSerializer(model)\n dictionary = Dictionary({\n \"complete_analysis\": analysis,\n \"data\": data\n })\n return Response(DictionarySerializer(dictionary).data)\n\n\ndef calculateScore(score, label):\n value = 50\n # score = score/100\n if label == 'neutral':\n if score < 0.6:\n value = value - 10 * score\n else:\n value = value + 10 * score\n elif label == 'positive':\n value = value + 30 * score\n elif label == 'negative':\n value = value - 30 * score\n\n return value\n# result = finbert(analysis_text)\n# return Response(analysis)\n\n# def get_queryset(self):\n# model = self.kwargs.get('model')\n# return getattr(models, model).objects.all()\n# def getGenericSerializer(model_arg):\n# class GenericSerializer(serializers.ModelSerializer):\n# class Meta:\n# model = model_arg\n# fields = '__all__'\n#\n# return GenericSerializer\n\n# finbert = pickle.load(open(os.path.join(PROJECT_ROOT+'/FinBert.pkl')))\n", "repo_name": "hurrairaa/stock-analysis", "sub_path": "api/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3430, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "20", "api": [{"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "api.settings.PROJECT_ROOT", "line_number": 16, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 17, "usage_type": "call"}, {"api_name": "api.util.SentimentData.SentimentData.get_raddit_data", "line_number": 18, "usage_type": "call"}, {"api_name": "api.util.SentimentData.SentimentData", "line_number": 18, "usage_type": "name"}, {"api_name": "api.util.SentimentData.SentimentData.get_stock_twit_data", "line_number": 19, "usage_type": "call"}, {"api_name": "api.util.SentimentData.SentimentData", "line_number": 19, "usage_type": "name"}, {"api_name": "api.serialziers.Dictionary", "line_number": 66, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 70, "usage_type": "call"}, {"api_name": "api.serialziers.DictionarySerializer", "line_number": 70, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 12, "usage_type": "call"}]}
+{"seq_id": "5634493262", "text": "# pip install opencv-python\n# pip install numpy\n# pip install torch\n# pip install transformers\n# pip install pillow\n\nimport cv2\nimport os\nimport numpy as np\nimport time\nfrom transformers import DetrFeatureExtractor, DetrForObjectDetection, pipeline, \\\n YolosFeatureExtractor, YolosForObjectDetection\nfrom PIL import Image, ImageDraw\n\n\n\ndef get_metadata(file_path):\n \"\"\"\n This function outputs the metadata of the input video file.\n\n :param file_path: Path to the input video file.\n :return: video_fps, total_frames, frames_height, frames_width\n \"\"\"\n\n # read video from file\n recording = cv2.VideoCapture(file_path)\n\n # Read metadata of the video\n video_fps = recording.get(cv2.CAP_PROP_FPS)\n total_frames = int(recording.get(cv2.CAP_PROP_FRAME_COUNT))\n frames_height = int(recording.get(cv2.CAP_PROP_FRAME_HEIGHT))\n frames_width = int(recording.get(cv2.CAP_PROP_FRAME_WIDTH))\n\n print(f\"Frame per second: {video_fps} \\nTotal Frames: {total_frames} \\nHeight: {frames_height} \\nWidth: {frames_width}\")\n return video_fps, total_frames, frames_height, frames_width\n\n\ndef process_frames(file_path, output_file_path):\n \"\"\"\n This function will take an input video and run every fram through the DETR-resnet-50 model to detect abjects in the frames.\n All objectes will be marked with a box and all frames will be put back together to the output video.\n\n :param file_path: Path to the input recording file\n :param output_file_path: Path and name of the output file\n :return: Returns the input video with boxes around the detected objects\n \"\"\"\n\n # initialize variables\n count = 0 # count to ensure all frames have been processed\n list_frames = [] # list of all processed frames\n recording = cv2.VideoCapture(file_path) # reading the input video with cv2\n\n # extract video metadata\n video_fps = recording.get(cv2.CAP_PROP_FPS) # check fps of input video\n total_frames = int(recording.get(cv2.CAP_PROP_FRAME_COUNT)) # count total frames of input video\n frames_height = int(recording.get(cv2.CAP_PROP_FRAME_HEIGHT)) # check pixel height of input video frame\n frames_width = int(recording.get(cv2.CAP_PROP_FRAME_WIDTH)) # check pixel width of input video frame\n codec = cv2.VideoWriter.fourcc(*'mp4v') # define format of output video\n\n # define the video output format and the values to resize the frames\n output_frames_height = 1080\n output_frames_width = 1920\n video_writer = cv2.VideoWriter(output_file_path, codec, video_fps, (output_frames_width, output_frames_height))\n\n # Initialize detr-resnet-50 model\n ##feature_extractor = DetrFeatureExtractor.from_pretrained(\"facebook/detr-resnet-50\") # Taken from HuggingFace\n ##model = DetrForObjectDetection.from_pretrained(\"facebook/detr-resnet-50\") # Taken from HuggingFace\n\n # Initialize Yolos-tiny model\n feature_extractor = YolosFeatureExtractor.from_pretrained('hustvl/yolos-tiny')\n model = YolosForObjectDetection.from_pretrained('hustvl/yolos-tiny')\n\n # create the object detection pipeline\n object_detection_pipe = pipeline(\"object-detection\",\n model=model,\n feature_extractor=feature_extractor)\n\n while count != 50:\n\n\n\n # Read video and retrieve individual frames\n ret, frame = recording.read()\n\n if frame is None:\n continue\n\n # Resize image for faster processing\n resizeFrame = cv2.resize(frame, (output_frames_width, output_frames_height))\n\n # Convert np array to PIL image - pipeline only works with PIL images\n pilFrame = Image.fromarray(np.uint8(resizeFrame))\n\n start = time.perf_counter()\n\n # Detect all objects in the frame\n results = object_detection_pipe(pilFrame)\n\n end = time.perf_counter()\n\n ms = (end-start) * 10**6\n seconds = ms / (10**6)\n print(f\"Elapsed {seconds:.03f} secs.\")\n\n # Add boxes and description of boxes to image\n im1 = ImageDraw.Draw(pilFrame)\n for result in results:\n box = result['box']\n xmin, xmax, ymin, ymax = box['xmin'], box['xmax'], box['ymin'], box['ymax']\n label = result['label']\n prob = result['score']\n shape = [xmin, ymin, xmax, ymax]\n text = f'{label}: {prob:0.2f}'\n if label == \"person\":\n im1.rectangle(shape, outline=\"red\", width=3)\n im1.text((xmin,ymax), text, fill=\"black\")\n elif label == \"sports ball\":\n im1.rectangle(shape, outline=\"blue\", width=3)\n im1.text((xmin, ymax), text, fill=\"black\")\n else:\n continue\n\n # Convert PIL image back to numpy array\n pilFrame = np.array(pilFrame)\n\n # Write the frame to the output video\n video_writer.write(pilFrame)\n\n # Print for every 50 frames processed\n if (count % 50 == 0):\n print('Processed ', count, ' frames')\n list_frames.append(frame)\n if len(list_frames) == int(total_frames):\n break\n\n # Increase count for every processed frame\n count += 1\n\n # Release to close all the resources that we have opened for reading and writing the video\n recording.release()\n video_writer.release()\n\n cv2.destroyAllWindows()\n\n\n# Define the input and output video location\ndir = ('/Users/christophmeier/Code_MastersThesis/AttackHeigtTracker/inputVideo')\nfile = os.listdir(dir)[0]\npath = str(dir) + '/' + str(file)\n\noutputDir = '/Users/christophmeier/Code_MastersThesis/AttackHeigtTracker/'\noutputFileName = 'processedVideo'\noutputFileType = 'mp4'\noutputFile = outputDir + outputFileName + '.' + outputFileType\n\n# check metadata\nget_metadata(path)\n\n# process every frame\nprocess_frames(path, outputFile)\n\n", "repo_name": "chrisP-cpmr/Object_Detection_BeachVolleyball", "sub_path": "AttackHeigtTracker/readRecording.py", "file_name": "readRecording.py", "file_ext": "py", "file_size_in_byte": 5975, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "cv2.VideoCapture", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 29, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_COUNT", "line_number": 30, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_HEIGHT", "line_number": 31, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_WIDTH", "line_number": 32, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 54, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_COUNT", "line_number": 55, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_HEIGHT", "line_number": 56, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_WIDTH", "line_number": 57, "usage_type": "attribute"}, {"api_name": "cv2.VideoWriter.fourcc", "line_number": 58, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 58, "usage_type": "attribute"}, {"api_name": "cv2.VideoWriter", "line_number": 63, "usage_type": "call"}, {"api_name": "transformers.YolosFeatureExtractor.from_pretrained", "line_number": 70, "usage_type": "call"}, {"api_name": "transformers.YolosFeatureExtractor", "line_number": 70, "usage_type": "name"}, {"api_name": "transformers.YolosForObjectDetection.from_pretrained", "line_number": 71, "usage_type": "call"}, {"api_name": "transformers.YolosForObjectDetection", "line_number": 71, "usage_type": "name"}, {"api_name": "transformers.pipeline", "line_number": 74, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 89, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 92, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 92, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 92, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 94, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 99, "usage_type": "call"}, {"api_name": "PIL.ImageDraw.Draw", "line_number": 106, "usage_type": "call"}, {"api_name": "PIL.ImageDraw", "line_number": 106, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 124, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 143, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 148, "usage_type": "call"}]}
+{"seq_id": "21910889579", "text": "# coding: utf-8\n\nimport six\n\nfrom huaweicloudsdkcore.utils.http_utils import sanitize_for_serialization\n\n\nclass ResourceInstanceReqBody:\n\n \"\"\"\n Attributes:\n openapi_types (dict): The key is attribute name\n and the value is attribute type.\n attribute_map (dict): The key is attribute name\n and the value is json key in definition.\n \"\"\"\n sensitive_list = []\n\n openapi_types = {\n 'without_any_tag': 'bool',\n 'tags': 'list[TagsDTO]',\n 'matches': 'list[Match]'\n }\n\n attribute_map = {\n 'without_any_tag': 'without_any_tag',\n 'tags': 'tags',\n 'matches': 'matches'\n }\n\n def __init__(self, without_any_tag=None, tags=None, matches=None):\n \"\"\"ResourceInstanceReqBody\n\n The model defined in huaweicloud sdk\n\n :param without_any_tag: 不包含任意一个标签,该字段为true时查询所有不带标签的资源。\n :type without_any_tag: bool\n :param tags: 包含标签,最多包含10个key,每个key下面的value最多10个,结构体不能缺失,key不能为空或者空字符串。Key不能重复,同一个key中values不能重复。返回包含所有标签的资源列表,key之间是与的关系,key-value结构中value是或的关系。无tag过滤条件时返回全量数据。\n :type tags: list[:class:`huaweicloudsdkorganizations.v1.TagsDTO`]\n :param matches: 要绑定到新创建的帐号的标签列表。\n :type matches: list[:class:`huaweicloudsdkorganizations.v1.Match`]\n \"\"\"\n \n \n\n self._without_any_tag = None\n self._tags = None\n self._matches = None\n self.discriminator = None\n\n if without_any_tag is not None:\n self.without_any_tag = without_any_tag\n if tags is not None:\n self.tags = tags\n if matches is not None:\n self.matches = matches\n\n @property\n def without_any_tag(self):\n \"\"\"Gets the without_any_tag of this ResourceInstanceReqBody.\n\n 不包含任意一个标签,该字段为true时查询所有不带标签的资源。\n\n :return: The without_any_tag of this ResourceInstanceReqBody.\n :rtype: bool\n \"\"\"\n return self._without_any_tag\n\n @without_any_tag.setter\n def without_any_tag(self, without_any_tag):\n \"\"\"Sets the without_any_tag of this ResourceInstanceReqBody.\n\n 不包含任意一个标签,该字段为true时查询所有不带标签的资源。\n\n :param without_any_tag: The without_any_tag of this ResourceInstanceReqBody.\n :type without_any_tag: bool\n \"\"\"\n self._without_any_tag = without_any_tag\n\n @property\n def tags(self):\n \"\"\"Gets the tags of this ResourceInstanceReqBody.\n\n 包含标签,最多包含10个key,每个key下面的value最多10个,结构体不能缺失,key不能为空或者空字符串。Key不能重复,同一个key中values不能重复。返回包含所有标签的资源列表,key之间是与的关系,key-value结构中value是或的关系。无tag过滤条件时返回全量数据。\n\n :return: The tags of this ResourceInstanceReqBody.\n :rtype: list[:class:`huaweicloudsdkorganizations.v1.TagsDTO`]\n \"\"\"\n return self._tags\n\n @tags.setter\n def tags(self, tags):\n \"\"\"Sets the tags of this ResourceInstanceReqBody.\n\n 包含标签,最多包含10个key,每个key下面的value最多10个,结构体不能缺失,key不能为空或者空字符串。Key不能重复,同一个key中values不能重复。返回包含所有标签的资源列表,key之间是与的关系,key-value结构中value是或的关系。无tag过滤条件时返回全量数据。\n\n :param tags: The tags of this ResourceInstanceReqBody.\n :type tags: list[:class:`huaweicloudsdkorganizations.v1.TagsDTO`]\n \"\"\"\n self._tags = tags\n\n @property\n def matches(self):\n \"\"\"Gets the matches of this ResourceInstanceReqBody.\n\n 要绑定到新创建的帐号的标签列表。\n\n :return: The matches of this ResourceInstanceReqBody.\n :rtype: list[:class:`huaweicloudsdkorganizations.v1.Match`]\n \"\"\"\n return self._matches\n\n @matches.setter\n def matches(self, matches):\n \"\"\"Sets the matches of this ResourceInstanceReqBody.\n\n 要绑定到新创建的帐号的标签列表。\n\n :param matches: The matches of this ResourceInstanceReqBody.\n :type matches: list[:class:`huaweicloudsdkorganizations.v1.Match`]\n \"\"\"\n self._matches = matches\n\n def to_dict(self):\n \"\"\"Returns the model properties as a dict\"\"\"\n result = {}\n\n for attr, _ in six.iteritems(self.openapi_types):\n value = getattr(self, attr)\n if isinstance(value, list):\n result[attr] = list(map(\n lambda x: x.to_dict() if hasattr(x, \"to_dict\") else x,\n value\n ))\n elif hasattr(value, \"to_dict\"):\n result[attr] = value.to_dict()\n elif isinstance(value, dict):\n result[attr] = dict(map(\n lambda item: (item[0], item[1].to_dict())\n if hasattr(item[1], \"to_dict\") else item,\n value.items()\n ))\n else:\n if attr in self.sensitive_list:\n result[attr] = \"****\"\n else:\n result[attr] = value\n\n return result\n\n def to_str(self):\n \"\"\"Returns the string representation of the model\"\"\"\n import simplejson as json\n if six.PY2:\n import sys\n reload(sys)\n sys.setdefaultencoding(\"utf-8\")\n return json.dumps(sanitize_for_serialization(self), ensure_ascii=False)\n\n def __repr__(self):\n \"\"\"For `print`\"\"\"\n return self.to_str()\n\n def __eq__(self, other):\n \"\"\"Returns true if both objects are equal\"\"\"\n if not isinstance(other, ResourceInstanceReqBody):\n return False\n\n return self.__dict__ == other.__dict__\n\n def __ne__(self, other):\n \"\"\"Returns true if both objects are not equal\"\"\"\n return not self == other\n", "repo_name": "huaweicloud/huaweicloud-sdk-python-v3", "sub_path": "huaweicloud-sdk-organizations/huaweicloudsdkorganizations/v1/model/resource_instance_req_body.py", "file_name": "resource_instance_req_body.py", "file_ext": "py", "file_size_in_byte": 6328, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 104, "dataset": "github-code", "pt": "20", "api": [{"api_name": "six.iteritems", "line_number": 128, "usage_type": "call"}, {"api_name": "six.PY2", "line_number": 154, "usage_type": "attribute"}, {"api_name": "sys.setdefaultencoding", "line_number": 157, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 158, "usage_type": "call"}, {"api_name": "huaweicloudsdkcore.utils.http_utils.sanitize_for_serialization", "line_number": 158, "usage_type": "call"}]}
+{"seq_id": "25763712056", "text": "\nfrom odoo import fields, models, api\nfrom datetime import datetime\nfrom odoo.exceptions import AccessError, UserError, ValidationError\nfrom odoo.tools.translate import _\n\n\n\nclass HrLeave(models.Model):\n _inherit = 'hr.leave'\n # delegations_id = fields.Many2one('employee.delegations')\n delegations_employee_id = fields.Many2one('hr.employee', 'delegations')\n \n def action_approve(self):\n\n # group_e.write({'users': [(1, self.env.user.id)]})\n\n super().action_approve()\n\n for rec in self :\n is_leave_user_test = rec.user_has_groups('hr_holidays.group_hr_holidays_user')\n if is_leave_user_test:\n print(is_leave_user_test, '//////////////////////////////////////////////////////////////////////')\n\n group_e = self.env.ref('hr_holidays.group_hr_holidays_user', False)\n group_e.write({'users': [(3, self.env.user.employee_id.id)]})\n group_e.write({'users': [(4, rec.delegations_employee_id.id)]})\n delegations_info = {'employee_id':self.env.user.employee_id.id,'delegated_employee_id':rec.delegations_employee_id.id,\n 'date_from':rec.date_from,'date_to':rec.date_to,'state':'draft','name':'test','date':datetime.now(),}\n filed = self.env['employee.delegations'].create(delegations_info)\n filed.onchange_method()\n filed.access_granted()\n notification_ids = [((0, 0, {\n 'res_partner_id': rec.delegations_employee_id.id,\n 'notification_type': 'inbox'}))]\n # user_id = self.env.user.id\n # message = (\"You have a assigned a delegations from %s from %s to %s\") % (self.env.user.employee_id.name,rec.date_from,rec.date_to)\n # channel = self.env['mail.channel'].channel_get([rec.delegations_employee_id.id])\n # channel_id = self.env['mail.channel'].browse(channel[\"id\"])\n # channel_id.message_post(author_id=user_id,\n # body=(message),\n # message_type='notification',\n # subtype_xmlid=\"mail.mt_comment\",\n # notification_ids=notification_ids,\n # partner_ids=[rec.delegations_employee_id.id],\n # notify_by_email=False,\n # )\n\n\n\n def _check_double_validation_rules(self, employees, state):\n super(HrLeave, self)._check_double_validation_rules(employees,state)\n print(\"here/************************************************************************************/\")\n is_leave_user = self.user_has_groups('hr_holidays.group_hr_holidays_user')\n if state == 'validate1':\n # print(is_leave_user , \"/*********************************************************************/\")\n employees2 = [employee.leave_manager_id.id for employee in employees ]#employees.filtered(lambda employee: employee.leave_manager_id != self.env.user)\n print(employees2, '/________________________________________-------------__========000____-----')\n #\n # if employees2 :\n # if self.env.user in employees:\n # raise AccessError(_('You cannot first approve a time off for %s, because you are not his time off manager', employees[0].name))\n # elif state == 'validate' and not is_leave_user:\n # print(\"test +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\")\n\n\n", "repo_name": "sideeg/nidlp", "sub_path": "employee_delegations/models/hr_leave.py", "file_name": "hr_leave.py", "file_ext": "py", "file_size_in_byte": 3609, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "odoo.models.Model", "line_number": 9, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 9, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 12, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 12, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 29, "usage_type": "name"}]}
+{"seq_id": "21284310742", "text": "from decimal import Decimal\nfrom django.http import HttpResponseRedirect\nfrom django.conf import settings\nfrom django.shortcuts import redirect, render, get_object_or_404\nfrom django.views.decorators.http import require_POST\nfrom movie_store.models import Movie\nfrom movie_store.forms import BasketAddProductForm \n\nclass Basket(object):\n \n def __init__(self, request):\n self.session = request.session\n basket = self.session.get(settings.BASKET_SESSION_ID)\n if not basket:\n # save an empty basket in the session\n basket = self.session[settings.BASKET_SESSION_ID] = {}\n self.basket = basket\n\n def __iter__(self):\n print(f'basket: { self.basket }')\n movie_ids = self.basket.keys()\n movies = Movie.objects.filter(id__in=movie_ids)\n\n basket = self.basket.copy()\n for movie in movies:\n basket[str(movie.id)]['movie'] = movie\n basket[str(movie.id)]['movie_id'] = movie.id\n\n for item in basket.values():\n item['price'] = Decimal(item['price'])\n item['total_price'] = item['price'] * item['quantity']\n yield item\n\n def __len__(self):\n return sum(item['quantity'] for item in self.basket.values())\n\n def add(self, movie, quantity=1, override_quantity=False):\n movie_id = str(movie.id)\n if movie_id not in self.basket:\n self.basket[movie_id] = {'quantity': 0,\n 'price': str(movie.price)}\n if override_quantity:\n self.basket[movie_id]['quantity'] = quantity\n else:\n self.basket[movie_id]['quantity'] += quantity\n self.save()\n\n def save(self):\n self.session.modified = True\n\n def remove(self, movie):\n movie_id = str(movie.id)\n if movie_id in self.basket:\n del self.basket[movie_id]\n self.save()\n\n def clear(self):\n del self.session[settings.BASKET_SESSION_ID]\n self.save()\n\n def get_total_price(self):\n return sum(Decimal(item['price']) * item['quantity'] for item in self.basket.values())\n\n\n@require_POST\ndef basket_add(request, movie_id):\n basket = Basket(request)\n movie = get_object_or_404(Movie, id=movie_id)\n form = BasketAddProductForm(request.POST)\n if form.is_valid():\n cd = form.cleaned_data\n basket.add(movie=movie,\n quantity=cd['quantity'],\n override_quantity=cd['override'])\n return HttpResponseRedirect(request.META.get('HTTP_REFERER'))\n\n@require_POST\ndef basket_remove(request, movie_id):\n basket = Basket(request)\n movie = get_object_or_404(Movie, id=movie_id)\n basket.remove(movie)\n return redirect('basket_detail')\n\ndef basket_detail(request):\n basket = Basket(request)\n total = len(basket)\n for item in basket:\n item['update_quantity_form'] = BasketAddProductForm(initial={'quantity': item['quantity'],\n 'override': True})\n return render(request, 'movie_store/basket.html', {'basket': basket, 'total':total})\n", "repo_name": "topherlee/movie-store", "sub_path": "movie_store/views/basket.py", "file_name": "basket.py", "file_ext": "py", "file_size_in_byte": 3124, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "django.conf.settings.BASKET_SESSION_ID", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 13, "usage_type": "name"}, {"api_name": "django.conf.settings.BASKET_SESSION_ID", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 16, "usage_type": "name"}, {"api_name": "movie_store.models.Movie.objects.filter", "line_number": 22, "usage_type": "call"}, {"api_name": "movie_store.models.Movie.objects", "line_number": 22, "usage_type": "attribute"}, {"api_name": "movie_store.models.Movie", "line_number": 22, "usage_type": "name"}, {"api_name": "decimal.Decimal", "line_number": 30, "usage_type": "call"}, {"api_name": "django.conf.settings.BASKET_SESSION_ID", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 58, "usage_type": "name"}, {"api_name": "decimal.Decimal", "line_number": 62, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 68, "usage_type": "call"}, {"api_name": "movie_store.models.Movie", "line_number": 68, "usage_type": "argument"}, {"api_name": "movie_store.forms.BasketAddProductForm", "line_number": 69, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 75, "usage_type": "call"}, {"api_name": "django.views.decorators.http.require_POST", "line_number": 65, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 80, "usage_type": "call"}, {"api_name": "movie_store.models.Movie", "line_number": 80, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 82, "usage_type": "call"}, {"api_name": "django.views.decorators.http.require_POST", "line_number": 77, "usage_type": "name"}, {"api_name": "movie_store.forms.BasketAddProductForm", "line_number": 88, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 90, "usage_type": "call"}]}
+{"seq_id": "1592649900", "text": "from django.shortcuts import render\nfrom django.http import JsonResponse, HttpResponse\nfrom rest_framework.decorators import api_view, renderer_classes\nfrom rest_framework.renderers import JSONRenderer\nfrom .serializers import ForecastSerializer\nfrom .predictor import get_forecast\nimport os\nfrom django.conf import settings\nimport pandas as pd\n\ndef index(request):\n return render(request, 'index.html')\n\ndata_path = os.path.join(settings.BASE_DIR, 'media', 'Cleaned_Balance_Stay.xlsx')\ndata = pd.read_excel(data_path)\n\n\n\n@api_view(['GET'])\n@renderer_classes([JSONRenderer])\ndef balance_forecast_view(request):\n profits_image, _, _ = get_forecast(plot_graphs=True) # Получаем изображение для прибыли\n\n if profits_image:\n response = HttpResponse(content_type=\"image/png\")\n response.write(profits_image.getvalue()) # Get the Profits forecast image\n return response\n\n return JsonResponse({\"message\": \"No graph generated.\"})\n\n@api_view(['GET'])\n@renderer_classes([JSONRenderer])\ndef profits_forecast_view(request):\n _, _, income_spends_image = get_forecast(plot_graphs=True) # Получаем изображение для доходов и расходов\n\n if income_spends_image:\n response = HttpResponse(content_type=\"image/png\")\n response.write(income_spends_image.getvalue()) # Get the Income & Spends forecast image\n return response\n\n return JsonResponse({\"message\": \"No graph generated.\"})\n\n@api_view(['GET'])\n@renderer_classes([JSONRenderer])\ndef income_spends_forecast_view(request):\n _, balance_image, _ = get_forecast(plot_graphs=True) # Получаем изображение для баланса\n\n if balance_image:\n response = HttpResponse(content_type=\"image/png\")\n response.write(balance_image.getvalue()) # Get the Balance forecast image\n return response\n\n return JsonResponse({\"message\": \"No graph generated.\"})\n", "repo_name": "enpure/stayresort-prediction", "sub_path": "stayresort/forecast/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1959, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.shortcuts.render", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.conf.settings.BASE_DIR", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 14, "usage_type": "name"}, {"api_name": "pandas.read_excel", "line_number": 15, "usage_type": "call"}, {"api_name": "predictor.get_forecast", "line_number": 22, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 25, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 29, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 19, "usage_type": "call"}, {"api_name": "rest_framework.decorators.renderer_classes", "line_number": 20, "usage_type": "call"}, {"api_name": "rest_framework.renderers.JSONRenderer", "line_number": 20, "usage_type": "name"}, {"api_name": "predictor.get_forecast", "line_number": 34, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 37, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 41, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 31, "usage_type": "call"}, {"api_name": "rest_framework.decorators.renderer_classes", "line_number": 32, "usage_type": "call"}, {"api_name": "rest_framework.renderers.JSONRenderer", "line_number": 32, "usage_type": "name"}, {"api_name": "predictor.get_forecast", "line_number": 46, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 49, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 53, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 43, "usage_type": "call"}, {"api_name": "rest_framework.decorators.renderer_classes", "line_number": 44, "usage_type": "call"}, {"api_name": "rest_framework.renderers.JSONRenderer", "line_number": 44, "usage_type": "name"}]}
+{"seq_id": "13503119811", "text": "import numpy as np\nfrom tqdm import tqdm\nfrom utils import digitize_datetime, decay_func\n\ndef get_k_accuracy(res_mat):\n res_mat = np.array(res_mat)\n return [np.mean(np.sum(res_mat[:,:k], axis=1) > 0) for k in [1,5,10,15,20]]\n\ndef get_recommandation_result(testset, args, poi2region, emb_set, mode, weight=[1,1,1]):\n if mode == 'GE':\n embeddings, region_embeddings, time_embeddings = emb_set\n elif mode == 'STSG':\n# sem_emb, embeddings, time_embeddings = emb_set\n sem_emb, geo_emb, time_embeddings = emb_set\n sem_dim = sem_emb.shape[1]; geo_dim = geo_emb.shape[1];\n embeddings = np.concatenate([sem_emb, geo_emb], axis=1)\n elif mode == 'Skipgram_wt':\n embeddings, time_embeddings = emb_set\n elif mode == 'Skipgram_wot':\n embeddings = emb_set\n elif mode == 'POI2VEC':\n embeddings = emb_set\n elif mode == 'PRME':\n embeddings, embeddings_u, user_embeddings = emb_set\n\n result_mat = list()\n for u, seq in tqdm(enumerate(testset), total=len(testset)):\n lenseq = len(seq)\n lenitem = len(seq[0])\n poiseq, dtseq = seq[:,0], seq[:,lenitem-1]\n tslot_seq = digitize_datetime(dtseq, args.pattern)\n seq = np.concatenate([seq, digitize_datetime(dtseq, args.pattern)[:,None]], axis=1)\n\n for i in range(lenseq-1):\n history = seq[:i]\n target = seq[i+1] \n if lenitem == 2:\n l_n, dt_n, tslot_n = seq[i] #now\n l_y, dt_y, tslot_y = target\n elif lenitem == 3:\n l_n, usr, dt_n, tslot_n = seq[i]\n l_y, usr, dt_y, tslot_y = target\n else: assert 0\n \n decays = list()\n l_xs = history[:,0].astype(int)\n for j, x in enumerate(history):\n decays.append(decay_func(y=dt_y, x=x[lenitem-1]))\n\n embs = embeddings[l_xs]\n if args.use_decay == True:\n user_profile = np.sum(embs*np.array(decays).reshape(-1,1), axis=0)\n else: \n user_profile = np.mean(embs, axis=0)\n if not mode in ['POI2VEC', 'Skipgram_wot', 'PRME']:\n t_emb_now = time_embeddings[tslot_n]\n\n if mode == 'GE':\n r_emb_now = region_embeddings[poi2region[l_n]]\n scores = np.matmul(user_profile, embeddings.T) + \\\n np.matmul(r_emb_now, embeddings.T) + \\\n np.matmul(t_emb_now, embeddings.T)\n elif mode == 'STSG':\n# scores = np.matmul(user_profile, embeddings.T) + \\\n# np.matmul(t_emb_now, sem_emb.T)\n \n profile_sem, profile_geo = user_profile[:sem_dim], user_profile[sem_dim:]\n score_sem = np.matmul(profile_sem, sem_emb.T) / float(sem_dim)\n score_geo = np.matmul(profile_geo, geo_emb.T) / float(geo_dim)\n score_time = np.matmul(t_emb_now, sem_emb.T) / float(sem_dim)\n scores = np.stack([score_sem, score_geo, score_time],axis=0)*np.array(weight).reshape(3,1)\n scores = np.sum(scores, axis=0)\n elif mode == 'Skipgram_wt':\n scores = np.matmul(user_profile, embeddings.T) + \\\n np.matmul(t_emb_now, embeddings.T)\n elif mode == 'Skipgram_wot':\n scores = np.matmul(user_profile, embeddings.T)\n elif mode == 'POI2VEC':\n scores = np.matmul(user_profile, embeddings.T)\n elif mode == 'PRME':\n user_profile = user_embeddings[usr]\n scores = np.matmul(user_profile, embeddings_u.T) + \\\n np.matmul(embeddings[l_n], embeddings.T)\n# scores = np.matmul(user_profile, embeddings.T)\n result_mat.append((-scores).argsort()[:20] == l_y)\n return result_mat", "repo_name": "kunrenzhilu/recommand", "sub_path": "recommand_np.py", "file_name": "recommand_np.py", "file_ext": "py", "file_size_in_byte": 3874, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "numpy.array", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 16, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 27, "usage_type": "call"}, {"api_name": "utils.digitize_datetime", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 32, "usage_type": "call"}, {"api_name": "utils.digitize_datetime", "line_number": 32, "usage_type": "call"}, {"api_name": "utils.decay_func", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 83, "usage_type": "call"}]}
+{"seq_id": "15517289648", "text": "import os\nimport numpy as np\nimport pandas as pd\n\nfrom torchvision.transforms import ToTensor # transform PIL image to torch.Tensor\nimport torch\nfrom torch.utils.data import DataLoader # mini-batch loader\nfrom torch import nn\nfrom torch.utils.data import random_split\nfrom torchvision.models import resnet50\n\nimport hydra\nfrom hydra.core.config_store import ConfigStore\n\nfrom config import MNISTConfig\nfrom model import MNIST_lightning\nfrom data import CustomDataset\n\nfrom pytorch_lightning.callbacks.early_stopping import EarlyStopping\nimport pytorch_lightning as pl\n\n\ncs = ConfigStore.instance()\ncs.store(name=\"mnist_config\", node=MNISTConfig)\n\nlabels_item = {\n '0' : 0,\n '1' : 1,\n '2' : 2,\n '3' : 3,\n '4' : 4,\n '5' : 5,\n '6' : 6,\n '7' : 7,\n '8' : 8,\n '9' : 9,\n \n }\n\ndef toCSVfile(input_dir, output_dir, file_name):\n dir_list = os.listdir(input_dir)\n img_dir = []\n labels = []\n\n for dir in dir_list:\n current_path = os.path.join(input_dir, dir)\n idir = os.listdir(current_path)\n lb = [labels_item[dir]]*len(idir)\n\n img_dir += np.core.defchararray.add(current_path + \"/\", np.array(idir)).tolist()\n labels += lb\n df = pd.DataFrame({'filename': img_dir, 'label':labels})\n \n out_dir = output_dir + '/' + file_name\n if not os.path.exists(output_dir):\n os.mkdir(out_dir)\n if os.path.exists(out_dir):\n os.remove(out_dir)\n df.to_csv(out_dir, index = False)\n\n\n@hydra.main(config_path='../configs', config_name='train', version_base=None)\ndef main(cfg: MNISTConfig):\n toCSVfile(cfg.paths.data + '/' + cfg.files.train_folder, \n cfg.paths.data,\n cfg.files.train_file)\n\n \n data = CustomDataset(cfg.paths.data + '/' + cfg.files.train_file, dir = None, transform = ToTensor())\n\n # use 20% of training data for validation\n train_set_size = int(len(data) * cfg.train_size)\n valid_set_size = len(data) - train_set_size\n\n # split the train set into two\n seed = torch.Generator().manual_seed(cfg.params.seed)\n train_set, valid_set = random_split(data, [train_set_size, valid_set_size], generator=seed)\n\n # define dataloader\n train_loader = DataLoader(train_set, batch_size = cfg.params.batch_size, shuffle = True)\n valid_loader = DataLoader(valid_set, batch_size = cfg.params.batch_size, shuffle = False)\n\n # defind model\n model = MNIST_lightning(resnet50(num_classes = 10))\n trainer = pl.Trainer(devices=1, \n accelerator=\"gpu\", \n callbacks=[EarlyStopping(monitor=\"train_loss\", mode=\"min\")], \n max_epochs = cfg.params.n_epoch)\n trainer.fit(model, train_dataloaders=train_loader, val_dataloaders=valid_loader)\n\n\nif __name__ == '__main__':\n main()\n", "repo_name": "haydenshimada/MNIST-hydra-lightning", "sub_path": "src/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2818, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "hydra.core.config_store.ConfigStore.instance", "line_number": 23, "usage_type": "call"}, {"api_name": "hydra.core.config_store.ConfigStore", "line_number": 23, "usage_type": "name"}, {"api_name": "config.MNISTConfig", "line_number": 24, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.core.defchararray.add", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.core", "line_number": 50, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 50, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 58, "usage_type": "call"}, {"api_name": "config.MNISTConfig", "line_number": 63, "usage_type": "name"}, {"api_name": "data.CustomDataset", "line_number": 69, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.Generator", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.utils.data.random_split", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 81, "usage_type": "call"}, {"api_name": "model.MNIST_lightning", "line_number": 84, "usage_type": "call"}, {"api_name": "torchvision.models.resnet50", "line_number": 84, "usage_type": "call"}, {"api_name": "pytorch_lightning.Trainer", "line_number": 85, "usage_type": "call"}, {"api_name": "pytorch_lightning.callbacks.early_stopping.EarlyStopping", "line_number": 87, "usage_type": "call"}, {"api_name": "hydra.main", "line_number": 62, "usage_type": "call"}]}
+{"seq_id": "34888076529", "text": "######################################\n# Kaihua Tang\n######################################\n\nimport os\nimport json\nimport math\nimport torch\nimport numpy as np \nimport torch.nn as nn\nimport torch.optim as optim\nimport torch.optim.lr_scheduler as lr_scheduler\nimport torch.nn.functional as F \nfrom torch.optim.optimizer import Optimizer, required\n\nfrom utils.general_utils import *\n\nfrom torch import Tensor\nfrom typing import List, Optional\n\n\ndef create_optimizer(model, classifier, logger, config):\n training_opt = config['training_opt']\n lr = training_opt['optim_params']['lr']\n weight_decay = training_opt['optim_params']['weight_decay']\n\n # IMPORTANT\n # when the deadline is approaching, I suddenly found that I forgot to add momentum into my SGD optimizer.\n # therefore, I have to just accept the setting of 0 momentum, but since all the methods are replemented \n # under the same optimizer, our conclusions and analyses still hold\n # For the follower, please remember to add momentum here.\n\n logger.info('=====> Create optimizer')\n all_params = []\n\n for _, val in model.named_parameters():\n if not val.requires_grad:\n continue\n all_params += [{\"params\": [val], \"lr\": lr, \"weight_decay\": weight_decay}]\n for _, val in classifier.named_parameters():\n if not val.requires_grad:\n continue\n all_params += [{\"params\": [val], \"lr\": lr, \"weight_decay\": weight_decay}]\n \n if training_opt['optimizer'] == 'Adam':\n return optim.Adam(all_params)\n elif training_opt['optimizer'] == 'SGD':\n return optim.SGD(all_params)\n else:\n logger.info('********** ERROR: unidentified optimizer **********')\n\n\ndef create_optimizer_stage2(model, classifier, logger, config):\n training_opt = config['training_opt']\n lr = training_opt['optim_params']['lr']\n weight_decay = training_opt['optim_params']['weight_decay']\n\n # IMPORTANT\n # when the deadline is approaching, I suddenly found that I forgot to add momentum into my SGD optimizer.\n # therefore, I have to just accept the setting of 0 momentum, but since all the methods are replemented \n # under the same optimizer, our conclusions and analyses still hold\n # For the follower, please remember to add momentum here.\n\n logger.info('=====> Create optimizer')\n all_params = []\n\n # in two-stage training, the second stage should freeze the backbone\n logger.info('========= Freeze Backbone Parameters ===========')\n for _, val in model.named_parameters():\n val.requires_grad = False\n\n for _, val in classifier.named_parameters():\n if not val.requires_grad:\n continue\n all_params += [{\"params\": [val], \"lr\": lr, \"weight_decay\": weight_decay}]\n \n if training_opt['optimizer'] == 'Adam':\n return optim.Adam(all_params)\n elif training_opt['optimizer'] == 'SGD':\n return optim.SGD(all_params)\n else:\n logger.info('********** ERROR: unidentified optimizer **********')\n\n\ndef create_scheduler(optimizer, logger, config):\n training_opt = config['training_opt']\n\n logger.info('=====> Create Scheduler')\n scheduler_params = training_opt['scheduler_params']\n\n if training_opt['scheduler'] == 'cosine':\n return optim.lr_scheduler.CosineAnnealingLR(optimizer, training_opt['num_epochs'], eta_min=scheduler_params['endlr'])\n elif training_opt['scheduler'] == 'step':\n return optim.lr_scheduler.StepLR(optimizer, gamma=scheduler_params['gamma'], step_size=scheduler_params['step_size'])\n elif training_opt['scheduler'] == 'multistep':\n return optim.lr_scheduler.MultiStepLR(optimizer, gamma=scheduler_params['gamma'], milestones=scheduler_params['milestones'])\n else:\n logger.info('********** ERROR: unidentified optimizer **********')\n\n\n\ndef create_loss(logger, config, train_loader):\n training_opt = config['training_opt']\n\n if training_opt['loss'] == 'CrossEntropy':\n loss = nn.CrossEntropyLoss()\n elif training_opt['loss'] == 'Focal':\n loss = FocalLoss(gamma=2.0)\n elif training_opt['loss'] == 'BalancedSoftmax':\n loss = BlSoftmaxLoss(train_loader)\n elif training_opt['loss'] == 'LDAM':\n loss = LDAMLoss(train_loader, total_epoch=training_opt['num_epochs'])\n elif training_opt['loss'] == 'RIDE':\n loss = RIDELoss(train_loader, additional_diversity_factor=config['algorithm_opt']['diversity_factor'])\n elif training_opt['loss'] == 'TADE':\n loss = TADELoss(train_loader, tau=config['algorithm_opt']['tau'])\n else:\n logger.info('********** ERROR: unidentified optimizer **********')\n logger.info('====== Set Loss Function to {} ======='.format(training_opt['loss']))\n return loss\n\n\n\nclass CenterLoss(nn.Module):\n def __init__(self, num_classes=10, feat_dim=2, use_gpu=True):\n super(CenterLoss, self).__init__()\n self.num_class = num_classes\n self.num_feature = feat_dim\n if use_gpu:\n self.centers = nn.Parameter(torch.randn(self.num_class, self.num_feature).cuda())\n else:\n self.centers = nn.Parameter(torch.randn(self.num_class, self.num_feature))\n\n def forward(self, x, labels):\n center = self.centers[labels]\n dist = (x-center).pow(2).sum(dim=-1)\n loss = torch.clamp(dist, min=1e-12, max=1e+12).mean(dim=-1)\n\n return loss\n\n\nclass CenterCosLoss(nn.Module):\n def __init__(self, num_classes=10, feat_dim=2, use_gpu=True):\n super(CenterCosLoss, self).__init__()\n self.num_class = num_classes\n self.num_feature = feat_dim\n if use_gpu:\n self.centers = nn.Parameter(torch.randn(self.num_class, self.num_feature).cuda())\n else:\n self.centers = nn.Parameter(torch.randn(self.num_class, self.num_feature))\n\n def l2_norm(self, x):\n normed_x = x / torch.norm(x, 2, 1, keepdim=True)\n return normed_x\n\n def forward(self, x, labels):\n center = self.centers[labels]\n norm_c = self.l2_norm(center)\n norm_x = self.l2_norm(x)\n similarity = (norm_c * norm_x).sum(dim=-1)\n dist = 1.0 - similarity\n loss = torch.clamp(dist, min=1e-12, max=1e+12).mean(dim=-1)\n\n return loss\n\n\nclass CenterTripletLoss(nn.Module):\n def __init__(self, num_classes=10, feat_dim=2, use_gpu=True):\n super(CenterTripletLoss, self).__init__()\n self.num_class = num_classes\n self.num_feature = feat_dim\n if use_gpu:\n self.centers = nn.Parameter(torch.randn(self.num_class, self.num_feature).cuda())\n else:\n self.centers = nn.Parameter(torch.randn(self.num_class, self.num_feature))\n self.triplet_loss = nn.TripletMarginLoss(margin=1.0, p=2)\n\n def forward(self, x, preds, labels):\n # use most likely categories as negative samples\n preds = preds.softmax(-1)\n batch_size = x.shape[0]\n idxs = torch.arange(batch_size).to(x.device)\n preds[idxs, labels] = -1\n adv_labels = preds.max(-1)[1]\n\n anchor = x # num_batch, num_dim\n positive = self.centers[labels] # num_batch, num_dim\n negative = self.centers[adv_labels] # num_batch, num_dim\n\n output = self.triplet_loss(anchor, positive, negative)\n return output\n\n\n\nclass BlSoftmaxLoss(nn.Module):\n def __init__(self, train_loader, reduction=\"mean\"):\n super(BlSoftmaxLoss, self).__init__()\n # reduction: string. One of \"none\", \"mean\", \"sum\"\n label_count_array = count_dataset(train_loader)\n label_count_array = np.array(label_count_array) / np.sum(label_count_array)\n adjustments = np.log(label_count_array + 1e-12)\n adjustments = torch.from_numpy(adjustments).view(1, -1)\n self.adjustments = adjustments\n self.reduction = reduction\n\n def forward(self, logits, target):\n logits = logits + self.adjustments.to(logits.device)\n loss = F.cross_entropy(input=logits, target=target, reduction=self.reduction)\n return loss\n\nclass FocalLoss(nn.Module):\n def __init__(self, gamma=2.0, alpha=None, size_average=True):\n super(FocalLoss, self).__init__()\n self.gamma = gamma\n self.alpha = alpha\n self.size_average = size_average\n\n def forward(self, input, target):\n if input.dim()>2:\n input = input.view(input.size(0),input.size(1),-1) # N,C,H,W => N,C,H*W\n input = input.transpose(1,2) # N,C,H*W => N,H*W,C\n input = input.contiguous().view(-1,input.size(2)) # N,H*W,C => N*H*W,C\n target = target.view(-1,1)\n\n logpt = F.log_softmax(input, dim=-1)\n logpt = logpt.gather(1,target)\n logpt = logpt.view(-1)\n pt = logpt.detach().exp()\n\n if self.alpha is not None:\n assert False\n\n loss = -1 * (1-pt)**self.gamma * logpt\n if self.size_average: \n return loss.mean()\n else: \n return loss.sum()\n\nclass LDAMLoss(nn.Module):\n def __init__(self, dataloader, total_epoch, max_m=0.5, s=30):\n super(LDAMLoss, self).__init__()\n self.cls_num_list = count_dataset(dataloader)\n m_list = 1.0 / np.sqrt(np.sqrt(self.cls_num_list))\n m_list = m_list * (max_m / np.max(m_list))\n m_list = torch.FloatTensor(m_list)\n self.m_list = m_list\n assert s > 0\n self.s = s\n self.total_epoch = total_epoch\n\n def set_weight(self, epoch):\n idx = epoch // int(self.total_epoch * 0.8)\n betas = [0, 0.9999]\n effective_num = 1.0 - np.power(betas[idx], self.cls_num_list)\n per_cls_weights = (1.0 - betas[idx]) / np.array(effective_num)\n per_cls_weights = per_cls_weights / np.sum(per_cls_weights) * len(self.cls_num_list)\n self.weight = torch.FloatTensor(per_cls_weights)\n\n def forward(self, x, target):\n index = torch.zeros_like(x, dtype=torch.uint8)\n index.scatter_(1, target.data.view(-1, 1), 1)\n \n index_float = index.float().to(x.device)\n batch_m = torch.matmul(self.m_list.to(x.device)[None, :], index_float.transpose(0,1))\n batch_m = batch_m.view((-1, 1))\n x_m = x - batch_m\n \n output = torch.where(index, x_m, x)\n return F.cross_entropy(self.s*output, target, weight=self.weight.to(x.device))\n\n\nclass TADELoss(nn.Module):\n def __init__(self, dataloader, tau=2):\n super().__init__()\n self.base_loss = F.cross_entropy \n cls_num_list = count_dataset(dataloader)\n prior = np.array(cls_num_list) / np.sum(cls_num_list)\n self.prior = torch.tensor(prior).float().cuda()\n self.C_number = len(cls_num_list) # class number\n self.tau = tau \n\n def inverse_prior(self, prior): \n value, idx0 = torch.sort(prior)\n _, idx1 = torch.sort(idx0)\n idx2 = prior.shape[0]-1-idx1 # reverse the order\n inverse_prior = value.index_select(0,idx2)\n \n return inverse_prior\n\n def forward(self, output_logits, target, extra_info=None):\n if extra_info is None:\n return self.base_loss(output_logits, target) # output_logits indicates the final prediction\n\n loss = 0\n\n # Obtain logits from each expert \n expert1_logits = extra_info['logits'][0]\n expert2_logits = extra_info['logits'][1] \n expert3_logits = extra_info['logits'][2] \n \n # Softmax loss for expert 1 \n loss += self.base_loss(expert1_logits, target)\n \n # Balanced Softmax loss for expert 2 \n expert2_logits = expert2_logits + torch.log(self.prior + 1e-9) \n loss += self.base_loss(expert2_logits, target)\n \n # Inverse Softmax loss for expert 3\n inverse_prior = self.inverse_prior(self.prior)\n expert3_logits = expert3_logits + torch.log(self.prior + 1e-9) - self.tau * torch.log(inverse_prior+ 1e-9) \n loss += self.base_loss(expert3_logits, target)\n \n return loss\n\n\nclass RIDELoss(nn.Module):\n '''\n Copy from https://github.com/frank-xwang/RIDE-LongTailRecognition/blob/main/model/loss.py\n '''\n def __init__(self, dataloader=None, base_diversity_temperature=1.0, max_m=0.5, s=30, reweight=True, reweight_epoch=80, \n base_loss_factor=1.0, additional_diversity_factor=-0.2, reweight_factor=0.02):\n super().__init__()\n self.base_loss = F.cross_entropy\n self.base_loss_factor = base_loss_factor\n if not reweight:\n self.reweight_epoch = -1\n else:\n self.reweight_epoch = reweight_epoch\n\n # LDAM is a variant of cross entropy and we handle it with self.m_list.\n if dataloader is None:\n # No cls_num_list is provided, then we cannot adjust cross entropy with LDAM.\n\n self.m_list = None\n self.per_cls_weights_enabled = None\n self.per_cls_weights_enabled_diversity = None\n else:\n # We will use LDAM loss if we provide cls_num_list.\n cls_num_list = count_dataset(dataloader)\n m_list = 1.0 / np.sqrt(np.sqrt(cls_num_list))\n m_list = m_list * (max_m / np.max(m_list))\n m_list = torch.tensor(m_list, dtype=torch.float, requires_grad=False)\n self.m_list = m_list\n self.s = s\n assert s > 0\n \n if reweight_epoch != -1:\n idx = 1 # condition could be put in order to set idx\n betas = [0, 0.9999]\n effective_num = 1.0 - np.power(betas[idx], cls_num_list)\n per_cls_weights = (1.0 - betas[idx]) / np.array(effective_num)\n per_cls_weights = per_cls_weights / np.sum(per_cls_weights) * len(cls_num_list)\n self.per_cls_weights_enabled = torch.tensor(per_cls_weights, dtype=torch.float, requires_grad=False)\n else:\n self.per_cls_weights_enabled = None\n\n cls_num_list = np.array(cls_num_list) / np.sum(cls_num_list)\n C = len(cls_num_list)\n per_cls_weights = C * cls_num_list * reweight_factor + 1 - reweight_factor\n\n # Experimental normalization: This is for easier hyperparam tuning, the effect can be described in the learning rate so the math formulation keeps the same.\n # At the same time, the 1 - max trick that was previously used is not required since weights are already adjusted.\n per_cls_weights = per_cls_weights / np.max(per_cls_weights)\n\n assert np.all(per_cls_weights > 0), \"reweight factor is too large: out of bounds\"\n # save diversity per_cls_weights\n self.per_cls_weights_enabled_diversity = torch.tensor(per_cls_weights, dtype=torch.float, requires_grad=False).cuda()\n\n self.base_diversity_temperature = base_diversity_temperature\n self.additional_diversity_factor = additional_diversity_factor\n\n def to(self, device):\n super().to(device)\n if self.m_list is not None:\n self.m_list = self.m_list.to(device)\n \n if self.per_cls_weights_enabled is not None:\n self.per_cls_weights_enabled = self.per_cls_weights_enabled.to(device)\n\n if self.per_cls_weights_enabled_diversity is not None:\n self.per_cls_weights_enabled_diversity = self.per_cls_weights_enabled_diversity.to(device)\n\n return self\n\n def set_epoch(self, epoch):\n if self.reweight_epoch != -1:\n self.epoch = epoch\n\n if epoch > self.reweight_epoch:\n self.per_cls_weights_base = self.per_cls_weights_enabled\n self.per_cls_weights_diversity = self.per_cls_weights_enabled_diversity\n else:\n self.per_cls_weights_base = None\n self.per_cls_weights_diversity = None\n\n def get_final_output(self, output_logits, target):\n x = output_logits\n\n index = torch.zeros_like(x, dtype=torch.uint8, device=x.device)\n index.scatter_(1, target.data.view(-1, 1), 1)\n \n index_float = index.float()\n batch_m = torch.matmul(self.m_list[None, :], index_float.transpose(0,1))\n \n batch_m = batch_m.view((-1, 1))\n x_m = x - batch_m * self.s\n\n final_output = torch.where(index, x_m, x)\n return final_output\n\n def forward(self, output_logits, target, extra_info=None):\n if extra_info is None:\n return self.base_loss(output_logits, target)\n\n loss = 0\n\n self.to(output_logits.device)\n # Adding RIDE Individual Loss for each expert\n for logits_item in extra_info['logits']:\n ride_loss_logits = logits_item\n # the following line of code is unfair (original implementation) for no diversity loss\n #ride_loss_logits = output_logits if self.additional_diversity_factor == 0 else logits_item\n if self.m_list is None:\n loss += self.base_loss_factor * self.base_loss(ride_loss_logits, target)\n else:\n final_output = self.get_final_output(ride_loss_logits, target)\n loss += self.base_loss_factor * self.base_loss(final_output, target, weight=self.per_cls_weights_base)\n \n base_diversity_temperature = self.base_diversity_temperature\n\n if self.per_cls_weights_diversity is not None:\n diversity_temperature = base_diversity_temperature * self.per_cls_weights_diversity.view((1, -1))\n temperature_mean = diversity_temperature.mean().item()\n else:\n diversity_temperature = base_diversity_temperature\n temperature_mean = base_diversity_temperature\n \n output_dist = F.log_softmax(logits_item / diversity_temperature, dim=1)\n with torch.no_grad():\n # Using the mean takes only linear instead of quadratic time in computing and has only a slight difference so using the mean is preferred here\n mean_output_dist = F.softmax(output_logits / diversity_temperature, dim=1)\n \n loss += self.additional_diversity_factor * temperature_mean * temperature_mean * F.kl_div(output_dist, mean_output_dist, reduction='batchmean')\n \n return loss\n\n\n\n\n\n\n\n", "repo_name": "KaihuaTang/Generalized-Long-Tailed-Benchmarks.pytorch", "sub_path": "utils/training_utils.py", "file_name": "training_utils.py", "file_ext": "py", "file_size_in_byte": 18258, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 106, "dataset": "github-code", "pt": "20", "api": [{"api_name": "torch.optim.Adam", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.optim.SGD", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 48, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.optim.SGD", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 80, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.CosineAnnealingLR", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 92, "usage_type": "attribute"}, {"api_name": "torch.optim", "line_number": 92, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 94, "usage_type": "attribute"}, {"api_name": "torch.optim", "line_number": 94, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.MultiStepLR", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 96, "usage_type": "attribute"}, {"api_name": "torch.optim", "line_number": 96, "usage_type": "name"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 106, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 124, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 124, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 130, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.nn.Parameter", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 132, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.clamp", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 142, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 142, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 148, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.nn.Parameter", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 150, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.norm", "line_number": 153, "usage_type": "call"}, {"api_name": "torch.clamp", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 167, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 167, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 173, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.nn.Parameter", "line_number": 175, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 175, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 175, "usage_type": "call"}, {"api_name": "torch.nn.TripletMarginLoss", "line_number": 176, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 176, "usage_type": "name"}, {"api_name": "torch.arange", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 195, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 195, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 201, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 202, "usage_type": "call"}, {"api_name": "torch.nn.functional.cross_entropy", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 208, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 211, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 211, "usage_type": "name"}, {"api_name": "torch.nn.functional.log_softmax", "line_number": 225, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 225, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 239, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 239, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 244, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 245, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 256, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 257, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 260, "usage_type": "call"}, {"api_name": "torch.uint8", "line_number": 260, "usage_type": "attribute"}, {"api_name": "torch.matmul", "line_number": 264, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 268, "usage_type": "call"}, {"api_name": "torch.nn.functional.cross_entropy", "line_number": 269, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 269, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 272, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 272, "usage_type": "name"}, {"api_name": "torch.nn.functional.cross_entropy", "line_number": 275, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 275, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 277, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 278, "usage_type": "call"}, {"api_name": "torch.sort", "line_number": 283, "usage_type": "call"}, {"api_name": "torch.sort", "line_number": 284, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 305, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 310, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 316, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 316, "usage_type": "name"}, {"api_name": "torch.nn.functional.cross_entropy", "line_number": 323, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 323, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 341, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 342, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 342, "usage_type": "attribute"}, {"api_name": "numpy.power", "line_number": 350, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 351, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 352, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 353, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 353, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 357, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 357, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 363, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 365, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 367, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 367, "usage_type": "attribute"}, {"api_name": "torch.zeros_like", "line_number": 399, "usage_type": "call"}, {"api_name": "torch.uint8", "line_number": 399, "usage_type": "attribute"}, {"api_name": "torch.matmul", "line_number": 403, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 408, "usage_type": "call"}, {"api_name": "torch.nn.functional.log_softmax", "line_number": 438, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 438, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 439, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 441, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 441, "usage_type": "name"}, {"api_name": "torch.nn.functional.kl_div", "line_number": 443, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 443, "usage_type": "name"}]}
+{"seq_id": "36212403703", "text": "#! /usr/bin/python\n# coding: utf-8\nimport os\nimport io\nimport sys\nimport json\nimport time\nimport pprint\nimport argparse\n\npp = pprint.PrettyPrinter(indent=4)\n\nif not os.path.isfile('./config.json'):\n\tparser = argparse.ArgumentParser(description='IBM Bluemix Container \\\n\t\tManagement Instrument for Prototyping and Deployments',\n\t\tformatter_class=argparse.ArgumentDefaultsHelpFormatter)\n\tparser.add_argument(\n\t\t'--flag', '-f', required=False, type=int, \n\t\thelp='flag number current or init value\\n \\\n\t\tdefaults to 0', default=0)\n\tparser.add_argument(\n\t\t'--ip', '-p', required=False, type=str, \n\t\thelp='optional public IP address')\n\tparser.add_argument(\n\t\t'--namespace', '-s', required=True, type=str, \n\t\thelp='designated namespace')\n\tparser.add_argument(\n\t\t'--directory', '-d', required=True, type=str, \n\t\thelp='absolute directory suggested')\n\tparser.add_argument(\n\t\t'--ports', '-r', required=True, type=int, nargs='+',\n\t\thelp='space separated list of ports, include 80')\n\tparser.add_argument(\n\t\t'--image', '-i', required=False, type=str, \n\t\thelp='static image name\\n \\\n\t\tdefaults to \"xyz\"', default='xyz')\n\tparser.add_argument(\n\t\t'--container', '-n', required=False, type=str, \n\t\thelp='static base container name\\n \\\n\t\tdefaults to \"bmContainer', default='bmContainer')\n\tparser.add_argument(\n\t\t'--url', '-u', required=False, type=str, \n\t\thelp='Bluemix base url, defaults to:\\n \\\n\t\tregistry.ng.bluemix.net/',\n\t\tdefault='registry.ng.bluemix.net/')\n\n\targs = parser.parse_args()\n\n\tconfigs = {\n\t\t'flag': args.flag,\n\t\t'directory': args.directory,\n\t\t'ports': args.ports,\n\t\t'ip': args.ip,\n\t\t'namespace': args.namespace,\n\t\t'image': args.image,\n\t\t'container': args.container,\n\t\t'url': args.url\n\t}\n\n\tprint('*** CONFIRMATION REQUIRED ***\\\n\t\t\\nplease confirm writing the following configurations to file:\\n')\n\tpp.pprint(configs)\n\tprint('\\nReturn Y to proceed or anything else to exit')\n\tans = raw_input()\n\tif ans.lower() != 'y':\n\t\tsys.exit()\n\n\twith io.open('config.json', 'w', encoding='utf-8') as f:\n\t\tf.write(unicode(json.dumps(configs, \n\t\t\t\t\t\tensure_ascii=False, indent = 4)))\n\n\tprint('*' * 40)\n\tprint('Congratulations, configuration is completed. \\\n\t\t \\nTo delete cartridge and images, run script without args: \\\n\t\t $ python image.py')\n\tsys.exit()\n\n\nwith open('config.json') as data_file: \n data = json.load(data_file)\n\ncontainer = data['container']\ndirectory = data['directory']\nflag = data['flag']\nimage = data['image']\nip = data['ip']\nnamespace = data['namespace']\nports = data['ports']\nurl = data['url']\n\nimageAddr = namespace + '/' + image + ':'\n\n# 1 - stop old container\noldContainer = container + str(flag)\ncommand = \"ice stop \" + oldContainer\nos.system(command)\n\n# 2 - if ip, unbind\nif ip is not None:\n\tcommand = \"ice ip unbind \" + ip + \" \" + oldContainer\n\tos.system(command)\n\n\n#3 - delete previous images\npreviousImageShort = imageAddr + str(flag)\npreviousImageLong = url + previousImageShort\ncommand = \"ice --local rmi \" + previousImageShort\nos.system(command)\ncommand = \"ice --local rmi \" + previousImageLong\nos.system(command)\ncommand = \"ice --cloud rmi \" + previousImageShort\nos.system(command)\n\n#4 - build, tag, push new image\nnewImageShort = imageAddr + str(flag + 1)\nnewImageLong = url + newImageShort\ncommand = \"ice --local build -t \" + newImageShort + \" \" + directory\nos.system(command)\ncommand = \"ice --local tag \" + newImageShort + \" \" + newImageLong\nos.system(command)\ncommand = \"ice --local push \" + newImageLong\n\n#5 - run new container\nnewContainer = container + str(flag + 1)\nspecifiedPorts = ''\nfor port in ports:\n\tspecifiedPorts+='-p ' + str(port) + ' '\ncommand = \"ice run --name \" + newContainer + \" \" + specifiedPorts + \" \" + newImageShort\nos.system(command)\n\n#6 - if ip bind port to new container\nif ip is not None:\n\tcommand = \"ice ip bind \" + ip + \" \" + newContainer\n\tos.system(command)\n\n#7 - delete old container\ncommand = \"ice rm \" + oldContainer\nos.system(command)\n\ndata['flag']+=1\n\nwith io.open('config.json', 'w+', encoding='utf-8') as f:\n\tf.write(unicode(json.dumps(data, \n\t\t\t\t\tensure_ascii=False, indent = 4)))\n", "repo_name": "aug2uag/BlueImage", "sub_path": "src/script.py", "file_name": "script.py", "file_ext": "py", "file_size_in_byte": 4045, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "20", "api": [{"api_name": "pprint.PrettyPrinter", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 14, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 16, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 66, "usage_type": "call"}, {"api_name": "io.open", "line_number": 68, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 69, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 76, "usage_type": "call"}, {"api_name": "json.load", "line_number": 80, "usage_type": "call"}, {"api_name": "os.system", "line_number": 96, "usage_type": "call"}, {"api_name": "os.system", "line_number": 101, "usage_type": "call"}, {"api_name": "os.system", "line_number": 108, "usage_type": "call"}, {"api_name": "os.system", "line_number": 110, "usage_type": "call"}, {"api_name": "os.system", "line_number": 112, "usage_type": "call"}, {"api_name": "os.system", "line_number": 118, "usage_type": "call"}, {"api_name": "os.system", "line_number": 120, "usage_type": "call"}, {"api_name": "os.system", "line_number": 129, "usage_type": "call"}, {"api_name": "os.system", "line_number": 134, "usage_type": "call"}, {"api_name": "os.system", "line_number": 138, "usage_type": "call"}, {"api_name": "io.open", "line_number": 142, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 143, "usage_type": "call"}]}
+{"seq_id": "36488311623", "text": "import re\n\nfrom pyppeteer.browser import Browser, Page\nfrom pyppeteer.element_handle import ElementHandle\nfrom sqlalchemy.ext.asyncio import AsyncSession\n\nfrom app.db.curriculum import store_or_update_curriculum\nfrom app.db.models import Curriculum\nfrom app.db.program import get_program_by_sigaa_id, get_programs\nfrom app.scraper.constants import ELEMENT_INNER_TEXT, curricula_list_base_url\nfrom app.scraper.utils import get_page\n\n\nasync def get_cell_text_by_header_text(page: Page, th_text: str) -> str:\n expression = f\"//th[contains(., '{th_text}')]/following-sibling::td\"\n [cell_element, *_] = await page.Jx(expression)\n\n cell_inner_text: str = await page.evaluate(ELEMENT_INNER_TEXT, cell_element)\n\n return cell_inner_text\n\n\nasync def get_programs_sigaa_ids(\n session: AsyncSession, program_sigaa_id: int | None = None\n) -> set[int]:\n programs_sigaa_ids: set[int] = set()\n\n if program_sigaa_id:\n programs_sigaa_ids.add(program_sigaa_id)\n else:\n programs = await get_programs(session)\n programs_sigaa_ids.update(program.sigaa_id for program in programs)\n\n return programs_sigaa_ids\n\n\ndef get_program_curricula_url(program_sigaa_id: int) -> str:\n return f\"{curricula_list_base_url}?id={program_sigaa_id}\"\n\n\nasync def get_program_curricula_page(browser: Browser, program_sigaa_id: int) -> Page:\n program_curricula_url = get_program_curricula_url(program_sigaa_id)\n\n page = await get_page(browser, url=program_curricula_url)\n\n return page\n\n\nasync def get_curricula_tr_elements(\n page: Page, curriculum_sigaa_id: str | None = None\n) -> list[ElementHandle]:\n table_xpath_selector = \"//table[@id='table_lt']\"\n tr_xpath_selector = \"//tr[(@class='linha_par' or @class='linha_impar')]\"\n xpath_selector = f\"{table_xpath_selector}{tr_xpath_selector}\"\n\n if curriculum_sigaa_id:\n td_xpath_selector = f\"[descendant::td[contains(., '{curriculum_sigaa_id}')]]\"\n xpath_selector += td_xpath_selector\n\n curricula_tr = await page.xpath(xpath_selector)\n\n return curricula_tr\n\n\nasync def get_curriculum_status(curriculum_tr: ElementHandle) -> bool:\n active_elements = await curriculum_tr.xpath(\"td[contains(., 'Ativa')]\")\n\n return bool(active_elements)\n\n\nasync def get_curriculum_page(\n browser: Browser, program_sigaa_id: int, curriculum_sigaa_id: str\n) -> Page:\n program_curricula_url = get_program_curricula_url(program_sigaa_id)\n\n page = await browser.newPage()\n await page.goto(program_curricula_url)\n\n [curriculum_tr] = await get_curricula_tr_elements(page, curriculum_sigaa_id)\n button = await curriculum_tr.J(\"a[title='Relatório da Estrutura Curricular']\")\n\n if not button:\n raise Exception(\"Could not find button to open curriculum page\")\n\n await button.click()\n await page.waitForNavigation()\n\n return page\n\n\nasync def get_curriculum_sigaa_id_by_tr_element(curriculum_tr: ElementHandle) -> str:\n raw_sigaa_id = await curriculum_tr.Jeval(\"td:first-child\", ELEMENT_INNER_TEXT)\n\n sigaa_id_pattern = \"Detalhes da Estrutura Curricular (.*),\"\n sigaa_id_match = re.search(sigaa_id_pattern, raw_sigaa_id)\n\n if not sigaa_id_match:\n raise Exception(\"Could not find curriculum sigaa_id\")\n\n sigaa_id = sigaa_id_match.group(1).strip()\n\n return sigaa_id\n\n\nasync def get_curriculum_sigaa_id(curriculum_page: Page) -> str:\n return await get_cell_text_by_header_text(curriculum_page, \"Código\")\n\n\nasync def get_curriculum_start_period(curriculum_page: Page) -> tuple[int, int]:\n header_text = \"Período Letivo de Entrada em Vigor\"\n raw_start_period = await get_cell_text_by_header_text(curriculum_page, header_text)\n\n [start_year, start_period] = raw_start_period.split(\".\")\n start_year = int(start_year)\n start_period = int(start_period)\n\n return start_year, start_period\n\n\nasync def get_curriculum_min_periods(curriculum_page: Page) -> int:\n min_periods = await get_cell_text_by_header_text(curriculum_page, \"Mínimo:\")\n min_periods = int(min_periods)\n\n return min_periods\n\n\nasync def get_curriculum_max_periods(curriculum_page: Page) -> int:\n max_periods = await get_cell_text_by_header_text(curriculum_page, \"Máximo:\")\n max_periods = int(max_periods)\n\n return max_periods\n\n\nasync def get_curriculum_min_period_workload(curriculum_page: Page) -> int:\n header_text = \"Carga Horária Mínima por Período Letivo\"\n raw_workload = await get_cell_text_by_header_text(curriculum_page, header_text)\n\n min_period_workload = format_workload_to_number(raw_workload)\n\n return min_period_workload\n\n\ndef format_workload_to_number(raw_workload: str) -> int:\n workload = int(raw_workload.replace(\"h\", \"\"))\n\n return workload\n\n\nasync def get_curriculum_max_period_workload(curriculum_page: Page) -> int:\n header_text = \"Carga Horária Máxima por Período Letivo\"\n raw_workload = await get_cell_text_by_header_text(curriculum_page, header_text)\n\n max_period_workload = format_workload_to_number(raw_workload)\n\n return max_period_workload\n\n\nasync def get_curriculum_min_workload(curriculum_page: Page) -> int:\n header_text = \"Total Mínima\"\n raw_workload = await get_cell_text_by_header_text(curriculum_page, header_text)\n\n min_workload = format_workload_to_number(raw_workload)\n\n return min_workload\n\n\nasync def get_curriculum_mandatory_components_workload(curriculum_page: Page) -> int:\n header_text = \"Total:\"\n raw_workload = await get_cell_text_by_header_text(curriculum_page, header_text)\n\n mandatory_components_workload = format_workload_to_number(raw_workload)\n\n return mandatory_components_workload\n\n\nasync def get_curriculum_min_elective_components_workload(curriculum_page: Page) -> int:\n header_text = \"Carga Horária Optativa Mínima:\"\n raw_workload = await get_cell_text_by_header_text(curriculum_page, header_text)\n\n min_elective_components_workload = format_workload_to_number(raw_workload)\n\n return min_elective_components_workload\n\n\nasync def get_curriculum_min_complementary_components_workload(\n curriculum_page: Page,\n) -> int:\n header_text = \"Carga Horária Complementar Mínima:\"\n raw_workload = await get_cell_text_by_header_text(curriculum_page, header_text)\n\n min_complementary_components_workload = format_workload_to_number(raw_workload)\n\n return min_complementary_components_workload\n\n\nasync def get_curriculum_max_complementary_components_workload(\n curriculum_page: Page,\n) -> int:\n header_text = \"Carga Horária Máxima de Componentes Eletivos\"\n raw_workload = await get_cell_text_by_header_text(curriculum_page, header_text)\n\n max_complementary_components_workload = format_workload_to_number(raw_workload)\n\n return max_complementary_components_workload\n\n\nasync def get_curriculum(\n session: AsyncSession, curriculum_page: Page, program_sigaa_id: int, active: bool\n) -> Curriculum:\n sigaa_id = await get_curriculum_sigaa_id(curriculum_page)\n start_year, start_period = await get_curriculum_start_period(curriculum_page)\n min_periods = await get_curriculum_min_periods(curriculum_page)\n max_periods = await get_curriculum_max_periods(curriculum_page)\n min_period_workload = await get_curriculum_min_period_workload(curriculum_page)\n max_period_workload = await get_curriculum_max_period_workload(curriculum_page)\n min_workload = await get_curriculum_min_workload(curriculum_page)\n\n mandatory_components_workload = await get_curriculum_mandatory_components_workload(\n curriculum_page\n )\n\n min_elective_components_workload = (\n await get_curriculum_min_elective_components_workload(curriculum_page)\n )\n\n max_elective_components_workload = min_elective_components_workload\n\n min_complementary_components_workload = (\n await get_curriculum_min_complementary_components_workload(curriculum_page)\n )\n\n max_complementary_components_workload = (\n await get_curriculum_max_complementary_components_workload(curriculum_page)\n )\n\n program = await get_program_by_sigaa_id(session, program_sigaa_id)\n\n if not program:\n raise Exception(\"Program not found\")\n\n curriculum = Curriculum(\n sigaa_id=sigaa_id,\n active=active,\n start_year=start_year,\n start_period=start_period,\n min_periods=min_periods,\n max_periods=max_periods,\n min_period_workload=min_period_workload,\n max_period_workload=max_period_workload,\n min_workload=min_workload,\n mandatory_components_workload=mandatory_components_workload,\n min_elective_components_workload=min_elective_components_workload,\n max_elective_components_workload=max_elective_components_workload,\n min_complementary_components_workload=min_complementary_components_workload,\n max_complementary_components_workload=max_complementary_components_workload,\n program_id=program.id,\n )\n\n return curriculum\n\n\nasync def scrape_curricula(\n browser: Browser,\n session: AsyncSession,\n program_sigaa_id: int | None = None,\n only_active: bool = True,\n):\n \"\"\"Scrape and store (or update) curricula.\"\"\"\n\n programs_sigaa_ids: set[int]\n programs_sigaa_ids = await get_programs_sigaa_ids(session, program_sigaa_id)\n\n # There are too many programs (~150) to open all tabs at once\n for p_sigaa_id in programs_sigaa_ids:\n curricula_page = await get_program_curricula_page(browser, p_sigaa_id)\n curricula_tr = await get_curricula_tr_elements(curricula_page)\n\n for curriculum_tr in curricula_tr:\n active = await get_curriculum_status(curriculum_tr)\n\n if only_active and not active:\n continue\n\n sigaa_id = await get_curriculum_sigaa_id_by_tr_element(curriculum_tr)\n curriculum_page = await get_curriculum_page(browser, p_sigaa_id, sigaa_id)\n\n curriculum = await get_curriculum(\n session, curriculum_page, p_sigaa_id, active\n )\n\n await store_or_update_curriculum(session, curriculum)\n\n await curriculum_page.close()\n\n await curricula_page.waitFor(1000)\n await curricula_page.close()\n", "repo_name": "irwinschmitt/fluxo-agil-api", "sub_path": "app/scraper/curricula.py", "file_name": "curricula.py", "file_ext": "py", "file_size_in_byte": 10158, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "pyppeteer.browser.Page", "line_number": 14, "usage_type": "name"}, {"api_name": "app.scraper.constants.ELEMENT_INNER_TEXT", "line_number": 18, "usage_type": "argument"}, {"api_name": "sqlalchemy.ext.asyncio.AsyncSession", "line_number": 24, "usage_type": "name"}, {"api_name": "app.db.program.get_programs", "line_number": 31, "usage_type": "call"}, {"api_name": "app.scraper.constants.curricula_list_base_url", "line_number": 38, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Browser", "line_number": 41, "usage_type": "name"}, {"api_name": "app.scraper.utils.get_page", "line_number": 44, "usage_type": "call"}, {"api_name": "pyppeteer.browser.Page", "line_number": 41, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 50, "usage_type": "name"}, {"api_name": "pyppeteer.element_handle.ElementHandle", "line_number": 51, "usage_type": "name"}, {"api_name": "pyppeteer.element_handle.ElementHandle", "line_number": 65, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Browser", "line_number": 72, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 73, "usage_type": "name"}, {"api_name": "pyppeteer.element_handle.ElementHandle", "line_number": 91, "usage_type": "name"}, {"api_name": "app.scraper.constants.ELEMENT_INNER_TEXT", "line_number": 92, "usage_type": "argument"}, {"api_name": "re.search", "line_number": 95, "usage_type": "call"}, {"api_name": "pyppeteer.browser.Page", "line_number": 105, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 109, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 120, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 127, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 134, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 149, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 158, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 167, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 176, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 186, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 197, "usage_type": "name"}, {"api_name": "sqlalchemy.ext.asyncio.AsyncSession", "line_number": 208, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Page", "line_number": 208, "usage_type": "name"}, {"api_name": "app.db.program.get_program_by_sigaa_id", "line_number": 236, "usage_type": "call"}, {"api_name": "app.db.models.Curriculum", "line_number": 241, "usage_type": "call"}, {"api_name": "app.db.models.Curriculum", "line_number": 209, "usage_type": "name"}, {"api_name": "pyppeteer.browser.Browser", "line_number": 263, "usage_type": "name"}, {"api_name": "sqlalchemy.ext.asyncio.AsyncSession", "line_number": 264, "usage_type": "name"}, {"api_name": "app.db.curriculum.store_or_update_curriculum", "line_number": 291, "usage_type": "call"}]}
+{"seq_id": "74899025010", "text": "import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport matplotlib.cm as cm\n\n# name = './mandelbrot/data/mandelbrot_10000'\nname = './julia/data/julia_-0.134_0.250_100'\n\ndf= pd.read_csv(name+'.csv', header=None, dtype='float')\n\nsize = len(df)\n# drop last empty column\ndel df[df.columns[-1]] \n\nfractal_set = name.split('/')[1]\nif fractal_set == 'julia':\n cmap = cm.bone_r # julia\nelse:\n cmap = cm.viridis # mandelbrot\n\n# set colour for stable numbers / members\nmasked_array = np.ma.masked_where(df== 100.0, df)\ncmap.set_bad(color='black')\n\nplt.imshow(masked_array, cmap=cmap, interpolation='none')\nplt.gca().set_aspect(\"equal\")\nplt.axis(\"off\")\nplt.tight_layout()\n# plt.show()\n\npng_name = './'+fractal_set+'/plots/'+name.split('/')[-1]+'.png'\nplt.savefig(png_name, dpi=1000, bbox_inches=None)\n", "repo_name": "astroDimitrios/Fortran", "sub_path": "19_GDB/plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 818, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "20", "api": [{"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.cm.bone_r", "line_number": 17, "usage_type": "attribute"}, {"api_name": "matplotlib.cm", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.cm.viridis", "line_number": 19, "usage_type": "attribute"}, {"api_name": "matplotlib.cm", "line_number": 19, "usage_type": "name"}, {"api_name": "numpy.ma.masked_where", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 22, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}]}
+{"seq_id": "42556080100", "text": "import hashlib\nimport os\nimport uuid\n\nfrom django.db import models\n\nfrom .issue import Issue\nfrom .timestamped import TimestampedModel\n\n\ndef get_s3_key(file_upload, filename):\n \"\"\"\n Get S3 key for the file - use a hash of the file bytes to\n ensure that files are unique and that filenames are not easily guessed.\n \"\"\"\n file = file_upload.file\n if file._file:\n img_bytes = file._file.file.read()\n file._file.file.seek(0)\n filename_base = hashlib.md5(img_bytes).hexdigest()\n _, filename_ext = os.path.splitext(filename)\n filename = filename_base + filename_ext.lower()\n\n return f\"{file_upload.UPLOAD_KEY}/{filename}\"\n\n\nclass FileUpload(TimestampedModel):\n \"\"\"\n An image or document uploaded by a user as a part of a issue.\n \"\"\"\n\n UPLOAD_KEY = \"file-uploads\"\n\n id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)\n file = models.FileField(upload_to=get_s3_key)\n issue = models.ForeignKey(Issue, on_delete=models.SET_NULL, null=True, blank=True)\n", "repo_name": "AnikaLegal/clerk", "sub_path": "app/core/models/upload.py", "file_name": "upload.py", "file_ext": "py", "file_size_in_byte": 1042, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 31, "dataset": "github-code", "pt": "24", "api": [{"api_name": "hashlib.md5", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "timestamped.TimestampedModel", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.models.UUIDField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 34, "usage_type": "attribute"}, {"api_name": "django.db.models.FileField", "line_number": 35, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 36, "usage_type": "call"}, {"api_name": "issue.Issue", "line_number": 36, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 36, "usage_type": "attribute"}]}
+{"seq_id": "27184543162", "text": "import numpy as np\nimport tensorflow as tf\nfrom networks import generator\nfrom options.test_options import TestOptions\nfrom tqdm import tqdm\nimport os\nimport cv2\nimport glob\n\nopt = TestOptions().parse()\nsave_path = os.path.join(opt.checkpoint_path, opt.save_path, opt.image_save_path)\nif not os.path.exists(save_path):\n os.makedirs(save_path)\ncheckpoint_path = opt.weight_path\n\nclass get_evaluation(object):\n def __init__(self, opt):\n self.opt = opt\n self.test_list = glob.glob(os.path.join(self.opt.test_path, '*.jpg'))\n\n\n def open_image(self, path, width, height, angle, isDown=True, isCrop=False, isResize=True, isflip=False, isRotate=False):\n img = cv2.imread(path)\n img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)\n if isCrop:\n img = img[20:198, 0:178]\n if isResize:\n img = cv2.resize(img, (width, height), interpolation=cv2.INTER_LINEAR)\n if isflip: # horizontal flip\n img = cv2.flip(img, 1)\n if isRotate:\n img_waf = img\n img = cv2.getRotationMatrix2D((width / 2, height / 2), angle, 1)\n img = cv2.warpAffine(img_waf, img, (width, height))\n if isDown:\n img_lr_2 = cv2.resize(img, (64, 64), interpolation=cv2.INTER_LINEAR)\n img_lr_4 = cv2.resize(img_lr_2, (32, 32), interpolation=cv2.INTER_LINEAR)\n img_lr = cv2.resize(img_lr_4, (16, 16), interpolation=cv2.INTER_LINEAR)\n\n img_lr = img_lr.astype(np.float32)\n img = img.astype(np.float32)\n\n return img, img_lr\n\n def get_psnr_ssim(self,):\n\n for image in tqdm(sorted(self.test_list)):\n name = image.split('/')[-1]\n name = name.split('.')[0]\n imgs = []\n imgs_lr = []\n\n img, img_lr = self.open_image(image, width=self.opt.crop_size, height=self.opt.crop_size,\n isDown=True,\n isCrop=False,\n isResize=True,\n isflip=False,\n isRotate=False,\n angle=0)\n\n imgs.append(img)\n imgs_lr.append(img_lr)\n\n imgs_hr = np.array(imgs)\n imgs_lr = np.array(imgs_lr)\n\n imgs_sr, edgex2, edgex4, edgex8 = sess.run([RGB, Step1_edge, Step2_edge, Step3_edge],\n feed_dict={X_hr: imgs_hr,\n X_lr: imgs_lr})\n\n\n cv2.imwrite(os.path.join(save_path, str(name) + '_SR' + '.jpg'), cv2.cvtColor(imgs_sr[0] / 255, cv2.COLOR_RGB2BGR) * 255)\n imgs_lr = cv2.resize(imgs_lr[0], (128, 128), interpolation=cv2.INTER_LINEAR)\n cv2.imwrite(os.path.join(save_path, str(name) + '_LR' + '.jpg'), cv2.cvtColor(imgs_lr / 255, cv2.COLOR_RGB2BGR) * 255)\n\n\nX_lr = tf.placeholder(tf.float32, shape=[opt.batchSize, opt.crop_size/8, opt.crop_size/8, opt.output_nc])\nX_hr = tf.placeholder(tf.float32, shape=[opt.batchSize, opt.crop_size, opt.crop_size, opt.output_nc])\n\ntraining = False\n\nconfig = tf.ConfigProto()\nconfig.gpu_options.allow_growth = True\nconfig.gpu_options.visible_device_list = opt.gpu_ids\n\nRGB, Step1_edge, Step2_edge, Step3_edge = generator(X_lr)\n\nsess = tf.Session(config=config)\nsess.run(tf.global_variables_initializer())\nsaver = tf.train.Saver(max_to_keep=None)\nsave_file = os.path.join(checkpoint_path, 'G_weight.ckpt')\nsaver.restore(sess, save_file)\nevaluation = get_evaluation(opt)\nevaluation.get_psnr_ssim()\nprint('\\n HR images are generated !')\n", "repo_name": "BenjaminJonghyun/EIPNet", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 3591, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 13, "dataset": "github-code", "pt": "24", "api": [{"api_name": "options.test_options.TestOptions", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 13, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.flip", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.getRotationMatrix2D", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.warpAffine", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 36, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 37, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path", "line_number": 72, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 72, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 72, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 73, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 74, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 74, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 77, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 78, "usage_type": "attribute"}, {"api_name": "tensorflow.ConfigProto", "line_number": 82, "usage_type": "call"}, {"api_name": "networks.generator", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 89, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 90, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "attribute"}]}
+{"seq_id": "36500403307", "text": "from django.contrib.auth.decorators import login_required\nfrom django.shortcuts import render,redirect\nfrom django.http import HttpResponse,HttpResponseRedirect\nfrom django.urls import reverse\nimport datetime\nfrom .trajectoryPredictor import TrajectoryPredictor\nimport json\nfrom random import randrange\nfrom django.http import HttpResponse\nfrom pyecharts.charts import Line\nfrom pyecharts import options as opts\nfrom .models import TrajectoryData,Station,TrajectoryRecord,NextMinTrajectory\nfrom django.contrib.auth.models import User,Group\nfrom django.contrib.auth import authenticate,login,logout\nfrom .manageStaion import addStation\nfrom .serializers import NextMinTrajectorySerializer\nfrom rest_framework import generics,status\nfrom rest_framework.views import APIView\nfrom rest_framework.response import Response\n# Create your views here.\nclass TaskList(APIView):\n def get(self, request, format=None):\n tasks = NextMinTrajectory.objects.all()\n serializer = NextMinTrajectorySerializer(tasks, many=True)\n return Response(serializer.data)\n\n def post(self, request, format=None):\n print(request.data)\n nextMinPredictor = TrajectoryPredictor()\n try:\n longitude,latitude,altitude = float(request.POST[\"longitude\"]),float(request.POST[\"latitude\"]),float(request.POST[\"altitude\"])\n pressure, temperature, humid = float(request.POST[\"pressure\"]), float(request.POST[\"temperature\"]), float(request.POST[\"humid\"])\n nspeed, espeed, uspeed = float(request.POST[\"nspeed\"]), float(request.POST[\"espeed\"]), float(request.POST[\"uspeed\"])\n one_X = [pressure, temperature, humid, nspeed, espeed, uspeed, longitude, latitude, altitude]\n nextlongtitude,nextlatitude,nextaltitude=nextMinPredictor.makeNextMinPrediction(one_X)\n serializer = NextMinTrajectorySerializer(data={\"longitude\":nextlongtitude,\"latitude\":nextlatitude,\"altitude\":nextaltitude})\n if serializer.is_valid():\n serializer.save()\n return Response(serializer.data, status=status.HTTP_201_CREATED)\n else:\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n except:\n return Response(\"参数错误\", status=status.HTTP_400_BAD_REQUEST)\n\n\ndef index(request):\n content={}\n content['user']=request.user\n return render(request,'main/index.html',context=content)\n\ndef tologin(request):\n if request.method == 'POST':\n return login_check(request)\n else:\n return render(request, 'main/login.html',{'login_info':0})\n\ndef login_check(request):\n useremail = request.POST.get('username')\n password = request.POST.get('password')\n n=authenticate(username=useremail,password=password)\n print(n)\n if n:\n login(request,user=n)\n return HttpResponseRedirect(reverse('main:index'))\n return render(request, 'main/login.html')\n\ndef register(request):\n if request.method == 'POST':\n return register_check(request)\n else:\n return render(request, 'main/register.html')\n\ndef register_check(request):\n useremail = request.POST.get('email')\n password = request.POST.get('password')\n username=request.POST.get('username')\n u=User.objects.filter(email=useremail).first()\n if not u:\n User.objects.create_user(username=username,email=useremail,password=password)\n return HttpResponseRedirect(reverse('main:login'))\n return render(request, 'main/register.html')\n\ndef tologout(request):\n logout(request)\n return HttpResponseRedirect(reverse('main:index'))\n\n@login_required\ndef trajectory(request):\n content = {}\n content['user'] = request.user\n global trajectoryPredictor\n trajectoryPredictor=TrajectoryPredictor()\n\n stations = Station.objects.all()\n content[\"stations\"] = stations\n return render(request, 'main/trajectory.html',context=content)\n\n@login_required\ndef trajectorySeries(request):\n stationId, stationName = request.POST[\"stationId\"].split()\n stationId=int(stationId)\n pressure,temperature,humid=float(request.POST[\"pressure\"]),float(request.POST[\"temperature\"]),float(request.POST[\"humid\"])\n nspeed,espeed,uspeed=float(request.POST[\"nspeed\"]),float(request.POST[\"espeed\"]),float(request.POST[\"uspeed\"])\n stationObj=Station.objects.get(stationId=stationId)\n longitude,latitude=stationObj.longitude,stationObj.latitude\n altitude=float(request.POST[\"altitude\"])\n one_X=[pressure,temperature,humid,nspeed,espeed,uspeed,longitude,latitude,altitude]\n\n longitudes, latitudes, altitudes=trajectoryPredictor.makeAPrediction(one_X)\n longitudes.insert(0,longitude)\n latitudes.insert(0,latitude)\n altitudes.insert(0,latitude)\n content={}\n content['user'] = request.user\n content['longitudes']=longitudes\n content['latitudes']=latitudes\n content['altitudes']=altitudes\n content['stationName']=stationObj.stationName\n content['stationLongitude'],content['stationLatitude']=longitude,latitude\n return render(request, 'main/trajectoryMap.html',context=content)\n\n@login_required\ndef historyPage(request):\n return HttpResponseRedirect(reverse('main:manageStation'))\n\n@login_required\ndef manageStation(request):\n content = {}\n content['user'] = request.user\n stations = Station.objects.all()\n content[\"longitude\"] = [s.longitude for s in stations]\n content[\"latitude\"] = [s.latitude for s in stations]\n content[\"altitude\"] = [s.altitude for s in stations]\n content[\"name\"] = [str(s.stationId) + str(s.stationName) for s in stations]\n content[\"stations\"]=stations\n return render(request, 'main/manageStation.html', context=content)\n\n\n\n\n\n\n\n\n@login_required\ndef modelPage(request):\n return render(request, 'main/model.html')\n\n\n\n", "repo_name": "season0817/trajectory_prediction_web", "sub_path": "main/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 5785, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "rest_framework.views.APIView", "line_number": 21, "usage_type": "name"}, {"api_name": "models.NextMinTrajectory.objects.all", "line_number": 23, "usage_type": "call"}, {"api_name": "models.NextMinTrajectory.objects", "line_number": 23, "usage_type": "attribute"}, {"api_name": "models.NextMinTrajectory", "line_number": 23, "usage_type": "name"}, {"api_name": "serializers.NextMinTrajectorySerializer", "line_number": 24, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 25, "usage_type": "call"}, {"api_name": "trajectoryPredictor.TrajectoryPredictor", "line_number": 29, "usage_type": "call"}, {"api_name": "serializers.NextMinTrajectorySerializer", "line_number": 36, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 39, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 39, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 39, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 41, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 41, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 41, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 43, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 43, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 43, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 49, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 55, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 60, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 63, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 64, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 64, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 65, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 71, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.filter", "line_number": 77, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 77, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 77, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 79, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 79, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 79, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 80, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 80, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 81, "usage_type": "call"}, {"api_name": "django.contrib.auth.logout", "line_number": 84, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 85, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 85, "usage_type": "call"}, {"api_name": "trajectoryPredictor.TrajectoryPredictor", "line_number": 92, "usage_type": "call"}, {"api_name": "models.Station.objects.all", "line_number": 94, "usage_type": "call"}, {"api_name": "models.Station.objects", "line_number": 94, "usage_type": "attribute"}, {"api_name": "models.Station", "line_number": 94, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 96, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 87, "usage_type": "name"}, {"api_name": "models.Station.objects.get", "line_number": 104, "usage_type": "call"}, {"api_name": "models.Station.objects", "line_number": 104, "usage_type": "attribute"}, {"api_name": "models.Station", "line_number": 104, "usage_type": "name"}, {"api_name": "trajectoryPredictor.makeAPrediction", "line_number": 109, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 120, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 98, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 124, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 124, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 122, "usage_type": "name"}, {"api_name": "models.Station.objects.all", "line_number": 130, "usage_type": "call"}, {"api_name": "models.Station.objects", "line_number": 130, "usage_type": "attribute"}, {"api_name": "models.Station", "line_number": 130, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 136, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 126, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 147, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 145, "usage_type": "name"}]}
+{"seq_id": "22241275513", "text": "import collections\n\nimport numpy as np\nimport tensorflow.compat.v2 as tf\n\nfrom tensorflow_probability.python.internal import prefer_static as ps\nfrom tensorflow_probability.python.internal import tensor_util\n\nfrom tensorflow.python.util import nest # pylint: disable=g-direct-tensorflow-import\n\n__all__ = [\n 'broadcast_structure',\n 'call_fn',\n 'cast_structure',\n 'expand_as_args',\n 'map_structure_with_named_args'\n]\n\n_is_namedtuple = nest._is_namedtuple # pylint: disable=protected-access\n\n\nUNSPECIFIED = object()\n\n_STRUCTURES_HAVE_MISMATCHING_TYPES = (\n \"The two structures don't have the same sequence type. Input structure has \"\n 'type {input_type}, while shallow structure has type {shallow_type}.'\n)\n\n_STRUCTURES_HAVE_MISMATCHING_LENGTHS = (\n \"The two structures don't have the same sequence length. Input \"\n 'structure has length {input_length}, while shallow structure has length '\n '{shallow_length}.'\n)\n\n_SHALLOW_TREE_HAS_INVALID_KEYS = (\n \"The shallow_tree's keys are not a subset of the input_tree's keys. The \"\n 'shallow_tree has the following keys that are not in the input_tree: {}.'\n)\n\n_IF_SHALLOW_IS_SEQ_INPUT_MUST_BE_SEQ = (\n 'If shallow structure is a sequence, input must also be a sequence. '\n 'Input has type: {}.'\n)\n\n\ndef broadcast_structure(to_structure, from_structure):\n \"\"\"Broadcasts `from_structure` to `to_structure`.\n\n This is useful for downstream usage of `zip` or `tf.nest.map_structure`.\n\n If `from_structure` is a singleton, it is tiled to match the structure of\n `to_structure`. Note that the elements in `from_structure` are not copied if\n this tiling occurs.\n\n Args:\n to_structure: A structure.\n from_structure: A structure.\n\n Returns:\n new_from_structure: Same structure as `to_structure`.\n\n #### Example:\n\n ```python\n a_structure = ['a', 'b', 'c']\n b_structure = broadcast_structure(a_structure, 'd')\n # -> ['d', 'd', 'd']\n c_structure = tf.nest.map_structure(\n lambda a, b: a + b, a_structure, b_structure)\n # -> ['ad', 'bd', 'cd']\n ```\n \"\"\"\n from_parts = tf.nest.flatten(from_structure)\n if len(from_parts) == 1:\n from_structure = tf.nest.map_structure(lambda _: from_parts[0],\n to_structure)\n return from_structure\n\n\ndef cast_structure(value, structure):\n \"\"\"Cast a structure.\"\"\"\n if tf.nest.is_nested(structure):\n if _is_namedtuple(structure): # pylint: disable=protected-access\n return type(structure)(*value)\n else:\n return type(structure)(value)\n return value\n\n\ndef map_structure_with_named_args(func,\n *structures,\n _check_types=True, # pylint: disable=invalid-name\n _expand_composites=False, # pylint: disable=invalid-name\n _up_to=UNSPECIFIED, # pylint: disable=invalid-name\n **named_structures):\n \"\"\"Calls `nest.map_structure` with named args.\n\n Args:\n func: a callable that accepts one or more named arguments.\n *structures: Structures of arguments passed positionally to `func`.\n _check_types: Forwarded as `map_structure(..., check_types=_check_types)`.\n _expand_composites: Forwarded as\n `map_structure(..., expand_composites=_expand_composites)`.\n _up_to: Optional shallow structure to map up to. If provided,\n `nest.map_structure_up_to` is called rather than `nest.map_structure`.\n Default value: `UNSPECIFIED`.\n **named_structures: Structures of arguments passed by name to `func`.\n Returns:\n A new structure matching that of the input structures (or the shallow\n structure `_up_to`, if specified), in which each element is computed\n by applying `func` to the corresponding elements of the input structures.\n\n #### Examples\n\n ```python\n func = lambda x, y: 2 * x + 3 * y\n\n map_structure_with_named_args(func, [1, 2], [10, 11])\n # ==> [32, 37]\n\n map_structure_with_named_args(func, [1, 2], y=[10, 11])\n # ==> [32, 37]\n\n map_structure_with_named_args(func, x=[1, 2], y=[10, 11])\n # ==> [32, 37]\n\n map_structure_with_named_args(func, [10, 11], x=[1, 2])\n # ==> TypeError: () got multiple values for argument 'x'.\n ```\n\n \"\"\"\n names, named_values = (zip(*named_structures.items())\n if named_structures else ((), ()))\n # Wrapper function that takes positional args and passes keyword args.\n def kwarg_passing_fn(*leaf_values):\n return func(*leaf_values[:len(structures)],\n **dict(zip(names, leaf_values[len(structures):])))\n\n map_fn = (nest.map_structure if _up_to is UNSPECIFIED\n else lambda *a, **kw: nest.map_structure_up_to(_up_to, *a, **kw))\n return map_fn(kwarg_passing_fn,\n *(structures + named_values),\n check_types=_check_types,\n expand_composites=_expand_composites)\n\n\ndef map_structure_coroutine(coroutine,\n *structures,\n _expand_composites=False, # pylint: disable=invalid-name\n _up_to=UNSPECIFIED, # pylint: disable=invalid-name\n _with_tuple_paths=False, # pylint: disable=invalid-name\n **named_structures):\n # pylint: disable=g-doc-return-or-yield\n \"\"\"Invokes a coroutine multiple times with args from provided structures.\n\n This is semantically identical to `map_structure_with_named_args`, except\n that the first argument is a generator or coroutine (a callable whose body\n contains `yield` statements) rather than a function. This is invoked with\n arguments from the provided structure(s), thus defining an outer generator/\n coroutine that `yield`s values in sequence from each call to the inner\n `coroutine`.\n\n The argument structures are traversed, and the coroutine is invoked, in\n the order defined by `tf.nest.flatten`. A stripped-down implementation of\n the core logic is as follows:\n\n ```python\n def map_structure_coroutine(coroutine, *structures):\n flat_results = []\n for args in zip(*[tf.nest.flatten(s) for s in structures]):\n retval = yield from coroutine(*args)\n flat_results.append(retval)\n return tf.nest.pack_sequence_as(structures[0], flat_results)\n ```\n\n Args:\n coroutine: a generator/coroutine callable that accepts one or more named\n arguments.\n *structures: Structures of arguments passed positionally to `coroutine`.\n _expand_composites: Forwarded as\n `tf.nest.flatten(..., expand_composites=_expand_composites)`.\n _up_to: Optional shallow structure to map up to. If provided,\n `nest.map_structure_up_to` is called rather than `nest.map_structure`.\n Default value: `UNSPECIFIED`.\n _with_tuple_paths: Python bool. If `True`, the first argument to `coroutine`\n is a tuple path to the current leaf of the argument structure(s).\n Default value: `False`.\n **named_structures: Structures of arguments passed by name to `coroutine`.\n Yields:\n Values `yield`ed by each invocation of `coroutine`, with invocations in\n order corresponding to `tf.nest.flatten`.\n Returns:\n A new structure matching that of the input structures (or the shallow\n structure `_up_to`, if specified), in which each element is the return\n value from applying `coroutine` to the corresponding elements of the input\n structures.\n\n ## Examples\n\n A JointDistributionCoroutine may define a reusable submodel as its own\n coroutine, for example:\n\n ```python\n def horseshoe_prior(path, scale):\n # Auxiliary-variable representation of a horseshoe prior on sparse weights.\n name = ','.join(path)\n z = yield tfd.HalfCauchy(loc=0., scale=scale, name=name + '_z')\n w_noncentered = yield tfd.Normal(\n loc=0., scale=z, name=name + '_w_noncentered')\n return z * w_noncentered\n ```\n\n Note that this submodel yields two auxiliary random variables, and returns the\n sampled weight as a third value.\n\n Using `map_structure_coroutine` we can define a structure of such submodels,\n and collect their return values:\n\n ```\n @tfd.JointDistributionCoroutineAutoBatched\n def model():\n weights = yield from nest_util.map_structure_coroutine(\n horseshoe_prior,\n scale={'a': tf.ones([5]) * 100., 'b': tf.ones([2]) * 1e-2},\n _with_tuple_paths=True)\n # ==> `weights` is a dict of weight values.\n yield tfd.Deterministic(\n tf.sqrt(tf.norm(weights['a'])**2 + tf.norm(weights['b'])**2),\n name='weights_norm')\n\n print(model.event_shape)\n # ==> StructTuple(\n # a_z=TensorShape([5]),\n # a_w_noncentered=TensorShape([5]),\n # b_z=TensorShape([2]),\n # b_w_noncentered=TensorShape([2]),\n # weights_norm=TensorShape([]))\n ```\n \"\"\"\n # pylint: enable=g-doc-return-or-yield\n\n names, named_structure_values = (zip(*named_structures.items())\n if named_structures else ((), ()))\n all_structures = structures + named_structure_values\n result_structure = all_structures[0] if _up_to is UNSPECIFIED else _up_to\n flat_arg_structures = [\n nest.flatten_up_to(result_structure, s)\n for s in all_structures]\n\n if _with_tuple_paths:\n # Pass tuple paths as a first positional arg (before any provided args).\n flat_paths = nest.yield_flat_paths(result_structure,\n expand_composites=_expand_composites)\n flat_arg_structures = [list(flat_paths)] + flat_arg_structures\n num_positional_args = 1 + len(structures)\n else:\n num_positional_args = len(structures)\n\n flat_results = []\n for leaf_values in zip(*flat_arg_structures):\n result = yield from coroutine(\n *leaf_values[:num_positional_args],\n **dict(zip(names, leaf_values[num_positional_args:])))\n flat_results.append(result)\n\n return nest.pack_sequence_as(result_structure, flat_results)\n\n\ndef _force_leaf(struct):\n # Returns `True` if `struct` should be treated as a leaf, rather than\n # expanded/recursed into.\n return hasattr(struct, '_tfp_nest_expansion_force_leaf')\n\n\ndef _force_expand_as_args(struct):\n return hasattr(struct, '_tfp_nest_expansion_force_args')\n\n\ndef expand_as_args(args):\n \"\"\"Returns `True` if `args` should be expanded as `*args`.\"\"\"\n return ((isinstance(args, collections.abc.Sequence) and\n not _is_namedtuple(args) and not _force_leaf(args)) or\n _force_expand_as_args(args))\n\n\ndef _expand_as_kwargs(args):\n # Returns `True` if `args` should be expanded as `**args`.\n return isinstance(args, collections.abc.Mapping) and not _force_leaf(args)\n\n\ndef _maybe_convertible_to_tensor(struct):\n # Returns `True` if `struct` should be passed to `convert_to_tensor`.\n return not _is_namedtuple(struct) or _force_leaf(struct)\n\n\ndef _get_shallow_structure(struct):\n # Get a shallow version of struct where the children are replaced by\n # 'False'.\n return nest.get_traverse_shallow_structure(lambda s: s is struct, struct)\n\n\ndef _nested_convert_to_tensor(struct, dtype=None, name=None):\n \"\"\"Eagerly converts struct to Tensor, recursing upon failure.\"\"\"\n if dtype is not None or not tf.nest.is_nested(struct):\n return tf.convert_to_tensor(struct, dtype=dtype)\n\n if _maybe_convertible_to_tensor(struct):\n try:\n # Try converting the structure wholesale.\n return tf.convert_to_tensor(struct, name=name)\n except (ValueError, TypeError):\n # Unfortunately Eager/Graph mode don't agree on the error type.\n pass\n # Try converting all of its children.\n shallow_struct = _get_shallow_structure(struct)\n return nest.map_structure_up_to(\n shallow_struct, lambda s: _nested_convert_to_tensor(s, name=name), struct)\n\n\ndef convert_args_to_tensor(args, dtype=None, name=None):\n \"\"\"Converts `args` to `Tensor`s.\n\n Use this when it is necessary to convert user-provided arguments that will\n then be passed to user-provided callables.\n\n When `dtype` is `None` this function behaves as follows:\n\n 1A. If the top-level structure is a `list`/`tuple` but not a `namedtuple`,\n then it is left as is and only its elements are converted to `Tensor`s.\n\n 2A. The sub-structures are converted to `Tensor`s eagerly. E.g. if `args` is\n `{'arg': [[1], [2]]}` it is converted to\n `{'arg': tf.constant([[1], [2]])}`. If the conversion fails, it will\n attempt to recurse into its children.\n\n When `dtype` is specified, it acts as both a structural and numeric type\n constraint. `dtype` can be a single `DType`, `None` or a nested collection\n thereof. The conversion rule becomes as follows:\n\n 1B. The return value of this function will have the same structure as `dtype`.\n\n 2B. If the leaf of `dtype` is a concrete `DType`, then the corresponding\n sub-structure in `args` is converted to a `Tensor`.\n\n 3B. If the leaf of `dtype` is `None`, then the corresponding sub-structure is\n converted eagerly as described in the rule 2A above.\n\n Args:\n args: Arguments to convert to `Tensor`s.\n dtype: Optional structure/numeric type constraint.\n name: Optional name-scope to use.\n\n Returns:\n args: Converted `args`.\n\n #### Examples.\n\n This table shows some useful conversion cases. `T` means `Tensor`, `NT` means\n `namedtuple` and `CNT` means a `namedtuple` with a `Tensor`-conversion\n function registered.\n\n | args | dtype | output |\n |:------------:|:----------:|:------------------:|\n | `{\"a\": 1}` | `None` | `{\"a\": T(1)}` |\n | `T(1)` | `None` | `T(1)` |\n | `[1]` | `None` | `[T(1)]` |\n | `[1]` | `tf.int32` | `T([1])` |\n | `[[T(1)]]` | `None` | `[T([1])]` |\n | `[[T(1)]]` | `[[None]]` | `[[T(1)]]` |\n | `NT(1, 2)` | `None` | `NT(T(1), T(2))` |\n | `NT(1, 2)` | `tf.int32` | `T([1, 2])` |\n | `CNT(1, 2)` | `None` | `T(...)` |\n | `[[1, [2]]]` | `None` | `[[T(1), T([2])]]` |\n\n \"\"\"\n if dtype is None:\n if expand_as_args(args) or _expand_as_kwargs(args):\n shallow_args = _get_shallow_structure(args)\n return nest.map_structure_up_to(\n shallow_args, lambda s: _nested_convert_to_tensor(s, name=name), args)\n else:\n return _nested_convert_to_tensor(args, name=name)\n else:\n return nest.map_structure_up_to(\n dtype, lambda s, dtype: _nested_convert_to_tensor(s, dtype, name), args,\n dtype)\n\n\ndef call_fn(fn, args):\n \"\"\"Calls `fn` with `args`, possibly expanding `args`.\n\n Use this function when calling a user-provided callable using user-provided\n arguments.\n\n The expansion rules are as follows:\n\n `fn(*args)` if `args` is a `list` or a `tuple`, but not a `namedtuple`.\n `fn(**args)` if `args` is a `dict`.\n `fn(args)` otherwise.\n\n Args:\n fn: A callable that takes either `args` as an argument(s).\n args: Arguments to `fn`.\n\n Returns:\n result: Return value of `fn`.\n \"\"\"\n\n if expand_as_args(args):\n return fn(*args)\n elif _expand_as_kwargs(args):\n return fn(**args)\n else:\n return fn(args)\n\n\ndef convert_to_nested_tensor(value, dtype=None, dtype_hint=None,\n allow_packing=False, as_shape_tensor=False,\n convert_ref=True, name=None):\n \"\"\"Converts the given `value` to a (structure of) `Tensor`.\n\n This function converts Python objects of various types to a (structure of)\n `Tensor` objects. It accepts `Tensor` objects, numpy arrays, Python lists, and\n Python scalars.\n\n Args:\n value: An object whose structure matches that of `dtype` and for which each\n leaf has a registered `Tensor` conversion function.\n dtype: Optional structure of dtypes defining the structure of outputs and\n the `dtype` argument for nested calls to `convert_to_tensor`. If not\n nested, will be broadcasted to match the structure of `dtype_hint`.\n dtype_hint: Optional structure of dtypes defining the structure of outputs\n and the `dtype_hint` argument for nested calls to `convert_to_tensor`. If\n not nested, will be broadcasted to match the structure of `dtype`.\n allow_packing: Python `bool`, default `False`. If `True`, allow\n `convert_to_nested_tensor` to stack nested lists of Tensors along the\n leading dimension. Otherwise, raise.\n as_shape_tensor: Optional boolean when if `True` uses\n `prefer_static.convert_to_shape_tensor` instead of `tf.convert_to_tensor`\n for JAX compatibility.\n convert_ref: Python `bool`, default `True`. If `True`, convert objects with\n reference semantics to Tensor.\n name: Optional name to use if a new `Tensor` is created. If inputs are\n structured, elements are named accoring to '{name}/{path}.{to}.{elem}'.\n\n Returns:\n tensor: A (structure of) `Tensor` based on `value`.\n \"\"\"\n dtype_is_nested = nest.is_nested(dtype)\n hint_is_nested = nest.is_nested(dtype_hint)\n # If only one of dtype/dtype_hint is nested, broadcast the atom to match.\n if dtype_is_nested and hint_is_nested:\n nest.assert_same_structure(dtype, dtype_hint)\n elif dtype_is_nested:\n dtype_hint = broadcast_structure(dtype, dtype_hint)\n elif hint_is_nested:\n dtype = broadcast_structure(dtype_hint, dtype)\n\n # Call coerce_structure to force the argument structure to match dtype.\n value = coerce_structure(dtype, value)\n\n def convert_fn(path, value, dtype, dtype_hint, name=None):\n if not allow_packing and nest.is_nested(value) and any(\n # Treat arrays like Tensors for full parity in JAX backend.\n tf.is_tensor(x) or isinstance(x, np.ndarray)\n for x in nest.flatten(value)):\n raise NotImplementedError(('Cannot convert a structure of tensors to a '\n 'single tensor. Saw {} at path {}.'\n ).format(value, path))\n if as_shape_tensor:\n return ps.convert_to_shape_tensor(value, dtype, dtype_hint, name=name)\n elif 'KerasTensor' in str(type(value)):\n # This is a hack to detect symbolic Keras tensors to work around\n # b/206660667. The issue was that symbolic Keras tensors would\n # break the Bijector cache on forward/inverse log det jacobian,\n # because tf.convert_to_tensor is not a no-op thereon.\n return value\n elif convert_ref:\n return tf.convert_to_tensor(value, dtype, dtype_hint, name=name)\n else:\n return tensor_util.convert_nonref_to_tensor(\n value, dtype, dtype_hint, name=name)\n\n ### The following branches only affect naming.\n # For unstructured calls, just use the provided name.\n if not nest.is_nested(dtype):\n return convert_fn((), value, dtype, dtype_hint, name=name)\n # For structured calls where name is provided, include a scope and name\n # members according to \"{path}.{to}.{element}\".\n elif name is not None:\n with tf.name_scope(name):\n convert_with_name = lambda path, *args: convert_fn( # pylint: disable=g-long-lambda\n path, *args, name='.'.join(map(str, path)))\n return nest.map_structure_with_tuple_paths_up_to(\n dtype, convert_with_name, value, dtype, dtype_hint, check_types=False)\n # For structured calls without name, skip the scope and don't pass a\n # struct-path to convert-to-tensor.\n else:\n return nest.map_structure_with_tuple_paths_up_to(\n dtype, convert_fn, value, dtype, dtype_hint, check_types=False)\n\n\nclass _DotString(object):\n\n def __str__(self):\n return '.'\n\n def __repr__(self):\n return '.'\n\n\n_DOT = _DotString()\n\n\n# pylint: disable=protected-access\n# TODO(b/173044916): Support namedtuple interop in nest and remove this method.\ndef coerce_structure(shallow_tree, input_tree):\n \"\"\"Coerces the containers in `input_tree` to exactly match `shallow_tree`.\n\n This method largely parallels the behavior of `nest.assert_shallow_structure`,\n but allows `namedtuples` to be interpreted as either sequences or mappings.\n It returns a structure with the container-classes found in `shallow_tree`\n and the contents of `input_tree`, such that `shallow_tree` and `input_tree`\n may be used safely in downstream calls to `nest.map_structure_up_to`.\n\n Note: this method does not currently support `expand_composites`.\n\n Example Usage:\n ```python\n\n ab = collections.namedtuple('AB', 'a b')(0, 1)\n ba = collections.namedtuple('BA', 'b a')(2, 3)\n\n coerce_structure(ab, ba)\n # -> AB(a=3, b=2)\n ```\n\n Args:\n shallow_tree: A (shallow) structure to be populated.\n input_tree: A (parallel) structure of values.\n Returns:\n A structure with containers from shallow_tree and values from input_tree.\n Raises:\n ValueError: When nested sub-structures have differing lengths.\n ValueError: When nested sub-structures have different keys.\n TypeError: When `shallow_tree` is deeper than `input_tree`\n TypeError: When nested sub-structures are incompatible (e.g., list vs dict).\n \"\"\"\n try:\n return _coerce_structure(shallow_tree, input_tree)\n except (ValueError, TypeError) as e:\n str1 = str(nest.map_structure(lambda _: _DOT, shallow_tree))\n str2 = str(nest.map_structure(lambda _: _DOT, input_tree))\n raise type(e)(('{}\\n'\n 'Entire first structure:\\n{}\\n'\n 'Entire second structure:\\n{}'\n ).format(e, str1, str2))\n\n\ndef _coerce_structure(shallow_tree, input_tree):\n \"\"\"Implementation of coerce_structure.\"\"\"\n if not nest.is_nested(shallow_tree):\n return input_tree\n\n if not nest.is_nested(input_tree):\n raise TypeError(\n _IF_SHALLOW_IS_SEQ_INPUT_MUST_BE_SEQ.format(type(input_tree))\n )\n\n if len(input_tree) != len(shallow_tree):\n raise ValueError(\n _STRUCTURES_HAVE_MISMATCHING_LENGTHS.format(\n input_length=len(input_tree), shallow_length=len(shallow_tree)\n )\n )\n\n # Determine whether shallow_tree should be treated as a Mapping or a Sequence.\n # Namedtuples can be interpreted either way (but keys take precedence).\n _shallow_is_namedtuple = nest._is_namedtuple(shallow_tree) # pylint: disable=invalid-name\n _shallow_is_mapping = isinstance(shallow_tree, collections.abc.Mapping) # pylint: disable=invalid-name\n shallow_supports_keys = _shallow_is_namedtuple or _shallow_is_mapping\n shallow_supports_iter = _shallow_is_namedtuple or not _shallow_is_mapping\n\n # Branch-selection depends on both shallow and input container-classes.\n input_is_mapping = isinstance(input_tree, collections.abc.Mapping)\n if nest._is_namedtuple(input_tree):\n if shallow_supports_keys:\n lookup_branch = lambda k: getattr(input_tree, k)\n else:\n input_iter = nest._yield_value(input_tree)\n lookup_branch = lambda _: next(input_iter)\n elif shallow_supports_keys and input_is_mapping:\n lookup_branch = lambda k: input_tree[k]\n elif shallow_supports_iter and not input_is_mapping:\n input_iter = nest._yield_value(input_tree)\n lookup_branch = lambda _: next(input_iter)\n else:\n raise TypeError(\n _STRUCTURES_HAVE_MISMATCHING_TYPES.format(\n input_type=type(input_tree),\n shallow_type=(\n type(shallow_tree.__wrapped__)\n if hasattr(shallow_tree, '__wrapped__')\n else type(shallow_tree)\n ),\n )\n )\n\n flat_coerced = []\n needs_wrapping = type(shallow_tree) is not type(input_tree)\n for shallow_key, shallow_branch in nest._yield_sorted_items(shallow_tree):\n try:\n input_branch = lookup_branch(shallow_key)\n except (KeyError, AttributeError):\n # pylint: disable=raise-missing-from\n raise ValueError(_SHALLOW_TREE_HAS_INVALID_KEYS.format([shallow_key]))\n flat_coerced.append(_coerce_structure(shallow_branch, input_branch))\n # Keep track of whether nested elements have changed.\n needs_wrapping |= input_branch is not flat_coerced[-1]\n\n # Only create a new instance if containers differ or contents changed.\n return (nest._sequence_like(shallow_tree, flat_coerced)\n if needs_wrapping else input_tree)\n\n# pylint: enable=protected-access\n", "repo_name": "tensorflow/probability", "sub_path": "tensorflow_probability/python/internal/nest_util.py", "file_name": "nest_util.py", "file_ext": "py", "file_size_in_byte": 23975, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3997, "dataset": "github-code", "pt": "24", "api": [{"api_name": "tensorflow.python.util.nest._is_namedtuple", "line_number": 19, "usage_type": "attribute"}, {"api_name": "tensorflow.python.util.nest", "line_number": 19, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.nest.flatten", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.nest", "line_number": 73, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 73, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.nest.map_structure", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.nest", "line_number": 75, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 75, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.nest.is_nested", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.nest", "line_number": 82, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 82, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.map_structure", "line_number": 139, "usage_type": "attribute"}, {"api_name": "tensorflow.python.util.nest", "line_number": 139, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.map_structure_up_to", "line_number": 140, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 140, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.flatten_up_to", "line_number": 247, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 247, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.yield_flat_paths", "line_number": 252, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 252, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.pack_sequence_as", "line_number": 266, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 266, "usage_type": "name"}, {"api_name": "collections.abc", "line_number": 281, "usage_type": "attribute"}, {"api_name": "collections.abc", "line_number": 288, "usage_type": "attribute"}, {"api_name": "tensorflow.python.util.nest.get_traverse_shallow_structure", "line_number": 299, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 299, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.nest.is_nested", "line_number": 304, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.nest", "line_number": 304, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 304, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.convert_to_tensor", "line_number": 305, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 305, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.convert_to_tensor", "line_number": 310, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 310, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.map_structure_up_to", "line_number": 316, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 316, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.map_structure_up_to", "line_number": 379, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 379, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.map_structure_up_to", "line_number": 384, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 384, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.is_nested", "line_number": 449, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 449, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.is_nested", "line_number": 450, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 450, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.assert_same_structure", "line_number": 453, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 453, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.is_nested", "line_number": 463, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 463, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.is_tensor", "line_number": 465, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 465, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 465, "usage_type": "attribute"}, {"api_name": "tensorflow.python.util.nest.flatten", "line_number": 466, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 466, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.prefer_static.convert_to_shape_tensor", "line_number": 471, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.prefer_static", "line_number": 471, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.convert_to_tensor", "line_number": 479, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 479, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.tensor_util.convert_nonref_to_tensor", "line_number": 481, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.tensor_util", "line_number": 481, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.is_nested", "line_number": 486, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 486, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.name_scope", "line_number": 491, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 491, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.map_structure_with_tuple_paths_up_to", "line_number": 494, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 494, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.map_structure_with_tuple_paths_up_to", "line_number": 499, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 499, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.map_structure", "line_number": 552, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 552, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.map_structure", "line_number": 553, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 553, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.is_nested", "line_number": 562, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 562, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.is_nested", "line_number": 565, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 565, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest._is_namedtuple", "line_number": 579, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 579, "usage_type": "name"}, {"api_name": "collections.abc", "line_number": 580, "usage_type": "attribute"}, {"api_name": "collections.abc", "line_number": 585, "usage_type": "attribute"}, {"api_name": "tensorflow.python.util.nest._is_namedtuple", "line_number": 586, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 586, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest._yield_value", "line_number": 590, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 590, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest._yield_value", "line_number": 595, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 595, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest._yield_sorted_items", "line_number": 611, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 611, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest._sequence_like", "line_number": 622, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 622, "usage_type": "name"}]}
+{"seq_id": "24268932861", "text": "from clrprint import clrprint\nfrom datetime import datetime\nimport pandas as pd\n\ndef str_to_time(string):\n return datetime.strptime(string, \"%Y-%m-%d\")\n\nCOLUMN_TYPES = [int, int, None, None, str, str_to_time]\nCOLUMN_NAMES = [\"Delivery\", \"CO\", \"Quantity\", \"Logos\", \"Operator\", \"Date\"]\n\ndef search_for(path, target, column=0, chunksize=1000, max_col=6):\n results = pd.DataFrame()\n cols = range(0, max_col)\n for chunk in pd.read_csv(\n path,\n encoding=\"ISO-8859-1\",\n chunksize=chunksize,\n usecols=COLUMN_NAMES\n ):\n\n df = chunk.fillna(0)\n df[\"Date\"] = pd.to_datetime(df[\"Date\"].astype(str), format='mixed', errors=\"ignore\")\n df[\"Delivery\"] = df[\"Delivery\"].astype('int64', errors=\"ignore\")\n df[\"CO\"] = df[\"CO\"].astype('int64', errors=\"ignore\")\n try:\n df[\"Date\"] = df[\"Date\"].dt.strftime('%m/%d/%Y %I:%M %p')\n except:\n pass\n\n target_oftype = COLUMN_TYPES[column](target)\n found = df.loc[df.iloc[:, column] == target_oftype]\n results = pd.concat([results, found], ignore_index=True, sort=False)\n return results\n\nif __name__ == \"__main__\":\n try:\n column = int(input(\"column: \"))\n except Exception as e:\n clrprint(\"[Error]\", f\"{e}\", sep=\"\", clr=\"r,w\")\n\n try:\n target = int(input(\"target: \"))\n except Exception as e:\n clrprint(\"[Error]\", f\"{e}\", sep=\"\", clr=\"r,w\")\n\n start_time = datetime.now()\n results = search_for('res\\index.csv', target, column)\n\n print(results)\n clrprint(\"[DONE] \", \"Search completed in \", f\"{datetime.now() - start_time}\", \".\", sep=\"\", clr=\"g,w,y,w\")\n", "repo_name": "xavmcc3/order-search", "sub_path": "search_csv.py", "file_name": "search_csv.py", "file_ext": "py", "file_size_in_byte": 1651, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 6, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 6, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 32, "usage_type": "call"}, {"api_name": "clrprint.clrprint", "line_number": 39, "usage_type": "call"}, {"api_name": "clrprint.clrprint", "line_number": 44, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 46, "usage_type": "name"}, {"api_name": "clrprint.clrprint", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "name"}]}
+{"seq_id": "27714981665", "text": "# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\nimport logging\nimport os\nimport sys\n\nimport sushy\n\nfrom sushy_oem_idrac import utils\n\nUSERNAME = 'root'\nPASSWORD = 'calvin'\n\nSERVICE_ROOT = 'http://demo.snmplabs.com:80/redfish/v1'\n\nSYSTEM_ID = '437XR1138R2'\n\nBOOT_DEVICE = sushy.VIRTUAL_MEDIA_CD\nBOOT_MODE = sushy.BOOT_SOURCE_MODE_BIOS\n\nBOOT_IMAGE = 'http://demo.snmplabs.com/mini.iso'\n\nLOG = logging.getLogger(__name__)\n\n\ndef main():\n \"\"\"Boot Dell node from virtual media device\"\"\"\n\n LOG.setLevel(logging.INFO)\n handler = logging.StreamHandler()\n handler.setLevel(logging.INFO)\n LOG.addHandler(handler)\n\n authenticator = sushy.auth.BasicAuth(USERNAME, PASSWORD)\n\n conn = sushy.Sushy(SERVICE_ROOT, verify=False, auth=authenticator)\n\n LOG.info('connected to %s', SERVICE_ROOT)\n\n system = conn.get_system(\n os.path.join(SERVICE_ROOT, 'Systems', SYSTEM_ID))\n\n LOG.info('read system resource %s', system.identity)\n\n for manager in system.managers:\n\n LOG.info('trying manager %s', manager.identity)\n\n for v_media in manager.virtual_media.get_members():\n if BOOT_DEVICE not in v_media.media_types:\n continue\n\n LOG.info(\n 'device %s is present at %s', BOOT_DEVICE, manager.identity)\n\n try:\n manager_oem = manager.get_oem_extension('Dell')\n\n except sushy.exceptions.OEMExtensionNotFoundError:\n LOG.info('Dell OEM not found')\n continue\n\n LOG.info('found Dell OEM extension at %s', manager.identity)\n\n if v_media.inserted:\n v_media.eject_media()\n\n LOG.info('ejected virtual media')\n\n v_media.insert_media(BOOT_IMAGE, inserted=True,\n write_protected=True)\n\n LOG.info('inserted boot image %s into virtual media', BOOT_IMAGE)\n\n # the caller (e.g. ironic) sets boot mode first, boot device second\n system.set_system_boot_source(\n BOOT_DEVICE, enabled=sushy.BOOT_SOURCE_ENABLED_CONTINUOUS,\n mode=BOOT_MODE)\n\n # with Dell, patching System tree does not work as expected\n # we need to reboot for the new boot mode to take effect\n utils.reboot_system(system)\n\n LOG.info('set boot mode to %s', BOOT_MODE)\n\n manager_oem.set_virtual_boot_device(\n BOOT_DEVICE, persistent=False, manager=manager, system=system)\n\n LOG.info('set boot device to %s', BOOT_DEVICE)\n\n # real caller should better not use our way to reboot\n utils.reboot_system(system)\n\n LOG.info('system rebooted')\n\n return 0\n\n\nif __name__ == '__main__':\n sys.exit(main())\n", "repo_name": "etingof/sushy-oem-idrac", "sub_path": "sushy_oem_idrac/tests/functional/vmedia_boot.py", "file_name": "vmedia_boot.py", "file_ext": "py", "file_size_in_byte": 3264, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "sushy.VIRTUAL_MEDIA_CD", "line_number": 27, "usage_type": "attribute"}, {"api_name": "sushy.BOOT_SOURCE_MODE_BIOS", "line_number": 28, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 32, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 38, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 39, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 40, "usage_type": "attribute"}, {"api_name": "sushy.auth.BasicAuth", "line_number": 43, "usage_type": "call"}, {"api_name": "sushy.auth", "line_number": 43, "usage_type": "attribute"}, {"api_name": "sushy.Sushy", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "sushy.exceptions", "line_number": 68, "usage_type": "attribute"}, {"api_name": "sushy.BOOT_SOURCE_ENABLED_CONTINUOUS", "line_number": 86, "usage_type": "attribute"}, {"api_name": "sushy_oem_idrac.utils.reboot_system", "line_number": 91, "usage_type": "call"}, {"api_name": "sushy_oem_idrac.utils", "line_number": 91, "usage_type": "name"}, {"api_name": "sushy_oem_idrac.utils.reboot_system", "line_number": 101, "usage_type": "call"}, {"api_name": "sushy_oem_idrac.utils", "line_number": 101, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 109, "usage_type": "call"}]}
+{"seq_id": "14798373595", "text": "import functools\nimport math\nimport gdb\n\nfrom kgdb import *\n\ndef foreach_proc_in_system():\n allproc = gdb.lookup_global_symbol(\"allproc\").value()\n for p in list_foreach(allproc, \"p_list\"):\n yield p\n\n\ndef _vm_map_entry_succ(entry):\n after = entry['right']\n if after['left']['start'] > entry['start']:\n while True:\n after = after['left']\n if after['left'] == entry:\n break\n return after\n\n\ndef foreach_vm_map_entry(map):\n entry = map['header']['right']\n while entry != map['header'].address:\n yield entry\n entry = _vm_map_entry_succ(entry)\n\n\ndef cpu_foreach():\n all_cpus = gdb.lookup_global_symbol(\"all_cpus\").value()\n bitsz = gdb.lookup_type(\"long\").sizeof * 8\n maxid = gdb.lookup_global_symbol(\"mp_maxid\").value()\n\n cpu = 0\n while cpu <= maxid:\n upper = cpu >> int(math.log(bitsz, 2))\n lower = 1 << (cpu & (bitsz - 1))\n if (all_cpus['__bits'][upper] & lower) != 0:\n yield cpu\n cpu = cpu + 1\n\n\n# XXX-MJ doesn't handle \"struct vm_object *\"\n# XXX-MJ assumes there's a return value, assumes one parameter\ndef ctype(t):\n def thunk(f):\n @functools.wraps(f)\n def wrap(a):\n if a.type != gdb.lookup_type(t):\n raise gdb.GdbError(\"parameter type mismatch: expected {} have {}\".format(t, a.type))\n return f(a)\n return wrap\n return thunk\n\n@ctype(\"vm_object_t\")\ndef findobj(obj):\n \"\"\"Find all userspace vm_map entries referencing the specific VM object.\"\"\"\n for p in foreach_proc_in_system():\n for entry in foreach_vm_map_entry(p['p_vmspace']['vm_map'].address):\n if not entry['object']['vm_object']:\n continue\n if entry['object']['vm_object'] == obj:\n gdb.add_history(p)\n gdb.add_history(entry)\n return p\n", "repo_name": "markjdb/scripts", "sub_path": "kgdb/util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 1890, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 5, "dataset": "github-code", "pt": "24", "api": [{"api_name": "gdb.lookup_global_symbol", "line_number": 8, "usage_type": "call"}, {"api_name": "gdb.lookup_global_symbol", "line_number": 31, "usage_type": "call"}, {"api_name": "gdb.lookup_type", "line_number": 32, "usage_type": "call"}, {"api_name": "gdb.lookup_global_symbol", "line_number": 33, "usage_type": "call"}, {"api_name": "math.log", "line_number": 37, "usage_type": "call"}, {"api_name": "gdb.lookup_type", "line_number": 50, "usage_type": "call"}, {"api_name": "gdb.GdbError", "line_number": 51, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 48, "usage_type": "call"}, {"api_name": "gdb.add_history", "line_number": 64, "usage_type": "call"}, {"api_name": "gdb.add_history", "line_number": 65, "usage_type": "call"}]}
+{"seq_id": "31126851278", "text": "import torch\n\nimport unittest\nfrom utils.data.load_dataset import get_train_test_dataset\nfrom gpytorch.kernels import MaternKernel\nfrom hyperparameter_tuning.utils.gpytorch.models.cglb import CGLB\nfrom hyperparameter_tuning.utils.gpytorch.models.variational_gpr import OPTIMIZE_INDUCING_INPUTS, NUM_INDUCING_INPUTS, SELECTION_SCHEME, \\\n CONDITIONAL_VARIANCE, MAX_NUM_CG_STEPS\nfrom hyperparameter_tuning.utils.gpytorch.scipy import Scipy\n\n\nclass CGLBTestCase(unittest.TestCase):\n def test_gradient(self):\n device = \"cpu\"\n X, y, _, _ = get_train_test_dataset(\"wilson_pumadyn32nm\")\n X = torch.as_tensor(X)\n y = torch.as_tensor(y)\n X = X.to(device)\n y = y.to(device)\n k = MaternKernel()\n k = k.to(device)\n k.train()\n sn2 = lambda: torch.tensor(1, dtype=torch.float64, device=device)\n mu = lambda X: torch.zeros(X.shape[0], dtype=torch.float64, device=device)\n cglb = CGLB(X, y, k, sn2, mu, args={OPTIMIZE_INDUCING_INPUTS: True, NUM_INDUCING_INPUTS: 2,\n SELECTION_SCHEME: CONDITIONAL_VARIANCE, MAX_NUM_CG_STEPS: 100},\n device=device)\n\n loss = cglb.create_loss_closure()\n\n if False:\n t = time()\n tt = thread_time()\n l = loss()\n print(f\"time: {time() - t}\")\n print(f\"thread_time: {thread_time() - tt}\")\n\n t = time()\n tt = thread_time()\n l.backward()\n print(f\"time: {time() - t}\")\n print(f\"thread_time: {thread_time() - tt}\")\n\n variables = tuple([v for _, v in cglb.get_named_tunable_parameters()])\n\n if False:\n t = time()\n tt = thread_time()\n l = loss()\n print(f\"time: {time() - t}\")\n print(f\"thread_time: {thread_time() - tt}\")\n\n t = time()\n tt = thread_time()\n grads = torch.autograd.grad(l, variables)\n print(f\"time: {time() - t}\")\n print(f\"thread_time: {thread_time() - tt}\")\n\n # exit()\n\n x0 = Scipy.pack(variables)\n\n def _torch_eval(x):\n values = Scipy.unpack(variables, x)\n Scipy.assign(variables, values)\n return loss()\n\n torch.autograd.gradcheck(_torch_eval, x0)\n\n\nif __name__ == '__main__':\n unittest.main()\n", "repo_name": "SimonBartels/acgp", "sub_path": "experiments/tests/optimization/test_cglb.py", "file_name": "test_cglb.py", "file_ext": "py", "file_size_in_byte": 2374, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "20", "api": [{"api_name": "unittest.TestCase", "line_number": 12, "usage_type": "attribute"}, {"api_name": "utils.data.load_dataset.get_train_test_dataset", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.as_tensor", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.as_tensor", "line_number": 17, "usage_type": "call"}, {"api_name": "gpytorch.kernels.MaternKernel", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.float64", "line_number": 23, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.float64", "line_number": 24, "usage_type": "attribute"}, {"api_name": "hyperparameter_tuning.utils.gpytorch.models.cglb.CGLB", "line_number": 25, "usage_type": "call"}, {"api_name": "hyperparameter_tuning.utils.gpytorch.models.variational_gpr.OPTIMIZE_INDUCING_INPUTS", "line_number": 25, "usage_type": "name"}, {"api_name": "hyperparameter_tuning.utils.gpytorch.models.variational_gpr.NUM_INDUCING_INPUTS", "line_number": 25, "usage_type": "name"}, {"api_name": "hyperparameter_tuning.utils.gpytorch.models.variational_gpr.SELECTION_SCHEME", "line_number": 26, "usage_type": "name"}, {"api_name": "hyperparameter_tuning.utils.gpytorch.models.variational_gpr.MAX_NUM_CG_STEPS", "line_number": 26, "usage_type": "name"}, {"api_name": "hyperparameter_tuning.utils.gpytorch.models.variational_gpr.CONDITIONAL_VARIANCE", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.autograd.grad", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 55, "usage_type": "attribute"}, {"api_name": "hyperparameter_tuning.utils.gpytorch.scipy.Scipy.pack", "line_number": 61, "usage_type": "call"}, {"api_name": "hyperparameter_tuning.utils.gpytorch.scipy.Scipy", "line_number": 61, "usage_type": "name"}, {"api_name": "hyperparameter_tuning.utils.gpytorch.scipy.Scipy.unpack", "line_number": 64, "usage_type": "call"}, {"api_name": "hyperparameter_tuning.utils.gpytorch.scipy.Scipy", "line_number": 64, "usage_type": "name"}, {"api_name": "hyperparameter_tuning.utils.gpytorch.scipy.Scipy.assign", "line_number": 65, "usage_type": "call"}, {"api_name": "hyperparameter_tuning.utils.gpytorch.scipy.Scipy", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.autograd.gradcheck", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 68, "usage_type": "attribute"}, {"api_name": "unittest.main", "line_number": 72, "usage_type": "call"}]}
+{"seq_id": "31768658331", "text": "# boto3 is official AWS SDK for python\r\nimport boto3\r\n\r\n# Dynamo DB connection local connection\r\ndynamodb = boto3.resource('dynamodb', region_name='us-west-2', endpoint_url=\"http://localhost:8000\")\r\n\r\n\r\n# The Following data will be inserted in their respective tables\r\nBOARDS = [\r\n {\"board_id\": 1, \"name\": 'CIE'},\r\n {\"board_id\": 2, \"name\": 'Edexel'},\r\n {\"board_id\": 3, \"name\": 'IB'},\r\n]\r\n\r\nLEVELS = [\r\n {\"level_id\": 1, \"level_name\": 'A Levels'},\r\n {\"level_id\": 2, \"level_name\": 'GCE O Levels'},\r\n {\"level_id\": 3, \"level_name\": 'IG'},\r\n]\r\n\r\nQUESTION_TYPES = [\r\n {\"question_type_id\": 1, \"type\": 'MCQ'},\r\n {\"question_type_id\": 2, \"type\": 'Descriptive'},\r\n]\r\n\r\nSUBJECTS = [\r\n {\"subject_id\": 1, \"subject_name\": 'Chemistry'},\r\n {\"subject_id\": 2, \"subject_name\": 'Biology'},\r\n {\"subject_id\": 3, \"subject_name\": 'Physics'},\r\n {\"subject_id\": 4, \"subject_name\": 'cord_sci'},\r\n]\r\n\r\n# Creating Board Table\r\nboard_table = dynamodb.create_table(\r\n TableName='boards',\r\n KeySchema=[\r\n {\r\n 'AttributeName': 'board_id',\r\n 'KeyType': 'HASH' # Partition key\r\n },\r\n ],\r\n AttributeDefinitions=[\r\n {\r\n 'AttributeName': 'board_id',\r\n 'AttributeType': 'N'\r\n }\r\n ],\r\n # ProvisionedThroughput is ignored in Local Dynamo DB instance\r\n ProvisionedThroughput={\r\n 'ReadCapacityUnits': 10,\r\n 'WriteCapacityUnits': 10\r\n }\r\n)\r\n\r\nprint(\"Boards Table Created!\")\r\n\r\n# Board table data insertion\r\nfor i in range(len(BOARDS)):\r\n board_table.put_item(\r\n Item={\r\n \"board_id\": BOARDS[i][\"board_id\"],\r\n \"name\": BOARDS[i][\"name\"],\r\n }\r\n )\r\n pass\r\n\r\nprint(\"Data inserted in Boards Table\")\r\n\r\n# Creating Level Table\r\nlevel_table = dynamodb.create_table(\r\n TableName='levels',\r\n KeySchema=[\r\n {\r\n 'AttributeName': 'level_id',\r\n 'KeyType': 'HASH' # Partition key\r\n },\r\n ],\r\n AttributeDefinitions=[\r\n {\r\n 'AttributeName': 'level_id',\r\n 'AttributeType': 'N'\r\n }\r\n ],\r\n # ProvisionedThroughput is ignored in Local Dynamo DB instance\r\n ProvisionedThroughput={\r\n 'ReadCapacityUnits': 10,\r\n 'WriteCapacityUnits': 10\r\n }\r\n)\r\n\r\nprint(\"Levels Table Created!\")\r\n\r\n# Level table data insertion\r\nfor i in range(len(LEVELS)):\r\n level_table.put_item(\r\n Item={\r\n \"level_id\": LEVELS[i][\"level_id\"],\r\n \"level_name\": LEVELS[i][\"level_name\"],\r\n }\r\n )\r\n pass\r\n\r\nprint(\"Data inserted in Levels Table\")\r\n\r\n# Creating Level Table\r\nsubject_table = dynamodb.create_table(\r\n TableName='subjects',\r\n KeySchema=[\r\n {\r\n 'AttributeName': 'subject_id',\r\n 'KeyType': 'HASH' # Partition key\r\n },\r\n ],\r\n AttributeDefinitions=[\r\n {\r\n 'AttributeName': 'subject_id',\r\n 'AttributeType': 'N'\r\n }\r\n ],\r\n # ProvisionedThroughput is ignored in Local Dynamo DB instance\r\n ProvisionedThroughput={\r\n 'ReadCapacityUnits': 10,\r\n 'WriteCapacityUnits': 10\r\n }\r\n)\r\n\r\nprint(\"Levels Table Created!\")\r\n\r\n# Level table data insertion\r\nfor i in range(len(SUBJECTS)):\r\n subject_table.put_item(\r\n Item={\r\n \"subject_id\": SUBJECTS[i][\"subject_id\"],\r\n \"subject_name\": SUBJECTS[i][\"subject_name\"],\r\n }\r\n )\r\n pass\r\n\r\nprint(\"Data inserted in Subjects Table\")\r\n\r\n# Creating Question_Types Table\r\nquestion_type_table = dynamodb.create_table(\r\n TableName='question_types',\r\n KeySchema=[\r\n {\r\n 'AttributeName': 'question_type_id',\r\n 'KeyType': 'HASH' # Partition key\r\n },\r\n ],\r\n AttributeDefinitions=[\r\n {\r\n 'AttributeName': 'question_type_id',\r\n 'AttributeType': 'N'\r\n }\r\n ],\r\n # ProvisionedThroughput is ignored in Local Dynamo DB instance\r\n ProvisionedThroughput={\r\n 'ReadCapacityUnits': 10,\r\n 'WriteCapacityUnits': 10\r\n }\r\n)\r\n\r\nprint(\"Levels Table Created!\")\r\n\r\n# Level table data insertion\r\nfor i in range(len(QUESTION_TYPES)):\r\n question_type_table.put_item(\r\n Item={\r\n \"question_type_id\": QUESTION_TYPES[i][\"question_type_id\"],\r\n \"type\": QUESTION_TYPES[i][\"type\"],\r\n }\r\n )\r\n\r\nprint(\"Data inserted in QUESTION_TYPES Table\")\r\n\r\n# Creating Questions Table\r\nquestions_table = dynamodb.create_table(\r\n TableName='questions',\r\n KeySchema=[\r\n {\r\n 'AttributeName': 'question_id',\r\n 'KeyType': 'HASH' # Partition key\r\n },\r\n ],\r\n AttributeDefinitions=[\r\n {\r\n 'AttributeName': 'question_id',\r\n 'AttributeType': 'N'\r\n }\r\n ],\r\n # ProvisionedThroughput is ignored in Local Dynamo DB instance\r\n ProvisionedThroughput={\r\n 'ReadCapacityUnits': 10,\r\n 'WriteCapacityUnits': 10\r\n }\r\n)\r\n\r\nprint(\"Questions Table Created!\")\r\n", "repo_name": "mbshakoor/exam-paper-generator", "sub_path": "import-util/create_tables.py", "file_name": "create_tables.py", "file_ext": "py", "file_size_in_byte": 4984, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "boto3.resource", "line_number": 5, "usage_type": "call"}]}
+{"seq_id": "7385236727", "text": "# 1313. Decompress Run-Length Encoded List\n\nfrom typing import List\n\nclass Solution:\n def decompressRLElist(self, nums: List[int]) -> List[int]:\n encoded = []\n idx = 0\n while idx < len(nums):\n for i in range(nums[idx]):\n encoded.append(nums[idx+1])\n idx += 2\n\n return encoded\n\nsol = Solution().decompressRLElist(nums = [1,2,3,4])\nprint(sol)", "repo_name": "yunyunyang/leetcode-py", "sub_path": "algorithms/easy/decompress_run-length_encoded_list.py", "file_name": "decompress_run-length_encoded_list.py", "file_ext": "py", "file_size_in_byte": 408, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "typing.List", "line_number": 6, "usage_type": "name"}]}
+{"seq_id": "41702102030", "text": "from torchvision import transforms\nimport torchvision.datasets as datasets\nfrom PIL import Image\nimport torch.utils.data as data\nimport torch\nimport numpy as np\nimport cv2\nfrom tqdm import tqdm\nimport random\nimport albumentations\nimport json\n# from tuils.preprocess import *\nfrom torch.utils.data.dataloader import default_collate\n\nfpath = open('../configs/seg_path_configs.json', encoding='utf-8')\npath_data = json.load(fpath)\ntrain_img_path = path_data['train_img_path']\n\ndef generate_transforms(image_size):\n # MAX_SIZE = 448\n IMAGENET_SIZE = image_size\n\n train_transform = albumentations.Compose([\n \n albumentations.Resize(IMAGENET_SIZE, IMAGENET_SIZE),\n albumentations.OneOf([\n albumentations.RandomGamma(gamma_limit=(60, 120), p=0.9),\n albumentations.RandomBrightness(limit=0.2, p=0.9),\n albumentations.RandomContrast(limit=0.2, p=0.9),\n albumentations.CLAHE(clip_limit=4.0, tile_grid_size=(4, 4), p=0.9),\n ]),\n albumentations.OneOf([\n albumentations.Blur(blur_limit=4, p=1),\n albumentations.MotionBlur(blur_limit=4, p=1),\n albumentations.MedianBlur(blur_limit=4, p=1)\n ], p=0.5),\n albumentations.HorizontalFlip(p=0.5),\n albumentations.ShiftScaleRotate(shift_limit=0.2, scale_limit=0.2, rotate_limit=20, interpolation=cv2.INTER_LINEAR,border_mode=cv2.BORDER_CONSTANT, p=1),\n albumentations.Normalize(mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225), max_pixel_value=255.0, p=1.0)\n\n ])\n\n val_transform = albumentations.Compose([\n albumentations.Resize(IMAGENET_SIZE, IMAGENET_SIZE),\n albumentations.Normalize(mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225), max_pixel_value=255.0, p=1.0)\n ])\n\n return train_transform, val_transform\n\ndef rle2mask(rle, width, height):\n mask= np.zeros(width* height)\n array = np.asarray([int(x) for x in rle.split()])\n starts = array[0::2]\n lengths = array[1::2]\n\n current_position = 0\n for index, start in enumerate(starts):\n current_position += start\n mask[current_position:current_position+lengths[index]] = 255\n current_position += lengths[index]\n\n return mask.reshape(width, height)\n\n\n\nclass Siim_Dataset(data.Dataset):\n\n def __init__(self,\n df = None,\n name_list = None,\n transform = None\n ):\n self.df = df\n self.name_list = name_list\n self.transform = transform\n\n def __len__(self):\n return len(self.name_list)\n\n def __getitem__(self, idx):\n\n name = self.name_list[idx]\n image = cv2.imread(train_img_path + name)\n rle = self.df[self.df['ImageId']==(name.replace('.png', '').replace('.jpg', ''))]['EncodedPixels']\n if rle.values[0] == ' -1':\n masks = np.zeros((1024, 1024))\n else:\n masks = [np.expand_dims(rle2mask(x, 1024, 1024).T,axis=0) for x in rle]\n masks = np.sum(masks,0)\n masks[masks>1] = 1\n masks = masks[0, :, :]\n\n if self.transform is not None:\n augmented = self.transform(image=image, mask=masks)\n image = augmented['image'].transpose(2, 0, 1)\n masks = np.expand_dims(augmented['mask'], axis=0)\n\n return image, masks\n\nclass Siim_Dataset_cls_seg_train(data.Dataset):\n\n def __init__(self,\n df = None,\n name_list = None,\n transform = None\n ):\n self.df = df\n self.name_list = name_list\n self.transform = transform\n self.kernel = np.ones((3,3), np.uint8)\n\n def __len__(self):\n return len(self.name_list)\n\n def __getitem__(self, idx):\n\n name = self.name_list[idx]\n image = cv2.imread(train_img_path + name)\n rle = self.df[self.df['ImageId']==(name.replace('.png', '').replace('.jpg', ''))]['EncodedPixels']\n\n if rle.values[0] == '-1':\n masks = np.zeros((1024, 1024))\n cls_label = torch.FloatTensor([0])\n \n elif rle.values[0] == '2':\n masks = np.zeros((1024, 1024))\n cls_label = torch.FloatTensor([1]) \n else:\n\n masks = [np.expand_dims(rle2mask(x, 1024, 1024).T,axis=0) for x in rle]\n masks = np.sum(masks,0)\n masks[masks>1] = 1\n masks = masks[0, :, :]\n\n cls_label = torch.FloatTensor([1])\n\n if self.transform is not None:\n augmented = self.transform(image=image, mask=masks)\n image = augmented['image'].transpose(2, 0, 1)\n masks = np.expand_dims(augmented['mask'], axis=0)\n\n return image, masks, cls_label\n\n\nclass Siim_Dataset_cls_seg_val(data.Dataset):\n\n def __init__(self,\n df = None,\n name_list = None,\n transform = None\n ):\n self.df = df\n self.name_list = name_list\n self.transform = transform\n self.kernel = np.ones((3,3), np.uint8) \n\n def __len__(self):\n return len(self.name_list)\n\n def __getitem__(self, idx):\n\n name = self.name_list[idx]\n image = cv2.imread(train_img_path + name)\n rle = self.df[self.df['ImageId']==(name.replace('.png', '').replace('.jpg', ''))]['EncodedPixels']\n\n if rle.values[0] == '-1':\n masks = np.zeros((1024, 1024))\n cls_label = torch.FloatTensor([0])\n elif rle.values[0] == '2':\n masks = np.zeros((1024, 1024))\n cls_label = torch.FloatTensor([1]) \n else:\n masks = [np.expand_dims(rle2mask(x, 1024, 1024).T,axis=0) for x in rle]\n masks = np.sum(masks,0)\n masks[masks>1] = 1\n masks = masks[0, :, :]\n\n cls_label = torch.FloatTensor([1])\n\n if self.transform is not None:\n augmented = self.transform(image=image, mask=masks)\n image = augmented['image'].transpose(2, 0, 1)\n masks = np.expand_dims(augmented['mask'], axis=0)\n\n return image, masks, cls_label\n\n\n\ndef generate_dataset_loader_cls_seg(df_all, c_train, train_transform, train_batch_size, c_val, val_transform, val_batch_size, workers):\n\n train_dataset = Siim_Dataset_cls_seg_train(df_all, c_train, train_transform)\n val_dataset = Siim_Dataset_cls_seg_val(df_all, c_val, val_transform)\n\n train_loader = torch.utils.data.DataLoader(\n train_dataset,\n batch_size=train_batch_size, \n shuffle=True,\n num_workers=workers,\n pin_memory=True,\n drop_last=True)\n\n val_loader = torch.utils.data.DataLoader(\n val_dataset,\n batch_size=val_batch_size, \n shuffle=False,\n num_workers=workers,\n pin_memory=True,\n drop_last=False)\n\n return train_loader, val_loader\n", "repo_name": "SeuTao/kaggle-competition-solutions", "sub_path": "SIIM19_Pneumothorax_Segmentation_2nd_solution/src_unet_cls/dataset/dataset.py", "file_name": "dataset.py", "file_ext": "py", "file_size_in_byte": 6894, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 238, "dataset": "github-code", "pt": "24", "api": [{"api_name": "json.load", "line_number": 16, "usage_type": "call"}, {"api_name": "albumentations.Compose", "line_number": 23, "usage_type": "call"}, {"api_name": "albumentations.Resize", "line_number": 25, "usage_type": "call"}, {"api_name": "albumentations.OneOf", "line_number": 26, "usage_type": "call"}, {"api_name": "albumentations.RandomGamma", "line_number": 27, "usage_type": "call"}, {"api_name": "albumentations.RandomBrightness", "line_number": 28, "usage_type": "call"}, {"api_name": "albumentations.RandomContrast", "line_number": 29, "usage_type": "call"}, {"api_name": "albumentations.CLAHE", "line_number": 30, "usage_type": "call"}, {"api_name": "albumentations.OneOf", "line_number": 32, "usage_type": "call"}, {"api_name": "albumentations.Blur", "line_number": 33, "usage_type": "call"}, {"api_name": "albumentations.MotionBlur", "line_number": 34, "usage_type": "call"}, {"api_name": "albumentations.MedianBlur", "line_number": 35, "usage_type": "call"}, {"api_name": "albumentations.HorizontalFlip", "line_number": 37, "usage_type": "call"}, {"api_name": "albumentations.ShiftScaleRotate", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 38, "usage_type": "attribute"}, {"api_name": "cv2.BORDER_CONSTANT", "line_number": 38, "usage_type": "attribute"}, {"api_name": "albumentations.Normalize", "line_number": 39, "usage_type": "call"}, {"api_name": "albumentations.Compose", "line_number": 43, "usage_type": "call"}, {"api_name": "albumentations.Resize", "line_number": 44, "usage_type": "call"}, {"api_name": "albumentations.Normalize", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 66, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 66, "usage_type": "name"}, {"api_name": "cv2.imread", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 100, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 100, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 110, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 122, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 126, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 145, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 145, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 155, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 167, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 170, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 174, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 183, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 194, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 194, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 202, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 202, "usage_type": "attribute"}]}
+{"seq_id": "33032791636", "text": "from services.trakt import TraktService\nfrom services.plex import PlexService\nfrom plexapi.video import Video as PlexVideo\nfrom models import Movie, Availability\nfrom typing import List\nimport logging\n\n_logger = logging.getLogger(__name__)\n_logger.setLevel(logging.INFO)\n\nclass MediaService:\n def __init__(self, trakt_config: dict, plex_config: dict, secrets_manager_endpoint: str) -> None:\n self.trakt_service = TraktService(trakt_config.get('secret_name'), secrets_manager_endpoint)\n self.plex_service = PlexService(plex_config)\n\n def recommend_movie(self) -> Movie:\n trakt_movie = self.trakt_service.get_recommended_movie()\n movie: Movie = Movie.from_trakt(trakt_movie)\n movie.get_availability(self)\n return movie\n \n def search(self, query: str, media_type: str, limit: int) -> List[Movie]:\n results: List[PlexVideo] = self.plex_service.search_media(query, media_type, limit)\n media_list: List[Movie] = []\n if (len(results) == 0):\n return media_list\n\n platform_exclusions = ['netflix-basic-with-ads']\n for media in results:\n availability_list: List[Availability] = self.plex_service.get_media_availability(media)\n availability: List[Availability] = list(filter(lambda x: x.platform not in platform_exclusions, availability_list))\n movie: Movie = Movie.from_plex(media)\n movie.availability = availability\n media_list.append(movie)\n return media_list\n \n def get_media_availability(self, media: Movie) -> None:\n results: List[Movie] = self.search(f'{media.title} + ({media.year})', 1)\n if results:\n media.availability = results[0].availability\n\n\n", "repo_name": "mdinicola/watch-wizard", "sub_path": "watch_wizard/services/media.py", "file_name": "media.py", "file_ext": "py", "file_size_in_byte": 1739, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 9, "usage_type": "attribute"}, {"api_name": "services.trakt.TraktService", "line_number": 13, "usage_type": "call"}, {"api_name": "services.plex.PlexService", "line_number": 14, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 18, "usage_type": "name"}, {"api_name": "models.Movie.from_trakt", "line_number": 18, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 16, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 23, "usage_type": "name"}, {"api_name": "plexapi.video.Video", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 24, "usage_type": "name"}, {"api_name": "models.Movie", "line_number": 24, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 30, "usage_type": "name"}, {"api_name": "models.Availability", "line_number": 30, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 31, "usage_type": "name"}, {"api_name": "models.Availability", "line_number": 31, "usage_type": "name"}, {"api_name": "models.Movie", "line_number": 32, "usage_type": "name"}, {"api_name": "models.Movie.from_plex", "line_number": 32, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 22, "usage_type": "name"}, {"api_name": "models.Movie", "line_number": 22, "usage_type": "name"}, {"api_name": "models.Movie", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 38, "usage_type": "name"}, {"api_name": "models.Movie", "line_number": 38, "usage_type": "name"}]}
+{"seq_id": "33199894256", "text": "#!/usr/bin/env python\n\nfrom __future__ import print_function\nimport os.path\nimport sys\nfrom optparse import OptionParser\nfrom pyfiglet import Figlet\nfrom subprocess import Popen, PIPE\ntry:\n from colorama import init\n init(strip=not sys.stdout.isatty())\n from termcolor import cprint\nexcept:\n def cprint(text, color):\n print(text)\n\n__version__ = '0.1'\n\ndef fail(text):\n cprint(text, 'red')\n\ndef win(text):\n cprint(text, 'green')\n\ndef dump(text):\n for line in text.split('\\n'):\n print(repr(line))\n\nclass Test(object):\n def __init__(self, opts):\n self.opts = opts\n self.ok = 0\n self.fail = 0\n self.failed = []\n self.oked = []\n # known bugs...\n self.skip = ['runic', 'pyramid', 'eftifont', 'DANC4', 'dietcola']\n # Toilet fonts that we don't handle identically, yet\n self.skip += ['emboss', 'emboss2', 'future', 'letter', 'pagga',\n 'smblock', 'smbraille', 'wideterm']\n # fonts that throw Unicode decoding errors\n self.skip += ['dosrebel', 'konto', 'kontoslant']\n # zip fonts we don't support\n self.skip += ['ascii12', 'ascii9', 'bigascii12', 'bigascii9',\n 'bigmono12', 'bigmono9', 'mono12', 'mono9', 'smascii12',\n 'smascii9', 'smmono12', 'smmono9']\n # what looks like the same bug, but in non-zip fonts\n self.skip += ['dwhistled', 'gradient']\n # failing tests:\n self.skip += ['crawford2', 'konto_slant', 'danc4', 'diet_cola',\n 'stronger_than_all']\n\n self.f = Figlet()\n\n def outputUsingFigletorToilet(self, text, font, fontpath):\n if os.path.isfile(fontpath + '.flf'):\n cmd = ('figlet', '-d', 'pyfiglet/fonts', '-f', font, text)\n elif os.path.isfile(fontpath + '.tlf'):\n cmd = ('toilet', '-d', 'pyfiglet/fonts', '-f', font, text)\n else:\n raise Exception('Missing font file: {}'.format(fontpath))\n\n p = Popen(cmd, bufsize=4096, stdout=PIPE)\n try:\n outputFiglet = p.communicate()[0].decode('utf8')\n except UnicodeDecodeError as e:\n print(\"Unicode Error handling font {}\".format(font))\n outputFiglet = ''\n return outputFiglet\n\n def validate_font_output(self, font, outputFiglet, outputPyfiglet):\n if outputPyfiglet == outputFiglet:\n win('[OK] %s' % font)\n self.ok += 1\n self.oked.append(font)\n return\n\n fail('[FAIL] %s' % font)\n self.fail += 1\n self.failed.append(font)\n self.show_result(outputFiglet, outputPyfiglet, font)\n\n def show_result(self, outputFiglet, outputPyfiglet, font):\n if self.opts.show is True:\n print('[PYTHON] *** %s\\n\\n' % font)\n dump(outputPyfiglet)\n print('[FIGLET] *** %s\\n\\n' % font)\n dump(outputFiglet)\n raw_input()\n\n def check_font(self, text, font):\n if font in self.skip:\n return\n fontpath = os.path.join('pyfiglet', 'fonts', font)\n\n self.f.setFont(font=font)\n\n outputPyfiglet = self.f.renderText(text)\n outputFiglet = self.outputUsingFigletorToilet(text, font, fontpath)\n\n # Our TLF rendering isn't perfect, yet\n strict = os.path.isfile(fontpath + '.flf')\n if not strict:\n outputPyfiglet = outputPyfiglet.strip('\\n')\n outputFiglet = outputFiglet.strip('\\n')\n\n self.validate_font_output(font, outputFiglet, outputPyfiglet)\n\n\n def check_text(self, text):\n for font in self.f.getFonts():\n self.check_font(text, font)\n\n def check_result(self):\n print('OK = %d, FAIL = %d' % (self.ok, self.fail))\n if len(self.failed) > 0:\n print('FAILED = %s' % repr(self.failed))\n\n return self.failed, self.oked\n\ndef banner(text):\n cprint(Figlet().renderText(text), \"blue\")\n\ndef main():\n parser = OptionParser(version=__version__)\n\n parser.add_option('-s', '--show', action='store_true', default=False,\n help='pause at each failure and compare output '\n '(default: %default)')\n\n opts, args = parser.parse_args()\n test = Test(opts)\n banner(\"TESTING one word\")\n test.check_text(\"foo\")\n banner(\"TESTING cut at space\")\n test.check_text(\"This is a very long text with many spaces and little words\")\n banner(\"TESTING cut at last char\")\n test.check_text(\"Averylongwordthatwillbecutatsomepoint I hope\")\n banner(\"TESTING explicit new line\")\n test.check_text(\"this text\\nuse new line\")\n if len(test.check_result()[0]) == 0:\n return 0\n else:\n return 1\n\n\nif __name__ == '__main__':\n sys.exit(main())\n", "repo_name": "pwaller/pyfiglet", "sub_path": "pyfiglet/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 4766, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1241, "dataset": "github-code", "pt": "24", "api": [{"api_name": "colorama.init", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.stdout.isatty", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 11, "usage_type": "attribute"}, {"api_name": "termcolor.cprint", "line_number": 20, "usage_type": "call"}, {"api_name": "termcolor.cprint", "line_number": 23, "usage_type": "call"}, {"api_name": "pyfiglet.Figlet", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.path.isfile", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 56, "usage_type": "name"}, {"api_name": "os.path.path.isfile", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 58, "usage_type": "name"}, {"api_name": "subprocess.Popen", "line_number": 63, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 63, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 94, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 94, "usage_type": "name"}, {"api_name": "os.path.path.isfile", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 102, "usage_type": "name"}, {"api_name": "termcolor.cprint", "line_number": 122, "usage_type": "call"}, {"api_name": "pyfiglet.Figlet", "line_number": 122, "usage_type": "call"}, {"api_name": "optparse.OptionParser", "line_number": 125, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 148, "usage_type": "call"}]}
+{"seq_id": "74549552381", "text": "import pandas as pd\nfrom datetime import datetime, timedelta\n\ndef change_wrongly_formatted_csv(input_file):\n with open(input_file) as file:\n data = file.read()\n data = data.replace(\";\", \",\")\n with open(input_file, \"w\") as file:\n file.write(data)\n\n# Converting datetime\ndef convert_timestamp_to_datetime(timestamp, before_or_after):\n date = timestamp.split(\"T\")[0]\n splitted_date = date.split(\"-\")\n time = timestamp.split(\"T\")[1]\n splitted_time = time.split(\"+\")[0].split(\":\")\n t = datetime(int(splitted_date[0]), int(splitted_date[1]), int(splitted_date[2]), int(splitted_time[0]), int(splitted_time[1]))\n data = {\"Timestamp\": t}\n data[before_or_after + \"hour\"] = t.hour\n data[before_or_after + \"month\"] = t.month\n data[before_or_after + \"year\"] = t.year\n data[before_or_after + \"day_of_week\"] = t.weekday()\n data[before_or_after + \"day_of_month\"] = t.day\n data[before_or_after + \"day_of_year\"] = t.timetuple().tm_yday\n return data\n\ntrafikk_data_path = \"trafikk.csv\"\n# change_wrongly_formatted_csv(trafikk_data_path)\ntrafikk_data = pd.read_csv(trafikk_data_path)\ntrafikk_data[\"output_volum\"] = 0\n\ntest_timestamp = \"2019-02-27T01:00+01:00\"\nfor i in trafikk_data.index:\n fra_time = trafikk_data.at[i, \"Fra\"]\n data_dict = convert_timestamp_to_datetime(fra_time, \"fra_\")\n for key, value in data_dict.items():\n trafikk_data.at[i, key] = value\n\ntrafikk_data = trafikk_data.drop([\"Navn\", \"Vegreferanse\", \"Fra\", \"Til\",\n \"Ikke gyldig lengde\", \"Lengdedekningsgrad (%)\", \"Felt\", \"Felt gyldig fra\",\n \"Felt gyldig til\", \"< 5\", \"6m\", \"> 5\", \"6m.1\", \"5\", \"6m - 7\", \"6m.2\", \"7\",\n \"6m - 12\", \"5m\", \"12\", \"5m - 16\", \"0m\", \"16\", \"0m - 24\", \"0m.1\", \"> 24\", \"0m.2\"], axis=1)\n\nnumber_of_hours_to_subtract = 6\n\nfor i in trafikk_data.index:\n for j in trafikk_data.index:\n before_data = trafikk_data.at[i, \"Timestamp\"]\n after_data = trafikk_data.at[j, \"Timestamp\"]\n t= datetime()\n timedelta(hours=6)\n before_data\n\n all_match = True\n for column in trafikk_data.columns.values:\n if(column == \"hour\" and trafikk_data.at[i, column] - 6 == trafikk_data.at[j, column]):\n pass\n elif(trafikk_data.at[i, column] == trafikk_data.at[j, column]):\n pass\n", "repo_name": "magmkri/hakathon-brain", "sub_path": "trafikk_data_with_own_model.py", "file_name": "trafikk_data_with_own_model.py", "file_ext": "py", "file_size_in_byte": 2303, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "datetime.datetime", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 51, "usage_type": "call"}]}
+{"seq_id": "72951116861", "text": "\"\"\"\nThis module contains the Branch class (one branch of the tree)\nand the Nodes class name\n\"\"\"\nfrom __future__ import annotations\nimport numpy as np\nimport logging\nfrom scipy.spatial import cKDTree\nfrom .mesh import InvalidNodeError, Mesh\n\nlogger = logging.getLogger(__name__)\n\n\nclass Branch:\n \"\"\"Class that contains a branch of the fractal tree\n\n Args:\n mesh:\n an object of the mesh class, where the fractal tree will grow\n initial_node (int):\n initial node to grow the branch. This is an index that refers\n to a node in the nodes.nodes array.\n initial_direction (array):\n initial direction to grow the branch. In general, it refers to\n the direction of the last segment of the mother brach.\n initial_triangle (int):\n the index of triangle of the mesh where the initial_node sits.\n l (float):\n total length of the branch\n angle (float):\n angle (rad) with respect to the initial_direction\n in the plane of the initial_triangle triangle\n repulsitivity (float):\n repulsitivity parameter. Controls how much the branches repel each other.\n nodes:\n the object of the class nodes that contains all the\n nodes of the existing branches.\n brother_nodes (list):\n the nodes of the brother and mother branches, to be excluded\n from the collision detection between branches.\n num_segments (int):\n number of segments to divide the branch.\n\n\n Attributes:\n child (list):\n contains the indexes of the child branches.\n It is not assigned when created.\n dir (array):\n vector direction of the last segment of the branch.\n nodes (list):\n contains the node indices of the branch. The node coordinates can\n be retrieved using nodes.nodes[i]\n triangles (list):\n contains the indices of the triangles from the mesh where every\n node of the branch lies.\n tri (int):\n triangle index where last node sits.\n growing (bool):\n False if the branch collide or is out of the surface. True otherwise.\n\n \"\"\"\n\n def __init__(\n self,\n mesh: Mesh,\n initial_node: int,\n initial_direction: np.ndarray,\n initial_triangle: int,\n length: float,\n angle: float,\n repulsitivity: float,\n nodes: \"Nodes\",\n brother_nodes: list[int],\n num_segments: int,\n ):\n self.child = [0, 0]\n self.dir = np.array([0.0, 0.0, 0.0])\n self.nodes = []\n self.triangles = []\n\n self.queue = []\n self.growing = True\n\n nodes.update_collision_tree(brother_nodes)\n self.nodes.append(initial_node)\n self.queue.append(nodes.nodes[initial_node])\n self.triangles.append(initial_triangle)\n grad = nodes.gradient(self.queue[0])\n\n self._initialize_direction(\n init_normal=mesh.normals[initial_triangle],\n initial_direction=initial_direction,\n angle=angle,\n )\n self._update_direction(repulsitivity=repulsitivity, grad=grad)\n\n for i in range(1, num_segments):\n self._grow(mesh, length, i, num_segments, nodes, repulsitivity)\n if not self.growing:\n break\n\n self.nodes += nodes.add_nodes(self.queue[1:])\n if not self.growing:\n nodes.end_nodes.append(self.nodes[-1])\n\n self.tri = self.triangles[-1]\n\n def _initialize_direction(\n self, init_normal: np.ndarray, initial_direction: np.ndarray, angle: float\n ) -> None:\n inplane = -np.cross(initial_direction, init_normal)\n self.dir = np.cos(angle) * initial_direction + np.sin(angle) * inplane\n self.dir /= np.linalg.norm(self.dir)\n\n def _grow(\n self,\n mesh: Mesh,\n length: float,\n i: int,\n num_segments: int,\n nodes: \"Nodes\",\n repulsitivity: float,\n ) -> None:\n intriangle = self.add_node_to_queue(\n mesh, self.queue[i - 1], self.dir * length / num_segments\n )\n if not intriangle:\n logger.debug(f\"Point {i} not in triangle\")\n self.growing = False\n return\n\n collision = nodes.collision(self.queue[i])\n if collision[1] < length / 5.0:\n logger.debug(f\"Collision {i}: {collision}\")\n self.growing = False\n self.queue.pop()\n self.triangles.pop()\n return\n\n grad = nodes.gradient(self.queue[i])\n normal = mesh.normals[self.triangles[i], :]\n # Project the gradient to the surface\n grad = grad - (np.dot(grad, normal)) * normal\n self._update_direction(repulsitivity=repulsitivity, grad=grad)\n\n def _update_direction(self, repulsitivity: float, grad: np.ndarray) -> None:\n self.dir = (self.dir + repulsitivity * grad) / np.linalg.norm(\n self.dir + repulsitivity * grad\n )\n\n def add_node_to_queue(\n self, mesh: Mesh, initial_node: np.ndarray, dir: np.ndarray\n ) -> bool:\n \"\"\"Functions that projects a node in the mesh surface\n and it to the queue is it lies in the surface.\n\n Args:\n mesh:\n an object of the mesh class, where the fractal tree will grow\n initial_node (array):\n vector that contains the coordinates of the\n last node added in the branch.\n dir (array):\n vector that contains the direction from the initial_node\n to the node to project.\n\n Return:\n success (bool):\n true if the new node is in the triangle.\n\n \"\"\"\n try:\n point, triangle = mesh.project_new_point(initial_node + dir)\n except InvalidNodeError:\n return False\n\n success = False\n if triangle >= 0:\n self.queue.append(point)\n self.triangles.append(triangle)\n success = True\n\n return success\n\n\nclass Nodes:\n \"\"\"A class containing the nodes of the branches plus some\n functions to compute distance related quantities.\n\n Args:\n initial_node (array):\n an array with the coordinates of the initial node of the first branch.\n\n Attributes:\n nodes (list):\n list of arrays containing the coordinates of the nodes\n last_node (int):\n last added node.\n end_nodes (list):\n a list containing the indices of all end nodes\n (nodes that are not connected) of the tree.\n tree (scipy.spatial.cKDTree):\n a k-d tree to compute the distance from any point\n to the closest node in the tree. It is updated once a branch is finished.\n collision_tree (scipy.spatial.cKDTree):\n a k-d tree to compute the distance from any point to the closest node\n in the tree, except from the brother and mother branches.\n It is used to check collision between branches.\n\n \"\"\"\n\n def __init__(self, initial_node: np.ndarray) -> None:\n self.nodes = [initial_node]\n self.last_node = 0\n self.end_nodes: list[int] = []\n self.tree = cKDTree(self.nodes)\n\n def add_nodes(self, queue: list[np.ndarray]) -> list[int]:\n \"\"\"This function stores a list of nodes of a branch and\n returns the node indices. It also updates the tree to compute distances.\n\n Args:\n queue (list):\n a list of arrays containing the coordinates of the nodes of one branch.\n\n Returns:\n nodes_id (list):\n the indices of the added nodes.\n \"\"\"\n nodes_id = []\n for point in queue:\n self.nodes.append(point)\n self.last_node += 1\n nodes_id.append(self.last_node)\n\n self.tree = cKDTree(self.nodes)\n return nodes_id\n\n def distance_from_point(self, point: np.ndarray) -> float:\n \"\"\"This function returns the distance from any\n point to the closest node in the tree.\n\n Args:\n point (array):\n the coordinates of the point to calculate the distance from.\n\n Returns:\n d (float):\n the distance between point and the closest node in the tree.\n \"\"\"\n return self.tree.query(point)[0]\n\n def distance_from_node(self, node: int) -> float:\n \"\"\"This function returns the distance from any\n node to the closest node in the tree.\n\n Args:\n node (int):\n the index of the node to calculate the distance from.\n\n Returns:\n d (float):\n the distance between specified node and the closest node in the tree.\n \"\"\"\n return self.distance_from_point(self.nodes[node])\n\n def update_collision_tree(self, nodes_to_exclude: list[int]) -> None:\n \"\"\"This function updates the collision_tree excluding a\n list of nodes from all the nodes in the tree. If all the\n existing nodes are excluded, one distant node is added.\n\n Args\n nodes_to_exclude (list):\n contains the nodes to exclude from the tree.\n Usually it should be the mother and the brother branch nodes.\n\n Returns:\n none\n \"\"\"\n nodes = set(range(len(self.nodes))).difference(nodes_to_exclude)\n nodes_to_consider = [self.nodes[x] for x in nodes]\n self.nodes_to_consider_keys = list(nodes)\n if len(nodes_to_consider) == 0:\n nodes_to_consider = [\n np.array([-100000000000.0, -100000000000.0, -100000000000.0])\n ]\n self.nodes_to_consider_keys = [100000000]\n logger.debug(\"no nodes to consider\")\n\n self.collision_tree = cKDTree(nodes_to_consider)\n\n def collision(self, point: np.ndarray):\n \"\"\"This function returns the distance between one point and\n the closest node in the tree and the index of the closest\n node using the collision_tree.\n\n Args:\n point (array):\n the coordinates of the point to calculate the distance from.\n\n Returns:\n collision (tuple):\n (distance to the closest node, index of the closest node)\n \"\"\"\n d, node = self.collision_tree.query(point)\n collision = (self.nodes_to_consider_keys[node], d)\n return collision\n\n def gradient(self, point: np.ndarray, delta: float = 0.01):\n \"\"\"This function returns the gradient of the distance from\n the existing points of the tree from any point. It uses a\n central finite difference approximation.\n\n Args:\n point (array):\n the coordinates of the point to calculate the gradient of the distance from.\n\n Returns:\n grad (array):\n (x,y,z) components of gradient of the distance.\n \"\"\"\n\n dx = np.array([delta, 0, 0])\n dy = np.array([0.0, delta, 0.0])\n dz = np.array([0.0, 0.0, delta])\n\n diff = [\n point - dx,\n point + dx,\n point - dy,\n point + dy,\n point - dz,\n point + dz,\n ]\n xm, xp, ym, yp, zm, zp = self.tree.query(diff)[0]\n\n grad = np.array(\n [\n (xp - xm) / (2 * delta),\n (yp - ym) / (2 * delta),\n (zp - zm) / (2 * delta),\n ]\n )\n return grad\n", "repo_name": "fsahli/fractal-tree", "sub_path": "src/fractal_tree/branch.py", "file_name": "branch.py", "file_ext": "py", "file_size_in_byte": 11616, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 17, "dataset": "github-code", "pt": "24", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "mesh.Mesh", "line_number": 66, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 68, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 78, "usage_type": "call"}, {"api_name": "mesh.normals", "line_number": 92, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 110, "usage_type": "attribute"}, {"api_name": "numpy.cross", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 114, "usage_type": "attribute"}, {"api_name": "mesh.Mesh", "line_number": 118, "usage_type": "name"}, {"api_name": "mesh.normals", "line_number": 142, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 147, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 148, "usage_type": "attribute"}, {"api_name": "mesh.Mesh", "line_number": 153, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 153, "usage_type": "attribute"}, {"api_name": "mesh.project_new_point", "line_number": 174, "usage_type": "call"}, {"api_name": "mesh.InvalidNodeError", "line_number": 175, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 213, "usage_type": "attribute"}, {"api_name": "scipy.spatial.cKDTree", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 219, "usage_type": "attribute"}, {"api_name": "scipy.spatial.cKDTree", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 240, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 286, "usage_type": "call"}, {"api_name": "scipy.spatial.cKDTree", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 293, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 310, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 324, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 325, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 326, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 338, "usage_type": "call"}]}
+{"seq_id": "1584563767", "text": "from django.db import models\nfrom django.urls import reverse\nfrom .file_reader import read\nimport datetime\nimport django.utils\nimport datefinder\n\nclass Event(models.Model):\n class_year = (\n (\"FR\", \"Freshman\"),\n (\"SO\", \"Sophomore\"),\n (\"JR\", \"Junior\"),\n (\"SR\", \"Senior\"),\n (\"NA\", \"No Associated Class Year\"),\n )\n title = models.CharField(max_length=200, blank = True)\n description = models.TextField(blank=True)\n date = models.DateField(null=True,blank=True)\n course = models.CharField(max_length=10)\n course_year = models.CharField(max_length=2, choices=class_year, default=\"NA\")\n\n # Override the save function to allow input saves and auto-generation\n # of model instances\n def save(self):\n if (not self.title or not self.description or\n not self.date or not self.course):\n lessons = read()\n bulk_lessons = []\n for lesson in lessons:\n new_lesson = Event()\n new_lesson.title = ' '.join(lesson[0])\n new_lesson.description = ' '.join(lesson[2] + lesson[3])\n l = datefinder.find_dates(' '.join(lesson[1]))\n for d in l:\n new_lesson.date = d.date()\n self.date = d.date()\n new_lesson.course = 'MilArt'\n new_lesson.course_year = \"Junior\"\n bulk_lessons.append(new_lesson)\n Event.objects.bulk_create(bulk_lessons)\n else:\n super(Event, self).save()\n\n # Allow a user to click in an event and see it's details in a separate page\n @property\n def get_url(self):\n url = reverse('planner:event_view', args=(self.id,))\n return f' {self.course}: {self.title} '\n\nclass File(models.Model):\n description = models.CharField(max_length=255, blank=True)\n document = models.FileField(upload_to='documents/')\n uploaded_at = models.DateTimeField(auto_now_add=True)\n", "repo_name": "ShadowWolf7027/Django-Project", "sub_path": "planner/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1992, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.db.models.Model", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "file_reader.read", "line_number": 27, "usage_type": "call"}, {"api_name": "datefinder.find_dates", "line_number": 33, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 47, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 50, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 50, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.models.FileField", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 53, "usage_type": "name"}]}
+{"seq_id": "41094630136", "text": "from time import * # Used for timing\r\nfrom serial import * # Used for serial communication with the arduino\r\nimport json # Parsing of data from Arduino\r\nimport threading\r\nfrom SETTINGS import *\r\nfrom collision_avoidance_algorithm import *\r\nfrom gps_functions import *\r\nfrom navigation import *\r\nimport os\r\nimport signal\r\nimport sys\r\nimport board # https://learn.adafruit.com/circuitpython-libraries-on-linux-and-the-nvidia-jetson-nano/digital-i-o\r\nimport digitalio\r\n\r\nbutton = digitalio.DigitalInOut(board.D4)\r\nbutton.direction = digitalio.Direction.INPUT\r\n\r\ne_stop = digitalio.DigitalInOut(board.D17)\r\ne_stop.direction = digitalio.Direction.INPUT\r\n\r\ndirection_data = False\r\nlocation_data = False\r\n\r\nMOTOR_PORT_INITALIZED = \"...\"\r\n\r\ndef get_motor_port():\r\n return MOTOR_PORT_INITALIZED\r\n\r\ndef get_coordinate(data_in):\r\n lat = getDecimal(data_in['lat'])\r\n lon = getDecimal(data_in['lon'])\r\n\r\n return lat, lon\r\n\r\n\r\ndef get_degree(data_in):\r\n value = data_in['dir'] + COMPASS_ANGLE_ADJUST\r\n return (value + 360) % 360\r\n\r\n\r\ndef get_direction():\r\n return direction_data\r\n\r\n\r\ndef get_location():\r\n return location_data\r\n\r\n\r\ndef usb_address_define():\r\n ports = [\"/dev/ttyUSB0\", \"/dev/ttyUSB1\", \"/dev/ttyUSB2\"]\r\n\r\n global NAV_PORT\r\n global SENSOR_PORT\r\n global MOTOR_PORT\r\n \r\n for port in ports:\r\n print(\"Testing port: \", port)\r\n ser_state = False # Connection state\r\n while not ser_state: # If not connected, try again\r\n if not button.value:\r\n state_handler.shutdown_flag = True\r\n #os.kill(os.getpid(), signal.SIGINT)\r\n if state_handler.shutdown_flag:\r\n print(\"Switch turned off\")\r\n sys.exit()\r\n try: # Set up error catch\r\n serial_conn = Serial(port, 9600, timeout=4) # Trying to connect serial\r\n serial_conn.setDTR(False)\r\n sleep(1)\r\n serial_conn.flushInput()\r\n serial_conn.setDTR(True)\r\n print(\"Successfully connected to \", port) # Print for the console\r\n try:\r\n print(\"Going to read data...\")\r\n data = serial_conn.readline()\r\n print(\"Inital data read: \", data)\r\n decoded = unicode(data, \"utf-8\")\r\n print(\"Inital data decoded: \", decoded)\r\n if decoded == \"NAV\\n\":\r\n print(\"Found the Navigation controller on port: \", port)\r\n NAV_PORT = port\r\n print(\"This is the nav port in settings: \", NAV_PORT)\r\n ser_state = True # If successful set the connection state true\r\n elif decoded == \"SENS\\n\":\r\n print(\"Found the Sensor controller on port: \", port)\r\n SENSOR_PORT = port\r\n print(\"This is the sensor port in settings: \", SENSOR_PORT)\r\n ser_state = True # If successful set the connection state true\r\n elif decoded == \"MOTOR\\r\\n\":\r\n print(\"Found the Motor controller on port: \", port)\r\n MOTOR_PORT = port\r\n MOTOR_PORT_INITALIZED = port\r\n state_handler.motor_port = port\r\n print(\"This is the motor port in settings: \", MOTOR_PORT)\r\n ser_state = True # If successful set the connection state true\r\n else:\r\n print(\"Port match not found for:\", data)\r\n except SerialException:\r\n print(\"Nothing read\")\r\n except SerialException: # Error handling\r\n if not button.value:\r\n os.kill(os.getpid(), signal.SIGINT)\r\n print('Trying to connect ', port) # Inform the user that we are trying to connect\r\n sleep(0.5) # Wait 0.5 second before trying again\r\n print(\"Done finding ports\")\r\n\r\n\r\ndef serial_connect(port): # This function connects the Navigation Controller\r\n ser_state = False # Connection state\r\n while not ser_state: # If not connected, try again\r\n if not button.value:\r\n state_handler.shutdown_flag = True\r\n #os.kill(os.getpid(), signal.SIGINT)\r\n if state_handler.shutdown_flag:\r\n print(\"Switch turned off\")\r\n sys.exit()\r\n try: # Set up error catch\r\n serial_conn = Serial(port, 9600, timeout=4) # Trying to connect serial\r\n serial_conn.setDTR(False)\r\n sleep(1)\r\n serial_conn.flushInput()\r\n serial_conn.setDTR(True)\r\n print(\"Successfully connected to \", port) # Print for the console\r\n try:\r\n print(\"Going to read data...\")\r\n data = serial_conn.readline()\r\n print(\"Inital data read: \", data)\r\n decoded = unicode(data, \"utf-8\")\r\n print(\"Inital data decoded: \", decoded)\r\n ser_state = True\r\n except SerialException:\r\n print(\"Nothing read\")\r\n\r\n return serial_conn\r\n\r\n except SerialException: # Error handling\r\n if not button.value:\r\n os.kill(os.getpid(), signal.SIGINT)\r\n print('Trying to connect ', port) # Inform the user that we are trying to connect\r\n sleep(0.5) # Wait 0.5 second before trying again\r\n\r\n\r\ndef decode_json_data(data):\r\n try:\r\n return json.loads(data.decode('utf-8'))\r\n except:\r\n return False\r\n\r\n\r\ndef get_navigation_data():\r\n ser = serial_connect(NAV_PORT) # Define ser as the serial connection.\r\n ser.readline() # Read the buffer to clear any unwanted bytes\r\n while True:\r\n if not button.value:\r\n state_handler.shutdown_flag = True\r\n #os.kill(os.getpid(), signal.SIGINT)\r\n if state_handler.shutdown_flag:\r\n sys.exit()\r\n\r\n state_handler.emergency_stop_flag = not e_stop.value\r\n\r\n try:\r\n data = ser.readline() # Read the serial data from the Arduino\r\n if data.decode('utf-8') != '': # If there is data\r\n # test = json.loads(data.decode('utf-8')) # Decode the data from the Arduino\r\n test_data = decode_json_data(data)\r\n if test_data != False:\r\n state_handler.navigation_disconnect_flag = False\r\n if test_data['data'] == 0:\r\n global direction_data\r\n direction_data = test_data\r\n if test_data['data'] == 1: # 1 means gps data, a 0 would be the direction data\r\n global location_data\r\n location_data = test_data\r\n # Print all the data\r\n else:\r\n print(\"Data Error on navigation port, raw data: \", data)\r\n # sleep(0.2)\r\n # state_handler.navigation_disconnect_flag = True\r\n else:\r\n print(\"Timeout on Navigation - Trying again...\")\r\n # state_handler.navigation_disconnect_flag = True\r\n except SerialException:\r\n # Disconnect of USB->UART occurred\r\n print(\"Device disconnected\")\r\n state_handler.navigation_disconnect_flag = True\r\n ser = serial_connect(NAV_PORT)\r\n\r\ndef sensor_bool_convert(sensor_value):\r\n if sensor_value == 1:\r\n return True\r\n else:\r\n return False\r\n\r\n\r\ndef get_sensor_data():\r\n\r\n print(\"SENSOR LOOP STARTED\")\r\n\r\n ser = serial_connect(SENSOR_PORT) # Define ser as the serial connection.\r\n ser.flushInput()\r\n ser.readline() # Read the buffer to clear any unwanted bytes\r\n while True:\r\n if not button.value:\r\n state_handler.shutdown_flag = True\r\n #os.kill(os.getpid(), signal.SIGINT)\r\n if state_handler.shutdown_flag:\r\n sys.exit()\r\n\r\n state_handler.emergency_stop_flag = not e_stop.value\r\n\r\n try:\r\n data = ser.readline() # Read the serial data from the Arduino\r\n if data.decode('utf-8') != '': # If there is data\r\n # test = json.loads(data.decode('utf-8')) # Decode the data from the Arduino\r\n test_data = decode_json_data(data)\r\n if test_data != False:\r\n state_handler.sensor_disconnect_flag = False\r\n # Got the data, pass it and set different states of control system etc etc\r\n \r\n state_handler.front_collision_flag = sensor_bool_convert(test_data['T_F'])\r\n state_handler.rear_collision_flag = sensor_bool_convert(test_data['T_B'])\r\n\r\n ir_data = test_data['IR']\r\n sharp_data = test_data['SHARP']\r\n\r\n state_handler.ir_front_left = sensor_bool_convert(ir_data[0])\r\n state_handler.ir_front_right = sensor_bool_convert(ir_data[1])\r\n state_handler.ir_rear_left = sensor_bool_convert(ir_data[2])\r\n state_handler.ir_rear_right = sensor_bool_convert(ir_data[3])\r\n\r\n state_handler.json_string_recieved = test_data\r\n\r\n state_handler.sharp_front = sharp_data[0]\r\n \r\n if (state_handler.motor_driver_disconnect_flag is False and state_handler.navigation_disconnect_flag\r\n is False and state_handler.emergency_stop_flag is False):\r\n print(\"Sending ACK\")\r\n ser.write(69) # ACK to the sensor controller\r\n print(\"Did send ACK\")\r\n\r\n else:\r\n print(\"Data Error on sensor controller, raw data: \", data)\r\n state_handler.sensor_disconnect_flag = True\r\n else:\r\n print(\"Timeout on Sensor - Trying again...\")\r\n state_handler.sensor_disconnect_flag = True\r\n except SerialException:\r\n # Disconnect of USB->UART occurred\r\n print(\"Device disconnected\")\r\n state_handler.sensor_disconnect_flag = True\r\n ser = serial_connect(SENSOR_PORT)\r\n\r\n\r\n\r\nnav_loop = threading.Thread(name=\"background\", target=get_navigation_data)\r\nsensor_loop = threading.Thread(name='background', target=get_sensor_data)\r\n\r\ndef serial_init():\r\n nav_loop.start()\r\n sensor_loop.start()\r\n", "repo_name": "rasmushedeager/spro3grp4-E19", "sub_path": "serial_cummunications.py", "file_name": "serial_cummunications.py", "file_ext": "py", "file_size_in_byte": 10677, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "digitalio.DigitalInOut", "line_number": 15, "usage_type": "call"}, {"api_name": "board.D4", "line_number": 15, "usage_type": "attribute"}, {"api_name": "digitalio.Direction", "line_number": 16, "usage_type": "attribute"}, {"api_name": "digitalio.DigitalInOut", "line_number": 18, "usage_type": "call"}, {"api_name": "board.D17", "line_number": 18, "usage_type": "attribute"}, {"api_name": "digitalio.Direction", "line_number": 19, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 65, "usage_type": "call"}, {"api_name": "os.kill", "line_number": 102, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 102, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 102, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 116, "usage_type": "call"}, {"api_name": "os.kill", "line_number": 138, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 138, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 138, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 145, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 158, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 208, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 256, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 257, "usage_type": "call"}]}
+{"seq_id": "17126084581", "text": "__author__ = \"Ivan Begtin (ibegtin@gmail.com)\"\n__version__ = \"3.0.4\"\n__copyright__ = \"Copyright (c) 2008 Ivan Begtin\"\n__license__ = \"Proprietary\"\n\nfrom setuptools import setup, find_packages\n\nsetup(name='opendata',\n version='1.0',\n description='Python OpenGovData (OpenData)',\n author='Ivan Begtin',\n author_email='ibegtin@gmail.com',\n url='',\n download_url='',\n packages=find_packages(),\n license='Creative Commons',\n keywords='opendata',\n classifiers=[\"Development Status :: 3 - Alpha\",\n \"Intended Audience :: Developers\",\n \"Intended Audience :: Science/Research\",\n \"License :: OSI Approved :: MIT License\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python\",\n \"Topic :: Software Development :: Libraries :: Python Modules\"\n ],\n )\n", "repo_name": "ivbeg/opengovdataru", "sub_path": "opendata/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 935, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 6, "dataset": "github-code", "pt": "24", "api": [{"api_name": "setuptools.setup", "line_number": 8, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 15, "usage_type": "call"}]}
+{"seq_id": "71305125824", "text": "#-*- coding:utf-8 -*-\nimport logging\nimport cv2\nimport numpy as np\n\nfrom lxml import html\n\nfrom django.core.management.base import BaseCommand\nfrom django.conf import settings\n\nlogger = logging.getLogger(__name__)\n\nclass Command(BaseCommand):\n def add_arguments(self, parser):\n # Named (optional) arguments\n parser.add_argument('--demo_mode',\n action = 'store_true',\n dest = 'demo_mode',\n default = False,\n help = 'Set demo mode')\n parser.add_argument('--cat_id',\n action = 'store',\n dest = 'cat_id',\n type = str,\n default = False,\n help = 'Set cat tag for update')\n def handle(self, *args, **options):\n path = '/home/jocker/Downloads/tb50413_1.png'\n output = '/home/jocker/Downloads/tb50413_1_CLEAN.png'\n\ndef back_rm(filename):\n # Load the image\n img = cv2.imread(filename)\n\n # Convert the image to grayscale\n gr = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)\n\n # Make a copy of the grayscale image\n bg = gr.copy()\n\n # Apply morphological transformations\n for i in range(5):\n kernel2 = cv2.getStructuringElement(cv2.MORPH_ELLIPSE,\n (2 * i + 1, 2 * i + 1))\n bg = cv2.morphologyEx(bg, cv2.MORPH_CLOSE, kernel2)\n bg = cv2.morphologyEx(bg, cv2.MORPH_OPEN, kernel2)\n\n # Subtract the grayscale image from its processed copy\n dif = cv2.subtract(bg, gr)\n\n # Apply thresholding\n bw = cv2.threshold(dif, 0, 255, cv2.THRESH_BINARY_INV | cv2.THRESH_OTSU)[1]\n dark = cv2.threshold(bg, 0, 255, cv2.THRESH_BINARY_INV | cv2.THRESH_OTSU)[1]\n\n # Extract pixels in the dark region\n darkpix = gr[np.where(dark > 0)]\n\n # Threshold the dark region to get the darker pixels inside it\n darkpix = cv2.threshold(darkpix, 0, 255, cv2.THRESH_BINARY | cv2.THRESH_OTSU)[1]\n\n # Paste the extracted darker pixels in the watermark region\n bw[np.where(dark > 0)] = darkpix.T\n\n cv2.imwrite('final.jpg', bw)\n\n\n#back_rm('watermark.jpg')", "repo_name": "dkramorov/astwobytes", "sub_path": "apps/upload_tasks/management/commands/try_remove_watermark.py", "file_name": "try_remove_watermark.py", "file_ext": "py", "file_size_in_byte": 2057, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "django.core.management.base.BaseCommand", "line_number": 13, "usage_type": "name"}, {"api_name": "cv2.imread", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 36, "usage_type": "attribute"}, {"api_name": "cv2.getStructuringElement", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.MORPH_ELLIPSE", "line_number": 43, "usage_type": "attribute"}, {"api_name": "cv2.morphologyEx", "line_number": 45, "usage_type": "call"}, {"api_name": "cv2.MORPH_CLOSE", "line_number": 45, "usage_type": "attribute"}, {"api_name": "cv2.morphologyEx", "line_number": 46, "usage_type": "call"}, {"api_name": "cv2.MORPH_OPEN", "line_number": 46, "usage_type": "attribute"}, {"api_name": "cv2.subtract", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 52, "usage_type": "attribute"}, {"api_name": "cv2.THRESH_OTSU", "line_number": 52, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 53, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 53, "usage_type": "attribute"}, {"api_name": "cv2.THRESH_OTSU", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 56, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 59, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 59, "usage_type": "attribute"}, {"api_name": "cv2.THRESH_OTSU", "line_number": 59, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 62, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 64, "usage_type": "call"}]}
+{"seq_id": "19202525168", "text": "from flask import Flask, render_template, redirect, request\nfrom loginform import LoginForm\nfrom my_configure import my_configure\n\napp = Flask(__name__)\n\n\n@app.route('/')\n@app.route('/index')\ndef index():\n user = \"Ученик Яндекс.Лицея\"\n return render_template('index.html', title='Домашняя страница', username=user)\n\n\n@app.route('/form_sample', methods=['GET', 'POST'])\ndef form():\n form = LoginForm()\n if form.validate_on_submit():\n print(request.form['username_first'])\n print(request.form['real_name'])\n print(request.form['password'])\n print(request.form['sex'])\n print(request.form['commentary'])\n print(request.form['language'])\n print(request.form['remember_me'])\n return redirect('/success')\n return render_template('login.html', title='Форма', form=form)\n\n\n@app.route('/success')\ndef success():\n return render_template('success.html')\n\n \nif __name__ == '__main__':\n my_configure(app)\n app.run(port=8080, host='127.0.0.1')\n", "repo_name": "gri-gri/web-server", "sub_path": "Urok 2/Elementy formy/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 1048, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 12, "usage_type": "call"}, {"api_name": "loginform.LoginForm", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 19, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 22, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 22, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 23, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 24, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 25, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 32, "usage_type": "call"}, {"api_name": "my_configure.my_configure", "line_number": 36, "usage_type": "call"}]}
+{"seq_id": "33194272253", "text": "import unicodedata\nfrom bs4 import BeautifulSoup\n\nSKIDPASE_BASEURL = 'http://skidpaste.org/'\n\nclass MyHTMLParser:\n def __init__(self, html_doc, requestHandler):\n self.soup = BeautifulSoup(html_doc, 'html.parser')\n self.requestHandler = requestHandler\n self.pastes = []\n self.parsedJson = {\"url\":[], \"title\":[],\"content\":[]}\n self.getdetails()\n\n def getdetails(self):\n limit = 10\n leftSideTag = self.soup.find(id=\"content_left\")\n for tr in leftSideTag.find_all('tr'):\n if limit == 0:\n return\n limit -= 1\n print(\"parsing paste\")\n a = tr.find('a')\n if a != None:\n title = a.get_text()\n url = SKIDPASE_BASEURL+a['href']+'.txt'\n textChunc = self.requestHandler.request(url)\n contentText = self.convertByteToString(textChunc)\n paste = {\"title\": title, 'url': url, 'content': contentText}\n self.pastes.append(paste)\n\n def getJson(self):\n return self.pastes\n\n def convertByteToString(self, text):\n txt = text.decode(encoding='UTF-8')\n print('convert')\n return txt\n", "repo_name": "danabig/skidPasteCrawler", "sub_path": "MyHTMLParser.py", "file_name": "MyHTMLParser.py", "file_ext": "py", "file_size_in_byte": 1209, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "bs4.BeautifulSoup", "line_number": 8, "usage_type": "call"}]}
+{"seq_id": "41716730155", "text": "import math\n\nfrom pathlib import Path\n\nimport pytest\nimport torch\nimport torch.nn as nn\n\nfrom pytorch3d.transforms import random_rotation\n\nfrom deepmc.models.components.vn_dgcnn_pose_net import VN_DGCNN_pose, VN_DGCNN_pose_seg\nfrom deepmc.models.components.vnn import VNMaxPool\n\n\n@pytest.mark.parametrize(\"batch_size\", [2, 8])\ndef test_vn(batch_size):\n model = VN_DGCNN_pose().to(\"cuda:7\")\n\n inp = torch.rand(batch_size, 3, 1024).to(\"cuda:7\")\n out = model(inp).transpose(-1, -2)\n print(out.shape)\n\n R = random_rotation().to(\"cuda:7\")\n inp_r = R @ inp\n out2 = model(inp_r).transpose(-1, -2)\n out_r = R @ out\n \n print(out_r[0,0,0], out2[0,0,0])\n print(torch.abs(out_r - out2).max())\n\n@pytest.mark.parametrize(\"batch_size\", [1, 8])\ndef test_vnmaxpool(batch_size):\n model = VNMaxPool(682)\n\n inp = torch.rand(batch_size, 682, 3, 1024)\n out = model(inp)\n\n print(out.shape)\n\n # R = random_rotation()\n # inp_r = R @ inp.unsqueeze(-1)\n # inp_r = inp_r.squeeze()\n # out2 = m_enc(inp_r)\n # out2 = m_dec(out2)\n \n # print(torch.max(out1_r - out2))\n\n\n\ndef main():\n test_vn(4)\n # test_vnmaxpool(1)\n\nif __name__ == \"__main__\":\n main()", "repo_name": "beyaldiz/DeepMC", "sub_path": "tests/test_vn_dgcnn.py", "file_name": "test_vn_dgcnn.py", "file_ext": "py", "file_size_in_byte": 1193, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "deepmc.models.components.vn_dgcnn_pose_net.VN_DGCNN_pose", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 19, "usage_type": "call"}, {"api_name": "pytorch3d.transforms.random_rotation", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.abs", "line_number": 29, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 15, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 15, "usage_type": "attribute"}, {"api_name": "deepmc.models.components.vnn.VNMaxPool", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 35, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 31, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 31, "usage_type": "attribute"}]}
+{"seq_id": "28958200042", "text": "from gpiozero import Device\n# from gpiozero.pins.pigpio import PiGPIOFactory\nfrom gpiozero import DigitalOutputDevice, PWMOutputDevice\nimport time\n# import RPi.GPIO as GPIO\n# Device.pin_factory = PiGPIOFactory()\n\n\nclass Motor:\n \"\"\"\n The class takes three pin numbers as the input to control one of the motor connected to TB6612FNG module.\n \"\"\"\n\n def __init__(self, in1, in2, pwm):\n self.in1 = DigitalOutputDevice(in1)\n self.in1.off()\n\n self.in2 = DigitalOutputDevice(in2)\n self.in2.on()\n\n self.pwm = PWMOutputDevice(pwm, frequency=1000)\n\n def set_throttle(self, val):\n \"\"\"Control the orientation and the speed of the motor.\n Arguments:\n val: a number between -1.0 and 1.0. The motor rotates in forward direction if val > 1, otherwise in reverse direction.\n Setting val to None will set the motor to stop mode.\n \"\"\"\n\n # Set the motor to stop mode.\n if val is None:\n self.in1.off()\n self.in2.off()\n self.pwm.value = 1.0\n\n else:\n # Determine the orientation of the motor.\n if val > 0.0:\n self.in1.off()\n self.in2.on()\n else:\n self.in1.on()\n self.in2.off()\n\n # Clamp the pwm signal (throttle) to [0, 1].\n pwm = max(0.0, min(abs(val), 1.0))\n\n # Note that setting PWM to low will brake the motor no matter what\n # in1 and in2 input is.\n self.pwm.value = pwm\n\n\n def close(self):\n self.in1.close()\n self.in2.close()\n self.pwm.close()\n\n\ndef test_left_wheel():\n # 22 -- pin 15 \n # 23 -- ping 16\n # 19 -- pin 35\n motor = Motor(22, 23, 19)\n # [-1, -0.5, 0, 0.5, 1]\n for val in [0.5, 1]:\n motor.set_throttle(val)\n time.sleep(1)\n\n # Set motor to stop mode.\n motor.set_throttle(None)\n motor.close()\n\ndef test_right_wheel():\n # 17 -- pin 11\n # 27 -- pin 13\n # 18 --- pin 12\n motor = Motor(17, 27, 18)\n for val in [0.5, 1]:\n motor.set_throttle(val)\n time.sleep(1)\n\n # Set motor to stop mode.\n motor.set_throttle(None)\n motor.close()\n\n\ndef test_two_wheel():\n l_motor = Motor(22, 23, 19)\n r_motor = Motor(17, 27, 18)\n\n for val in [0.5, 1]:\n l_motor.set_throttle(val)\n r_motor.set_throttle(val)\n time.sleep(0.5)\n\n\n\nif __name__ == \"__main__\":\n # test_left_wheel()\n # test_right_wheel()\n test_two_wheel()\n", "repo_name": "iamfaith/picar", "sub_path": "device.py", "file_name": "device.py", "file_ext": "py", "file_size_in_byte": 2520, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "gpiozero.DigitalOutputDevice", "line_number": 15, "usage_type": "call"}, {"api_name": "gpiozero.DigitalOutputDevice", "line_number": 18, "usage_type": "call"}, {"api_name": "gpiozero.PWMOutputDevice", "line_number": 21, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 67, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 80, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 94, "usage_type": "call"}]}
+{"seq_id": "35151884863", "text": "import os\nfrom pprint import pprint\nimport torch\nfrom ml_research.train.PriceRangeTrainer import ProcessModel\nimport pandas as pd\nimport glob\nfrom torch.cuda.amp.autocast_mode import autocast\nimport torchmetrics\nfrom tqdm import tqdm\nimport numpy as np\n\n\ndef CrossValPred(iteration, datasetGenerator):\n gen = datasetGenerator\n iter = f\"iter_{iteration}\"\n taskName = gen.srcDfPath.split(\"/\")[-1]\n allFoldsLoader = gen.genDataloader()\n trainedModelDir = \"/home/alextay96/Desktop/workspace/mrm_workspace/dmg_consistent_detection/data/auto_select\"\n outputDir = \"/home/alextay96/Desktop/workspace/mrm_workspace/dmg_consistent_detection/data/cross_val_pred\"\n taskName = iter + \"_\" + taskName\n device = torch.device(\"cuda\")\n allDfWithPreds = []\n for _, valLoader in allFoldsLoader:\n srcDf: pd.DataFrame = valLoader.dataset.df\n allPredLogit = []\n foldId = srcDf[\"kfold\"].unique().item()\n search = f\"{trainedModelDir}/{foldId}/**/*.ckpt\"\n accMetrics = torchmetrics.Accuracy(num_classes=2).to(device)\n confMatMetrics = torchmetrics.ConfusionMatrix(\n num_classes=2, normalize=\"true\"\n ).to(device)\n allmodelPath = glob.glob(search, recursive=True)\n modelPath = allmodelPath[0]\n trainedModel = ProcessModel.load_from_checkpoint(modelPath)\n trainedModel = trainedModel.to(device)\n trainedModel.eval()\n with torch.no_grad():\n for img, targets in tqdm(valLoader):\n img = img.to(device)\n targets = targets.to(device)\n with autocast():\n logit = trainedModel(img)\n preds = torch.argmax(logit, dim=1)\n logitNp = logit.cpu().numpy().tolist()\n allPredLogit.extend(logitNp)\n accMetrics.update(preds, targets)\n confMatMetrics.update(preds, targets)\n assert len(allPredLogit) == len(srcDf)\n srcDf[\"logit\"] = allPredLogit\n srcDf[\"model_version\"] = modelPath\n allDfWithPreds.append(srcDf)\n print(accMetrics.compute())\n pprint(confMatMetrics.compute())\n accMetrics.reset()\n confMatMetrics.reset()\n allDf = pd.concat(allDfWithPreds)\n assert len(allDf) == len(allDf[\"dst_filename\"].unique())\n outputFile = f\"{outputDir}/preds_{taskName}\"\n allDf.to_csv(outputFile)\n return outputFile\n\n\nif __name__ == \"__main__\":\n CrossValPred(1)\n", "repo_name": "MLMonkATGY/dmg_consistent_detection", "sub_path": "ml_research/analysis/CrossValPred.py", "file_name": "CrossValPred.py", "file_ext": "py", "file_size_in_byte": 2455, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "torch.device", "line_number": 21, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 24, "usage_type": "attribute"}, {"api_name": "torchmetrics.Accuracy", "line_number": 28, "usage_type": "call"}, {"api_name": "torchmetrics.ConfusionMatrix", "line_number": 29, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 32, "usage_type": "call"}, {"api_name": "ml_research.train.PriceRangeTrainer.ProcessModel.load_from_checkpoint", "line_number": 34, "usage_type": "call"}, {"api_name": "ml_research.train.PriceRangeTrainer.ProcessModel", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 37, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.cuda.amp.autocast_mode.autocast", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 43, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 53, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 56, "usage_type": "call"}]}
+{"seq_id": "30843185355", "text": "import sys\nfrom collections import deque\n\ninput = sys.stdin.readline\n\ndef solution(src, dest):\n need_visit = deque([(st, 0) for st in station_graph[src-1]])\n visited = {src}\n transfer_count = []\n\n while need_visit:\n station: int\n station, count = need_visit.popleft()\n # 목적지\n if station == dest:\n transfer_count.append(count)\n # 목적지 X\n if station not in visited:\n # 환승역 check\n visited.add(station)\n need_visit.extend([(st, count+1) if (count_station[st-1] >= 2) else (st, count) for st in station_graph[station-1]])\n else: continue\n\n print(transfer_count)\n if len(transfer_count) != 0:\n return min(transfer_count)\n return -1\n\n\nN, L = map(int, input().split())\n# {역 번호: [이동 가능한 역 리스트]} 초기화\nstation_graph = [[] for _ in range(N)]\ncount_station = [0 for _ in range(N)]\n\nfor _ in range(L):\n line = list(map(int, input().split()))\n for idx in range(len(line)-1):\n curr, nxt = line[idx], line[idx + 1]\n if nxt != -1:\n station_graph[curr-1].append(nxt)\n station_graph[nxt - 1].append(curr)\n count_station[curr-1] += 1\n\n# src, dest\nS, D = map(int, input().split())\n\nprint(solution(S, D))\n\n\n\"\"\"\n10 3\n1 2 3 4 5 -1\n9 7 10 -1\n7 6 3 8 -1\n1 10\n\n\n2\n\"\"\"", "repo_name": "jjaen0823/codingTest", "sub_path": "5_DFSBFS/BJ_최소환승경로_2021.py", "file_name": "BJ_최소환승경로_2021.py", "file_ext": "py", "file_size_in_byte": 1355, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "sys.stdin", "line_number": 4, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 7, "usage_type": "call"}]}
+{"seq_id": "27288180792", "text": "from django.shortcuts import render\nfrom rest_framework import status\nfrom rest_framework.decorators import api_view\nfrom rest_framework.response import Response\n\nfrom .models import Task\nfrom .serializer import TaskSerializer\n\n# This is an example of function based view for serializer\n\n@api_view(['GET', 'POST'])\ndef task_list(request):\n \"\"\"\n lists all tasks or creates a task\n :param request:\n :return:\n \"\"\"\n if request.method == \"GET\":\n tasks = Task.objects.all()\n\n # tasks here is a query set. So we are essentially passing the entire query set into the serializer\n # the many=True attribute here is super important. Without this attribute an error would be raised\n\n serializer = TaskSerializer(tasks, many=True)\n return Response(serializer.data)\n\n elif request.method == \"POST\":\n serializer = TaskSerializer(data=request.DATA)\n if serializer.is_valid():\n serializer.save()\n return Response(serializer.data, status=status.HTTP_201_CREATED)\n\n else:\n # there is a validation error and hence there is a problem with the data in the request\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n@api_view(['GET', 'PUT', 'DELETE'])\ndef task_detail(request, pk):\n \"\"\"\n get update or delete a specific task\n :param request:\n :param pk:\n :return:\n \"\"\"\n try:\n task = Task.objects.get(pk=pk)\n except Task.DoesNotExist:\n return Response(status=status.HTTP_404_NOT_FOUND)\n\n if request.method == 'GET':\n serializer = TaskSerializer(task)\n return Response(serializer.data)\n\n elif request.method == 'PUT':\n serializer = TaskSerializer(task, data=request.DATA)\n if serializer.is_valid():\n serializer.save()\n # returning the serializer data after saving it to the database\n return Response(serializer.data)\n\n else:\n # there were some validation errors with the data\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n elif request.method == 'DELETE':\n # recall we already have the task present\n task.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)", "repo_name": "RiflerRick/Django-Notes", "sub_path": "django-rest-framework/DjangoRestFrameworkDemo/api_demo/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2268, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "models.Task.objects.all", "line_number": 19, "usage_type": "call"}, {"api_name": "models.Task.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "models.Task", "line_number": 19, "usage_type": "name"}, {"api_name": "serializer.TaskSerializer", "line_number": 24, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 25, "usage_type": "call"}, {"api_name": "serializer.data", "line_number": 25, "usage_type": "attribute"}, {"api_name": "serializer.TaskSerializer", "line_number": 28, "usage_type": "call"}, {"api_name": "serializer.is_valid", "line_number": 29, "usage_type": "call"}, {"api_name": "serializer.save", "line_number": 30, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 31, "usage_type": "call"}, {"api_name": "serializer.data", "line_number": 31, "usage_type": "attribute"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 31, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 31, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 35, "usage_type": "call"}, {"api_name": "serializer.errors", "line_number": 35, "usage_type": "attribute"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 35, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 35, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Task.objects.get", "line_number": 46, "usage_type": "call"}, {"api_name": "models.Task.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "models.Task", "line_number": 46, "usage_type": "name"}, {"api_name": "models.Task.DoesNotExist", "line_number": 47, "usage_type": "attribute"}, {"api_name": "models.Task", "line_number": 47, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 48, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 48, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 48, "usage_type": "name"}, {"api_name": "serializer.TaskSerializer", "line_number": 51, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 52, "usage_type": "call"}, {"api_name": "serializer.data", "line_number": 52, "usage_type": "attribute"}, {"api_name": "serializer.TaskSerializer", "line_number": 55, "usage_type": "call"}, {"api_name": "serializer.is_valid", "line_number": 56, "usage_type": "call"}, {"api_name": "serializer.save", "line_number": 57, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 59, "usage_type": "call"}, {"api_name": "serializer.data", "line_number": 59, "usage_type": "attribute"}, {"api_name": "rest_framework.response.Response", "line_number": 63, "usage_type": "call"}, {"api_name": "serializer.errors", "line_number": 63, "usage_type": "attribute"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 63, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 63, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 68, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_204_NO_CONTENT", "line_number": 68, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 68, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 37, "usage_type": "call"}]}
+{"seq_id": "35295853579", "text": "import sqlite3\nimport sys\n\n# MÉTODO QUE CREA CONEXIÓN CON LA BASE DE DATOS\ndef conectar():\n conexion = sqlite3.connect(\"EJERCICIOS_BD/articulos/articulosBD.db\") #realiza la conexión\n cursor = conexion.cursor() \n return conexion, cursor\n\n# MÉTODO QUE CIERRA LA CONEXIÓN CON LA BASE DE DATOS\ndef cerrar_conexion(conexion):\n conexion.close()\n\n# MÉTODO QUE CREA LA TABLA ARTÍCULOS\ndef crearTabla():\n conexion, cursor = conectar()\n sql = \"\"\"\n CREATE TABLE IF NOT EXISTS articulos(\n id INTEGER PRIMARY KEY AUTOINCREMENT NOT NULL,\n nombre VARCHAR(20) NOT NULL,\n cantidad INT NOT NULL,\n importe FLOAT NOT NULL\n )\n \"\"\"\n if (cursor.execute(sql)):\n print(\"Tabla creada\")\n else: \n print(\"No se pudo crear la tabla\")\n cerrar_conexion(conexion)\n\n# MÉTODO QUE NOS CARGA EL MENÚ Y NOS DIRIGE A LA OPCIÓN SELECCIONADA\ndef menu():\n print(\"\\nINDIQUE QUE ACCIÓN DESEA REALIZAR\")\n print(\"1. Alta\")\n print(\"2. Listar\")\n print(\"3. Modificacion\")\n print(\"4. Borrado\")\n print(\"0. Salir\")\n\n opcion = input('--> ')\n if opcion not in ['1','2','3','4','0']:\n print('\\n********* Selecciona una opcion valida *********\\n')\n else:\n if opcion == '1':\n anadir_datos()\n if opcion == '2':\n listar_datos()\n if opcion == '3':\n modificar_datos()\n if opcion == '4':\n eliminar_datos()\n if opcion == '0':\n print('HASTA PRONTO !!')\n sys.exit() \n\n# MÉTODO QUE AÑADE DATOS A LA BASE DE DATOS\ndef anadir_datos():\n conexion, cursor = conectar()\n\n nombre = input(\"\\nINSERTA NOMBRE: \")\n cantidad = input(\"INSERTA CANTIDAD: \")\n importe = input(\"INSERTA IMPORTE: \")\n datos = (nombre, cantidad, importe)\n\n sql = \"\"\"INSERT INTO articulos(nombre, cantidad, importe) VALUES (?, ?, ?)\"\"\"\n if(cursor.execute(sql, datos)):\n print('\\n --- Datos guardados ---')\n else:\n print('*** No se pudieron guardar los datos ***')\n\n conexion.commit()\n conexion.close()\n menu()\n\n# MÉTODO QUE LISTA LOS DATOS DE LA BASE DE DATOS\ndef listar_datos():\n conexion, cursor = conectar()\n sql = \"SELECT * FROM articulos\"\n cursor.execute(sql)\n for fila in cursor:\n print(\"\\n\")\n print(\"*\"*25)\n print(f\"ID = {fila[0]}\")\n print(f\"NOMBRE = {fila[1]}\")\n print(f\"CANTIDAD = {fila[2]}\")\n print(f\"IMPORTE = {fila[3]}€\")\n print(\"*\"*25)\n\n conexion.close()\n menu()\n\n\ndef seleccionarId():\n conexion, cursor = conectar()\n idSeleccioando = input(\"\\nINSERTA ID DEL ARTÍCULO QUE QUIERAS MODIFICAR: \")\n\n seleccionar = \"SELECT * FROM articulos WHERE id=\"+idSeleccioando\n cursor.execute(seleccionar)\n\n for fila in cursor:\n print(\"\\n\")\n print(\"*\"*25)\n print(f\"ID = {fila[0]}\")\n print(f\"NOMBRE = {fila[1]}\")\n print(f\"CANTIDAD = {fila[2]}\")\n print(f\"IMPORTE = {fila[3]}€\")\n print(\"*\"*25)\n\n nombre = input(\"\\nINSERTA NUEVO NOMBRE: \")\n cantidad = input(\"INSERTA NUEVO CANTIDAD: \")\n importe = input(\"INSERTA NUEVO IMPORTE: \")\n\n conexion.close()\n return nombre,cantidad,importe,idSeleccioando\n \n\n# MÉTODO QUE MODIFICA LOS DATOS DE LA BASE DE DATOS\ndef modificar_datos():\n conexion, cursor = conectar()\n\n nombre,cantidad,importe,idSeleccioando = seleccionarId()\n\n sql = \"UPDATE articulos SET NOMBRE='\"+nombre+\"',CANTIDAD='\"+cantidad+\"',IMPORTE='\"+importe+\"' WHERE ID='\"+idSeleccioando+\"'\"\n cursor.execute(sql)\n cursor.close()\n conexion.commit()\n conexion.close()\n menu()\n\n# MÉTODO QUE ELIMINA LOS DATOS DE LA BASE DE DATOS\ndef eliminar_datos():\n conexion, cursor = conectar()\n\n idSeleccioando = input(\"\\nINSERTA ID DEL ARTÍCULO QUE QUIERAS MODIFICAR: \")\n sql = \"DELETE FROM articulos WHERE ID=\"+idSeleccioando\n if(cursor.execute(sql)):\n print(f'Se ha eliminado correctamente el articulo con id: {idSeleccioando}')\n else:\n print(f'Error al borrar el id: {idSeleccioando}')\n\n conexion.commit()\n conexion.close()\n menu()\n\n\n\nif __name__ == '__main__':\n crearTabla()\n menu()", "repo_name": "aleefb9/CURSO_AVANZADO_PYTHON", "sub_path": "EJERCICIOS_BD/articulos/ejercicio_articulos.py", "file_name": "ejercicio_articulos.py", "file_ext": "py", "file_size_in_byte": 4185, "program_lang": "python", "lang": "es", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "sqlite3.connect", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 54, "usage_type": "call"}]}
+{"seq_id": "30746213627", "text": "import json\nimport os\n\nimport numpy as np\nimport shapely\nimport torch\n\nfrom plankassembly.datasets.data_utils import add_noise, quantize_values\n\n\nclass Sideface():\n def __init__(self, linestring, line_width, line_type):\n self.linestring = linestring\n self.line_width = line_width\n self.line_type = line_type\n\n def to_polygon(self):\n return shapely.buffer(self.linestring, self.line_width / 2, cap_style=\"flat\")\n\n\ndef parse_sideface_from_polygons(polygons, max_thickness):\n\n lines = []\n for polygon in polygons:\n bounds = shapely.bounds(polygon).reshape(-1, 2)\n diffs = np.diff(bounds, axis=0).flatten()\n center = np.mean(bounds, 0)\n\n if diffs[1] < max_thickness:\n line = shapely.linestrings([bounds[0][0], bounds[1][0]], [center[1], center[1]])\n lines.append(Sideface(line, diffs[1], 1))\n\n if diffs[0] < max_thickness:\n line = shapely.linestrings([center[0], center[0]], [bounds[0][1], bounds[1][1]])\n lines.append(Sideface(line, diffs[0], 0))\n\n return lines\n\n\ndef merge_colinaer_sidefaces(lines, merge_tolerance, min_thickness):\n\n merged_lines = [lines[0], ]\n\n for query_line in lines[1:]:\n\n tree = shapely.STRtree([line.linestring for line in merged_lines])\n indices = tree.query(query_line.linestring, predicate='intersects')\n\n colinear_indices = []\n\n for index in np.sort(indices):\n # find colinear case and merge\n coords = shapely.get_coordinates([query_line.linestring, merged_lines[index].linestring])\n\n if ((np.max(coords[:, 0]) - np.min(coords[:, 0])) < merge_tolerance or\n (np.max(coords[:, 1]) - np.min(coords[:, 1])) < merge_tolerance) \\\n and np.abs(query_line.line_width - merged_lines[index].line_width) < merge_tolerance \\\n and query_line.line_type == merged_lines[index].line_type:\n colinear_indices.append(index)\n\n if len(colinear_indices) > 0:\n # merge colinear lines\n multilinestrings = shapely.multilinestrings(\n [query_line.linestring, ] + [merged_lines[i].linestring for i in colinear_indices])\n bounds = shapely.bounds(multilinestrings)\n bounds = np.array(bounds).reshape(2, 2)\n linestring = shapely.linestrings(*bounds.T)\n query_line = Sideface(linestring, query_line.line_width, query_line.line_type)\n\n # remove merged lines\n for i in reversed(colinear_indices):\n merged_lines.pop(i)\n\n # append new line\n merged_lines.append(query_line)\n\n merged_lines = [line.to_polygon() for line in merged_lines if line.line_width >= min_thickness]\n\n return merged_lines\n\n\nclass SidefaceDataset(torch.utils.data.Dataset):\n\n def __init__(self, root, info_files, token, cfg, augmentation=False):\n\n self.root = root\n self.info_files = info_files\n self.augmentation = augmentation\n self.token = token\n\n self.vocab_size = cfg.VOCAB_SIZE\n self.num_input_dof = cfg.NUM_INPUT_DOF\n self.max_input_length = cfg.MAX_INPUT_LENGTH\n self.max_output_length = cfg.MAX_OUTPUT_LENGTH\n self.num_bits = cfg.NUM_BITS\n\n self.aug_ratio = cfg.AUG_RATIO\n self.noise_ratio = cfg.NOISE_RATIO\n self.noise_length = cfg.NOISE_LENGTH\n\n self.max_thickness = cfg.MAX_THICKNESS / cfg.SCALE\n self.min_thickness = cfg.MIN_THICKNESS / cfg.SCALE\n self.merge_tolerance = cfg.MERGE_TOLERANCE / cfg.SCALE\n\n def __len__(self):\n return len(self.info_files)\n\n def extract_sideface(self, linestrings, views):\n\n sidefaces = []\n faceviews = []\n\n for view_index in range(3):\n\n line = [l_i for l_i, v_i in zip(linestrings, views) if v_i == view_index]\n\n if len(line) == 0:\n continue\n\n polygon = shapely.get_parts(shapely.polygonize(line))\n\n sideface = parse_sideface_from_polygons(polygon, self.max_thickness)\n\n if len(sideface) == 0:\n continue\n\n merged_sideface = merge_colinaer_sidefaces(sideface, self.merge_tolerance, self.min_thickness)\n\n sidefaces.extend(merged_sideface)\n faceviews.extend([view_index, ] * len(merged_sideface))\n\n sidefaces = shapely.bounds(sidefaces)\n\n return sidefaces, faceviews\n\n def prepare_input_sequence(self, faces, views):\n # input\n input_value = quantize_values(np.array(faces), self.num_bits)\n input_view = np.array(views, dtype='long')\n \n if len(faces) != 0:\n # sort faces by first by view, then by lines\n face_with_view = np.concatenate((input_value, input_view[..., np.newaxis]), axis=1)\n sort_inds = np.lexsort(face_with_view.T[[3, 1, 2, 0, 4]])\n\n input_value = input_value[sort_inds].flatten()\n input_view = input_view[sort_inds]\n\n # position\n _, counts = np.unique(input_view, return_counts=True)\n input_pos = np.concatenate([np.arange(count) for count in counts])\n\n # coordinate\n input_coord = np.arange(len(input_value)) % self.num_input_dof\n\n # repeat for each token\n input_pos = np.repeat(input_pos, 4)\n input_view = np.repeat(input_view, 4)\n\n else:\n # deal with empty sidefaces\n input_pos = np.zeros_like(input_view, dtype='long')\n input_coord = np.zeros_like(input_view, dtype='long')\n\n # add stop token\n input_value = np.append(input_value, self.token.END)\n num_input = len(input_value)\n\n # add pad tokens\n pad_length = self.max_input_length - num_input\n\n input_value = np.pad(input_value, (0, pad_length-1), constant_values=self.token.PAD)\n input_pos = np.pad(input_pos, (0, pad_length))\n input_coord = np.pad(input_coord, (0, pad_length))\n input_view = np.pad(input_view, (0, pad_length))\n input_mask = (input_value == self.token.PAD)\n\n inputs = {\n 'input_value': input_value,\n 'input_pos': input_pos,\n 'input_coord': input_coord,\n 'input_view': input_view,\n 'input_mask': input_mask\n }\n\n return inputs\n\n def prepare_output_sequence(self, planks, attach):\n # output\n value = quantize_values(planks, self.num_bits)\n\n # add stop token\n value = np.append(value, self.token.END)\n num_output = len(value)\n\n # add pad tokens\n value = np.pad(value, (0, self.max_output_length - num_output), constant_values=self.token.PAD)\n mask = (value == self.token.PAD)\n\n # label\n label = np.pad(attach, (0, self.max_output_length - len(attach)), constant_values=-1)\n\n label[label != -1] += self.vocab_size\n label[label == -1] = value[label == -1]\n\n outputs = {\n 'output_value': value,\n 'output_label': label,\n 'output_mask': mask\n }\n\n return outputs\n\n def __getitem__(self, index):\n \"\"\" Load data for data i\"\"\"\n with open(os.path.join(self.root, self.info_files[index]), \"r\") as f:\n info = json.loads(f.read())\n\n name = info['name']\n svgs = info['svgs']\n\n linestrings = [shapely.from_geojson(svg) for svg in svgs]\n\n views = np.array(info['views'], dtype='long')\n types = np.array(info['types'], dtype='long')\n\n planks = np.array(info['coords']).flatten()\n attach = np.array(info['attach']).flatten()\n\n sidefaces, faceviews = [], []\n\n if self.augmentation and np.random.random() < self.aug_ratio:\n\n noisy_linestrings, noisy_views, _ = add_noise(\n linestrings, views, types, self.noise_ratio, self.noise_length)\n\n sidefaces, faceviews = self.extract_sideface(noisy_linestrings, noisy_views)\n\n # detect degenerated case\n if len(sidefaces) == 0:\n linestrings = [shapely.from_geojson(svg) for svg in svgs]\n views = np.array(info['views'], dtype='long')\n\n sidefaces, faceviews = self.extract_sideface(linestrings, views)\n\n inputs = self.prepare_input_sequence(sidefaces, faceviews)\n\n outputs = self.prepare_output_sequence(planks, attach)\n\n # construct batch data\n batch = {'name': name, **inputs, **outputs}\n\n return batch\n", "repo_name": "manycore-research/PlankAssembly", "sub_path": "plankassembly/datasets/sideface_data.py", "file_name": "sideface_data.py", "file_ext": "py", "file_size_in_byte": 8500, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 49, "dataset": "github-code", "pt": "24", "api": [{"api_name": "shapely.buffer", "line_number": 18, "usage_type": "call"}, {"api_name": "shapely.bounds", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 27, "usage_type": "call"}, {"api_name": "shapely.linestrings", "line_number": 30, "usage_type": "call"}, {"api_name": "shapely.linestrings", "line_number": 34, "usage_type": "call"}, {"api_name": "shapely.STRtree", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 51, "usage_type": "call"}, {"api_name": "shapely.get_coordinates", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 57, "usage_type": "call"}, {"api_name": "shapely.multilinestrings", "line_number": 63, "usage_type": "call"}, {"api_name": "shapely.bounds", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 66, "usage_type": "call"}, {"api_name": "shapely.linestrings", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 82, "usage_type": "attribute"}, {"api_name": "shapely.get_parts", "line_number": 120, "usage_type": "call"}, {"api_name": "shapely.polygonize", "line_number": 120, "usage_type": "call"}, {"api_name": "shapely.bounds", "line_number": 132, "usage_type": "call"}, {"api_name": "plankassembly.datasets.data_utils.quantize_values", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 143, "usage_type": "attribute"}, {"api_name": "numpy.lexsort", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 175, "usage_type": "call"}, {"api_name": "plankassembly.datasets.data_utils.quantize_values", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 216, "usage_type": "call"}, {"api_name": "os.path", "line_number": 216, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 217, "usage_type": "call"}, {"api_name": "shapely.from_geojson", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 232, "usage_type": "attribute"}, {"api_name": "plankassembly.datasets.data_utils.add_noise", "line_number": 234, "usage_type": "call"}, {"api_name": "shapely.from_geojson", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 242, "usage_type": "call"}]}
+{"seq_id": "33043540698", "text": "import urllib.request\nfrom bs4 import BeautifulSoup\nfrom datetime import date\nimport re\nimport csv\n\nfrom backend import createGlobalEstatesCsv, updateOffers\nfrom consts import OFFICE_PROPERTY, NEWLINE, WRITING_MODE, DELIMITER, ENCODING, HEADERS\nfrom progressBar import ProgressBar\nfrom scrapers.american.myhelpers import TEMP_ARR, TEMPLATE, LINKS, FIELD_NAMES, APARTMENT, PLOT, HOUSE, TYPES\nfrom helpers import getFileName\n\n\nclass Searcher:\n progressBar = None\n today = date.today()\n result = []\n offers_len = 0\n offers = {\n 'mieszkanie': {\n 'count': 1,\n 'links': []\n },\n 'dom': {\n 'count': 1,\n 'links': []\n },\n 'dzialka': {\n 'count': 1,\n 'links': []\n },\n 'lokal': {\n 'count': 2,\n 'links': []\n },\n }\n\n def getValue(self, value):\n value = dict(zip(range(len(value)), value))\n if value.get(1):\n temp_text = str(value.get(1))\n temp_text = temp_text.replace('', '')\n temp_text = temp_text.replace('', '')\n return str(value.get(0)) + temp_text\n else:\n return str(value.get(0))\n\n def findValues(self, offer, key, temp_arr):\n temp = temp_arr.copy()\n r = urllib.request.urlopen(offer)\n soup = BeautifulSoup(r, \"html.parser\")\n photos = ''\n count_photos = 0\n temp['typ'] = key\n temp['link'] = offer\n temp['nazwa_biura'] = 'American Home'\n for values in soup.findAll('div', class_='area'):\n if values.strong.previous_sibling.find(''):\n if 'Piętro' == values.strong.previous_sibling.get_text():\n if re.findall(\"(.*)/.\", self.getValue(values.strong.contents)):\n temp['pietro'] = re.findall(\"(.*)/.\", self.getValue(values.strong.contents))[0]\n if re.findall(\".*/(.*)\", self.getValue(values.strong.contents)):\n temp['budynek_pietra'] = re.findall(\".*/(.*)\", self.getValue(values.strong.contents))[0]\n elif 'dzialka' == key and 'Powierzchnia' == values.strong.previous_sibling.get_text():\n temp['powierzchnia_dzialki'] = self.getValue(values.strong.contents)\n elif 'Cena' == values.strong.previous_sibling.get_text():\n temp['cena'] = ''.join(re.findall(\"(\\d*\\d)\", self.getValue(values.strong.contents)))\n elif 'Numer oferty' == values.strong.previous_sibling.get_text():\n temp['numer_oferty'] = self.getValue(values.strong.contents)\n temp['nr_oferty'] = self.getValue(values.strong.contents)\n elif values.strong.previous_sibling.get_text() in FIELD_NAMES:\n temp[FIELD_NAMES[values.strong.previous_sibling.get_text()]] = self.getValue(\n values.strong.contents)\n else:\n if values.strong.previous_sibling.previous_sibling + values.strong.previous_sibling.get_text() in FIELD_NAMES:\n temp[FIELD_NAMES[\n values.strong.previous_sibling.previous_sibling + values.strong.previous_sibling.get_text()]] = self.getValue(\n values.strong.contents)\n\n for values in soup.findAll('div', class_='tab-pane fade active in'):\n temp['opis'] = values.find_next(class_='property-detail_overview').get_text()\n for values in soup.findAll('i', class_='fa fa-phone'):\n temp['telefon'] = values.find_next('a').get_text()\n for values in soup.findAll('i', class_='fa fa-envelope-o'):\n temp['email'] = values.find_next('a').get_text()\n for values in soup.findAll('a', class_='blueimp'):\n photos += \"https://www.americanhome.pl/\" + values['href'] + ','\n count_photos = count_photos + 1\n temp['zdjecia_linki'] = photos\n temp['zdjecie_glowne'] = photos.split(',')[0]\n temp['zdjecie_glowne_link'] = photos.split(',')[0]\n temp['liczba_zdjec'] = count_photos\n temp['data_skanowania'] = self.today.strftime(\"%d/%m/%Y\")\n self.progressBar.progress()\n return temp\n\n def searchOffers(self):\n for link in LINKS:\n if len(self.offers[TYPES[(link.split('/'))[4]]]['links']) >= 5:\n continue\n r = urllib.request.urlopen(link)\n soup = BeautifulSoup(r, \"html.parser\", from_encoding=\"iso-8859-1\")\n for offer in soup.findAll('a', href=True):\n if len(self.offers[TYPES[(link.split('/'))[4]]]['links']) == 5:\n continue\n if TEMPLATE in offer['href']:\n if offer['href'] not in self.offers[TYPES[(link.split('/'))[4]]]['links']:\n self.offers[TYPES[(link.split('/'))[4]]]['links'].append(offer['href'])\n\n def getOffers(self):\n for key in self.offers.keys():\n for offer in self.offers[key]['links']:\n self.result.append(self.findValues(offer, key, TEMP_ARR))\n\n def saveToFile(self):\n with open(getFileName(OFFICE_PROPERTY['american']), WRITING_MODE, newline=NEWLINE, encoding=ENCODING) as f:\n writer = csv.DictWriter(f, delimiter=DELIMITER, fieldnames=HEADERS)\n writer.writerows(self.result)\n\n def run(self, root):\n self.progressBar = ProgressBar(root, 20)\n self.searchOffers()\n self.getOffers()\n self.saveToFile()\n\n\ndef startAmerican(root, loader, updateOfferLabel, updateNewOffers):\n searcher = Searcher()\n searcher.run(root)\n updateOffers(root, loader, updateOfferLabel, updateNewOffers)\n\n", "repo_name": "dilejt/Scraper", "sub_path": "scrapers/american/american.py", "file_name": "american.py", "file_ext": "py", "file_size_in_byte": 5724, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "datetime.date.today", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 16, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 50, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 50, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 50, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 51, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 60, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 61, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 62, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 63, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 67, "usage_type": "call"}, {"api_name": "scrapers.american.myhelpers.FIELD_NAMES", "line_number": 71, "usage_type": "name"}, {"api_name": "scrapers.american.myhelpers.FIELD_NAMES", "line_number": 72, "usage_type": "name"}, {"api_name": "scrapers.american.myhelpers.FIELD_NAMES", "line_number": 75, "usage_type": "name"}, {"api_name": "scrapers.american.myhelpers.FIELD_NAMES", "line_number": 76, "usage_type": "name"}, {"api_name": "scrapers.american.myhelpers.LINKS", "line_number": 98, "usage_type": "name"}, {"api_name": "scrapers.american.myhelpers.TYPES", "line_number": 99, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 101, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 101, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 101, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 102, "usage_type": "call"}, {"api_name": "scrapers.american.myhelpers.TYPES", "line_number": 104, "usage_type": "name"}, {"api_name": "scrapers.american.myhelpers.TEMPLATE", "line_number": 106, "usage_type": "name"}, {"api_name": "scrapers.american.myhelpers.TYPES", "line_number": 107, "usage_type": "name"}, {"api_name": "scrapers.american.myhelpers.TYPES", "line_number": 108, "usage_type": "name"}, {"api_name": "scrapers.american.myhelpers.TEMP_ARR", "line_number": 113, "usage_type": "argument"}, {"api_name": "consts.WRITING_MODE", "line_number": 116, "usage_type": "argument"}, {"api_name": "helpers.getFileName", "line_number": 116, "usage_type": "call"}, {"api_name": "consts.OFFICE_PROPERTY", "line_number": 116, "usage_type": "name"}, {"api_name": "consts.NEWLINE", "line_number": 116, "usage_type": "name"}, {"api_name": "consts.ENCODING", "line_number": 116, "usage_type": "name"}, {"api_name": "csv.DictWriter", "line_number": 117, "usage_type": "call"}, {"api_name": "consts.DELIMITER", "line_number": 117, "usage_type": "name"}, {"api_name": "consts.HEADERS", "line_number": 117, "usage_type": "name"}, {"api_name": "progressBar.ProgressBar", "line_number": 121, "usage_type": "call"}, {"api_name": "backend.updateOffers", "line_number": 130, "usage_type": "call"}]}
+{"seq_id": "40174553337", "text": "#!/usr/bin/env python\n\"\"\"\nexample creation of PNGs for use with this program:\n./FigureMaker.py in/apr14T085454.ini\n\nFor a real video root directory, finds all the time results and makes multipage GIFs or TIFFs as you like from the PNGs\nif you have very many PNGs, you might like to use FFV1 in an .avi instead of .gif.\n\nMotivation: I have two separate programs (histfeas, histutils) that each create complicated figures. There is NOT currently\nhttp://stackoverflow.com/questions/22521560/how-to-combine-several-matplotlib-figures-into-one-figure\na good way to combine two figures into one, even by grabbing axes.\n\nRecommendation was to do this--create PNGs and smash together in post-processing.\n\"\"\"\nfrom pathlib import Path\nfrom wand.image import Image\nfrom tempfile import NamedTemporaryFile\nimport subprocess\n\n# from wand.exceptions import WandError\n# from wand.display import display\n\nMINHEIGHT = 500 # each element of canvas will be scaled up to at least this height--avoids widely disparate image sizes\n\n\ndef pngsmash(rdir, impat, fpat, outfn):\n rdir = Path(rdir).expanduser()\n outfn = Path(outfn).expanduser()\n\n if not outfn.suffix:\n raise ValueError(\n \"you must specify a suffix e.g. .gif to the output filename so that Wand knows what format to write\"\n )\n\n ilist = sorted(rdir.glob(impat))\n flist = sorted(rdir.glob(fpat))\n\n assert len(ilist) == len(flist), \"unequal len() {} & {} {} != {}\".format(\n ilist, flist, len(ilist), len(flist)\n )\n\n with Image() as anim:\n for i, f in zip(ilist, flist):\n with Image(filename=str(i)) as I, Image(filename=str(f)) as F, Image() as J:\n #%% enforce minimum height (aspect-preserving scale increase if needed)\n for im in (I, F):\n if im.height < MINHEIGHT:\n im.transform(resize=\"x\" + str(MINHEIGHT))\n #%% compose composite canvas\n J.blank(\n max(I.width, F.width), I.height + F.height\n ) # blank canvas on which to place images\n J.composite(F, 0, 0) # add data to canvas\n J.composite(I, 0, F.height) # add video frame to canvas\n anim.sequence.append(J)\n\n if outfn.suffix == \".gif\":\n print(\"writing\", outfn)\n anim.save(filename=str(outfn))\n elif outfn.suffix in (\".avi\", \".mp4\", \".ogv\", \".webm\"):\n with NamedTemporaryFile(\n suffix=\".gif\"\n ) as f: # forcing .gif temp since it's what Wand can handle\n print(\"using tempfile\", f.name)\n anim.save(filename=f.name) # the handle didn't work for some reason\n print(\"writing\", outfn)\n # NOTE: -c:v ffv\n subprocess.call([\"ffmpeg\", \"-i\", f.name, \"-c:v\", \"ffv1\", outfn])\n\n\nif __name__ == \"__main__\":\n from argparse import ArgumentParser\n\n p = ArgumentParser()\n p.add_argument(\"rdir\", help=\"directory where the sim/inversion output PNGs are\")\n p.add_argument(\"ofn\", help=\"output GIF filename\")\n p.add_argument(\"-i\", \"--impat\", help=\"glob pattern for raw image PNG\", default=\"rawFrame*.png\")\n p.add_argument(\"-f\", \"--fpat\", help=\"glob pattern for inversion PNG\", default=\"est*.png\")\n p = p.parse_args()\n\n pngsmash(p.rdir, p.impat, p.fpat, p.ofn)\n\n\n\"\"\"\nold way\n#!/bin/sh\n\ntmpfn=/tmp/$(printf \"%03d\" $i)tmp.png\necho \"${flist[i]} ${glist[i]} -> $tmpfn\"\nconvert -append \"${flist[i]}\" \"${glist[i]}\" tmpfn\ndone\n\nconvert -convert -delay 30 /tmp/*tmp.png $rdir/$outfn\nrm /tmp/*tmp.png\n\"\"\"\n", "repo_name": "space-physics/histfeas", "sub_path": "Plots/CombineTimePlots.py", "file_name": "CombineTimePlots.py", "file_ext": "py", "file_size_in_byte": 3598, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "pathlib.Path", "line_number": 27, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 28, "usage_type": "call"}, {"api_name": "wand.image.Image", "line_number": 42, "usage_type": "call"}, {"api_name": "wand.image.Image", "line_number": 44, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 61, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 68, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 74, "usage_type": "call"}]}
+{"seq_id": "11610500510", "text": "import collections\nimport math\ndef numberOfBoomerangs(points):\n nums = 0\n for x1, y1 in points:\n distance = collections.defaultdict(int)\n for x2, y2 in points:\n \tdx = abs(x2-x1)\n \tdy = abs(y2-y1)\n \td = dx*dx+dy*dy\n \tdistance[d]+=1\n\n print(distance)\n nums += sum(n * (n-1) for n in distance.values())\n\n return nums\n\n\n\nprint(numberOfBoomerangs([[0,0],[1,0],[1,0]]))\n\n\n\n\t\t\n\n\n", "repo_name": "ted801008/Practice", "sub_path": "leetcode/numberOfBoomerangs.py", "file_name": "numberOfBoomerangs.py", "file_ext": "py", "file_size_in_byte": 437, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "collections.defaultdict", "line_number": 6, "usage_type": "call"}]}
+{"seq_id": "9868402619", "text": "import gym\nfrom gym import spaces\nimport numpy as np\n\nLEN_LOOKBACK = 10\nLEN_EPISODE = 500\n\n\n\nclass InsuranceEnv(gym.Env):\n\n def __init__(self, num_agents, num_insurances):\n super(InsuranceEnv, self).__init__()\n\n self.NUM_AGENTS = num_agents\n self.NUM_INSURANCES = num_insurances\n\n self.safe_mu = 0\n self.safe_sigma = 0.1\n self.risky_mu = 0\n self.risky_sigma = 1\n self.insurance_return = 0\n\n self.action_space = spaces.Discrete(2+2*self.NUM_INSURANCES)\n \"\"\"\n 0: Safe non-insured\n 1: Risky non-insured\n 2: Safe insured\n 3: Risky insured\n 4: Safe insured2\n ...\n \"\"\"\n\n self.action_switcher = {\n 0: (self.safe_mu, self.safe_sigma, 0.),\n 1: (self.risky_mu, self.risky_sigma, 0.),\n }\n\n for i in range(self.NUM_INSURANCES*2):\n if i % 2 == 0:\n self.action_switcher[i+2] = (self.safe_mu, self.safe_sigma, 1.)\n else:\n self.action_switcher[i+2] = (self.risky_mu, self.risky_sigma, 1.)\n\n\n self.observation_space = spaces.Dict({\n 'was_insured': spaces.MultiBinary(LEN_LOOKBACK),\n 'insurance_costs': spaces.Box(low=0.0, high=1.0, shape=(LEN_LOOKBACK,), dtype=np.float32),\n 'new_cost': spaces.Box(low=0.0, high=1.0, shape=(1,), dtype=np.float32)\n })\n\n self.reset()\n\n def step(self, agent_actions):\n agent_rewards = [0.0 for _ in range(self.NUM_AGENTS)]\n insurance_rewards = [-.01 for _ in range(self.NUM_INSURANCES)]\n\n if not type(agent_actions) == list or type(agent_actions) == np.array:\n agent_actions = [agent_actions]\n\n for agent_id, agent_action in enumerate(agent_actions):\n\n mu, sigma, insured = self.action_switcher.get(agent_action, (0, 0, 0.0))\n\n insurance_id = agent_action//2 - 1\n insurance_cost = self.current_cost[insurance_id]\n\n self.action_counter[agent_id, agent_action] += 1\n\n agent_reward = np.random.normal(mu, sigma)\n insurer_reward = 0\n if insured == 1.:\n if agent_reward < self.insurance_return:\n insurer_reward = agent_reward - self.insurance_return + insurance_cost\n agent_reward = self.insurance_return - insurance_cost\n else:\n agent_reward -= insurance_cost\n insurer_reward = insurance_cost\n\n self.was_insured[agent_id].pop(0)\n self.was_insured[agent_id].append(insured)\n self.insurance_costs[insurance_id].pop(0)\n self.insurance_costs[insurance_id].append(self.get_insurance_cost(insurance_id))\n\n agent_rewards[agent_id] = agent_reward\n insurance_rewards[insurance_id] += insurer_reward\n\n self.num_trials -= 1\n\n done = self.num_trials == 0\n\n observation = self.get_obs()\n\n \"\"\"Try to force insurance to make price competitive\"\"\"\n # insurer_reward -= .01\n\n return [observation for _ in range(self.NUM_INSURANCES+self.NUM_AGENTS)], (*insurance_rewards, *agent_rewards),\\\n done, None\n\n \"\"\"Get action from insurance\"\"\"\n\n \"\"\"Get action from agent\"\"\"\n\n \"\"\"\"Calculate rewards\"\"\"\n\n \"\"\"Save rewards in replay memory\"\"\"\n\n \"\"\"Train insurance and agent\"\"\"\n\n def get_obs(self):\n obs = np.zeros((self.NUM_INSURANCES, 21), dtype=np.float32)\n\n obs[:, :10] = self.was_insured\n obs[:, 10:20] = self.insurance_costs\n obs[:, 20] = self.current_cost\n return obs\n\n def reset(self):\n self.was_insured = [[0.0 for _ in range(LEN_LOOKBACK)] for _ in range(self.NUM_AGENTS)]\n self.insurance_costs = [[0.0 for _ in range(LEN_LOOKBACK)] for _ in range(self.NUM_INSURANCES)]\n self.set_insurance_cost()\n self.num_trials = LEN_EPISODE\n\n self.action_counter = np.zeros((self.NUM_AGENTS, self.NUM_INSURANCES*2+2))\n\n return [self.get_obs() for _ in range(self.NUM_INSURANCES+self.NUM_AGENTS)]\n\n def set_insurance_cost(self, insurance_cost: float = 0.0, insurance_id=None):\n if insurance_id is None:\n self.current_cost = [insurance_cost for _ in range(self.NUM_INSURANCES)]\n else:\n if insurance_id > self.NUM_INSURANCES:\n raise Exception('Insurance ID higher than number of insurances (ID: {}, insurances: {}'.format(\n insurance_id, self.NUM_INSURANCES\n ))\n self.current_cost[insurance_id] = insurance_cost\n\n def get_insurance_cost(self, insurance_id=None):\n if insurance_id is None:\n return self.current_cost\n else:\n if insurance_id > self.NUM_INSURANCES:\n raise Exception('Insurance ID higher than number of insurances (ID: {}, insurances: {}'.format(\n insurance_id, self.NUM_INSURANCES\n ))\n return self.current_cost[insurance_id]\n\n def render(self, mode='human', close=False):\n pass\n", "repo_name": "valko073/rl_insurance", "sub_path": "insurance_env.py", "file_name": "insurance_env.py", "file_ext": "py", "file_size_in_byte": 5107, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "gym.Env", "line_number": 10, "usage_type": "attribute"}, {"api_name": "gym.spaces.Discrete", "line_number": 24, "usage_type": "call"}, {"api_name": "gym.spaces", "line_number": 24, "usage_type": "name"}, {"api_name": "gym.spaces.Dict", "line_number": 46, "usage_type": "call"}, {"api_name": "gym.spaces", "line_number": 46, "usage_type": "name"}, {"api_name": "gym.spaces.MultiBinary", "line_number": 47, "usage_type": "call"}, {"api_name": "gym.spaces", "line_number": 47, "usage_type": "name"}, {"api_name": "gym.spaces.Box", "line_number": 48, "usage_type": "call"}, {"api_name": "gym.spaces", "line_number": 48, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 48, "usage_type": "attribute"}, {"api_name": "gym.spaces.Box", "line_number": 49, "usage_type": "call"}, {"api_name": "gym.spaces", "line_number": 49, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 49, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 58, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 70, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 111, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 124, "usage_type": "call"}]}
+{"seq_id": "27877853094", "text": "import io\nimport os\nfrom setuptools import setup\n\nimport monkeypatch # pylint: disable=W0611\n\ndef read_contents(*names, **kwargs):\n return io.open(\n os.path.join(*names),\n encoding=kwargs.get(\"encoding\", \"utf8\")\n ).read()\n\ndescription = 'Environment variable editor.'\ntry:\n long_description = read_contents(os.path.dirname(__file__), 'README.rst')\nexcept:\n long_description = description\n\nsetup(name='enveditor',\n version='0.1',\n description=description,\n long_description=long_description,\n author='Simon Kennedy',\n author_email='sffjunkie+code@gmail.com',\n url=\"https://github.com/sffjunkie/astral\",\n license='Apache-2.0',\n classifiers=[\n \"Intended Audience :: Developers\",\n \"Programming Language :: Python :: 3\",\n ],\n package_dir={'': 'src'},\n packages=['enveditor'],\n tests_require=['pytest-runner'],\n)\n", "repo_name": "sffjunkie/enveditor", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 907, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "io.open", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "setuptools.setup", "line_number": 19, "usage_type": "call"}]}
+{"seq_id": "3069640037", "text": "from django.shortcuts import render,get_object_or_404,reverse,redirect\nfrom .models import Post\nfrom .froms import CreatePostForm, PostEditForm\n\n# Create your views here.\n\n\n\n\ndef PostList(request):\n postlist = Post.objects.all()\n context = {\n \"postlist\":postlist\n }\n template_name = \"postlist.html\"\n return render(request,template_name,context)\n\n\n\n\ndef PostDetail(request,id):\n post = get_object_or_404(Post,id=id)\n context = {\n 'post':post\n }\n template_name = \"postdetail.html\"\n return render(request,template_name,context)\n\n\ndef PostUpdateis(request,id):\n post = get_object_or_404(Post,id=id)\n if request.user != post.Author :\n return redirect(\"home\")\n else :\n if request.method == \"POST\" :\n form = PostEditForm(request.POST ,instance=post)\n if form.is_valid():\n form.save()\n return redirect(\"home\")\n else:\n form = PostEditForm(instance=post)\n return render(request,\"update-post.html\",context={\"form\":form})\n\ndef CreatPost(request):\n if request.method == \"POST\":\n form = CreatePostForm(request.POST or None)\n if form.is_valid():\n new_post= form.save(commit=False)\n new_post.Author = request.user\n new_post.save()\n return redirect(\"home\")\n else:\n form = CreatePostForm()\n context = {\"form\":form }\n template_name=\"CreatePostForm.html\"\n return render(request,template_name,context)\n\n\n\n\n", "repo_name": "riddick1368/Create_shop_online_Postgres", "sub_path": "ecommerce/blog/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1505, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "models.Post.objects.all", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 11, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 16, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 22, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 22, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 27, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 31, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 31, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 33, "usage_type": "call"}, {"api_name": "froms.PostEditForm", "line_number": 36, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 39, "usage_type": "call"}, {"api_name": "froms.PostEditForm", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 42, "usage_type": "call"}, {"api_name": "froms.CreatePostForm", "line_number": 46, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 51, "usage_type": "call"}, {"api_name": "froms.CreatePostForm", "line_number": 53, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 56, "usage_type": "call"}]}
+{"seq_id": "412468980", "text": "import logging\nfrom datetime import datetime\nimport os\n\n# Creating logs directory to store log in files\nLOG_DIR = \"Insurance_log\"\n\n# Creating file name for log file based on current timestamp\nCURRENT_TIME_STAMP = f\"{datetime.now().strftime('%Y-%m-%d-%H-%M-%S')}\"\n\n# Here, We are going to define the path to store log with folder_name\nLOG_FIlE_NAME = f\"log_{CURRENT_TIME_STAMP}.log\"\n\n\n#Creating LOG_DIR if it does not exists.\nos.makedirs(LOG_DIR, exist_ok=True)\n\n#Creating file path for projects.\nLOG_FIlE_PATH = os.path.join(LOG_DIR, LOG_FIlE_NAME)\n\n# If you want to read log select baseconfig and press f12 from your system.\nlogging.basicConfig(filename=LOG_FIlE_PATH,\nfilemode = \"w\",\nformat = '[%(asctime)s] %(name)s - %(levelname)s - %(message)s',\n#level=logging.INFO,\nlevel=logging.DEBUG,\n)", "repo_name": "Shivan118/Project-EWB", "sub_path": "Insurance/logger/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 794, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 8, "dataset": "github-code", "pt": "24", "api": [{"api_name": "datetime.datetime.now", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 9, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 22, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 26, "usage_type": "attribute"}]}
+{"seq_id": "34460065110", "text": "import serial\nimport serial.tools.list_ports\nimport os\nimport subprocess\nimport time\nimport stream\nfrom utils import parsers\nfrom tkinter import messagebox\n\n\ndef pre_test():\n \"\"\" Runs the pre-test to check if what is required for etching is ready\n\n This pre-test runs the homing.gcode file, which attemps to home the stage\n first, then home the stage to the origin point of the 3D glass cube\n \"\"\"\n mcu_serial = parsers.get_from_config('serial_number', os.path.dirname(os.path.realpath(__file__)))\n device_path = _get_device_path(mcu_serial)\n reason = ''\n\n # Some spaghetti logic up in here\n if device_path is False:\n reason = 'Could not find correct serial port'\n result = False\n else:\n result = _home_stage(device_path)\n if result is False:\n reason = 'Error in streaming'\n\n ret_dict = dict()\n ret_dict['device'] = device_path\n ret_dict['result'] = result\n ret_dict['reason'] = reason\n return ret_dict\n\n\ndef _get_device_path(target_serial=None):\n \"\"\" Returns the device path for first match of given serial \"\"\"\n if target_serial is None:\n return False\n\n ports = list(serial.tools.list_ports.comports())\n for p in ports:\n if p.serial_number == target_serial:\n return p.device\n\n return False\n\n\ndef _home_stage(device_path):\n \"\"\" Homes the stage to begin etching process \"\"\"\n homing_gcode = os.path.dirname(os.path.realpath(__file__)) + '\\\\homing.gcode'\n# streaming_file = os.path.dirname(os.path.realpath(__file__)) + '\\\\stream.py'\n return stream.start_stream(homing_gcode, device_path)\n\n\ndef full_test(filename, device_path):\n \"\"\" Initiates the stream for the given gcode file\n\n Requires the gcode file to be passed in and the device path\n \"\"\"\n result = stream.start_stream(filename, device_path)\n if result is False:\n reason = 'CRITICAL ERROR STREAMING GCODE'\n else:\n reason =''\n\n return result, reason\n\n\ndef start_laser():\n \"\"\" Establishes a connection and fires the laser \"\"\"\n laser_serial = parsers.get_from_config('laser_number', os.path.dirname(os.path.realpath(__file__)))\n device_path = _get_device_path(laser_serial)\n\n if device_path is False:\n message = 'Could not find laser to turn on! Turn it on manually!\\n' \\\n 'Etching process will continue after this is closed'\n messagebox.showerror(title='Error', message=message)\n return\n\n s = serial.Serial(device_path, 9600, parity=serial.PARITY_NONE, stopbits=serial.STOPBITS_ONE, bytesize=serial.EIGHTBITS)\n s.write('$FIRE 01\\r'.encode())\n s.close()\n\n\ndef stop_laser():\n \"\"\" Stops the laser \"\"\"\n laser_serial = parsers.get_from_config('laser_number', os.path.dirname(os.path.realpath(__file__)))\n device_path = _get_device_path(laser_serial)\n\n if device_path is False:\n message = 'Could not find laser to turn off! Turn it off manually!\\n'\n messagebox.showerror(title='Error', message=message)\n return\n\n s = serial.Serial(device_path, 9600, parity=serial.PARITY_NONE, stopbits=serial.STOPBITS_ONE, bytesize=serial.EIGHTBITS)\n s.write('$STOP 00\\r'.encode())\n s.close()\n\n\nif __name__ == \"__main__\":\n print('starting')\n start_laser()\n print('stopping')\n time.sleep(2)\n stop_laser()\n\n", "repo_name": "Kommotion/Senior-Design", "sub_path": "gui/serials.py", "file_name": "serials.py", "file_ext": "py", "file_size_in_byte": 3323, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "24", "api": [{"api_name": "utils.parsers.get_from_config", "line_number": 17, "usage_type": "call"}, {"api_name": "utils.parsers", "line_number": 17, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 17, "usage_type": "call"}, {"api_name": "serial.tools.list_ports.comports", "line_number": 42, "usage_type": "call"}, {"api_name": "serial.tools", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 52, "usage_type": "call"}, {"api_name": "stream.start_stream", "line_number": 54, "usage_type": "call"}, {"api_name": "stream.start_stream", "line_number": 62, "usage_type": "call"}, {"api_name": "utils.parsers.get_from_config", "line_number": 73, "usage_type": "call"}, {"api_name": "utils.parsers", "line_number": 73, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 73, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 79, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 79, "usage_type": "name"}, {"api_name": "serial.Serial", "line_number": 82, "usage_type": "call"}, {"api_name": "serial.PARITY_NONE", "line_number": 82, "usage_type": "attribute"}, {"api_name": "serial.STOPBITS_ONE", "line_number": 82, "usage_type": "attribute"}, {"api_name": "serial.EIGHTBITS", "line_number": 82, "usage_type": "attribute"}, {"api_name": "utils.parsers.get_from_config", "line_number": 89, "usage_type": "call"}, {"api_name": "utils.parsers", "line_number": 89, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 89, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 94, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 94, "usage_type": "name"}, {"api_name": "serial.Serial", "line_number": 97, "usage_type": "call"}, {"api_name": "serial.PARITY_NONE", "line_number": 97, "usage_type": "attribute"}, {"api_name": "serial.STOPBITS_ONE", "line_number": 97, "usage_type": "attribute"}, {"api_name": "serial.EIGHTBITS", "line_number": 97, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 106, "usage_type": "call"}]}
+{"seq_id": "3039574006", "text": "from __future__ import print_function\nimport os.path\nfrom googleapiclient.discovery import build\nfrom google_auth_oauthlib.flow import InstalledAppFlow\nfrom google.auth.transport.requests import Request\nfrom google.oauth2.credentials import Credentials\n\nimport os\nimport pickle\n# Gmail API utils\n# for encoding/decoding messages in base64\nfrom base64 import urlsafe_b64decode, urlsafe_b64encode\n# for dealing with attachement MIME types\nfrom email.mime.text import MIMEText\nfrom email.mime.multipart import MIMEMultipart\nfrom email.mime.image import MIMEImage\nfrom email.mime.audio import MIMEAudio\nfrom email.mime.base import MIMEBase\nfrom mimetypes import guess_type as guess_mime_type\n\n# Request all access (permission to read/send/receive emails, manage the inbox, and more)\nSCOPES = ['https://mail.google.com/']\nour_email = 'xandernaumenko@gmail.com'\n\n\ndef search_messages(service, query):\n result = service.users().messages().list(userId='me',q=query).execute()\n messages = [ ]\n if 'messages' in result:\n messages.extend(result['messages'])\n while 'nextPageToken' in result:\n page_token = result['nextPageToken']\n result = service.users().messages().list(userId='me',q=query, pageToken=page_token).execute()\n if 'messages' in result:\n messages.extend(result['messages'])\n return messages\n\n# utility functions\ndef get_size_format(b, factor=1024, suffix=\"B\"):\n \"\"\"\n Scale bytes to its proper byte format\n e.g:\n 1253656 => '1.20MB'\n 1253656678 => '1.17GB'\n \"\"\"\n for unit in [\"\", \"K\", \"M\", \"G\", \"T\", \"P\", \"E\", \"Z\"]:\n if b < factor:\n return f\"{b:.2f}{unit}{suffix}\"\n b /= factor\n return f\"{b:.2f}Y{suffix}\"\n\n\ndef clean(text):\n # clean text for creating a folder\n return \"\".join(c if c.isalnum() else \"_\" for c in text)\n\ndef parse_parts(service, parts, folder_name, message):\n \"\"\"\n Utility function that parses the content of an email partition\n \"\"\"\n ret = ''\n if parts:\n for part in parts:\n filename = part.get(\"filename\")\n mimeType = part.get(\"mimeType\")\n body = part.get(\"body\")\n data = body.get(\"data\")\n file_size = body.get(\"size\")\n part_headers = part.get(\"headers\")\n if part.get(\"parts\"):\n # recursively call this function when we see that a part\n # has parts inside\n parse_parts(service, part.get(\"parts\"), folder_name, message)\n if mimeType == \"text/plain\":\n # if the email part is text plain\n if data:\n text = urlsafe_b64decode(data).decode()\n ret += '\\n'.join([line for line in text.splitlines() if 'http' not in line]) + '\\n'\n elif mimeType == \"text/html\":\n # if the email part is an HTML content\n # save the HTML file and optionally open it in the browser\n if not filename:\n filename = \"index.html\"\n filepath = os.path.join(folder_name, filename)\n # print(\"Saving HTML to\", filepath)\n # with open(filepath, \"wb\") as f:\n # f.write(urlsafe_b64decode(data))\n else:\n # attachment other than a plain text or HTML\n for part_header in part_headers:\n part_header_name = part_header.get(\"name\")\n part_header_value = part_header.get(\"value\")\n if part_header_name == \"Content-Disposition\":\n if \"attachment\" in part_header_value:\n # we get the attachment ID \n # and make another request to get the attachment itself\n print(\"Saving the file:\", filename, \"size:\", get_size_format(file_size))\n attachment_id = body.get(\"attachmentId\")\n attachment = service.users().messages() \\\n .attachments().get(id=attachment_id, userId='me', messageId=message['id']).execute()\n data = attachment.get(\"data\")\n # filepath = os.path.join(folder_name, filename)\n # if data:\n # with open(filepath, \"wb\") as f:\n # f.write(urlsafe_b64decode(data))\n return ret\n\ndef read_message(service, message):\n \"\"\"\n This function takes Gmail API `service` and the given `message_id` and does the following:\n - Downloads the content of the email\n - Prints email basic information (To, From, Subject & Date) and plain/text parts\n - Creates a folder for each email based on the subject\n - Downloads text/html content (if available) and saves it under the folder created as index.html\n - Downloads any file that is attached to the email and saves it in the folder created\n \"\"\"\n msg = service.users().messages().get(userId='me', id=message['id'], format='full').execute()\n # parts can be the message body, or attachments\n payload = msg['payload']\n headers = payload.get(\"headers\")\n parts = payload.get(\"parts\")\n folder_name = \"email\"\n has_subject = False\n ret = ''\n if headers:\n # this section prints email basic info & creates a folder for the email\n for header in headers:\n name = header.get(\"name\")\n value = header.get(\"value\")\n if name.lower() == 'from':\n # we print the From address\n ret += \"From: \" + value + '\\n'\n if name.lower() == \"to\":\n # we print the To address\n ret += \"To:\" + value + '\\n'\n if name.lower() == \"subject\":\n # make our boolean True, the email has \"subject\"\n has_subject = True\n # make a directory with the name of the subject\n folder_name = clean(value)\n # we will also handle emails with the same subject name\n folder_counter = 0\n while os.path.isdir(folder_name):\n folder_counter += 1\n # we have the same folder name, add a number next to it\n if folder_name[-1].isdigit() and folder_name[-2] == \"_\":\n folder_name = f\"{folder_name[:-2]}_{folder_counter}\"\n elif folder_name[-2:].isdigit() and folder_name[-3] == \"_\":\n folder_name = f\"{folder_name[:-3]}_{folder_counter}\"\n else:\n folder_name = f\"{folder_name}_{folder_counter}\"\n # os.mkdir(folder_name)\n ret += \"Subject: \" + value + '\\n'\n if name.lower() == \"date\":\n # we print the date when the message was sent\n ret += \"Date:\" + value + '\\n'\n if not has_subject:\n # if the email does not have a subject, then make a folder with \"email\" name\n # since folders are created based on subjects\n if not os.path.isdir(folder_name):\n os.mkdir(folder_name)\n ret += parse_parts(service, parts, folder_name, message)\n ret += \"=\"*50 + '\\n'\n return ret\n\ndef mark_as_read(service, query):\n messages_to_mark = search_messages(service, query)\n return service.users().messages().batchModify(\n userId='me',\n body={\n 'ids': [ msg['id'] for msg in messages_to_mark ],\n 'removeLabelIds': ['UNREAD']\n }\n ).execute()\n\n\ndef register():\n creds = None\n # The file token.json stores the user's access and refresh tokens, and is\n # created automatically when the authorization flow completes for the first\n # time.\n if os.path.exists('token.json'):\n creds = Credentials.from_authorized_user_file('token.json', SCOPES)\n # If there are no (valid) credentials available, let the user log in.\n if not creds or not creds.valid:\n if creds and creds.expired and creds.refresh_token:\n creds.refresh(Request())\n else:\n flow = InstalledAppFlow.from_client_secrets_file(\n 'credentials.json', SCOPES)\n creds = flow.run_local_server(port=0)\n # Save the credentials for the next run\n with open('token.json', 'w') as token:\n token.write(creds.to_json())\n\n service = build('gmail', 'v1', credentials=creds)\n return service\n\ndef main():\n \"\"\"Shows basic usage of the Gmail API.\n Lists the user's Gmail labels.\n \"\"\"\n service = register()\n\n # Call the Gmail API\n results = service.users().labels().list(userId='me').execute()\n labels = results.get('labels', [])\n\n if not labels:\n print('No labels found.')\n else:\n print('Labels:')\n for label in labels:\n print(label['name'])\n\n # get emails that match the query you specify\n results = search_messages(service, \"in:UNREAD\")\n # for each email matched, read it (output plain/text to console & save HTML and attachments)\n for msg in results:\n print(read_message(service, msg))\n\n mark_as_read(service, 'in:UNREAD')\n\n\n\nif __name__ == '__main__':\n main()", "repo_name": "misprit7/messenger-spam", "sub_path": "gmail.py", "file_name": "gmail.py", "file_ext": "py", "file_size_in_byte": 9239, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "base64.urlsafe_b64decode", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path", "line_number": 160, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 182, "usage_type": "call"}, {"api_name": "os.path", "line_number": 182, "usage_type": "attribute"}, {"api_name": "google.oauth2.credentials.Credentials.from_authorized_user_file", "line_number": 183, "usage_type": "call"}, {"api_name": "google.oauth2.credentials.Credentials", "line_number": 183, "usage_type": "name"}, {"api_name": "google.auth.transport.requests.Request", "line_number": 187, "usage_type": "call"}, {"api_name": "google_auth_oauthlib.flow.InstalledAppFlow.from_client_secrets_file", "line_number": 189, "usage_type": "call"}, {"api_name": "google_auth_oauthlib.flow.InstalledAppFlow", "line_number": 189, "usage_type": "name"}, {"api_name": "googleapiclient.discovery.build", "line_number": 196, "usage_type": "call"}]}
+{"seq_id": "73112115583", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# ## Параллельная обработка изображений в многозадачной среде\n\n# In[13]:\n\n\nimport os\nfrom concurrent.futures import ThreadPoolExecutor\nfrom PIL import Image, ImageEnhance, ImageOps\n\n\n# ### Фильтры\n\n# 1. Фильтр сепии\n\n# In[14]:\n\n\ndef sepia(inp, out_path):\n try:\n image = Image.open(inp)\n sepia_image = ImageOps.colorize(image.convert('L'), '#704214', '#C0C080')\n sepia_image.save(out_path)\n except Exception as e:\n print(f'К файлу {inp} невозможно применить фильтр в виду ошибки: {str(e)}')\n\n\n# 2. Фильтр резкости\n\n# In[15]:\n\n\ndef sharpen(inp, out_path):\n try:\n image = Image.open(inp)\n sharpened_image = ImageEnhance.Sharpness(image).enhance(2.0)\n sharpened_image.save(out_path)\n except Exception as e:\n print(f'К файлу {inp} невозможно применить фильтр в виду ошибки: {str(e)}')\n\n\n# 3. Фильтр уменьшения размера\n\n# In[16]:\n\n\ndef resize(inp, out_path):\n try:\n image = Image.open(inp)\n resized_image = image.resize((300, 300))\n resized_image.save(out_path)\n except Exception as e:\n print(f'К файлу {inp} невозможно применить фильтр в виду ошибки: {str(e)}')\n\n\n# ### Обработка изображений\n\n# In[18]:\n\n\ndef process(inp, out_folder):\n image_name = os.path.basename(inp)\n filename, _ = os.path.splitext(image_name)\n\n sharpen(inp, os.path.join(out_folder, f'{filename}_f1.jpg'))\n sepia(inp, os.path.join(out_folder, f'{filename}_f2.jpg'))\n resize(inp, os.path.join(out_folder, f'{filename}_f3.jpg'))\n\ndef parallel_images(image_folder, out_folder):\n if not os.path.exists(out_folder):\n os.makedirs(out_folder)\n tasks = []\n for image_name in os.listdir(image_folder):\n image_path = os.path.join(image_folder, image_name)\n if os.path.isfile(image_path) and image_name.lower().endswith(('.png', '.jpg', '.jpeg')):\n tasks.append((image_path, out_folder))\n\n with ThreadPoolExecutor() as executor:\n executor.map(lambda args: process(*args), tasks)\n \nimage_folder = r'C:\\Users\\sibfl\\OneDrive\\photo'\nout_folder = r'C:\\Users\\sibfl\\OneDrive\\photo\\new'\n\nparallel_images(image_folder, out_folder)\n\n", "repo_name": "LisaNota/Parallel_processing", "sub_path": "parallel.py", "file_name": "parallel.py", "file_ext": "py", "file_size_in_byte": 2445, "program_lang": "python", "lang": "ru", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "PIL.Image.open", "line_number": 23, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 23, "usage_type": "name"}, {"api_name": "PIL.ImageOps.colorize", "line_number": 24, "usage_type": "call"}, {"api_name": "PIL.ImageOps", "line_number": 24, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 37, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 37, "usage_type": "name"}, {"api_name": "PIL.ImageEnhance.Sharpness", "line_number": 38, "usage_type": "call"}, {"api_name": "PIL.ImageEnhance", "line_number": 38, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 51, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 51, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path", "line_number": 72, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 73, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 80, "usage_type": "call"}]}
+{"seq_id": "31132787966", "text": "import cv2 as cv\nimport numpy as np\nfrom matplotlib import pyplot as plt\n\ndef measure_object(image):\n gray = cv.cvtColor(image,cv.COLOR_BGR2GRAY)\n ret, binary = cv.threshold(gray,0,255,cv.THRESH_BINARY_INV | cv.THRESH_OTSU)\n contours,hericahy = cv.findContours(binary,cv.RETR_EXTERNAL,cv.CHAIN_APPROX_SIMPLE)\n cv.imshow(\"binary image\",binary)\n for i,contour in enumerate(contours):\n arae = cv.contourArea(contour)\n x,y,w,h = cv.boundingRect(contour)\n rate = min(w,h)/max(w,h)\n #print(\"rectang rate : %s\"%rate)\n #输出宽高比\n #外接矩形的大小\n mm = cv.moments(contour)\n #求取几何矩\n print(type(mm))\n #是一个字典类型\n cx = mm['m10']/mm['m00']\n cy = mm['m01']/mm['m00']\n #即可得到矩形的中芯位置\n #cv.circle(image,(np.int(cx),np.int(cy)),3,(0,255,255),-1)\n #cv.rectangle(image,(x,y),(x+w,y+h),(0,0,255),2)\n #绘制外接矩形\n #print(\"contour arae :%s\"%arae)\n approxCurve = cv.approxPolyDP(contour,4,True)\n if approxCurve.shape[0] ==4:\n cv.drawContours(image,contours,i,(0,255,0),2)\n cv.imshow('measure',image)\n\nscr = cv.imread(\"D:/opencvtupian/12.jpg\")\n#读取一张图片\ncv.namedWindow(\"input image\",cv.WINDOW_AUTOSIZE)\n#创建GUI显示图片\ncv.imshow(\"input image\",scr)\nmeasure_object(scr)\ncv.waitKey(0)\ncv.destroyAllWindows()", "repo_name": "1YangTao1/YT_study-note", "sub_path": "opencv/tutorial_20对象测量.py", "file_name": "tutorial_20对象测量.py", "file_ext": "py", "file_size_in_byte": 1424, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "cv2.cvtColor", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 6, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 7, "usage_type": "attribute"}, {"api_name": "cv2.THRESH_OTSU", "line_number": 7, "usage_type": "attribute"}, {"api_name": "cv2.findContours", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 8, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 8, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.contourArea", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.boundingRect", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.moments", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.approxPolyDP", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.drawContours", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.WINDOW_AUTOSIZE", "line_number": 35, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 40, "usage_type": "call"}]}
+{"seq_id": "30655850968", "text": "\"\"\"\nSubmodule for creating Iogas XML templates from matplotlib axes.\n\"\"\"\nimport numpy as np\nfrom lxml.etree import ElementTree\nfrom pyrolite.util.text import int_to_alpha\nfrom pyrolite.util.plot import get_contour_paths\nfrom pyrolite.util.meta import subkwargs\n\nfrom ..util.xml import prettify_xml\nfrom .common import Poly, Polygon, RegionPolygon\nfrom . import freediagram\nfrom . import geochemdiagram\nfrom . import freeternary\n\nfrom pyrolite.geochem.ind import common_elements, common_oxides\n\n__els__ = common_elements(as_set=True)\n__ox__ = common_oxides(as_set=True)\n__chem__ = __els__ | __ox__\n\n\ndef contours_to_XYDiagram(\n ax,\n xvar=\"X\",\n yvar=\"Y\",\n logscalex=False,\n logscaley=False,\n logxdata=False,\n logydata=False,\n filename=\"element.xml\",\n contournames=None,\n allow_free_func=True,\n resolution=100,\n description_prefix=\"\",\n encoding=\"utf-8\",\n):\n \"\"\"\n Take the contour lines from an axis and convert them to an iogas xml diagram\n template.\n\n Parameters\n ------------\n\n Note\n ------\n\n The polygons need not return to the same point.\n\n If the diagram is for log, the coordinates need to be log.\n \"\"\"\n filename = str(filename)\n if all([i in __chem__ for i in xvar.split(\"/\")]) and all(\n [i in __chem__ for i in yvar.split(\"/\")]\n ):\n dg = geochemdiagram.XYDiagram\n poly = Poly\n else:\n dg = freediagram.XYDiagram\n if allow_free_func and any([\"/\" in v for v in [xvar, yvar]]):\n poly = RegionPolygon\n else:\n poly = Poly\n sk = subkwargs(\n dict(\n logscalex=logscalex,\n logscaley=logscaley,\n logxdata=logxdata,\n logydata=logydata,\n allow_free_func=allow_free_func,\n ),\n dg,\n )\n diagram = dg(xvar, yvar, **sk)\n cpaths, cnames, styles = get_contour_paths(ax, resolution=resolution)\n if contournames is not None:\n assert len(contournames) == len(cpaths)\n cnames = contournames\n # create contours\n contours = []\n for ix, (p, name, sty) in enumerate(zip(cpaths, cnames, styles)):\n for six, subpath in enumerate(p):\n if logxdata:\n subpath[0] = np.log(subpath[0])\n if logydata:\n subpath[1] = np.log(subpath[1])\n if len(p) != 1:\n suffix = \"-\" + int_to_alpha(six)\n else:\n suffix = \"\"\n cname = [\"Countour-{}\".format(name), \"Countour-{}\".format(ix)][\n name is None\n ] + suffix\n c = poly(\n subpath,\n color=sty[\"color\"],\n name=str(name),\n description=description_prefix,\n )\n contours.append(c)\n diagram.extend(contours)\n version_poly = Polygon(\n name=\"_ v6.1 required to open diagram _\",\n visible=\"true\",\n xpoints=[30, 200, 200, 30],\n ypoints=[30, 30, 40, 40],\n )\n\n diagram.extend([version_poly])\n ElementTree(diagram).write(filename, method=\"xml\", encoding=encoding)\n return prettify_xml(diagram)\n\n\ndef contours_to_FreeTernaryDiagram():\n pass\n", "repo_name": "morganjwilliams/gas-templates", "sub_path": "pyogas/writer/mpl2iogas.py", "file_name": "mpl2iogas.py", "file_ext": "py", "file_size_in_byte": 3186, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "pyrolite.geochem.ind.common_elements", "line_number": 18, "usage_type": "call"}, {"api_name": "pyrolite.geochem.ind.common_oxides", "line_number": 19, "usage_type": "call"}, {"api_name": "common.Poly", "line_number": 57, "usage_type": "name"}, {"api_name": "common.RegionPolygon", "line_number": 61, "usage_type": "name"}, {"api_name": "common.Poly", "line_number": 63, "usage_type": "name"}, {"api_name": "pyrolite.util.meta.subkwargs", "line_number": 64, "usage_type": "call"}, {"api_name": "pyrolite.util.plot.get_contour_paths", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 86, "usage_type": "call"}, {"api_name": "pyrolite.util.text.int_to_alpha", "line_number": 88, "usage_type": "call"}, {"api_name": "common.Polygon", "line_number": 102, "usage_type": "call"}, {"api_name": "lxml.etree.ElementTree", "line_number": 110, "usage_type": "call"}, {"api_name": "util.xml.prettify_xml", "line_number": 111, "usage_type": "call"}]}
+{"seq_id": "73775367741", "text": "# Imports\nimport itertools\nimport tarfile\nimport pandas\nimport pickle\n\n# Sklearn imports\nimport nltk\nimport nltk.collocations\n\n# LexNLP imports\nimport lexnlp.nlp.en.segments.sentences\nimport lexnlp.nlp.en.tokens\n\n# Setup default path for documents\ntar_input_path = \"agreements-text.tar.gz\"\n\n# Store tokens\ntokens = []\nbigrams = []\n\n# Set sample size in docs\nnum_samples = 1000\n\n# Set parameters\ncutoff_amount = 0.5\ntop_n_list = [100, 1000, 10000]\nmin_freq = 100\n\nif __name__ == \"__main__\":\n # Set number of samples\n num_samples = 10000\n\n # Initialize frequency distributions\n token_fd = nltk.FreqDist()\n wildcard_fd = nltk.FreqDist()\n bigram_fd = nltk.FreqDist()\n trigram_fd = nltk.FreqDist()\n\n # Iterate through files\n with tarfile.open(tar_input_path, \"r:gz\") as corpus_tar_file:\n # Get list of files\n member_list = corpus_tar_file.getmembers()[0:num_samples]\n num_members = len(member_list)\n\n # Iterate through all\n for i, tar_member in enumerate(member_list):\n # Output\n if i % 100 == 0:\n print((tar_input_path, i, float(i)/num_members * 100., tar_member.name, len(member_list)))\n\n # Read buffer\n member_file = corpus_tar_file.extractfile(tar_member.name)\n if member_file is None:\n print((tar_input_path, tar_member.name, \"invalid file\"))\n continue\n member_buffer = member_file.read().decode(\"utf-8\")\n if len(member_buffer.strip()) == 0:\n continue\n\n # Parse into sentence data\n try:\n for sentence in lexnlp.nlp.en.segments.sentences.get_sentence_list(member_buffer):\n sentence_tokens = lexnlp.nlp.en.tokens.get_token_list(sentence, lowercase=True)\n sentence_tokens = [t for t in sentence_tokens if t.isalpha()]\n\n for window in nltk.ngrams(sentence_tokens, 3, pad_right=True):\n w1 = window[0]\n for w2, w3 in itertools.combinations(window[1:], 2):\n token_fd[w1] += 1\n if w2 is None:\n continue\n bigram_fd[(w1, w2)] += 1\n if w3 is None:\n continue\n wildcard_fd[(w1, w3)] += 1\n trigram_fd[(w1, w2, w3)] += 1\n \n except Exception as e:\n print(e)\n continue\n \n # Create measure objects\n bigram_measures = nltk.collocations.BigramAssocMeasures()\n trigram_measures = nltk.collocations.TrigramAssocMeasures()\n\n for n in top_n_list:\n # Apply filter and output\n bigram_finder = nltk.collocations.BigramCollocationFinder(token_fd, bigram_fd)\n bigram_finder.apply_freq_filter(min_freq)\n bigram_collocations = list(bigram_finder.nbest(bigram_measures.pmi, n))\n print((n, len(bigram_collocations), bigram_collocations[-1]))\n\n # Save the tokenizer\n with open(\"collocation_bigrams_{0}.pickle\".format(n), \"wb\") as out_file:\n pickle.dump(bigram_collocations, out_file)\n\n # Apply filter and output\n trigram_finder = nltk.collocations.TrigramCollocationFinder(token_fd, bigram_fd, wildcard_fd, trigram_fd)\n trigram_finder.apply_freq_filter(min_freq)\n trigram_collocations = list(trigram_finder.nbest(trigram_measures.pmi, n))\n print((n, len(trigram_collocations), trigram_collocations[-1]))\n\n # Save the tokenizer\n with open(\"collocation_trigrams_{0}.pickle\".format(n), \"wb\") as out_file:\n pickle.dump(trigram_collocations, out_file)", "repo_name": "LexPredict/lexpredict-lexnlp", "sub_path": "notebooks/nlp/en/build_collocation_pickle.py", "file_name": "build_collocation_pickle.py", "file_ext": "py", "file_size_in_byte": 3922, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 654, "dataset": "github-code", "pt": "24", "api": [{"api_name": "nltk.FreqDist", "line_number": 35, "usage_type": "call"}, {"api_name": "nltk.FreqDist", "line_number": 36, "usage_type": "call"}, {"api_name": "nltk.FreqDist", "line_number": 37, "usage_type": "call"}, {"api_name": "nltk.FreqDist", "line_number": 38, "usage_type": "call"}, {"api_name": "tarfile.open", "line_number": 41, "usage_type": "call"}, {"api_name": "lexnlp.nlp.en.segments.sentences.nlp.en.segments.sentences.get_sentence_list", "line_number": 63, "usage_type": "call"}, {"api_name": "lexnlp.nlp.en.segments.sentences.nlp", "line_number": 63, "usage_type": "attribute"}, {"api_name": "lexnlp.nlp.en.segments.sentences", "line_number": 63, "usage_type": "name"}, {"api_name": "lexnlp.nlp.en.segments.sentences.nlp.en.tokens.get_token_list", "line_number": 64, "usage_type": "call"}, {"api_name": "lexnlp.nlp.en.segments.sentences.nlp", "line_number": 64, "usage_type": "attribute"}, {"api_name": "lexnlp.nlp.en.segments.sentences", "line_number": 64, "usage_type": "name"}, {"api_name": "nltk.ngrams", "line_number": 67, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 69, "usage_type": "call"}, {"api_name": "nltk.collocations.BigramAssocMeasures", "line_number": 84, "usage_type": "call"}, {"api_name": "nltk.collocations", "line_number": 84, "usage_type": "attribute"}, {"api_name": "nltk.collocations.TrigramAssocMeasures", "line_number": 85, "usage_type": "call"}, {"api_name": "nltk.collocations", "line_number": 85, "usage_type": "attribute"}, {"api_name": "nltk.collocations.BigramCollocationFinder", "line_number": 89, "usage_type": "call"}, {"api_name": "nltk.collocations", "line_number": 89, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 96, "usage_type": "call"}, {"api_name": "nltk.collocations.TrigramCollocationFinder", "line_number": 99, "usage_type": "call"}, {"api_name": "nltk.collocations", "line_number": 99, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 106, "usage_type": "call"}]}
+{"seq_id": "6479911924", "text": "import openpyxl as xl\nfrom openpyxl.chart import BarChart, Reference\n\n\ndef process_workbook(filename):\n workbook = xl.load_workbook(filename) # get Excel file\n sheet = workbook['Sheet1'] # get sheet from transactions.xlsx\n\n cell = sheet['a1']\n print(cell) # \n\n # the same as\n cell = sheet.cell(1, 1)\n print(cell) # \n print(cell.value) # transaction_id\n\n print(sheet.max_row) # 4\n\n for row in range(2, sheet.max_row + 1):\n cell = sheet.cell(row, 3)\n corrected_price = cell.value * 0.9 # calc new value\n corrected_price_cell = sheet.cell(row, 4) # get new cell in 4th column\n corrected_price_cell.value = corrected_price # set value to this cell\n\n values = Reference( # collect values\n sheet,\n min_row=2,\n max_row=sheet.max_row,\n min_col=4,\n max_col=4\n )\n chart = BarChart() # create chart\n chart.add_data(values) # added data to the chart\n sheet.add_chart(chart, 'e2') # added chart to the sheet\n\n workbook.save(filename) # save changes into file\n\n\nprocess_workbook('transactions.xlsx')\n", "repo_name": "vlfed-mage/python-excel-spreadsheets", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1147, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "openpyxl.load_workbook", "line_number": 6, "usage_type": "call"}, {"api_name": "openpyxl.chart.Reference", "line_number": 25, "usage_type": "call"}, {"api_name": "openpyxl.chart.BarChart", "line_number": 32, "usage_type": "call"}]}
+{"seq_id": "19632231149", "text": "'''\n Code adapted from https://gist.github.com/rusty1s/159eeff0d95e0786d220a164c6edd021\n'''\n\nimport os\nimport os.path as osp\nfrom six.moves import urllib\nimport errno\nimport tarfile\n\ndef makedirs(path):\n try:\n os.makedirs(osp.expanduser(osp.normpath(path)))\n except OSError as e:\n if e.errno != errno.EEXIST and osp.isdir(path):\n raise e\n\n\ndef download_url(url, folder, log=True):\n print('Downloading', url)\n makedirs(folder)\n\n data = urllib.request.urlopen(url)\n filename = url.rpartition('/')[2]\n path = osp.join(folder, filename)\n\n with open(path, 'wb') as f:\n f.write(data.read())\n\n return path\n\n\ndef extract_tar(path, folder, mode='r:gz', log=True):\n print('Extracting', path)\n with tarfile.open(path, mode) as f:\n f.extractall(folder)\n\n\ncwd = os.getcwd()\npath = os.path.join( cwd, 'datasets/QM9/qm9' )\nurl = 'http://deepchem.io.s3-website-us-west-1.amazonaws.com/' \\\n 'datasets/gdb9.tar.gz'\n\nfile_path = download_url(url, path)\nextract_tar(file_path, path, mode='r')", "repo_name": "olsson-group/transBG", "sub_path": "preprocessing/download_qm9.py", "file_name": "download_qm9.py", "file_ext": "py", "file_size_in_byte": 1052, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "os.makedirs", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "name"}, {"api_name": "os.path.normpath", "line_number": 13, "usage_type": "call"}, {"api_name": "errno.EEXIST", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "name"}, {"api_name": "six.moves.urllib.request.urlopen", "line_number": 23, "usage_type": "call"}, {"api_name": "six.moves.urllib.request", "line_number": 23, "usage_type": "attribute"}, {"api_name": "six.moves.urllib", "line_number": 23, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "name"}, {"api_name": "tarfile.open", "line_number": 35, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}]}
+{"seq_id": "42118414507", "text": "import glob\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom prettytable import PrettyTable\n\ndef sumzip(*items):\n return [sum(values) for values in zip(*items)]\n\nfiles = sorted(glob.glob('results/*.csv'))\nLOC = []\nNOFC = []\nLOF = []\nANDAVG = [] \nANDSTDEV = []\t\nSDEGMEAN = \t[]\nSDEGSTD = []\nTDEGMEAN = []\nTDEGSTD = []\nHOM\t = []\nHET = []\nHOHE = []\nNOFPFCMEAN = []\nNOFPFCSTD = []\nGRANGL = []\nGRANFL = []\nGRANBL = []\nGRANSL = []\nGRANEL = []\nGRANML = []\nGRANERR = []\nNDMAX = []\nFileCOUNT = []\n\n# Skip 2.2.11 Version\nfor file in files:\n\n\tlines = open(file, 'rt').readlines()\n\tl2 = lines[-1].split(',')\n\tdata2 = l2[2:13] + l2[21:23]\n\tdata2 = [d.rstrip() for d in data2]\n\tNOFC.append(float(data2[0]))\n\tLOF.append(float(data2[1]))\n\tANDAVG.append(float(data2[2]))\n\tANDSTDEV.append(float(data2[3]))\n\tSDEGMEAN.append(float(data2[4]))\n\tSDEGSTD.append(float(data2[5]))\n\tTDEGMEAN.append(float(data2[6]))\n\tTDEGSTD.append(float(data2[7]))\n\tHOM.append(float(data2[8]))\n\tHET.append(float(data2[9]))\n\tHOHE.append(float(data2[10]))\n\tNOFPFCMEAN.append(float(data2[11]))\n\tNOFPFCSTD.append(float(data2[12]))\n\tLOCT = []\n\tGRANGLT = []\n\tGRANFLT = []\n\tGRANBLT = []\n\tGRANSLT = []\n\tGRANELT = []\n\tGRANMLT = []\n\tGRANERRT = []\n\tNDMAXT = []\n\tFileCOUNTT = 0\n\tVP = []\n\t\n\tfor line in lines:\n\t\tif line.count(',') > 5 and line.split(',')[0][0] == '/':\n\t\t\tFileCOUNTT = FileCOUNTT + 1\n\t\t\tLOCT.append(float(line.split(',')[1]))\n\t\t\tGRANGLT.append(float(line.split(',')[13]))\n\t\t\tGRANFLT.append(float(line.split(',')[14]))\n\t\t\tGRANBLT.append(float(line.split(',')[15]))\n\t\t\tGRANSLT.append(float(line.split(',')[16]))\n\t\t\tGRANELT.append(float(line.split(',')[17]))\n\t\t\tGRANMLT.append(float(line.split(',')[18]))\n\t\t\tGRANERRT.append(float(line.split(',')[19]))\n\t\t\tNDMAXT.append(float(line.split(',')[20]))\n\tLOC.append(sum(LOCT))\n\tGRANGL.append(sum(GRANGLT))\n\tGRANFL.append(sum(GRANFLT))\n\tGRANBL.append(sum(GRANBLT))\n\tGRANSL.append(sum(GRANSLT))\n\tGRANEL.append(sum(GRANELT))\n\tGRANML.append(sum(GRANMLT))\n\tGRANERR.append(sum(GRANERRT))\n\tNDMAX.append(max(NDMAXT))\n\tFileCOUNT.append(FileCOUNTT)\n\t\n\t\n# Compare Originial results to ours\nt = PrettyTable(['name', 'LOC', 'NOFC', 'LOF', 'AND-MEAN', 'AND-STD', 'SD-MEAN', 'SD-STD', 'TD-MEAN', 'TD-STD'])\nt.add_row(['original', 214607, 1158, 40075, 1.18, 0.18, 5.58, 15.35, 1.73, 1.35])\nt.add_row(['ours', int(LOC[0]), int(NOFC[0]), int(LOF[0]), round(ANDAVG[0], 2), round(ANDSTDEV[0], 2), round(SDEGMEAN[0], 2), round(SDEGSTD[0], 2), round(TDEGMEAN[0], 2), round(TDEGMEAN[0], 2)])\n\nprint(t)\n\nt = PrettyTable(['name', 'HOM', 'HET', 'HOHE', 'GL', 'FL', 'BL', 'SL', 'EL', 'ML'])\nt.add_row(['original', 134, 2041, 98, 2015, 2186, 828, 43, 35, 13])\nt.add_row(['ours', int(HOM[0]), int(HET[0]), int(HOHE[0]), int(GRANGL[0]), int(GRANFL[0]), int(GRANBL[0]), int(GRANSL[0]), int(GRANEL[0]), int(GRANML[0])])\n\nprint(t)\n\t\n# Calculate some extra metrics\n\nPLOF = np.divide(LOF, LOC)\n\nVP = map(sum, zip(GRANGL, GRANFL, GRANBL, GRANSL, GRANEL, GRANML, GRANERR))\nLOFperVP = np.divide(LOF, VP)\n\nSD_FILE = np.divide(np.multiply(SDEGMEAN, NOFC), FileCOUNT)\nTD_FILE = np.divide(np.multiply(TDEGMEAN, NOFC), FileCOUNT)\n\n\nversions = ['2.4.25','2.4.26','2.4.27','2.4.28','2.4.29']\nnumberVersions = range(len(versions))\nwidthBar = 0.35\n\n# Global Metrics\n\nfig, ax1 = plt.subplots()\nax1.plot(LOC[1:], 'r-', label='LOC')\nax1.set_ylabel('LOC', color='r')\nax2 = ax1.twinx()\nax2.plot(LOF[1:], 'b-', label='LOF')\nax2.set_ylabel('LOF', color='b')\nfig.tight_layout()\nplt.xticks(numberVersions, versions)\nplt.title('LOC vs LOF')\nplt.show()\n\nplt.plot(PLOF[1:])\nplt.title('PLOF (LOF / LOC)')\nplt.xticks(numberVersions, versions)\nplt.show()\n\nplt.plot(NOFC[1:])\nplt.title('NOFC')\nplt.xticks(numberVersions, versions)\nplt.show()\n\nplt.plot(VP[1:])\nplt.title('VP (number of #ifdefs out of GRAN)')\nplt.xticks(numberVersions, versions)\nplt.show()\n\n\nplt.plot(LOFperVP[1:])\nplt.title('LOF per VP')\nplt.xticks(numberVersions, versions)\nplt.show()\n\n\n# SD, TD Metric\n\nplt.errorbar(numberVersions, SDEGMEAN[1:], SDEGSTD[1:], linestyle='None', marker='s')\nx1,x2,y1,y2 = plt.axis()\nplt.axis((x1,x2,-10,20))\nplt.title('SD Mean + SD')\nplt.xticks(numberVersions, versions)\nplt.show()\n\nplt.plot(SDEGMEAN[1:])\nplt.title('SDEGMEAN')\nplt.xticks(numberVersions, versions)\nplt.show()\n\nplt.plot(SDEGSTD[1:])\nplt.title('SDEGSTD')\nplt.xticks(numberVersions, versions)\nplt.show()\n\nplt.plot(SD_FILE[1:])\nplt.title('SD per File')\nplt.xticks(numberVersions, versions)\nplt.show()\n\nplt.errorbar(numberVersions, TDEGMEAN[1:], TDEGSTD[1:], linestyle='None', marker='s')\nx1,x2,y1,y2 = plt.axis()\nplt.axis((x1,x2,-4,4))\nplt.title('TD Mean + SD')\nplt.xticks(numberVersions, versions)\nplt.show()\n\nplt.plot(TDEGMEAN[1:])\nplt.title('TDEGMEAN')\nplt.xticks(numberVersions, versions)\nplt.show()\n\nplt.plot(TDEGSTD[1:])\nplt.title('TDEGSTD')\nplt.xticks(numberVersions, versions)\nplt.show()\n\nplt.plot(TD_FILE[1:])\nplt.title('TD per File')\nplt.xticks(numberVersions, versions)\nplt.show()\n\n# TYPE\n\np1 = plt.bar(numberVersions, HOM[1:], widthBar, color='#d62728')\np2 = plt.bar(numberVersions, HOHE[1:], widthBar, bottom=HOM[1:])\np3 = plt.bar(numberVersions, HET[1:], widthBar, bottom=list(map(sum, zip(HOM[1:], HOHE[1:]))))\n\nplt.title('TYPE')\nplt.legend((p1[0], p2[0], p3[0]), ('HOM', 'HOHE', 'HET'))\nplt.xticks(numberVersions, versions)\nplt.show()\n\n# GRAN\n\np1 = plt.bar(numberVersions, GRANGL[1:], widthBar, color='#d62728')\np2 = plt.bar(numberVersions, GRANFL[1:], widthBar, bottom=GRANGL[1:])\np3 = plt.bar(numberVersions, GRANBL[1:], widthBar, bottom=list(map(sum, zip(GRANGL[1:], GRANFL[1:]))))\np4 = plt.bar(numberVersions, GRANSL[1:], widthBar, bottom=list(map(sum, zip(GRANGL[1:], GRANFL[1:], GRANBL[1:]))))\np5 = plt.bar(numberVersions, GRANEL[1:], widthBar, bottom=list(map(sum, zip(GRANGL[1:], GRANFL[1:], GRANBL[1:], GRANSL[1:]))))\np6 = plt.bar(numberVersions, GRANML[1:], widthBar, bottom=list(map(sum, zip(GRANGL[1:], GRANFL[1:], GRANBL[1:], GRANSL[1:], GRANEL[1:]))))\np7 = plt.bar(numberVersions, GRANERR[1:], widthBar, bottom=list(map(sum, zip(GRANGL[1:], GRANFL[1:], GRANBL[1:], GRANSL[1:], GRANEL[1:], GRANML[1:]))))\n\nplt.title('GRAN')\nplt.legend((p1[0], p2[0], p3[0], p4[0], p5[0], p6[0], p7[0]), ('GL', 'FL', 'BL', 'SL', 'EL', 'ML', 'ERR'))\nplt.xticks(numberVersions, versions)\nplt.show()\n\n# AND\n\nplt.errorbar(numberVersions, ANDAVG[1:], ANDSTDEV[1:], linestyle='None', marker='s')\nplt.title('AND Mean + SD')\nx1,x2,y1,y2 = plt.axis()\nplt.axis((x1,x2,0,4))\nplt.xticks(numberVersions, versions)\nplt.show()\n\n\nplt.plot(ANDAVG[1:])\nplt.title('ANDAVG')\nplt.xticks(numberVersions, versions)\nplt.show()\n\nplt.plot(ANDSTDEV[1:])\nplt.title('ANDSTDEV')\nplt.xticks(numberVersions, versions)\nplt.show()\n\nplt.plot(NDMAX[1:])\nplt.title('NDMAX')\nplt.xticks(numberVersions, versions)\nplt.show()\n\n", "repo_name": "jodokae/cmput663-cpp-replic", "sub_path": "plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 6707, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "glob.glob", "line_number": 9, "usage_type": "call"}, {"api_name": "prettytable.PrettyTable", "line_number": 91, "usage_type": "call"}, {"api_name": "prettytable.PrettyTable", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 143, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 168, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 169, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 170, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 170, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 172, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 172, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 173, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 177, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 181, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 187, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 192, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 194, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 197, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 201, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 201, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 202, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 206, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 208, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 212, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 212, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 213, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 213, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 214, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 214, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 215, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 215, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 216, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 217, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 218, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 218, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 220, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 222, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 223, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 227, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 228, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 229, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 229, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 230, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 231, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 231, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 232, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 235, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 235, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 236, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 237, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 238, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 238, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 240, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 240, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 241, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 241, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 242, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 243, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 245, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 245, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 246, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 248, "usage_type": "name"}]}
+{"seq_id": "2831612790", "text": "\nimport numpy as np\nfrom matplotlib import pyplot as plt\n\n\nif __name__ == '__main__':\n x = np.arange(0, 11) # x轴数据\n # y = x * x + 5 # 函数关系\n y = x * x # 函数关系\n # plt.title(\"y=x*x+5\") # 图像标题\n plt.title(\"y=x*x\") # 图像标题\n plt.xlabel(\"x\") # x轴标签\n plt.ylabel(\"y\") # y轴标签\n plt.plot(x, y) # 生成图像\n plt.show() # 显示图像", "repo_name": "ArtistRuan/webspide", "sub_path": "csdnParactise/6_math_func/math_funct.py", "file_name": "math_funct.py", "file_ext": "py", "file_size_in_byte": 405, "program_lang": "python", "lang": "zh", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "numpy.arange", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}]}
+{"seq_id": "6905910909", "text": "from django import forms\nfrom .models import ChatMessage\n\nclass ChatMessageForm(forms.ModelForm):\n class Meta:\n model = ChatMessage\n fields = ('content', )\n widgets = {\n 'content': forms.Textarea(attrs={\n 'class': 'w-full py-4 px-6 rounded-xl border'\n })\n }", "repo_name": "TharunKumarReddyPolu/Graduate-Market-Place", "sub_path": "gmp_env/gmp/chat/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 325, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 12, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.forms.ModelForm", "line_number": 4, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 4, "usage_type": "name"}, {"api_name": "models.ChatMessage", "line_number": 6, "usage_type": "name"}, {"api_name": "django.forms.Textarea", "line_number": 9, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 9, "usage_type": "name"}]}
+{"seq_id": "75238501182", "text": "import csv\n\nfrom stop_words import StopWords\n\nclass GetTranscriptions:\n def __init__(self):\n self.stop_words = StopWords()\n self.transcriptions = []\n self.transcriptions_NOstopwords = []\n\n # ======================================================================\n\n '''\n * Agrega mas stop words desde un archivo\n\n Input:\n self.stop_words.stop_words = [de, su]\n path_stopwords = ./stopwords.txt\n \n stopwords.txt:\n el\n la\n ...\n\n Output:\n self.stop_words.stop_words = [de, su, el, la, ...]\n '''\n\n def AddStopWords(self, path_stopwords):\n self.stop_words.ReadFile(path_stopwords)\n\n # ======================================================================\n\n '''\n Input:\n sentence = ' el perro come su comida'\n \n Output:\n return 'el perro come su comida'\n '''\n\n def DeleteExtraSeparations(self, sentence):\n words = sentence.split()\n separator = ' '\n\n return separator.join(words)\n\n # ======================================================================\n\n '''\n Input:\n path_data = './data.csv'\n \n data.csv:\n ,conid,sequence,etiqueta,transcription\n 0,31321,0,label1,el perro come su comida\n 1,15216,1,label2,el gato duerme\n \n Output:\n self.transcriptions = ['el perro come su comida', 'el gato duerme']\n self.transcriptions_NOstopwords = ['perro come comida', 'gato duerme']\n '''\n\n def ReadData(self, path_data):\n with open(path_data, 'r') as file_csv:\n data_csv = csv.reader(file_csv, delimiter=',')\n\n count_row = 0\n index_transcription = 4\n \n for row in data_csv:\n if count_row != 0:\n transcription = row[index_transcription]\n self.transcriptions.append(transcription)\n\n transcription_NOstopwords = self.stop_words.DeleteStopWords(row[index_transcription])\n self.transcriptions_NOstopwords.append(transcription_NOstopwords)\n\n count_row += 1\n\n # ======================================================================\n\n def WriteFile(self, path, transcriptions):\n with open(path, 'w') as file_txt:\n for transcription in transcriptions:\n transcription = self.DeleteExtraSeparations(transcription)\n\n if transcription != '':\n file_txt.write(transcription + '\\n')\n\n # ======================================================================\n\n def ExportData(self):\n name_transcriptions = 'transcriptions.txt'\n path_transcriptions = './data/' + name_transcriptions\n self.WriteFile(path_transcriptions, self.transcriptions)\n\n name_transcriptions_NOstopwords = 'transcriptions_NOstopwords.txt'\n path_transcriptions_NOstopwords = './data/' + name_transcriptions_NOstopwords\n self.WriteFile(path_transcriptions_NOstopwords, self.transcriptions_NOstopwords)\n", "repo_name": "JonMollUCSP/AT-BiLSTM", "sub_path": "utilities/train_fasttext/get_transcriptions.py", "file_name": "get_transcriptions.py", "file_ext": "py", "file_size_in_byte": 2822, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "stop_words.StopWords", "line_number": 7, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 66, "usage_type": "call"}]}
+{"seq_id": "27979244570", "text": "import argparse\nimport os\nimport numpy as np\nimport tensorflow as tf\nfrom matplotlib import pyplot as plt\nfrom mpl_toolkits.axes_grid1 import make_axes_locatable\nfrom easydict import EasyDict as edict\nfrom PIL import Image\nimport cv2\nimport time\nfrom . import models\n\ndef load_model():\n height = 228\n width = 304\n channels = 3\n batch_size = 1\n # Create a placeholder for the input image\n\n input_node = tf.placeholder(tf.float32, shape=(None, height, width, channels))\n # Construct the network\n\n net = models.ResNet50UpProj({'data': input_node}, batch_size, 1, False)\n return net, input_node\n\ndef plot_image(img_ori, pred):\n fig = plt.figure()\n plt.subplot(121)\n ii = plt.imshow(img_ori, interpolation='nearest')\n\n plt.subplot(122)\n\n ax = plt.gca()\n ii = ax.imshow(pred, interpolation='nearest')\n divider = make_axes_locatable(ax)\n cax = divider.append_axes(\"right\", size=\"5%\", pad=0.05)\n plt.colorbar(ii, cax=cax)\n plt.show()\n \ndef predict_image(model_data_path, image_path, net, input_node):\n\n \n # Default input size\n height = 228\n width = 304\n \n # Read image\n img = Image.open(image_path)\n img = img.resize([width,height], Image.ANTIALIAS)\n img_ori = img.copy() \n img = np.array(img).astype('float32')\n img = np.expand_dims(np.asarray(img), axis = 0)\n\n with tf.Session() as sess:\n\n # Load the converted parameters\n print('Loading the model')\n\n # Use to load from ckpt file\n saver = tf.train.Saver() \n saver.restore(sess, model_data_path)\n\n # Use to load from npy file\n# net.load(model_data_path, sess) \n\n # Evalute the network for the given image\n pred = sess.run(net.get_output(), feed_dict={input_node: img})\n \n # Plot result\n print(img.shape, pred.shape)\n plot_image(img_ori, pred[0,:,:,0])\n \n return pred\n \ndef read_video(filename):\n cap = cv2.VideoCapture(filename)\n i = 0\n while(cap.isOpened()):\n ret, frame = cap.read()\n print('get frame %d'%i)\n yield frame\n i+=1\n\n cap.release()\n\ndef predict_video(model_data_path, video_path, net, input_node):\n\n \n # Default input size\n height = 228\n width = 304\n fig = plt.figure(figsize=(20, 16))\n ax1 = fig.add_subplot(121)\n ax2 = fig.add_subplot(122)\n\n plt.ion()\n fig.show()\n fig.canvas.draw()\n for data in read_video(video_path):\n # Read image\n# img = Image.open(image_path)\n \n# import pdb\n# pdb.set_trace()\n img = Image.fromarray(data, 'RGB')\n img = img.resize([width,height], Image.ANTIALIAS)\n img_ori = img.copy()\n\n img = np.array(img).astype('float32')\n img = np.expand_dims(np.asarray(img), axis = 0)\n\n with tf.Session() as sess:\n\n # Load the converted parameters\n print('Loading the model')\n\n # Use to load from ckpt file\n saver = tf.train.Saver() \n saver.restore(sess, model_data_path)\n\n # Use to load from npy file\n # net.load(model_data_path, sess) \n\n # Evalute the network for the given image\n pred = sess.run(net.get_output(), feed_dict={input_node: img})\n out = np.zeros((height, width * 2, 3), dtype=np.uint8)\n out[:height,:width,:] = img_ori\n pic = cv2.resize(pred[0, :, :, 0], (width, height), interpolation=cv2.INTER_CUBIC)\n npic = np.minimum(pic / 10 * 255, 255).astype(np.uint8)\n# out[:height,width:, 0] = npic\n# out[:height,width:, 1] = npic\n# out[:height,width:, 2] = npic\n\n # Plot result\n ax1.imshow(img_ori)\n ii = ax2.imshow(npic)\n divider = make_axes_locatable(ax2)\n cax = divider.append_axes(\"right\", size=\"5%\", pad=0.05)\n bar = plt.colorbar(ii, cax=cax)\n fig.canvas.draw()\n ax1.clear()\n ax2.clear()\n bar.remove()\n\n \ndef main():\n # Parse arguments\n# parser = argparse.ArgumentParser()\n# parser.add_argument('model_path', help='Converted parameters for the model')\n# parser.add_argument('image_paths', help='Directory of images to predict')\n# args = parser.parse_args()\n args = edict()\n tf.reset_default_graph()\n args.model_path = 'model/NYU_FCRN.ckpt'\n args.image_paths = 'images/test3.jpg'\n args.video_paths = 'images/test.mov'\n\n # Predict the image\n net, input_node = load_model()\n# pred = predict_image(args.model_path, args.image_paths, net, input_node)\n pred = predict_video(args.model_path, args.video_paths, net, input_node)\n# os._exit(0)\n\nif __name__ == '__main__':\n main()\n\n \n\n\n\n\n", "repo_name": "THVi-xTHU/xthu", "sub_path": "fcrn_depth_prediction/predict_video.py", "file_name": "predict_video.py", "file_ext": "py", "file_size_in_byte": 4781, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "24", "api": [{"api_name": "tensorflow.placeholder", "line_number": 20, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 20, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "mpl_toolkits.axes_grid1.make_axes_locatable", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 48, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 48, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 49, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 49, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 60, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "PIL.Image.fromarray", "line_number": 105, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 105, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 106, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 106, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 112, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 118, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 118, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 126, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 128, "usage_type": "call"}, {"api_name": "cv2.INTER_CUBIC", "line_number": 128, "usage_type": "attribute"}, {"api_name": "numpy.minimum", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 129, "usage_type": "attribute"}, {"api_name": "mpl_toolkits.axes_grid1.make_axes_locatable", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "easydict.EasyDict", "line_number": 152, "usage_type": "call"}, {"api_name": "tensorflow.reset_default_graph", "line_number": 153, "usage_type": "call"}]}
+{"seq_id": "1046222132", "text": "import json\nimport logging\nimport shutil\nimport signal\nfrom abc import abstractmethod\nfrom collections import defaultdict\nfrom copy import deepcopy\nfrom enum import Enum\nfrom multiprocessing import Process, Manager\nfrom multiprocessing.managers import SyncManager\nfrom pathlib import Path\nfrom typing import Dict, Tuple, Optional\n\nimport cv2\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nimport torch\nimport torch.nn as nn\nfrom clearml import Task, Logger, OutputModel\nfrom clearml.storage.helper import StorageHelper\nfrom torch.optim.optimizer import Optimizer\nfrom tqdm import tqdm\n\nfrom ..config import BaseExperimentConfig\nfrom ...dataset import WolfDataset\nfrom ...logger import logger\nfrom ...losses import Loss\nfrom ...plotting_service.classification import plot_confusion_matrix\n\n__all__ = ['BaseExperimentLogger']\n\nlogging.getLogger(\"clearml.storage\").setLevel(\"WARNING\")\n\n\nclass ConfigEncoder(json.JSONEncoder):\n \"\"\"Special class for encoding Config object.\"\"\"\n\n def default(self, obj):\n if isinstance(obj, Enum):\n return obj.name\n return json.JSONEncoder.default(self, obj)\n\n\nclass NumpyEncoder(json.JSONEncoder):\n \"\"\"Special json encoder for numpy types.\"\"\"\n\n def default(self, obj):\n if isinstance(obj, np.integer):\n return int(obj)\n elif isinstance(obj, np.floating):\n return float(obj)\n elif isinstance(obj, np.bool_):\n return bool(obj)\n elif isinstance(obj, np.ndarray):\n return obj.tolist()\n return json.JSONEncoder.default(self, obj)\n\n\ndef _mgr_init():\n \"\"\"Special initializer for SyncManager to ignore signals.\"\"\"\n signal.signal(signal.SIGINT, signal.SIG_IGN)\n signal.signal(signal.SIGQUIT, signal.SIG_IGN)\n\n\nclass BaseExperimentLogger:\n \"\"\"Base Class for storing and saving experiments results.\"\"\"\n\n RUN_CONFIG_FILENAME = 'run_config.JSON'\n INFERENCE_CONFIG_FILENAME = 'inference_config.JSON'\n RESULTS_FILENAME = 'results.JSON'\n SELECTION_METRIC_FILENAME = 'selection_metric.JSON'\n MODEL_FILENAME = 'model.pt'\n MODEL_WEIGHTS_FILENAME = 'model_weights.pt'\n IMAGES_FOLDER = 'images'\n IMAGES_FILENAME = 'images.csv'\n DATASET_FILENAME = 'dataset.pkl'\n\n # TENSORBOARD_FOLDER = 'tensorboard'\n\n @property\n @abstractmethod\n def FIG_SIZE(self) -> Tuple[int, int]:\n \"\"\"Defines the figure size for saving debug images.\"\"\"\n\n def __init__(self, saving_dir: str or Path, n_debug_samples: int, clearml_config: Optional[Dict] = None):\n \"\"\"\n Args:\n saving_dir: the directory to store all results\n n_debug_samples: the number of debug samples to save for each epoch\n clearml_config: the config for clearml\n \"\"\"\n\n self.saving_dir = Path(saving_dir)\n self.n_debug_samples = n_debug_samples\n self._cleaned_up = False\n\n self.saving_dir.mkdir(parents=True, exist_ok=True)\n\n self.images_dir = self.saving_dir.joinpath(self.IMAGES_FOLDER)\n self.images_dir.mkdir(exist_ok=True)\n\n self._metrics_keeper = defaultdict(list)\n\n # multiprocessing staff\n manager = SyncManager()\n manager.start(_mgr_init) # fire up the child manager process\n self._debug_samples_mp_manager = manager.list()\n self._debug_samples_processes = []\n\n self.clearml_task: Task = None\n self.clearml_logger: Logger = None\n self.clearml_model: OutputModel = None\n if clearml_config:\n clearml_config['auto_connect_frameworks'] = False\n Task.force_requirements_env_freeze(force=True, requirements_file='requirements.txt')\n self.clearml_task = Task.init(task_name=self.saving_dir.name, **clearml_config)\n self.clearml_logger = self.clearml_task.get_logger()\n self.clearml_model = OutputModel(self.clearml_task)\n destination = self.clearml_task.get_output_destination()\n self.clearml_logger.set_default_upload_destination(f'{destination}')\n\n @property\n def folder(self) -> Path:\n \"\"\"Returns the folder where the results are stored.\"\"\"\n return self.saving_dir\n\n def clean_up(self, stopped_by_user: bool = False):\n \"\"\"Removes logged data due to errors.\"\"\"\n if not self.results_path.exists():\n shutil.rmtree(self.saving_dir)\n self._cleaned_up = True\n if self.clearml_task is not None:\n self.clearml_task.delete(raise_on_error=True)\n else:\n if not stopped_by_user:\n logger.warning(f\"Failed to remove experiment folder, since it contains results data.\")\n\n @property\n def cleaned_up(self) -> bool:\n \"\"\"If clean_up is called returns True\"\"\"\n return self._cleaned_up\n\n @property\n def run_config_path(self) -> Path:\n \"\"\"Returns the run config file path.\"\"\"\n return self.saving_dir.joinpath(self.RUN_CONFIG_FILENAME)\n\n @property\n def inference_config_path(self) -> Path:\n \"\"\"Returns the inference config file path.\"\"\"\n return self.saving_dir.joinpath(self.INFERENCE_CONFIG_FILENAME)\n\n @property\n def results_path(self) -> Path:\n \"\"\"Returns the JSON file path, where the metrics will be stored.\"\"\"\n return self.saving_dir.joinpath(self.RESULTS_FILENAME)\n\n @property\n def selection_metric_path(self) -> Path:\n \"\"\"Returns the JSON file path, where the selection metric values will be stored.\"\"\"\n return self.saving_dir.joinpath(self.SELECTION_METRIC_FILENAME)\n\n @property\n def model_path(self) -> Path:\n \"\"\"Returns the file path, where the model will be stored.\"\"\"\n return self.saving_dir.joinpath(self.MODEL_FILENAME)\n\n @property\n def model_weights_path(self) -> Path:\n \"\"\"Returns the file path, where the model will be stored.\"\"\"\n return self.saving_dir.joinpath(self.MODEL_WEIGHTS_FILENAME)\n\n @property\n def train_dataset_path(self) -> Path:\n \"\"\"Returns the pickle file path, where the dataset will be saved.\"\"\"\n return self.saving_dir.joinpath(f'train_{self.DATASET_FILENAME}')\n\n @property\n def valid_dataset_path(self) -> Path:\n \"\"\"Returns the pickle file path, where the dataset will be saved.\"\"\"\n return self.saving_dir.joinpath(f'valid_{self.DATASET_FILENAME}')\n\n def _save_debug_samples_data(self):\n \"\"\"After each epoch saves saved images info.\"\"\"\n df = pd.DataFrame(list(self._debug_samples_mp_manager))\n if df.empty:\n logger.debug(f\"No information about saved images\")\n else:\n if self.saving_dir.exists():\n df = df.sort_values(by=['epoch', 'mode', 'batch', 'image_index'])\n saving_path = self.saving_dir.joinpath(self.IMAGES_FILENAME)\n df.to_csv(saving_path)\n logger.debug(f\"Information about saved images is saved in {saving_path} file.\")\n if self.clearml_logger is not None:\n self.clearml_logger.report_table(\"DebugSamples\", \"all\", csv=saving_path.as_posix())\n\n def finalize(self):\n \"\"\"Waits until all processes are finished.\"\"\"\n for p in self._debug_samples_processes:\n p.join()\n\n self._save_debug_samples_data()\n\n if self.clearml_model is not None:\n try:\n self.register_best_epoch_results()\n\n self.clearml_model.update_weights(\n weights_filename=self.model_weights_path.as_posix(),\n auto_delete_file=False\n )\n # waiting to finish all threads\n pool = StorageHelper._upload_pool\n if pool:\n pool.close()\n pool.join()\n except Exception as e:\n logger.warning(f\"Failed to log final results into clearml, due to error: {e}\")\n\n try:\n self.clearml_task.close()\n except Exception as e:\n logger.warning(f\"Failed to close clearml task due to error: {e}.\", exc_info=True)\n else:\n logger.info(f\"clearml task is closed.\")\n\n def register_run_config(self, config: BaseExperimentConfig):\n \"\"\"Registers the experiment config.\"\"\"\n with open(self.run_config_path, 'w') as f:\n json.dump(config.all_configs, f, cls=ConfigEncoder)\n\n if self.clearml_task is not None:\n self.clearml_task.connect_configuration(name='run_config', configuration=config.all_configs)\n self.clearml_model.update_design(config_dict=config.model)\n hparams = config.get_hyper_params_to_log()\n self.clearml_task.set_parameters_as_dict(hparams)\n\n epochs = hparams['early_stopping_patience']\n self.clearml_logger.set_default_debug_sample_history(epochs)\n\n print()\n for k, v in config.all_configs.items():\n print(f'{k} config: {v}')\n print()\n\n logger.debug(f\"Experiment run config is saved in {self.run_config_path} file.\")\n\n def register_inference_config(self, config: Dict):\n \"\"\"Registers the inference config.\"\"\"\n if not self._cleaned_up:\n with open(self.inference_config_path, 'w') as f:\n json.dump(config, f, cls=ConfigEncoder)\n\n if self.clearml_task is not None:\n self.clearml_task.upload_artifact(name='inference_config', artifact_object=config)\n\n logger.debug(f\"Experiment inference config is saved in {self.inference_config_path} file.\")\n\n @staticmethod\n def _format_results(results: Dict[str, float or int]) -> str:\n return ' | '.join(\n [\n f'{k} - {v:.4}' for k, v in results.items()\n if isinstance(v, (float, int, np.floating, np.integer)) and not np.isnan(v)\n ]\n )\n\n def _clearml_log_metrics(self, metrics: Dict, epoch_mode: str, epoch: int):\n \"\"\"Uses clearml to log metrics.\"\"\"\n if self.clearml_logger is not None:\n for k, v in metrics.items():\n if isinstance(v, (float, int, np.floating, np.integer)):\n self.clearml_logger.report_scalar(k, epoch_mode, v, epoch)\n elif isinstance(v, (list, np.ndarray)) and k in ('confusion_matrix', 'class_names'):\n pass\n else:\n raise TypeError(f\"Metrics should be scalar or matrix, got {type(v)}\")\n\n cm = metrics.get('confusion_matrix')\n class_names = metrics.get('class_names')\n if cm is not None:\n fig = plot_confusion_matrix(np.array(cm), class_names, annotate_samples=False, return_figure=True)\n self.clearml_logger.report_matplotlib_figure(\n title=f'{epoch_mode}/CFM',\n series=f'epoch={epoch}',\n figure=fig,\n iteration=epoch,\n report_image=False,\n report_interactive=True,\n )\n plt.close(fig=fig)\n\n def register_epoch_results(self,\n epoch: int,\n epoch_mode: str,\n epoch_loss_metrics: Dict,\n lr: float = None,\n progress_bar: tqdm = None):\n \"\"\"Registers the experiment single epoch results.\"\"\"\n if progress_bar is not None:\n progress_bar.write(\"\")\n progress_bar.write(f' Epoch {epoch_mode.upper()} Results: {self._format_results(epoch_loss_metrics)}')\n\n serializable = {k: (v.tolist() if isinstance(v, np.ndarray) else v) for k, v in epoch_loss_metrics.items()}\n if lr:\n serializable['lr'] = lr\n\n self._clearml_log_metrics(serializable, epoch_mode, epoch)\n\n serializable['epoch'] = epoch\n self._metrics_keeper[epoch_mode].append(serializable)\n\n with open(self.results_path, 'w') as f:\n json.dump(dict(self._metrics_keeper), f, cls=NumpyEncoder)\n logger.debug(f\"Experiment results are saved in {self.results_path} file.\")\n\n def register_selection_metric(self, selection_metric: Dict):\n \"\"\"Registers the selection_metric data.\"\"\"\n with open(self.selection_metric_path, 'w') as f:\n json.dump(selection_metric, f, cls=NumpyEncoder)\n logger.debug(f\"Experiment selection metric data is saved in {self.selection_metric_path} file.\")\n\n def register_best_epoch_results(self):\n \"\"\"Registers the best epoch results.\"\"\"\n with open(self.selection_metric_path) as f:\n selection_metric = json.load(f)\n\n selection_metric = pd.DataFrame(selection_metric['history'])\n best_epoch = int(selection_metric['epoch'].max())\n\n logger.info(f\"The best_epoch={best_epoch}. Logging its values.\")\n self.clearml_logger.report_text(f'BestEpoch={best_epoch}', print_console=False)\n for mode, mode_metrics in self._metrics_keeper.items():\n [best_epoch_metrics] = [item for item in mode_metrics if item['epoch'] == best_epoch]\n best_epoch_metrics = deepcopy(best_epoch_metrics)\n best_epoch_metrics.pop('epoch')\n serializable = {k: (v.tolist() if isinstance(v, np.ndarray) else v) for k, v in best_epoch_metrics.items()}\n serializable.pop('lr', None)\n self._clearml_log_metrics(serializable, f'best_{mode}', best_epoch)\n logger.info(\"Best Epoch results are logged.\")\n\n def register_model(self, epoch: int, network: nn.Module, optimizer: Optimizer, loss: Loss):\n \"\"\"Registers the model.\"\"\"\n try:\n state_dict = network.module.state_dict()\n except AttributeError:\n state_dict = network.state_dict()\n\n torch.save(\n {\n 'epoch': epoch,\n 'model_state_dict': state_dict,\n 'optimizer_state_dict': optimizer.state_dict(),\n 'loss': loss\n },\n self.model_path,\n )\n torch.save(state_dict, self.model_weights_path)\n logger.debug(f\"Training checkpoint is saved in {self.model_path} file.\")\n logger.debug(f\"Model weights are saved in {self.model_weights_path} file.\")\n\n def register_dataset(self, dataset: WolfDataset, mode: str):\n \"\"\"Registers dataset DataFrame.\"\"\"\n if mode == 'train':\n saving_path = str(self.train_dataset_path)\n elif mode == 'valid':\n saving_path = str(self.valid_dataset_path)\n else:\n raise ValueError(f\"mode must be train or valid, got: '{mode}'\")\n\n dataset.df.to_pickle(saving_path)\n logger.debug(f\"{mode.upper()} dataset pickle is saved in {saving_path} file.\")\n\n if self.clearml_task is not None:\n self.clearml_task.upload_artifact(f\"{mode}_dataset\", saving_path)\n\n def register_epoch_debug_samples(self, epoch: int, epoch_mode: str, epoch_images: Dict):\n \"\"\"Registers the experiment single epoch results.\"\"\"\n if self.n_debug_samples <= 0:\n return\n\n for batch, batch_images in epoch_images.items():\n if 'debug_image' not in batch_images or batch_images['debug_image'] is None:\n continue\n else:\n p = Process(\n target=self.save_epoch_images,\n kwargs=dict(\n images_dir=self.images_dir,\n epoch_mode=epoch_mode,\n epoch_number=epoch,\n batch_number=batch,\n n_images_to_save=self.n_debug_samples,\n info=dict(mode=epoch_mode, epoch=epoch, batch=batch),\n mp_manager=self._debug_samples_mp_manager,\n clearml_logger=self.clearml_logger,\n **batch_images\n )\n )\n p.start()\n self._debug_samples_processes.append(p)\n\n @classmethod\n def save_epoch_images(cls,\n input_image: np.ndarray,\n y_true: np.ndarray,\n y_pred: np.ndarray,\n info: Dict,\n mp_manager: Manager,\n mode: str = None,\n epoch_number: int = None,\n batch_number: int = None,\n images_dir: str = None,\n clearml_logger: Optional[Logger] = None,\n n_images_to_save: int = 10):\n \"\"\"Please implement this method in subclasses, depending on the task.\"\"\"\n raise NotImplementedError\n\n @staticmethod\n def save_image(image: np.ndarray, saving_path: str or Path, as_uint8: bool = True) -> None:\n \"\"\"Saves single debug image.\"\"\"\n if as_uint8:\n cv2.imwrite(str(saving_path), (image * 255).astype(np.uint8))\n else:\n cv2.imwrite(str(saving_path), image)\n", "repo_name": "razmikmelikbekyan/WolfPyTorch", "sub_path": "wolf/experiment/experiment_logging/base_logger.py", "file_name": "base_logger.py", "file_ext": "py", "file_size_in_byte": 17098, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "24", "api": [{"api_name": "logging.getLogger", "line_number": 33, "usage_type": "call"}, {"api_name": "json.JSONEncoder", "line_number": 36, "usage_type": "attribute"}, {"api_name": "enum.Enum", "line_number": 40, "usage_type": "argument"}, {"api_name": "json.JSONEncoder.default", "line_number": 42, "usage_type": "call"}, {"api_name": "json.JSONEncoder", "line_number": 42, "usage_type": "attribute"}, {"api_name": "json.JSONEncoder", "line_number": 45, "usage_type": "attribute"}, {"api_name": "numpy.integer", "line_number": 49, "usage_type": "attribute"}, {"api_name": "numpy.floating", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.bool_", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 55, "usage_type": "attribute"}, {"api_name": "json.JSONEncoder.default", "line_number": 57, "usage_type": "call"}, {"api_name": "json.JSONEncoder", "line_number": 57, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 62, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 62, "usage_type": "attribute"}, {"api_name": "signal.SIG_IGN", "line_number": 62, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 63, "usage_type": "call"}, {"api_name": "signal.SIGQUIT", "line_number": 63, "usage_type": "attribute"}, {"api_name": "signal.SIG_IGN", "line_number": 63, "usage_type": "attribute"}, {"api_name": "abc.abstractmethod", "line_number": 82, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 83, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 86, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 86, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 86, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 94, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 103, "usage_type": "call"}, {"api_name": "multiprocessing.managers.SyncManager", "line_number": 106, "usage_type": "call"}, {"api_name": "clearml.Task", "line_number": 111, "usage_type": "name"}, {"api_name": "clearml.Logger", "line_number": 112, "usage_type": "name"}, {"api_name": "clearml.OutputModel", "line_number": 113, "usage_type": "name"}, {"api_name": "clearml.Task.force_requirements_env_freeze", "line_number": 116, "usage_type": "call"}, {"api_name": "clearml.Task", "line_number": 116, "usage_type": "name"}, {"api_name": "clearml.Task.init", "line_number": 117, "usage_type": "call"}, {"api_name": "clearml.Task", "line_number": 117, "usage_type": "name"}, {"api_name": "clearml.OutputModel", "line_number": 119, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 124, "usage_type": "name"}, {"api_name": "shutil.rmtree", "line_number": 131, "usage_type": "call"}, {"api_name": "logger.logger.warning", "line_number": 137, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 137, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 145, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 150, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 155, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 160, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 165, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 170, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 175, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 180, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 186, "usage_type": "call"}, {"api_name": "logger.logger.debug", "line_number": 188, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 188, "usage_type": "name"}, {"api_name": "logger.logger.debug", "line_number": 194, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 194, "usage_type": "name"}, {"api_name": "clearml.storage.helper.StorageHelper._upload_pool", "line_number": 214, "usage_type": "attribute"}, {"api_name": "clearml.storage.helper.StorageHelper", "line_number": 214, "usage_type": "name"}, {"api_name": "logger.logger.warning", "line_number": 219, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 219, "usage_type": "name"}, {"api_name": "logger.logger.warning", "line_number": 224, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 224, "usage_type": "name"}, {"api_name": "logger.logger.info", "line_number": 226, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 226, "usage_type": "name"}, {"api_name": "config.BaseExperimentConfig", "line_number": 228, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 231, "usage_type": "call"}, {"api_name": "config.all_configs", "line_number": 231, "usage_type": "attribute"}, {"api_name": "config.all_configs", "line_number": 234, "usage_type": "attribute"}, {"api_name": "config.model", "line_number": 235, "usage_type": "attribute"}, {"api_name": "config.get_hyper_params_to_log", "line_number": 236, "usage_type": "call"}, {"api_name": "config.all_configs.items", "line_number": 243, "usage_type": "call"}, {"api_name": "config.all_configs", "line_number": 243, "usage_type": "attribute"}, {"api_name": "logger.logger.debug", "line_number": 247, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 247, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 249, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 253, "usage_type": "call"}, {"api_name": "logger.logger.debug", "line_number": 258, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 258, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 261, "usage_type": "name"}, {"api_name": "numpy.floating", "line_number": 265, "usage_type": "attribute"}, {"api_name": "numpy.integer", "line_number": 265, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 265, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 269, "usage_type": "name"}, {"api_name": "numpy.floating", "line_number": 273, "usage_type": "attribute"}, {"api_name": "numpy.integer", "line_number": 273, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 275, "usage_type": "attribute"}, {"api_name": "plotting_service.classification.plot_confusion_matrix", "line_number": 283, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 283, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 292, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 292, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 297, "usage_type": "name"}, {"api_name": "tqdm.tqdm", "line_number": 299, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 305, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 315, "usage_type": "call"}, {"api_name": "logger.logger.debug", "line_number": 316, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 316, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 318, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 321, "usage_type": "call"}, {"api_name": "logger.logger.debug", "line_number": 322, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 322, "usage_type": "name"}, {"api_name": "json.load", "line_number": 327, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 329, "usage_type": "call"}, {"api_name": "logger.logger.info", "line_number": 332, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 332, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 336, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 338, "usage_type": "attribute"}, {"api_name": "logger.logger.info", "line_number": 341, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 341, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 343, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 343, "usage_type": "name"}, {"api_name": "torch.optim.optimizer.Optimizer", "line_number": 343, "usage_type": "name"}, {"api_name": "losses.Loss", "line_number": 343, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 350, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 359, "usage_type": "call"}, {"api_name": "logger.logger.debug", "line_number": 360, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 360, "usage_type": "name"}, {"api_name": "logger.logger.debug", "line_number": 361, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 361, "usage_type": "name"}, {"api_name": "dataset.WolfDataset", "line_number": 363, "usage_type": "name"}, {"api_name": "dataset.df.to_pickle", "line_number": 372, "usage_type": "call"}, {"api_name": "dataset.df", "line_number": 372, "usage_type": "attribute"}, {"api_name": "logger.logger.debug", "line_number": 373, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 373, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 378, "usage_type": "name"}, {"api_name": "multiprocessing.Process", "line_number": 387, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 406, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 407, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 408, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 409, "usage_type": "name"}, {"api_name": "multiprocessing.Manager", "line_number": 410, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 415, "usage_type": "name"}, {"api_name": "clearml.Logger", "line_number": 415, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 421, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 421, "usage_type": "name"}, {"api_name": "cv2.imwrite", "line_number": 424, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 424, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 426, "usage_type": "call"}]}
+{"seq_id": "8915502306", "text": "from flask import request, make_response, jsonify\nfrom flask_expects_json import expects_json\nfrom flask_jwt_extended import jwt_required, get_jwt_identity\n\nfrom api.v1.common.constants import get_indicator_type_by_id\nfrom api.v1.researches.schemas import create_research_schema, ResearchDataSchema, create_indicator_schema, \\\n IndicatorDataSchema\nfrom api.v1.researches.services import get_research, create_research, get_researches, update_research, create_indicator, \\\n update_indicator, get_my_research\nfrom api.v1.users.services import get_user_by_email\n\n\n@jwt_required()\n@expects_json(create_research_schema)\ndef create_research_view():\n data = request.json\n research = create_research(data)\n return make_response(jsonify(ResearchDataSchema().dump(research, many=False)), 200)\n\n\ndef get_place_view(place_id):\n pass\n\n\n@jwt_required()\ndef get_my_places_view():\n user_email = get_jwt_identity()\n user = get_user_by_email(user_email)\n # places = places.get_my_places(user)\n researches = get_researches()\n return make_response(jsonify(ResearchDataSchema().dump(researches, many=True)), 200)\n\n\n@jwt_required()\ndef get_places_view():\n places = get_researches()\n\n return make_response(jsonify(ResearchDataSchema().dump(places, many=True)), 200)\n\n\n@jwt_required()\ndef get_research_view(research_id=None):\n user_email = get_jwt_identity()\n user = get_user_by_email(user_email)\n research = get_research(research_id, owner_id=user.id)\n return make_response(jsonify(ResearchDataSchema().dump(research, many=False)), 200)\n\n\n@jwt_required()\ndef update_research_view(research_id=None):\n user_email = get_jwt_identity()\n user = get_user_by_email(user_email)\n\n data = request.json\n update_research(research_id, data)\n research = get_research(research_id, owner_id=user.id)\n return make_response(jsonify(ResearchDataSchema().dump(research, many=False)), 200)\n\n\n@jwt_required()\n@expects_json(create_indicator_schema)\ndef create_indicator_view():\n data = request.json\n indicator = create_indicator(data)\n indicator.type = get_indicator_type_by_id(indicator.type_id)\n\n return make_response(jsonify(IndicatorDataSchema().dump(indicator, many=False)), 200)\n\n\n@jwt_required()\n@expects_json(create_indicator_schema)\ndef update_indicator_view(indicator_id=None):\n data = request.json\n update_indicator(indicator_id, data)\n\n return make_response(jsonify({\"status\": \"ok\"}), 200)\n\n\n@jwt_required()\ndef get_my_researches_view():\n user_email = get_jwt_identity()\n user = get_user_by_email(user_email)\n research = get_my_research(owner_id=user.id)\n return make_response(jsonify(ResearchDataSchema().dump(research, many=True)), 200)\n", "repo_name": "boooogyman/fish_king", "sub_path": "api/v1/researches/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2706, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "flask.request.json", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 16, "usage_type": "name"}, {"api_name": "api.v1.researches.services.create_research", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 18, "usage_type": "call"}, {"api_name": "api.v1.researches.schemas.ResearchDataSchema", "line_number": 18, "usage_type": "call"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 13, "usage_type": "call"}, {"api_name": "flask_expects_json.expects_json", "line_number": 14, "usage_type": "call"}, {"api_name": "api.v1.researches.schemas.create_research_schema", "line_number": 14, "usage_type": "argument"}, {"api_name": "flask_jwt_extended.get_jwt_identity", "line_number": 27, "usage_type": "call"}, {"api_name": "api.v1.users.services.get_user_by_email", "line_number": 28, "usage_type": "call"}, {"api_name": "api.v1.researches.services.get_researches", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 31, "usage_type": "call"}, {"api_name": "api.v1.researches.schemas.ResearchDataSchema", "line_number": 31, "usage_type": "call"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 25, "usage_type": "call"}, {"api_name": "api.v1.researches.services.get_researches", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 38, "usage_type": "call"}, {"api_name": "api.v1.researches.schemas.ResearchDataSchema", "line_number": 38, "usage_type": "call"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 34, "usage_type": "call"}, {"api_name": "flask_jwt_extended.get_jwt_identity", "line_number": 43, "usage_type": "call"}, {"api_name": "api.v1.users.services.get_user_by_email", "line_number": 44, "usage_type": "call"}, {"api_name": "api.v1.researches.services.get_research", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 46, "usage_type": "call"}, {"api_name": "api.v1.researches.schemas.ResearchDataSchema", "line_number": 46, "usage_type": "call"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 41, "usage_type": "call"}, {"api_name": "flask_jwt_extended.get_jwt_identity", "line_number": 51, "usage_type": "call"}, {"api_name": "api.v1.users.services.get_user_by_email", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 54, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 54, "usage_type": "name"}, {"api_name": "api.v1.researches.services.update_research", "line_number": 55, "usage_type": "call"}, {"api_name": "api.v1.researches.services.get_research", "line_number": 56, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 57, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 57, "usage_type": "call"}, {"api_name": "api.v1.researches.schemas.ResearchDataSchema", "line_number": 57, "usage_type": "call"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 63, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 63, "usage_type": "name"}, {"api_name": "api.v1.researches.services.create_indicator", "line_number": 64, "usage_type": "call"}, {"api_name": "api.v1.common.constants.get_indicator_type_by_id", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 67, "usage_type": "call"}, {"api_name": "api.v1.researches.schemas.IndicatorDataSchema", "line_number": 67, "usage_type": "call"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 60, "usage_type": "call"}, {"api_name": "flask_expects_json.expects_json", "line_number": 61, "usage_type": "call"}, {"api_name": "api.v1.researches.schemas.create_indicator_schema", "line_number": 61, "usage_type": "argument"}, {"api_name": "flask.request.json", "line_number": 73, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 73, "usage_type": "name"}, {"api_name": "api.v1.researches.services.update_indicator", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 76, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 76, "usage_type": "call"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 70, "usage_type": "call"}, {"api_name": "flask_expects_json.expects_json", "line_number": 71, "usage_type": "call"}, {"api_name": "api.v1.researches.schemas.create_indicator_schema", "line_number": 71, "usage_type": "argument"}, {"api_name": "flask_jwt_extended.get_jwt_identity", "line_number": 81, "usage_type": "call"}, {"api_name": "api.v1.users.services.get_user_by_email", "line_number": 82, "usage_type": "call"}, {"api_name": "api.v1.researches.services.get_my_research", "line_number": 83, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 84, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 84, "usage_type": "call"}, {"api_name": "api.v1.researches.schemas.ResearchDataSchema", "line_number": 84, "usage_type": "call"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 79, "usage_type": "call"}]}
+{"seq_id": "71515663742", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCopies s3 kv store into local directory structure\n\n(c) 2021 by aidan@latch.bio\n(c) copied and modified from https://stackoverflow.com/questions/31918960/boto3-to-download-all-files-from-a-s3-bucket\n\"\"\"\nimport os\n\nimport boto3\n\ns3_client = boto3.client(\"s3\")\n\n\ndef download_dir(prefix, local, bucket, client=s3_client):\n \"\"\"\n params:\n - prefix: pattern to match in s3\n - local: local path to folder in which to place files\n - bucket: s3 bucket with target contents\n - client: initialized s3 client object\n \"\"\"\n keys = []\n dirs = []\n next_token = \"\"\n base_kwargs = {\n \"Bucket\": bucket,\n \"Prefix\": prefix,\n }\n while next_token is not None:\n kwargs = base_kwargs.copy()\n if next_token != \"\":\n kwargs.update({\"ContinuationToken\": next_token})\n results = client.list_objects_v2(**kwargs)\n contents = results.get(\"Contents\")\n if contents is not None:\n for i in contents:\n k = i.get(\"Key\")\n if k[-1] != \"/\":\n keys.append(k)\n else:\n dirs.append(k)\n next_token = results.get(\"NextContinuationToken\")\n for d in dirs:\n dest_pathname = local + d.replace(prefix, \"\")\n if not os.path.exists(os.path.dirname(dest_pathname)):\n os.makedirs(os.path.dirname(dest_pathname))\n for k in keys:\n dest_pathname = local + k.replace(prefix, \"\")\n if not os.path.exists(os.path.dirname(dest_pathname)):\n os.makedirs(os.path.dirname(dest_pathname))\n client.download_file(bucket, k, dest_pathname)\n\n\ndef ensure_dir(file_path):\n directory = os.path.dirname(file_path)\n if not os.path.exists(directory):\n os.makedirs(directory)\n", "repo_name": "latchbio/wf-core-guideseq", "sub_path": "latch/s3_dir_download.py", "file_name": "s3_dir_download.py", "file_ext": "py", "file_size_in_byte": 1802, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "boto3.client", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 46, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 50, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 58, "usage_type": "call"}]}
+{"seq_id": "11005815325", "text": "# Abstract graph of 3D transformations through time\n# Has a list of nodes and a list of edges\n# Allows getting transforms (edges) for a given node in a given time (frame)\n# Edge could be:\n# - rigid and known (so its pre-calibrated and given)\n# - rigid and unknown (needs solving, and can use multiple frames for better results)\n# - non-rigid and known (live tracking, so its given)\n# - non-rigid and unknown (needs solving, only possible if it is a loop)\n\n# Abstract class for the above, we will implement a test class and a real class\n# Transformations are 4x4 matrices\n# The whole transformation graph can be solved by solving each node's transformation\n\n# In the future we can use networkx to perform graph operations, but for now we use matrices for representation\n\n# We will first need to define the type of each edge: \"rigid-unknown\", \"rigid-known\", \"non-rigid-unknown\", \"non-rigid-known\"\n# We will also need to define the noise of each edge\n# Then, we create a empty matrix of matrices, where each matrix is a 4x4 transformation matrix\n\n# === Implementation ===\n# We will create a custom class for the matrix of matrices, we will implement so that it can be indexed by a tuple (node, node)\n# Each edge will also be associated with a type and a noise\n# Also, setting a value will also set the inverse of the value in the other direction\n\n# We will wrap around the matrix of matrices to create a graph class, which will have a list of nodes and a list of edges\n# We model the transformation noise as a gaussian distribution\n\n# ======================\n\n# Then, we populate the matrix with the known transformations (rigid and known, non-rigid and known)\n\n# Then, we estimate the non-rigid and unknown transformations, which will only work if the graph is a loop\n# We combine multiple potential paths and use weighted least squares to estimate the transformation. The optimal weight will be the one that minimizes the error.\n# The unsolvable nodes will be marked as such\n\n# Next, we estimate the rigid and unknown transformations with a time window of n frames\n# We can then similarly use weighted least squares to estimate the transformation. The optimal weight will be the one that minimizes the error.\n# The unsolvable nodes will be marked as such\n# This will also allow for potentially solving problems where the graph is not a loop, like AX = XB, where X is unknown\n\n# Finally, we can use the graph to solve for the transformations of any node in any frame\n\nimport json\nimport math\nfrom typing import Literal\nimport numpy as np\nfrom tools import *\n\n# These implementations are used for testing mostly, and online variants will be implemented later that uses only an iterative interface for solving; since in the online case, we will not know the whole graph at once\n\nclass TransformationGraph:\n def __init__(self, num_nodes: int, edges: list[tuple[int, int, Literal[\"rigid-unknown\", \"rigid-known\", \"non-rigid-unknown\", \"non-rigid-known\"], float]], frames: int=1):\n \"\"\"Initialize the transformation graph\n\n Parameters:\n num_nodes (int): Number of nodes in the graph\n edges (list[tuple[int, int, Literal[\"rigid-unknown\", \"rigid-known\", \"non-rigid-unknown\", \"non-rigid-known\"], float]]): List of edges, where each edge is a tuple of (node1, node2, type, noise)\n frames (int): Number of frames in the graph\n \"\"\"\n if frames < 1:\n raise Exception(\"Number of frames must be at least 1\")\n # TODO: Have better representation of noise. Look into Kalman filters.\n\n # Create a matrix of matrices\n self._matrix = np.nan * np.ones((num_nodes, num_nodes, frames, 4, 4))\n\n # Create a seperate np object matrix for the edge types, set to None (not np.nan) initially\n self._types = np.empty((num_nodes, num_nodes), dtype=object)\n\n # Create a seperate matrix for the noise of each edge\n self._noise = np.nan * np.ones((num_nodes, num_nodes))\n\n # Set the matrix edge types and noise\n for edge in edges:\n # Set edge type\n self._types[edge[0], edge[1]] = edge[2]\n self._types[edge[1], edge[0]] = edge[2]\n # Set noise\n self._noise[edge[0], edge[1]] = edge[3]\n self._noise[edge[1], edge[0]] = edge[3]\n\n # TODO: This should be in test class instead\n # # Validate intra-group edges are rigid\n # for group in self._groups:\n # for node1 in group:\n # for node2 in group:\n # if self._types[node1, node2, 0] != \"rigid-known\" and self._types[node1, node2, 0] != \"rigid-unknown\":\n # raise Exception(\"Inconsistent edge type in group\")\n\n # Fill the identity matrices\n for i in range(num_nodes):\n for j in range(frames):\n self._matrix[i, i, j] = np.eye(4)\n\n def __getitem__(self, key: tuple[int, int, int]) -> np.ndarray:\n \"\"\"Get the transformation matrix for a given edge and frame\n\n Parameters:\n key (tuple[int, int, int]): Tuple of (node1, node2, frame)\n\n Returns:\n np.ndarray: Transformation matrix\n \"\"\"\n # Throw an error if there is no connection type\n if self._types[key[0], key[1]] == None:\n raise Exception(\"No connection type defined for edge\")\n return self._matrix[key[0], key[1], key[2]]\n\n def __setitem__(self, key: tuple[int, int, int], value: np.ndarray):\n \"\"\"Set the transformation matrix for a given edge and frame\n\n Parameters:\n key (tuple[int, int, int]): Tuple of (node1, node2, frame)\n value (np.ndarray): Transformation matrix\n \"\"\"\n # Throw an error if there is no connection type\n if self._types[key[0], key[1]] == None:\n raise Exception(\"No connection type defined for edge\")\n\n # Set the value\n self._matrix[key[0], key[1], key[2]] = value\n\n # Set the inverse\n self._matrix[key[1], key[0], key[2]] = np.linalg.inv(value)\n\n def get_type(self, key: tuple[int, int]) -> Literal[\"rigid-unknown\", \"rigid-known\", \"non-rigid-unknown\", \"non-rigid-known\"]:\n \"\"\"Get the type of a given edge\n\n Parameters:\n key (tuple[int, int]): Tuple of (node1, node2)\n\n Returns:\n Literal[\"rigid-unknown\", \"rigid-known\", \"non-rigid-unknown\", \"non-rigid-known\"]: Type of the edge\n \"\"\"\n return self._types[key[0], key[1]]\n\n def get_noise(self, key: tuple[int, int]) -> float:\n \"\"\"Get the noise of a given edge\n\n Parameters:\n key (tuple[int, int]): Tuple of (node1, node2)\n\n Returns:\n float: Noise of the edge\n \"\"\"\n return self._noise[key[0], key[1]]\n\n def get_nodes(self) -> list[int]:\n \"\"\"Get the list of nodes in the graph\n\n Returns:\n list[int]: List of nodes\n \"\"\"\n return list(range(self._matrix.shape[0]))\n\n def get_edges(self, node:int) -> list[tuple[int, Literal[\"rigid-unknown\", \"rigid-known\", \"non-rigid-unknown\", \"non-rigid-known\"], float]]:\n \"\"\"Get the list of edges for a given node\n\n Parameters:\n node (int): Node\n\n Returns:\n list[tuple[int, Literal[\"rigid-unknown\", \"rigid-known\", \"non-rigid-unknown\", \"non-rigid-known\"], float]]: List of edges, where each edge is a tuple of (node, type, noise)\n \"\"\"\n # print(f'Checking edges for node {node}')\n edges = []\n for i in range(self._matrix.shape[0]):\n edge_type = self._types[node, i]\n if edge_type != None:\n edges.append((i, self._types[node, i], self._noise[node, i]))\n # print(f'Found edges {edges}')\n return edges\n\n @property\n def num_nodes(self) -> int:\n \"\"\"Get the number of nodes in the graph\n\n Returns:\n int: Number of nodes\n \"\"\"\n return self._matrix.shape[0]\n\n @property\n def num_frames(self) -> int:\n \"\"\"Get the number of frames in the graph\n\n Returns:\n int: Number of frames\n \"\"\"\n return self._matrix.shape[2]\n\nclass TestTransformationGraph(TransformationGraph):\n \"\"\"\n Specialized transformation graph for testing. Generates random ground truth transformations for each node and frame.\n \"\"\"\n\n def __init__(self, num_nodes: int, edges: list[tuple[int, int, Literal[\"rigid-unknown\", \"rigid-known\", \"non-rigid-unknown\", \"non-rigid-known\"], float]], frames: int):\n \"\"\"Initialize the transformation graph\n\n Parameters:\n num_nodes (int): Number of nodes in the graph\n edges (list[tuple[int, int, Literal[\"rigid-unknown\", \"rigid-known\", \"non-rigid-unknown\", \"non-rigid-known\"], float]]): List of edges, where each edge is a tuple of (node1, node2, type, noise)\n frames (int): Number of frames in the graph\n \"\"\"\n super().__init__(num_nodes, edges, frames)\n # Identify node groups that are rigidly connected\n nodes = set(range(num_nodes))\n self._groups = []\n self._groupMap = dict()\n # For each node, use BFS to find all nodes that are rigidly connected. After that, remove all those nodes from the set. Repeat until all nodes are removed.\n while len(nodes) > 0:\n # Get a random node\n node = nodes.pop()\n\n # Add the node to the group\n group = set([node])\n\n # Add all rigidly connected nodes to the group\n queue = [node]\n while len(queue) > 0:\n # Get the next node\n node = queue.pop(0)\n\n # Add all rigidly connected nodes to the group\n for i in range(num_nodes):\n if self._types[node, i] == \"rigid-known\" or self._types[node, i] == \"rigid-unknown\":\n if i not in group:\n group.add(i)\n queue.append(i)\n\n # Remove all nodes in the group from the set\n nodes = nodes - group\n self._groups.append(group)\n for node in group:\n self._groupMap[node] = len(self._groups) - 1\n\n # Generate ground truth transformations \n self._worldTransforms = np.zeros((num_nodes, frames, 4, 4))\n # Generate random transformations within groups, which stays the same for all frames\n # Since we assume a node can only be in one group, we can generate a random transform for every node\n # Individual nodes can be considered as a group of size 1\n intra_group_transforms = []\n for node in range(num_nodes):\n intra_group_transforms.append(generate_random_transform())\n # Generate random group transforms per frame\n group_transforms = []\n for i in range(frames):\n group_transforms.append([]) # Add a new frame\n for group in self.groups:\n group_transforms[i].append(generate_random_transform())\n # Now we can generate per node transforms for each frame\n for frame in range(frames):\n for node in range(num_nodes):\n group_id = self.get_group_id(node)\n # Set the ground truth transform\n self._worldTransforms[node, frame] = group_transforms[frame][group_id] @ intra_group_transforms[node]\n # print(f'Ground truth transform for node {node} in frame {frame}:')\n # print(self._worldTransforms[node, frame])\n # Second pass to generate relative transforms\n for frame in range(frames):\n for node in range(num_nodes):\n # For each edge, derive the relative transform from the ground truth\n for (neighbor, edge_type, noise) in self.get_edges(node):\n # print(f'Calculating relative transform for edge ({node}, {neighbor}) in frame {frame}')\n # Skip if edge is not known\n if edge_type == \"rigid-unknown\" or edge_type == \"non-rigid-unknown\":\n continue\n # Get the relative transform using calc_relative_transform(from, to)\n # print(node, neighbor, frame)\n # print(self._worldTransforms[node, frame], self._worldTransforms[neighbor, frame])\n relative_transform = calc_relative_transform(self._worldTransforms[node, frame], self._worldTransforms[neighbor, frame])\n # print(relative_transform)\n self[node, neighbor, frame] = relative_transform\n # Now we should be done!\n # Note, mentally we can think of the world origin as a separate node with an unknown non-rigid transformation to every other node\n\n def get_group_id(self, node: int) -> int:\n \"\"\"Get the group id of a node\n\n Parameters:\n node (int): Node\n\n Returns:\n int: Group id\n \"\"\"\n return self._groupMap[node]\n\n def get_group(self, group_id: int) -> set[int]:\n \"\"\"Get the group of a node\n\n Parameters:\n group_id (int): Group id\n\n Returns:\n set[int]: Group\n \"\"\"\n return self._groups[group_id]\n\n @property\n def groups(self) -> list[set[int]]:\n \"\"\"Gets a copy of the list of groups\n\n Returns:\n list[set[int]]: List of groups\n \"\"\"\n return self._groups.copy()\n\n def world_transform_to_dict(self) -> dict[int, dict[int, np.ndarray]]:\n \"\"\"Converts the world transforms to a json compatible dictionary\n\n Returns:\n dict[int, dict[int, np.ndarray]]: Dictionary of world transforms\n \"\"\"\n world_transforms = dict()\n for node in range(self.num_nodes):\n world_transforms[node] = dict()\n for frame in range(self.num_frames):\n # Remember to convert nan to None\n world_transforms[node][frame] = self._worldTransforms[node, frame].tolist()\n world_transforms[node][frame] = [[None if math.isnan(x) else x for x in row] for row in world_transforms[node][frame]]\n \n return world_transforms\n # Local transform matrix and world transforms could be merged into a single matrix if we treat the world origin as a node\n\n def local_transforms_to_dict(self) -> dict[int, dict[int, dict[int, np.ndarray]]]:\n \"\"\"Converts the local transforms to a json compatible dictionary\n\n Returns:\n dict[int, dict[int, dict[int, np.ndarray]]]: Dictionary of local transforms\n \"\"\"\n local_transforms = dict()\n for node in range(self.num_nodes):\n local_transforms[node] = dict()\n for neighbor in range(self.num_nodes):\n local_transforms[node][neighbor] = dict()\n for frame in range(self.num_frames):\n try:\n local_transforms[node][neighbor][frame] = self[node, neighbor, frame].tolist()\n local_transforms[node][neighbor][frame] = [[None if math.isnan(x) else x for x in row] for row in local_transforms[node][neighbor][frame]]\n except Exception:\n # No connection\n pass\n return local_transforms\n\n # Local transform also serializes unknown transforms since they are meant to be solved\n # World transforms on the other hand only serializes known transforms\n\n def edges_to_dict(self) -> dict[int, dict[int, dict[str, str]]]:\n \"\"\"Converts the edges to a json compatible dictionary\n\n Returns:\n dict[int, dict[int, dict[str, str]]]: Dictionary of edges\n \"\"\"\n edges = dict()\n for node in range(self.num_nodes):\n edges[node] = dict()\n for neighbor in range(self.num_nodes):\n edges[node][neighbor] = dict()\n edges[node][neighbor][\"type\"] = self._types[node, neighbor]\n edges[node][neighbor][\"noise\"] = self._noise[node, neighbor]\n if np.isnan(edges[node][neighbor][\"noise\"]):\n edges[node][neighbor][\"noise\"] = None\n return edges\n\n def to_dict(self) -> dict[str, dict]:\n \"\"\"Converts the graph to a json compatible dictionary\n\n Returns:\n dict[str, dict]: Dictionary of graph\n \"\"\"\n graph = dict()\n graph[\"world_transforms\"] = self.world_transform_to_dict()\n graph[\"local_transforms\"] = self.local_transforms_to_dict()\n graph[\"edges\"] = self.edges_to_dict()\n return graph\n\n def to_json_string(self) -> str:\n \"\"\"Converts the graph to a json string\n\n Returns:\n str: Json string\n \"\"\"\n return json.dumps(self.to_dict())\n\n# TODO: Reimplement TransformationGraphs using iterators so that we can use the same solvers for real time and offline", "repo_name": "shiukaheng/transforms-solver", "sub_path": "server/graphs.py", "file_name": "graphs.py", "file_ext": "py", "file_size_in_byte": 16807, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "typing.Literal", "line_number": 52, "usage_type": "name"}, {"api_name": "numpy.nan", "line_number": 65, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 95, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 109, "usage_type": "attribute"}, {"api_name": "numpy.linalg.inv", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 124, "usage_type": "attribute"}, {"api_name": "typing.Literal", "line_number": 126, "usage_type": "name"}, {"api_name": "typing.Literal", "line_number": 156, "usage_type": "name"}, {"api_name": "typing.Literal", "line_number": 197, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 238, "usage_type": "call"}, {"api_name": "math.isnan", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 308, "usage_type": "attribute"}, {"api_name": "math.isnan", "line_number": 339, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 325, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 361, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 383, "usage_type": "call"}]}
+{"seq_id": "70165285182", "text": "# imports\nimport cv2 as cv\nimport numpy as np\nimport time\n\n# definitions\nwriter = None\nh, w, = None, None\n\n# load yolo labels\nwith open('yolo-coco-data/coco.names') as f:\n labels = [line.strip() for line in f]\n\nprint('List with labels names:')\nprint(labels)\n\n# load yolo network\nnetwork = cv.dnn.readNetFromDarknet(\"./yolo-coco-data/yolov3.cfg\", \"./yolo-coco-data/yolov3.weights\")\n\n# create a list with all the output layers names\nlayers_names_all = network.getLayerNames()\n\nlayers_names_output = [\\\n layers_names_all[i - 1] for i in network.getUnconnectedOutLayers()]\n\n# setting minimum probability to eliminate weak predictions\nprobability_minimum = 0.5\n# setting threshold for non-maximum suppression\nthreshold = 0.3\n\n# generate colors for representing every detected object\ncolors = np.random.randint(0, 255, size=(len(labels), 3), dtype='uint8')\n\nf = 0 # Counting frames\nt = 0 # Counting total time\n\nvideo = cv.VideoCapture(\"./videos/traffic-cars.mp4\")\nwhile True:\n ret, frame = video.read()\n \n if not ret:\n break;\n \n if w is None or h is None:\n h, w = frame.shape[:2]\n \n # the image has to be converted to a blob \n blob = cv.dnn.blobFromImage(\n frame, # input image\n 1/255.0, # normalization factor\n (416, 416), # size (416, 416) is recommended for yolo\n swapRB=True,# swap red and blue channels\n crop=False # no cropping\n )\n \n network.setInput(blob) # set the blob as input to the network\n start = time.time() # start the timer\n # You can get all outputs from the network by calling forward() method\n # and passing names of the layers you want to get outputs from\n # In this case, we will get the names of the output layers\n output_from_network = network.forward(layers_names_output) # forward pass\n end = time.time() # stop the timer\n \n f += 1\n f += end - start\n \n print('Frame number {0} took {1:.5f} seconds'.format(f, end - start))\n\n bounding_boxes = []\n confidences = []\n class_numbers = []\n \n# the YOLO model has multiple output layers because it detects objects at multiple scales. The output from each of these layers includes information on bounding boxes, objectness score, and class probabilities for multiple grid cells of a particular scale.\n# Specifically, for YOLOv3, there are three output layers corresponding to three different scales. For each scale, the model divides the input image into a grid (e.g., 13x13, 26x26, 52x52), and each grid cell predicts a fixed number of bounding boxes. For each bounding box, the model predicts coordinates (x, y, width, height), an objectness score, and class probabilities for all classes (e.g., 80 classes in the COCO dataset).\n# After the forward pass through the network, the output includes the predictions from all of these layers. The script processes these outputs to extract the class with the highest probability for each bounding box and uses this information to draw bounding boxes and labels on the video frames.\n \n for result in output_from_network:\n for detected_objects in result:\n scores = detected_objects[5:]\n class_current = np.argmax(scores)\n confidence_current = scores[class_current]\n \n if confidence_current > probability_minimum:\n box_current = detected_objects[0:4] * np.array([w, h, w, h])\n x_center, y_center, box_width, box_height = box_current\n x_min = int(x_center - (box_width / 2))\n y_min = int(y_center - (box_height / 2))\n \n bounding_boxes.append([x_min, y_min, int(box_width), int(box_height)])\n confidences.append(float(confidence_current))\n class_numbers.append(class_current)\n \n# Suppression of non-maximum boxes\n results = cv.dnn.NMSBoxes(bounding_boxes, confidences, probability_minimum, threshold)\n\n if len(results) > 0:\n for i in results.flatten():\n x_min, y_min = bounding_boxes[i][0], bounding_boxes[i][1]\n box_width, box_height = bounding_boxes[i][2], bounding_boxes[i][3]\n \n color_box_current = colors[class_numbers[i]].tolist()\n \n cv.rectangle(frame, (x_min, y_min), (x_min + box_width, y_min + box_height), color_box_current, 2)\n \n text_box_current = '{}: {:.4f}'.format(labels[int(class_numbers[i])], confidences[i])\n \n cv.putText(frame, text_box_current, (x_min, y_min - 5), cv.FONT_HERSHEY_SIMPLEX, 0.5, color_box_current, 2)\n \n # Initializing writer\n # we do it only once from the very beginning\n # when we get spatial dimensions of the frames\n if writer is None:\n # Constructing code of the codec\n # to be used in the function VideoWriter\n fourcc = cv.VideoWriter_fourcc(*'mp4v')\n\n # Writing current processed frame into the video file\n # Pay attention! If you're using Windows, yours path might looks like:\n # r'videos\\result-traffic-cars.mp4'\n # or:\n # 'videos\\\\result-traffic-cars.mp4'\n writer = cv.VideoWriter('videos/result-traffic-cars.mp4', fourcc, 30,\n (frame.shape[1], frame.shape[0]), True)\n\n # Write processed current frame to the file\n writer.write(frame)\n \n # Printing final results\nprint()\nprint('Total number of frames', f)\nprint('Total amount of time {:.5f} seconds'.format(t))\nprint('FPS:', round((f / t), 1))\n\n\n# Releasing video reader and writer\nvideo.release()\nwriter.release()", "repo_name": "felipedepauli/crz-docs", "sub_path": "docs-ai/computer-vision/courses/UY_03_Train_YOLO_for_Object_Detection/02_Class/02_Yolo_on_Video_2.py", "file_name": "02_Yolo_on_Video_2.py", "file_ext": "py", "file_size_in_byte": 5634, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "cv2.dnn.readNetFromDarknet", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 32, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.dnn.blobFromImage", "line_number": 48, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 48, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 57, "usage_type": "call"}, {"api_name": "time.time", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 84, "usage_type": "call"}, {"api_name": "cv2.dnn.NMSBoxes", "line_number": 94, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 94, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 103, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 107, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 107, "usage_type": "attribute"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 115, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 122, "usage_type": "call"}]}
+{"seq_id": "74603593662", "text": "from itertools import combinations\n\nclass place:\n def __init__(self, r, c):\n self.r, self.c = r,c\n def cal_dist(self, other):\n return abs(other.r-self.r)+abs(other.c-self.c)\n\nn, m = map(int, input().split())\ntown = [list(map(int, input().split())) for _ in range(n)]\nchicken_list = []\nhome_list = []\n\nfor i in range(n):\n for j in range(n):\n if town[i][j] == 0x01:\n home_list.append(place(i, j))\n elif town[i][j] == 0x02:\n chicken_list.append(place(i, j))\n\ncom_list = combinations(chicken_list, m)\nchicken_dist = 10000\nfor com in com_list:\n ch = sum([min([h.cal_dist(x) for x in com]) for h in home_list])\n chicken_dist = ch if ch < chicken_dist else chicken_dist\n \nprint(chicken_dist)", "repo_name": "d2h10s/cote", "sub_path": "baekjoon/b_15686_치킨배달.py", "file_name": "b_15686_치킨배달.py", "file_ext": "py", "file_size_in_byte": 756, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "itertools.combinations", "line_number": 21, "usage_type": "call"}]}
+{"seq_id": "550377033", "text": "import large_image\nimport mysql.connector\nimport numpy as np\nimport cv2\nfrom scipy.misc import imsave\n\n\ninputImageFile =\"/home/sanghoon/docker-py/hmlWeb/TCGA-3C-AALJ-01Z-00-DX1.svs.dzi.tif\"\nslideName = 'TCGA-3C-AALJ-01Z-00-DX1'\n\nleft = 50000\ntop = 35000\nwidth = 2000\nheight = 2000\nbottom = top + height\nright = left + width\n\nbold = 512\nbold_left = left - bold\nbold_top = top - bold\nbold_bottom = bottom + bold\nbold_right = right + bold\nbold_width = width + 2*bold\nbold_height = height + 2*bold\n\nts = large_image.getTileSource(inputImageFile)\n\nregion = dict(\n left=left, top=top,\n width=width, height=height,\n)\n\nim_region = ts.getRegion(\n region=region, format=large_image.tilesource.TILE_FORMAT_NUMPY\n)[0]\n\nmydb = mysql.connector.connect(\n host=\"localhost\",\n user=\"guest\",\n passwd=\"guest\",\n database=\"nuclei\",\n charset='utf8',\n use_unicode=True\n)\n\nboundaryTablename = 'sregionboundaries'\n\nruncursor = mydb.cursor()\n\nquery = 'SELECT boundary from ' + boundaryTablename + ' where slide=\"' + slideName + \\\n'\" AND centroid_x BETWEEN ' + str(left) + ' AND ' + str(right) + \\\n' AND centroid_y BETWEEN ' + str(top) + ' AND ' + str(bottom)\n\nruncursor.execute(query)\n\nboundarySet = runcursor.fetchall()\n\n# set an array for boundary points in a region to zero\n# boundaryPoints = np.zeros((1, 2), dtype=np.int32)\nboundaryPoints = []\n# b_index = 0\nfor b in boundarySet:\n object = b[0].encode('utf-8').split(' ')\n object_points = []\n for p in range(len(object)-1):\n intP = map(int, object[p].split(','))\n intP[0] = intP[0] - left + bold\n intP[1] = intP[1] - top + bold\n object_points.append(intP)\n boundaryPoints.append(np.asarray(object_points))\n\nim_bold = np.zeros((bold_width, bold_height), dtype=np.uint8)\n\ncv2.fillPoly(im_bold, boundaryPoints, 255)\n\nim_out = im_bold[bold:bold+width, bold:bold+width]\n\nimsave('./test.png', im_out)\n", "repo_name": "slee172/HistomicsML-TA-old", "sub_path": "predict-rest-api/validate.py", "file_name": "validate.py", "file_ext": "py", "file_size_in_byte": 1863, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "large_image.getTileSource", "line_number": 26, "usage_type": "call"}, {"api_name": "large_image.tilesource", "line_number": 34, "usage_type": "attribute"}, {"api_name": "mysql.connector.connector.connect", "line_number": 37, "usage_type": "call"}, {"api_name": "mysql.connector.connector", "line_number": 37, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 37, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 72, "usage_type": "attribute"}, {"api_name": "cv2.fillPoly", "line_number": 74, "usage_type": "call"}, {"api_name": "scipy.misc.imsave", "line_number": 78, "usage_type": "call"}]}
+{"seq_id": "25570126385", "text": "from PyQt5.QtGui import *\nfrom PyQt5.QtCore import *\nfrom PyQt5.QtWidgets import *\nfrom PyQt5.uic import loadUi\nfrom splitter_module import group, member\nfrom datetime import datetime\nimport matplotlib.pyplot as plt\nimport os, sys\nimport pickle\nimport pandas as pd\nimport numpy as np\nimport splitter_module as SM\n\n######################################################################################################################################################################################\n#GUI class for Change/Delete group window\nclass CDGwindow(QWidget):\n def __init__(self):\n super(CDGwindow,self).__init__();\n loadUi('./gui/qtgui/cdgw.ui',self);\n self.setWindowTitle(\"Select Group\");\n self.setStyleSheet(\"QWidget {background: \"+color+\";}\");\n######################################################################################################################################################################################\n#GUI class for payment window\nclass Pwindow(QWidget):\n def __init__(self):\n super(Pwindow,self).__init__();\n loadUi('./gui/qtgui/payment.ui',self);\n self.setWindowTitle(\"Payment\");\n self.setStyleSheet(\"QWidget {background: \"+color+\";}\");\n######################################################################################################################################################################################\n#GUI class for expense window\nclass Ewindow(QWidget):\n def __init__(self):\n super(Ewindow,self).__init__();\n loadUi('./gui/qtgui/expense.ui',self);\n self.setWindowTitle(\"New Expense\");\n self.setStyleSheet(\"QWidget {background: \"+color+\";}\");\n######################################################################################################################################################################################\n#GUI class for Empty window\nclass EmptyWindow(QWidget):\n def __init__(self):\n super(EmptyWindow,self).__init__();\n loadUi('./gui/qtgui/empty.ui',self);\n self.setStyleSheet(\"QWidget {background: \"+color+\";}\");\n######################################################################################################################################################################################\n#GUI class for Settings window\nclass SettingsWindow(QDialog):\n def __init__(self):\n super(SettingsWindow,self).__init__();\n loadUi('./gui/qtgui/settings.ui',self);\n self.setWindowTitle(\"Settings\");\n self.setStyleSheet(\"QDialog {background: \"+color+\";}\");\n######################################################################################################################################################################################\n#GUI class for main window\nclass mainWindow(QMainWindow):\n def __init__(self):\n #Load UI file\n super(mainWindow, self).__init__();\n loadUi('./gui/qtgui/splitter.ui',self);\n\n #Set startup text\n self.setWindowTitle(\"SpLiTtEr!!!\");\n self.disclaimer.setText(\"Welcome!!!\");\n\n #Mapping button to functions\n self.newGroupButton.clicked.connect(self.newGroup);\n self.changeGroupButton.clicked.connect(self.selectGroup);\n self.deleteGroupButton.clicked.connect(self.selectGroup);\n self.newMemberButton.clicked.connect(self.newMember);\n self.newExpenseButton.clicked.connect(self.newExpense);\n self.newPaymentButton.clicked.connect(self.newPayment);\n self.suggestedPaymentsButton.clicked.connect(self.suggPayments);\n self.statsButton.clicked.connect(self.stats);\n self.settingsButton.clicked.connect(self.settings);\n self.historyButton.clicked.connect(self.history);\n\n #Set current group name in display\n if(current_group != None):\n self.groupName.setText(current_group.name.upper());\n \n #Pre-exit function before closing\n def closeEvent(self, event):\n global color, currency, current_group;\n setting_data = {};\n #Confirmation message box\n reply = QMessageBox.question(self, 'Quit', 'Are You Sure to Quit?', QMessageBox.No | QMessageBox.Yes);\n if reply == QMessageBox.Yes:\n #Store the objects into memory\n with open('./data/record.pkl', 'wb') as output:\n pickle.dump(record, output, pickle.HIGHEST_PROTOCOL);\n\n #Get the objects to be stored in a dict\n setting_data['current_group'] = current_group;\n setting_data['currency'] = currency;\n setting_data['color'] = color;\n\n #Store the settings data into memory\n with open(\"./data/settings.pkl\",'wb') as output:\n pickle.dump(setting_data, output, pickle.HIGHEST_PROTOCOL);\n #Close all windows\n plt.close();\n event.accept();\n else:\n #Ignore if No is clicked\n event.ignore();\n\n #Group creation button API\n def newGroup(self):\n window.disclaimer.setText(\"\");\n input_window = QInputDialog();\n input_window.setOkButtonText(\"Create\");\n name, okPressed = input_window.getText(self,\"New Group\",\"Group Name: \");\n if(okPressed == True):\n #Condition check for empty input\n if (bool(name) == False):\n self.disclaimer.setText(\"Please enter a name\");\n #Condition check for same group name\n elif SM.isGroupPresent(name, group_list) == True:\n self.disclaimer.setText(\"Group Already Exists\");\n else:\n #Create group API call\n createGroup(name.lower());\n self.disclaimer.setText(\"Group \"+name+\" Added Successfully\");\n \n #Group change/delete button API\n def selectGroup(self):\n global group_list;\n window.disclaimer.setText(\"\");\n #Check if no groups are added\n if (SM.isGroupListEmpty(group_list) == True):\n self.disclaimer.setText(\"No Groups found!!!\");\n else:\n #Load change group UI\n self.nw = CDGwindow();\n #Map buttons to functions depending on the button clicked\n if (self.sender().objectName() == \"changeGroupButton\"):\n self.nw.buttonBox.accepted.connect(changeGroup);\n if (self.sender().objectName() == \"deleteGroupButton\"):\n self.nw.buttonBox.accepted.connect(delGroup);\n self.nw.buttonBox.rejected.connect(self.nw.close);\n #Add elements dynamically to the list box\n for ele in group_list:\n self.nw.cb.addItem(ele.name);\n #Display GUI box\n self.nw.show();\n\n #Member addition button API\n def newMember(self):\n window.disclaimer.setText(\"\");\n #Check if no groups are added\n if (SM.isGroupListEmpty(group_list) == True):\n self.disclaimer.setText(\"No Groups found!!!\");\n else:\n input_window = QInputDialog();\n input_window.setOkButtonText(\"Add\");\n name, okPressed = input_window.getText(self,\"New Member\",\"Member Name: \");\n if(okPressed == True):\n #Condition check for empty input\n if (bool(name) == False):\n self.disclaimer.setText(\"Please enter a name\");\n #Condition check for same member name\n elif (SM.isMemberPresent(name.lower(),current_group)):\n self.disclaimer.setText(\"Member Already Exists\");\n else:\n #Member addition API call\n current_group.addMember(name.lower());\n #Set display element text\n self.disclaimer.setText(name+\" added to Group \"+current_group.name);\n self.displaySummary();\n\n #New expense addition button API\n def newExpense(self):\n window.disclaimer.setText(\"\");\n #List to hold all credits of all members\n creds = {};\n #Function for placeholder text\n def updatePlaceholder():\n #Read amount entered\n amt = self.nw.amtInput.text();\n #Amount validation\n if (SM.isAmountValid(amt) == True):\n amt = float(amt);\n equal = round(amt/(current_group.size),2);\n #Check if checkbox is selected\n if (window.nw.edCheckBox.isChecked() == True):\n for (k,v) in self.lines.items():\n v.setPlaceholderText(str(equal)+\"+\");\n elif (window.nw.perCheckBox.isChecked() == True):\n for (k,v) in self.lines.items():\n v.setPlaceholderText(str(round(100.0/(current_group.size),2))+\"%\");\n else:\n for (k,v) in self.lines.items():\n v.setPlaceholderText(str(equal));\n \n\n #Function to change the label when check box is selected\n def state_changed(self):\n #Update plcaeholder text if selection is changed\n updatePlaceholder();\n #If equal+delta is selected\n if (window.nw.sender().objectName() == \"edCheckBox\"):\n if (window.nw.edCheckBox.isChecked() == True):\n window.nw.perCheckBox.setChecked(False);\n window.nw.layoutTitle.setText(\"Delta Difference\");\n #Set the placeholder text\n window.nw.amtInput.setPlaceholderText(\"Amount to be shared equally\");\n return;\n elif (window.nw.sender().objectName() == \"perCheckBox\"):\n if (window.nw.perCheckBox.isChecked() == True):\n window.nw.edCheckBox.setChecked(False);\n window.nw.layoutTitle.setText(\"Percentage Share\");\n #Clear the placeholder text\n window.nw.amtInput.setPlaceholderText(\"\");\n return;\n #If both of the options are not selected\n window.nw.layoutTitle.setText(\"Individual Share\");\n #Clear the placeholder text\n window.nw.amtInput.setPlaceholderText(\"\");\n return;\n \n #Function to perform the pre condition checks and prepare data to send to add expense API\n def getDebits(creds, amt):\n #List to hold all debits of all members\n debts = {};\n #Get the description\n des = self.nw.descripInput.text();\n #Loop through all Line Edits\n for mem in current_group.members:\n #Read data in each Line Edit\n ind_amt = self.lines[mem.name].text();\n #Check for empty LineEdit \n if (ind_amt == \"\"):\n debts[mem.name+\"_debit\"] = 0;\n #Check for numeric value of the LineEdit\n elif (ind_amt.isnumeric() == True):\n debts[mem.name+\"_debit\"] = (-1*float(ind_amt));\n #Invalid amount disclaimer\n else:\n self.disclaimer.setText(\"Invalid amount\");\n del(self.nw);\n return;\n\n #Check if all LineEdits are empty, if empty then equal share\n if (all(val == 0 for val in list(debts.values()))):\n #Check if the percentage option is selected\n if (self.nw.perCheckBox.isChecked() == True):\n debts = debts.fromkeys(debts.keys(),(-1*(100/current_group.size)));\n else:\n debts = debts.fromkeys(debts.keys(),(-1*(amt/current_group.size)));\n\n #Check if the equal+delta option is selected\n if (self.nw.edCheckBox.isChecked() == True):\n #Create a temp dict with equal shares\n temp_debts = dict.fromkeys(debts.keys(),(-1*(amt/current_group.size)));\n #Add all the delta differences to respective members\n for k in debts.keys():\n temp_debts[k] += debts[k]; \n #Copy the temp dict to debts\n debts = temp_debts.copy();\n #Modify the amount\n amt = abs(sum(debts.values()));\n #Check if the percentage option is selected\n elif (self.nw.perCheckBox.isChecked() == True):\n #Share the amount percentage-wise\n for key in debts.keys():\n debts[key] = (debts[key]*amt)/100;\n\n #Check if all values add up to the entered amount\n if ((abs(abs(sum(list(debts.values()))) - amt) > 0.1) or (abs(sum(list(creds.values())) - amt) > 0.1)):\n msg = QMessageBox();\n msg.setIcon(QMessageBox.Critical)\n msg.setText(\"Shares do not add up to the total amount given.\");\n msg.setWindowTitle(\"Error!\")\n msg.exec_();\n else:\n #Call add expense API by passing all the required processed data\n current_group.addExpense(des, creds, amt, debts);\n self.disclaimer.setText(\"Expense added successfully.\");\n self.displaySummary();\n del(self.nw);\n return;\n\n def getPayers():\n #Lists to hold all checkbox and lineedit objects\n self.checkboxes = [];\n self.lineedits = [];\n #Create a settle window\n self.pw = EmptyWindow();\n self.pw.setWindowTitle(\"Paid by\");\n #Add checkbox for each member\n for mem in current_group.members:\n #Create a checkbox\n self.pw.cb = QCheckBox(self.pw);\n self.pw.cb.setText(mem.name);\n #Create a LineEdit\n self.pw.le = QLineEdit(self.pw);\n self.pw.formLayout.addRow(self.pw.cb,self.pw.le);\n #Collect the checkboxes and lineedits for future reference\n self.checkboxes.append(self.pw.cb);\n self.lineedits.append(self.pw.le);\n\n #Map buttons and signals to the APIs\n self.pw.buttonBox.accepted.connect(getCreds);\n self.pw.buttonBox.rejected.connect(self.pw.close);\n #Display GUI box\n self.pw.show();\n\n def getCreds():\n #Get the amount\n amt = self.nw.amtInput.text();\n #Validate the amount\n if (SM.isAmountValid(amt) == False):\n self.disclaimer.setText(\"Invalid amount\");\n del(self.nw);\n return;\n #Convert amount to float\n amt = float(amt);\n\n #Create a credit library\n for mem in current_group.members:\n creds[mem.name+\"_credit\"] = 0;\n\n #Choose only the checkboxes which are selected\n for ele in self.checkboxes:\n if (ele.isChecked() == True):\n #Get and validate the individual amount\n num = self.lineedits[self.checkboxes.index(ele)].text();\n if (SM.isAmountValid(num) == False):\n self.disclaimer.setText(\"Invalid amount\");\n del(self.pw);\n return;\n else:\n #Convert the number to float\n num = float(num);\n #Update the credit dictionary\n creds[ele.text()+\"_credit\"] = num;\n\n #Get debits API\n getDebits(creds, amt); \n del(self.pw);\n return; \n\n #Variable to hold all lineedits\n self.lines = {};\n #Check if no groups are added\n if (SM.isGroupListEmpty(group_list) == True):\n self.disclaimer.setText(\"No Groups found!!!\");\n #Check if no members are present in a group\n elif SM.isMemberListEmpty(current_group):\n self.disclaimer.setText(\"No Members in the group\");\n else:\n #Load change group UI\n self.nw = Ewindow();\n self.nw.buttonBox.button(QDialogButtonBox.Ok).setText(\"Add\");\n #Dynamically add row for each member\n for mem in current_group.members:\n #Add LineEdit for each member\n self.lines[mem.name] = QLineEdit();\n self.nw.formLayout.addRow(QLabel(mem.name),self.lines[mem.name]);\n #Assigning placeholder text\n self.nw.amtInput.textChanged.connect(updatePlaceholder);\n\n #Map button to APIs\n self.nw.buttonBox.accepted.connect(getPayers);\n self.nw.buttonBox.rejected.connect(self.nw.close);\n self.nw.perCheckBox.stateChanged.connect(state_changed);\n self.nw.edCheckBox.stateChanged.connect(state_changed);\n \n #Display GUI box\n self.nw.show();\n\n #Method to call payment API\n def newPayment(self):\n window.disclaimer.setText(\"\");\n #Pre process data to call payment API\n def processPayment():\n #Get the payer and check if the member exists\n frm = self.nw.fromInput.currentText().lower();\n if (SM.isMemberPresent(frm, current_group) == False):\n self.disclaimer.setText(frm+\" not found in group \"+current_group.name);\n del(self.nw);\n return;\n\n #Get the payee and check if the member exists\n to = self.nw.toInput.currentText().lower();\n if (SM.isMemberPresent(to, current_group) == False):\n self.disclaimer.setText(to+\" not found in group \"+current_group.name);\n del(self.nw);\n return;\n\n #validate the amount entered\n amt = self.nw.amtInput.text();\n if (SM.isAmountValid(amt) == False):\n self.disclaimer.setText(\"Invalid amount\");\n del(self.nw);\n return;\n #Convert amount to float\n amt = float(amt);\n\n #Call payment addition API\n current_group.addPayment(frm,to,amt);\n self.displaySummary();\n #Set display elements after successful transaction\n self.disclaimer.setText(\"Payment from \"+frm+\" to \"+to+\" successfull\");\n del(self.nw);\n\n #Method to update the comboBox with members other than from member\n def updateComboBox():\n #Clear the combo box\n self.nw.toInput.clear();\n #Include all members except the selected one\n for mem in current_group.members:\n if(self.nw.fromInput.currentText().lower() != mem.name):\n self.nw.toInput.addItem(mem.name);\n \n #Check if no groups are added\n if (SM.isGroupListEmpty(group_list) == True):\n self.disclaimer.setText(\"No Groups found!!!\");\n #Check if no members are present in a group\n elif SM.isMemberListEmpty(current_group):\n self.disclaimer.setText(\"No Members in the group\");\n else:\n #Open the payment window\n self.nw = Pwindow();\n self.nw.buttonBox.button(QDialogButtonBox.Ok).setText(\"Pay\");\n #Add all member names into the from combobox\n for mem in current_group.members:\n self.nw.fromInput.addItem(mem.name);\n #Initial call to the function\n updateComboBox();\n #Map buttons and signals to the APIs\n self.nw.fromInput.currentIndexChanged.connect(updateComboBox);\n self.nw.buttonBox.accepted.connect(processPayment);\n self.nw.buttonBox.rejected.connect(self.nw.close);\n #Display GUI box\n self.nw.show();\n\n #Summary display function\n def displaySummary(self):\n disp_data = current_group.summary(); \n disp = \"\";\n for (k,v) in disp_data.items():\n disp += (k+\"\\t\\t\\t\\t \"+str(v)+\" \"+currency+\"\\n\");\n self.displayText.setText(\"\");\n self.displayText.setText(disp);\n\n #Settle API\n def suggPayments(self):\n window.disclaimer.setText(\"\");\n global settle_data;\n self.settle_lines = [];\n #Check if no groups are added\n if (SM.isGroupListEmpty(group_list) == True):\n self.disclaimer.setText(\"No Groups found!!!\");\n #Check if no members are present in a group\n elif SM.isMemberListEmpty(current_group):\n self.disclaimer.setText(\"No Members in the group\");\n else:\n #Get the settle status\n message, settle_data = current_group.suggestedPayments();\n #Condition check if the accounts are settled\n if (message == None):\n msg = QMessageBox();\n msg.setWindowTitle(\"Suggested Payments\");\n msg.setText(\"Dues already settled\");\n msg.exec_();\n else:\n #Create a settle window\n self.nw = EmptyWindow();\n self.nw.setWindowTitle(\"Suggested Payments\");\n #Add checkbox for each of the transactions\n for ele in settle_data:\n self.nw.cb = QCheckBox(self.nw);\n self.nw.cb.setText(ele['From'].upper()+ \" pays to \"+ele['To'].upper()+\"\\t \"+currency+\" \"+str(ele['amount']));\n self.nw.formLayout.addRow(self.nw.cb);\n #Collect the checkboxes for future reference\n self.settle_lines.append(self.nw.cb);\n\n #Map buttons and signals to the APIs\n self.nw.buttonBox.button(QDialogButtonBox.Ok).setText(\"Settle\");\n self.nw.buttonBox.accepted.connect(settleUp);\n self.nw.buttonBox.rejected.connect(self.nw.close);\n #Display GUI box\n self.nw.show();\n \n #Function to plot statistics\n def stats(self):\n window.disclaimer.setText(\"\");\n #Check if no groups are added\n if (SM.isGroupListEmpty(group_list) == True):\n self.disclaimer.setText(\"No Groups found!!!\");\n else:\n global file;\n #Read the csv file\n csv_data = SM.getCsvFile(file);\n #Collect all members as labels\n labels = [mem.name for mem in current_group.members];\n #Collect expenses of each member\n expenses = [mem.expenses for mem in current_group.members];\n \n #Pie-Chart\n plt.subplot(221);\n #Add all relevant titles to the plot\n plt.title(\"Total spends\\n\"+str(sum(expenses)));\n #Collect the data to be visualised\n sizes = [((val/sum(expenses))*100) for val in expenses]; \n explode = [0.05 for s in sizes];\n #Plotting a pie chart\n plt.pie(sizes, labels=labels, explode=explode, autopct='%1.1f%%', startangle=90);\n\n #Bar-Graph\n plt.subplot(222);\n #Add all relevant titles to the plot\n plt.title('Expenses per member');\n plt.xlabel('Members');\n plt.ylabel('Expenses');\n #Collect the data to be visualised\n x = np.arange(len(labels));\n plt.xticks(x, labels);\n spends_all = [mem.spend_count for mem in current_group.members];\n #Plot the bars\n plt.bar(x, spends_all, width=0.4, edgecolor='white');\n\n #Multiple Bar-Graph\n plt.subplot(212);\n #Add all relevant titles to the plot\n plt.title(\"Personal expenditure\");\n plt.xlabel('Months');\n plt.ylabel('Amount');\n #Collect the data to be visualised\n months = ['Jan','Feb','Mar','Apr','May','Jun','Jul','Aug','Sep','Oct','Nov','Dec'];\n data = {};\n bars = [];\n x = np.arange(len(months));\n plt.xticks(x, months);\n width = 0.15;\n #Create a dictionary for each member for all months\n for mem in labels:\n data[mem] = {};\n for m in months:\n data[mem][m] = 0; \n \n #Loop through all rows and collect expenses for all members in a selected month\n for idx, row in csv_data.iterrows():\n #Get the current month\n curr_month = datetime.strptime(row['Timestamp'],\"%Y-%m-%d %H:%M:%S.%f\").strftime(\"%b\");\n #Check debit and add it to the dictionary for corresponding month\n for mem in labels:\n data[mem][curr_month] += abs(row[mem+\"_debit\"]);\n \n #Position adjust for plotting the bars\n pos_adjust = x-(int(len(labels)/2)*width);\n #Plot all bars and collect then in bars\n for (k,v) in data.items():\n bars.append(plt.bar(pos_adjust,list(v.values()),width,edgecolor='white'));\n pos_adjust += width;\n\n #Show legend\n plt.legend(bars,labels);\n\n #Display the plot\n plt.show();\n\n #Method to apply settings\n def settings(self):\n window.disclaimer.setText(\"\");\n self.nw = SettingsWindow();\n #Map buttons and signals to the APIs\n self.nw.buttonBox.button(QDialogButtonBox.Ok).setText(\"Apply\");\n self.nw.colorButton.clicked.connect(openColorDialog);\n self.nw.buttonBox.accepted.connect(applySettings);\n self.nw.buttonBox.rejected.connect(self.nw.close);\n #Display GUI box\n self.nw.show();\n\n #Method to show transaction history\n def history(self):\n #Define exiting method\n def closeHistory():\n self.nw.close;\n del(self.nw);\n\n window.disclaimer.setText(\"\");\n #Check if no groups are added\n if (SM.isGroupListEmpty(group_list) == True):\n self.disclaimer.setText(\"No Groups found!!!\");\n else:\n #Create a new window\n self.nw = EmptyWindow();\n self.nw.resize(830,500);\n self.nw.setWindowTitle(\"Transaction History\");\n #Delete unwanted elements\n del(self.nw.formLayout);\n del(self.nw.buttonBox);\n #Create a table widget and a button widget\n self.tableWidget = QTableWidget();\n self.button = QPushButton(\"OK\");\n #Read csv file\n csv_data = SM.getCsvFile(file);\n #Add columns - Date, Description, Amount, Paid by, Shares\n self.tableWidget.setRowCount(csv_data[\"Timestamp\"].count());\n self.tableWidget.setColumnCount(5);\n self.tableWidget.setHorizontalHeaderLabels([\"Date\",\"Description\",\"Amount\",\"Paid by\",\"Shares(\"+\",\".join([mem.name for mem in current_group.members])+\")\"]);\n\n #Loop through all rows and collect expenses for all members in a selected month\n for idx, row in csv_data.iterrows():\n #Local temp varaibles\n shares = [];\n payers = [];\n temp_row = dict(row);\n #Loop through all elements to get the share and amount\n for ele in row:\n #Enter only if the string is float\n try:\n ele = float(ele);\n #Check only for debit transactions\n if ele < 0:\n shares.append(abs(ele));\n #Check for credit transactions\n elif ele > 0:\n #Get payers names\n p = list(temp_row.keys())[list(temp_row.values()).index(ele)];\n #Set the element to 0 to avoid duplicate condition\n temp_row[p] = 0;\n #Append payers name to the list\n payers.append(p[:-7]);\n except:\n pass;\n\n #Add the data into the respective fields and set column width\n self.tableWidget.setColumnWidth(0, 100);\n self.tableWidget.setItem(idx,0, QTableWidgetItem(str(row[\"Timestamp\"][:10])));\n self.tableWidget.setColumnWidth(1, 200);\n self.tableWidget.setItem(idx,1, QTableWidgetItem(row[\"Description\"]));\n self.tableWidget.setColumnWidth(2, 100);\n self.tableWidget.setItem(idx,2, QTableWidgetItem(str(sum(shares))));\n self.tableWidget.setColumnWidth(3, 150);\n self.tableWidget.setItem(idx,3, QTableWidgetItem(\",\".join(payers)));\n self.tableWidget.setColumnWidth(4, 200);\n self.tableWidget.setItem(idx,4, QTableWidgetItem(\",\".join([str(s) for s in shares])));\n\n # Add box layout, add table to box layout and add box layout and button to widget\n self.nw.layout = QVBoxLayout();\n self.nw.layout.addWidget(self.tableWidget);\n self.nw.layout.addWidget(self.button);\n self.nw.setLayout(self.nw.layout);\n\n #Link button to function\n self.button.clicked.connect(closeHistory);\n\n # Show widget\n self.nw.show();\n\n######################################################################################################################################################################################\n#Method to create a new group\ndef createGroup(name):\n global object_count, current_group, file;\n #Creating new group object\n new_Group = group(name);\n #Add group object to the lists\n group_list.append(new_Group);\n #Get appropriate file name for the group\n file = \"./reports/\"+name+\".csv\";\n #Add group object and data file into a dictionary for storage\n record[object_count] = new_Group;\n #Update object_counter\n object_count += 1;\n #Create a fresh csv file for the group\n columns = ['Timestamp', 'Description'];\n df = pd.DataFrame(columns=columns);\n df.to_csv(file, index=False);\n #Load the new group as current group\n current_group = new_Group;\n #Set current group name in display text\n window.groupName.setText(current_group.name.upper());\n window.displayText.setText(\"\");\n\n#Method to carryout object loading functionality\ndef initilize_data():\n global group_list, current_group, record, object_count, file, currency, color;\n #Condition check for file existence\n if(os.path.exists('./data/record.pkl')):\n #Read the pickle file\n with open('./data/record.pkl', 'rb') as infile:\n #Try-catch block to handle empty file errors\n try:\n loaded_data = pickle.load(infile);\n except EOFError:\n return;\n #Condition check to handle empty record dictionary\n if bool(loaded_data):\n #Load all data to current session\n group_list = [val for (key,val) in loaded_data.items()];\n record = loaded_data;\n object_count = len(record.keys());\n\n #Condition check for file existence\n if(os.path.exists('./data/settings.pkl')):\n #Read the pickle file\n with open('./data/settings.pkl', 'rb') as infile:\n #Try-catch block to handle empty file errors\n try:\n loaded_data = pickle.load(infile);\n except EOFError:\n return;\n #Condition check to handle empty record dictionary\n if bool(loaded_data):\n currency = loaded_data['currency'];\n current_group = loaded_data['current_group'];\n color = loaded_data['color'];\n if (current_group != None):\n file = \"./reports/\"+current_group.name+\".csv\"; \n\n#Method to change group context\ndef changeGroup():\n #Get group name from the GUI\n name = window.nw.cb.currentText().lower();\n global current_group, file, group_list;\n #Change group context\n for obj in group_list:\n if (obj.name == name):\n #Load the respective object\n current_group = obj;\n #Load the working csv file\n file = \"./reports/\"+name+\".csv\"; \n break;\n \n #Close the CDGwindow post operation\n window.nw.close();\n del(window.nw);\n #Set display elements in the GUI\n window.groupName.setText(current_group.name.upper());\n window.disclaimer.setText(\"Group changed to \"+current_group.name);\n #Call Summary to update the info\n window.displaySummary();\n\n#Method to delete a group\ndef delGroup():\n global current_group, file, group_list, record, object_count;\n #Get group name from the GUI\n name = window.nw.cb.currentText().lower();\n #Delete group confirmation message box\n reply = QMessageBox.question(window, 'Delete Group', 'Are You Sure to Delete the group?', QMessageBox.No | QMessageBox.Yes);\n if reply == QMessageBox.Yes:\n #Deleting the group and respective data\n for obj in group_list:\n if (obj.name == name):\n #Remove object from the list\n deleted_obj = group_list.pop(group_list.index(obj));\n #Update the record dictionary\n record = {key: value for key, value in record.items() if value is not deleted_obj};\n #Re-assign appropriate key values\n object_count = 0;\n keys = list(record.keys());\n for i in range(len(keys)):\n record[object_count] = record.pop(keys[i]);\n object_count += 1;\n #Delete the group object and csv file\n os.remove(\"./reports/\"+name+\".csv\");\n del(deleted_obj);\n #Disclaimer if group is empty after deletion\n if (SM.isGroupListEmpty(group_list) == True):\n current_group = None;\n else: \n #Change active group\n current_group = group_list[0];\n file = \"./reports/\"+current_group.name+\".csv\"; \n #current_group.summary();\n break;\n #Close CDG window\n del(window.nw);\n #Set display element in GUI post deletion\n window.disclaimer.setText(\"Group \"+name+ \" deleted\");\n if(current_group != None):\n window.groupName.setText(current_group.name.upper());\n #Call Summary to update the info\n window.displaySummary();\n else:\n window.groupName.setText(\"\");\n window.displayText.setText(\"\");\n\n#Method to settle up\ndef settleUp():\n global settle_data;\n #Collect indices of payments made\n to_rem = [];\n #Add individual payments to settle all\n for line in window.settle_lines:\n #Add payments to only selected choices\n if (line.isChecked() == True):\n #Get the selected line\n selected_line = settle_data[window.settle_lines.index(line)];\n #Payment API call\n current_group.addPayment(selected_line['From'], selected_line['To'], selected_line['amount']);\n #Track the index to remove later\n to_rem.append(window.settle_lines.index(line));\n #Update the disclaimer text\n window.disclaimer.setText(\"Payment from \"+selected_line['From']+\" To \"+selected_line['To']+\" successful.\");\n #Remove the data after the accounts ahve been settled\n for i in range(len(to_rem)-1,-1,-1):\n del settle_data[to_rem[i]];\n del window.settle_lines[to_rem[i]];\n #Call Summary to update the info\n window.displaySummary();\n #Delete the new window\n del(window.nw);\n\n#Method to open color Dialog\ndef openColorDialog():\n global color;\n #Get the selected color\n color = (QColorDialog.getColor()).name();\n\n#Method to apply the settings\ndef applySettings():\n global currency, color;\n #Check if any group is selected\n if (current_group != None):\n #Get the amount debit/credit for each member\n amt_data = list(current_group.summary().values());\n #Check if all accounts are settled\n if (SM.duesSettled(amt_data) == True):\n #Get selected currency\n currency = window.nw.currencyBox.currentText();\n window.displaySummary();\n else:\n window.disclaimer.setText(\"Currency cannot be changed\");\n\n #Apply the color theme\n window.setStyleSheet(\"QMainWindow {background: \"+color+\";}\");\n window.disclaimer.setText(\"Settings Applied\");\n #Close and delete the window\n del(window.nw);\n\n######################################################################################################################################################################################\n#Main function\nif __name__ == '__main__':\n \"\"\"Global declarations\"\"\"\n #Dictionary to hold objects to be saved\n record = {};\n object_count = 0;\n #Group instances monitoring lists\n group_list = [];\n #Active group\n current_group = None;\n currency = \"\";\n color = \"#F0F0F0\";\n #Settle data list\n settle_data = [];\n #global object_count, current_group, group_list, group_name_list;\n initilize_data();\n \n #Start GUI\n app = QApplication(sys.argv);\n window = mainWindow();\n #Apply color to the GUI\n window.setStyleSheet(\"QMainWindow {background: \"+color+\";}\");\n #Display GUI\n window.show();\n #Call summary to initiate display\n if (current_group != None):\n window.displaySummary();\n sys.exit(app.exec_());", "repo_name": "sourabh061295/Splitter", "sub_path": "src/splitter_gui.py", "file_name": "splitter_gui.py", "file_ext": "py", "file_size_in_byte": 37567, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "PyQt5.uic.loadUi", "line_number": 19, "usage_type": "call"}, {"api_name": "PyQt5.uic.loadUi", "line_number": 27, "usage_type": "call"}, {"api_name": "PyQt5.uic.loadUi", "line_number": 35, "usage_type": "call"}, {"api_name": "PyQt5.uic.loadUi", "line_number": 43, "usage_type": "call"}, {"api_name": "PyQt5.uic.loadUi", "line_number": 50, "usage_type": "call"}, {"api_name": "PyQt5.uic.loadUi", "line_number": 59, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 90, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 90, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 99, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 99, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.close", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "splitter_module.isGroupPresent", "line_number": 118, "usage_type": "call"}, {"api_name": "splitter_module.isGroupListEmpty", "line_number": 130, "usage_type": "call"}, {"api_name": "splitter_module.isGroupListEmpty", "line_number": 151, "usage_type": "call"}, {"api_name": "splitter_module.isMemberPresent", "line_number": 162, "usage_type": "call"}, {"api_name": "splitter_module.isAmountValid", "line_number": 181, "usage_type": "call"}, {"api_name": "splitter_module.isAmountValid", "line_number": 312, "usage_type": "call"}, {"api_name": "splitter_module.isAmountValid", "line_number": 328, "usage_type": "call"}, {"api_name": "splitter_module.isGroupListEmpty", "line_number": 346, "usage_type": "call"}, {"api_name": "splitter_module.isMemberListEmpty", "line_number": 349, "usage_type": "call"}, {"api_name": "splitter_module.isMemberPresent", "line_number": 379, "usage_type": "call"}, {"api_name": "splitter_module.isMemberPresent", "line_number": 386, "usage_type": "call"}, {"api_name": "splitter_module.isAmountValid", "line_number": 393, "usage_type": "call"}, {"api_name": "splitter_module.isGroupListEmpty", "line_number": 417, "usage_type": "call"}, {"api_name": "splitter_module.isMemberListEmpty", "line_number": 420, "usage_type": "call"}, {"api_name": "splitter_module.isGroupListEmpty", "line_number": 453, "usage_type": "call"}, {"api_name": "splitter_module.isMemberListEmpty", "line_number": 456, "usage_type": "call"}, {"api_name": "splitter_module.isGroupListEmpty", "line_number": 490, "usage_type": "call"}, {"api_name": "splitter_module.getCsvFile", "line_number": 495, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 502, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 502, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 504, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 504, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pie", "line_number": 509, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 509, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 512, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 512, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 514, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 514, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 515, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 515, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 516, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 516, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 518, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 519, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 519, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 522, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 522, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 525, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 525, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 527, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 527, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 528, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 528, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 529, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 529, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 534, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 535, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 535, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 546, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 546, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 555, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 555, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 559, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 559, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 562, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 562, "usage_type": "name"}, {"api_name": "splitter_module.isGroupListEmpty", "line_number": 585, "usage_type": "call"}, {"api_name": "splitter_module.getCsvFile", "line_number": 599, "usage_type": "call"}, {"api_name": "splitter_module.group", "line_number": 659, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 670, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 682, "usage_type": "call"}, {"api_name": "os.path", "line_number": 682, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 687, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 698, "usage_type": "call"}, {"api_name": "os.path", "line_number": 698, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 703, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 759, "usage_type": "call"}, {"api_name": "splitter_module.isGroupListEmpty", "line_number": 762, "usage_type": "call"}, {"api_name": "splitter_module.duesSettled", "line_number": 822, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 854, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 863, "usage_type": "call"}]}
+{"seq_id": "73157270142", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sun Nov 8 18:09:39 2020\r\n\r\n@author: NISHANT\r\n\"\"\"\r\n\r\n\r\n#Face Detection using haarcascade file \r\nimport cv2\r\nimport numpy\r\nface=cv2.CascadeClassifier(\"Data\\\\cascades\\\\haarcascade_frontalface_default.xml\") #for detecting face\r\neye = cv2.CascadeClassifier('Data\\\\cascades\\\\haarcascade_eye.xml') #for detecting eyes\r\n\r\nimage=cv2.imread(\"Data\\\\a.jpg\")\r\ngray= cv2.cvtColor(image,cv2.COLOR_BGR2GRAY) #convert into gray \r\n\r\n#parameters(img,scale_factor[reduce image size],min_neighbour)\r\nfaces = face.detectMultiScale(gray,4,4) #for faces\r\n\r\nfor(x,y,w,h) in faces:\r\n \r\n image=cv2.rectangle(image,(x,y),(x+w,y+h),(127,0,205),3)\r\n \r\n #Now detect eyes\r\n roi_gray = gray[y:y+h, x:x+w]\r\n roi_color = image[y:y+h, x:x+w]\r\n eyes = eye.detectMultiScale(roi_gray,1.2,1)\r\n for (ex,ey,ew,eh) in eyes:\r\n cv2.rectangle(roi_color,(ex,ey),(ex+ew,ey+eh),(255,0,0),2)\r\n \r\nimage = cv2.resize(image,(800,700))\r\ncv2.imshow(\"Face Detected\",image)\r\ncv2.waitKey(0)\r\ncv2.destroyAllWindows() \r\n", "repo_name": "askitlouder/Image-Processing-Tutorials", "sub_path": "43 - Face and Eyes Detection using Image.py", "file_name": "43 - Face and Eyes Detection using Image.py", "file_ext": "py", "file_size_in_byte": 1042, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 42, "dataset": "github-code", "pt": "24", "api": [{"api_name": "cv2.CascadeClassifier", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.CascadeClassifier", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 16, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 35, "usage_type": "call"}]}
+{"seq_id": "4362104436", "text": "from os.path import basename, isdir, isfile\nfrom pathlib import Path\nfrom shutil import copy2, copytree\n\nfrom util.build_info import BuildInfo\nfrom util.config import get_config_value, get_list_from_config_yaml\nfrom util.logger import LOGGER\n\n\ndef extract_single_worded_key(dictionary, key):\n \"\"\" verify that key is in dictionary and its value is a single word \"\"\"\n if key in dictionary:\n value = dictionary[key]\n if len(value.split()) == 1:\n return value\n raise RuntimeError('\\'{}\\' of injected file must be a single word, but got {}: \\'{}\\''\n .format(key, len(value.split()), value))\n\n raise RuntimeError('\\'{}\\' is not specified for injected file \\'{}\\' !'\n .format(key, dictionary))\n\n\ndef read_injected_files(overall_dest_dir):\n \"\"\"\n Copy file that need to be injected to temporary location,\n which will be accessible during post-install.\n One mandatory argument: a path to initrd directory that will be available during post_install\n \"\"\"\n artifacts_dir = get_config_value(\"ARTIFACTS_DIR\")\n\n # location used by post-install, should be created only if there are files to inject\n injected_files = 'etc/injected_files' # location used by post-install\n overall_dest_dir = overall_dest_dir + '/' + injected_files\n LOGGER.info('temporary location for injected files: %s', overall_dest_dir)\n\n # include user-specified files\n files_to_inject = get_list_from_config_yaml('UPDATE_IMAGE_FILES')\n\n # include information about installed software on the build machine\n build_info_file_name = \"build_info.json\"\n build_info_source = artifacts_dir + \"/\" + build_info_file_name\n build_info_destination = \"/\" + build_info_file_name\n files_to_inject.append({'source': build_info_source, 'destination': build_info_destination})\n build_info = BuildInfo()\n build_info.to_file(build_info_source)\n\n # each injected file directory to be stored in a separate directory \"file\"\n count = 0\n LOGGER.trace(\"files_to_inject: %s\", files_to_inject)\n for file in files_to_inject:\n LOGGER.trace(\"file: %s\", file)\n src = extract_single_worded_key(file, 'source')\n dest = extract_single_worded_key(file, 'destination')\n LOGGER.info('inject %s to temporary location %s', src, dest)\n\n file_holder = overall_dest_dir + '/file' + str(count) + '/'\n # copy source to \"src\"\n # source file name does not need to be preserved;\n # it will be copied to destination path on BIG-IP\n source_holder = file_holder + 'src'\n if isfile(src):\n Path(file_holder).mkdir(parents=True, exist_ok=True)\n copy2(src, source_holder)\n elif isdir(src):\n copytree(src, source_holder)\n else:\n raise RuntimeError('\\'{}\\' is neither a file nor a directory, cannot inject it!'\n .format(src))\n\n # store destination\n if dest[0] != '/':\n raise RuntimeError('injected file destination \\'{}\\' must be an absolute path!'\n .format(dest))\n with open(file_holder + 'dest', 'w') as dest_holder:\n print(\"{}\".format(dest), file=dest_holder)\n\n count += 1\n # end of for loop\n\n LOGGER.debug('leaving %s', basename(__file__))\n return 0\n", "repo_name": "perbonielsen/f5-bigip-image-generator", "sub_path": "src/lib/python/util/injected_files.py", "file_name": "injected_files.py", "file_ext": "py", "file_size_in_byte": 3366, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "github-code", "pt": "24", "api": [{"api_name": "util.config.get_config_value", "line_number": 29, "usage_type": "call"}, {"api_name": "util.logger.LOGGER.info", "line_number": 34, "usage_type": "call"}, {"api_name": "util.logger.LOGGER", "line_number": 34, "usage_type": "name"}, {"api_name": "util.config.get_list_from_config_yaml", "line_number": 37, "usage_type": "call"}, {"api_name": "util.build_info.BuildInfo", "line_number": 44, "usage_type": "call"}, {"api_name": "util.logger.LOGGER.trace", "line_number": 49, "usage_type": "call"}, {"api_name": "util.logger.LOGGER", "line_number": 49, "usage_type": "name"}, {"api_name": "util.logger.LOGGER.trace", "line_number": 51, "usage_type": "call"}, {"api_name": "util.logger.LOGGER", "line_number": 51, "usage_type": "name"}, {"api_name": "util.logger.LOGGER.info", "line_number": 54, "usage_type": "call"}, {"api_name": "util.logger.LOGGER", "line_number": 54, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 61, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 62, "usage_type": "call"}, {"api_name": "shutil.copy2", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 64, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 65, "usage_type": "call"}, {"api_name": "util.logger.LOGGER.debug", "line_number": 80, "usage_type": "call"}, {"api_name": "util.logger.LOGGER", "line_number": 80, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 80, "usage_type": "call"}]}
+{"seq_id": "72396044543", "text": "import logging\nimport os\nimport sys\nimport json\nfrom dataclasses import dataclass, field\nfrom typing import Optional\n\nimport datasets\nimport nltk # Here to have a nice missing dependency error message early on\nimport numpy as np\nfrom datasets import load_dataset\n\nimport transformers\nfrom filelock import FileLock\nfrom transformers import (\n AutoConfig,\n AutoModelForSeq2SeqLM,\n AutoModelForCausalLM, # add\n AutoTokenizer,\n HfArgumentParser,\n Seq2SeqTrainingArguments,\n set_seed, )\nfrom transformers.file_utils import is_offline_mode\nfrom transformers.trainer_utils import get_last_checkpoint\n\nfrom model.bloom import BloomForCausalLM_WithLoss\nfrom model.codegen import CodeGenForCausalLM_WithLoss\nfrom model.gpt_neox import GPTNeoXForCausalLM_WithLoss\n\nfrom uie_collator import DataCollatorForUIE\nfrom uie_dataset import gen_cache_path\n\nfrom uie_trainer import UIETrainer, DenserEvalCallback, skip_instructions\nfrom compute_metrics import compute_metrics, compute_grouped_metrics\n\n# off wandb\nos.environ['WANDB_DISABLED'] = \"True\"\n# os.environ['CUDA_VISIBLE_DEVICES'] = '0'\nlogger = logging.getLogger(__name__)\nCURRENT_DIR = os.path.dirname(__file__)\n\ntry:\n nltk.data.find(\"tokenizers/punkt\")\nexcept (LookupError, OSError):\n if is_offline_mode():\n raise LookupError(\n \"Offline mode: run this script without TRANSFORMERS_OFFLINE first to download nltk data files\"\n )\n with FileLock(\".lock\") as lock:\n nltk.download(\"punkt\", quiet=True)\n\n\n@dataclass\nclass ModelArguments:\n \"\"\"\n Arguments pertaining to which model/config/tokenizer we are going to fine-tune from.\n \"\"\"\n\n model_name_or_path: str = field(\n metadata={\"help\": \"Path to pretrained model or model identifier from huggingface.co/models\"}\n )\n config_name: Optional[str] = field(\n default=None, metadata={\"help\": \"Pretrained config name or path if not the same as model_name\"}\n )\n tokenizer_name: Optional[str] = field(\n default=None, metadata={\"help\": \"Pretrained tokenizer name or path if not the same as model_name\"}\n )\n cache_dir: Optional[str] = field(\n default=None,\n metadata={\"help\": \"Where to store the pretrained models downloaded from huggingface.co\"},\n )\n use_fast_tokenizer: bool = field(\n default=True,\n metadata={\"help\": \"Whether to use one of the fast tokenizer (backed by the tokenizers library) or not.\"},\n )\n model_revision: str = field(\n default=\"main\",\n metadata={\"help\": \"The specific model version to use (can be a branch name, tag name or commit id).\"},\n )\n use_auth_token: bool = field(\n default=False,\n metadata={\n \"help\": \"Will use the token generated when running `transformers-cli login` (necessary to use this script \"\n \"with private models).\"\n },\n )\n resize_position_embeddings: Optional[bool] = field(\n default=None,\n metadata={\n \"help\": \"Whether to automatically resize the position embeddings if `max_source_length` exceeds \"\n \"the model's position embeddings.\"\n },\n )\n\n\n@dataclass\nclass DataTrainingArguments:\n \"\"\"\n Arguments pertaining to what data we are going to input our model for training and eval.\n \"\"\"\n lang: str = field(default=None, metadata={\"help\": \"Language id for multilingual model.\"})\n data_dir: str = field(\n default=None, metadata={\"help\": \"The directory for saving the UIE train/dev/test splits.\"}\n )\n task_config_dir: str = field(\n default=None, metadata={\"help\": \"The json file for config training and testing tasks\"}\n )\n instruction_file: str = field(\n default=None, metadata={\"help\": \"The instruction file for different tasks.\"}\n )\n instruction_strategy: Optional[str] = field(\n default='single', metadata={\n \"help\": \"How many different instructions to use? Support 'single' and 'multiple' mode.\"\n }\n )\n overwrite_cache: bool = field(\n default=False, metadata={\"help\": \"Overwrite the cached training and evaluation sets\"}\n )\n input_record_file: str = field(\n default=None, metadata={\"help\": \"file to record model input\"}\n )\n preprocessing_num_workers: Optional[int] = field(\n default=None,\n metadata={\"help\": \"The number of processes to use for the preprocessing.\"},\n )\n max_source_length: Optional[int] = field(\n default=512,\n metadata={\n \"help\": \"The maximum total input sequence length after tokenization. Sequences longer \"\n \"than this will be truncated, sequences shorter will be padded.\"\n },\n )\n # for decoder model, it means max_new_tokens\n max_target_length: Optional[int] = field(\n default=50,\n metadata={\n \"help\": \"The maximum total sequence length for target text after tokenization. Sequences longer \"\n \"than this will be truncated, sequences shorter will be padded.\"\n },\n )\n repetition_penalty: Optional[float] = field(\n default=1.0,\n metadata={\n \"help\": \"Penalty for repeat tokens in decode stage.\"\n },\n )\n num_beams: Optional[int] = field(\n default=1,\n metadata={\n \"help\": \"Number of beams to use for evaluation. This argument will be passed to ``model.generate``, \"\n \"which is used during ``evaluate`` and ``predict``.\"\n },\n )\n max_num_instances_per_task: int = field(\n default=10000, metadata={\"help\": \"The maximum number of instances we will consider for each training task.\"}\n )\n max_num_instances_per_eval_task: int = field(\n default=200,\n metadata={\"help\": \"The maximum number of instances we will consider for each validation/test task.\"}\n )\n max_train_samples: Optional[int] = field(\n default=None,\n metadata={\n \"help\": \"For debugging purposes or quicker training, truncate the number of training examples to this \"\n \"value if set.\"\n },\n )\n max_eval_samples: Optional[int] = field(\n default=None,\n metadata={\n \"help\": \"For debugging purposes or quicker training, truncate the number of evaluation examples to this \"\n \"value if set.\"\n },\n )\n max_predict_samples: Optional[int] = field(\n default=None,\n metadata={\n \"help\": \"For debugging purposes or quicker training, truncate the number of prediction examples to this \"\n \"value if set.\"\n },\n )\n num_examples: Optional[int] = field(\n default=0,\n metadata={\"help\": \"number of in-context positive examples.\"}\n )\n ignore_pad_token_for_loss: bool = field(\n default=True,\n metadata={\n \"help\": \"Whether to ignore the tokens corresponding to padded labels in the loss computation or not.\"\n },\n )\n add_task_name: Optional[bool] = field(\n default=False,\n metadata={\"help\": \"whether to preappend task name before the task input.\"}\n )\n add_dataset_name: Optional[bool] = field(\n default=False,\n metadata={\"help\": \"whether to preappend dataset name before the task input.\"}\n )\n common_dataset_name: Optional[str] = field(\n default=None,\n metadata={\"help\": \"common dataset name for zero shot.\"}\n )\n over_sampling: Optional[str] = field(\n default=False,\n metadata={\"help\": \"Whether to over sampling the dataset to max_num_instances_per_task\"}\n )\n\n\n@dataclass\nclass UIETrainingArguments(Seq2SeqTrainingArguments):\n gradient_checkpointing: Optional[bool] = field(\n default=False,\n metadata={\"help\": \"Whether to use computing time to gain more memory\"}\n )\n denser_evaluation: Optional[bool] = field(\n default=False,\n metadata={\"help\": \"If specifid, the model will do more evaluation at the beginning of training.\"}\n )\n do_demo: bool = field(default=False, metadata={\"help\": \"Whether to run the model as a demo in the terminal.\"})\n\n\ndef main():\n # See all possible arguments in src/transformers/training_args.py\n # or by passing the --help flag to this script.\n # We now keep distinct sets of args, for a cleaner separation of concerns.\n\n parser = HfArgumentParser((ModelArguments, DataTrainingArguments, UIETrainingArguments))\n if len(sys.argv) == 2 and sys.argv[1].endswith(\".json\"):\n # If we pass only one argument to the script and it's the path to a json file,\n # let's parse it to get our arguments.\n model_args, data_args, training_args = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1]))\n else:\n model_args, data_args, training_args = parser.parse_args_into_dataclasses()\n\n # Setup logging\n logging.basicConfig(\n format=\"%(asctime)s - %(levelname)s - %(name)s - %(message)s\",\n datefmt=\"%m/%d/%Y %H:%M:%S\",\n handlers=[logging.StreamHandler(sys.stdout)],\n )\n log_level = training_args.get_process_log_level()\n logger.setLevel(log_level)\n datasets.utils.logging.set_verbosity(log_level)\n transformers.utils.logging.set_verbosity(log_level)\n transformers.utils.logging.enable_default_handler()\n transformers.utils.logging.enable_explicit_format()\n\n # Log on each process the small summary:\n logger.warning(\n f\"Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}\"\n + f\"distributed training: {bool(training_args.local_rank != -1)}, 16-bits training: {training_args.fp16}\"\n )\n logger.info(f\"Training/evaluation parameters {training_args}\")\n\n # Detecting last checkpoint.\n last_checkpoint = None\n if os.path.isdir(training_args.output_dir) and training_args.do_train and not training_args.overwrite_output_dir:\n last_checkpoint = get_last_checkpoint(training_args.output_dir)\n if last_checkpoint is None and len(os.listdir(training_args.output_dir)) > 0:\n raise ValueError(\n f\"Output directory ({training_args.output_dir}) already exists and is not empty. \"\n \"Use --overwrite_output_dir to overcome.\"\n )\n elif last_checkpoint is not None and training_args.resume_from_checkpoint is None:\n logger.info(\n f\"Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change \"\n \"the `--output_dir` or add `--overwrite_output_dir` to train from scratch.\"\n )\n\n # Set seed before initializing model.\n set_seed(training_args.seed)\n data_cache_dir = gen_cache_path(training_args.output_dir, data_args)\n\n # Get the UIE dataset\n raw_datasets = load_dataset(\n os.path.join(CURRENT_DIR, \"uie_dataset.py\"),\n data_dir=data_args.data_dir,\n task_config_dir=data_args.task_config_dir,\n instruction_file=data_args.instruction_file,\n instruction_strategy=data_args.instruction_strategy,\n cache_dir=data_cache_dir, # for debug, change dataset size, otherwise open it\n max_num_instances_per_task=data_args.max_num_instances_per_task,\n max_num_instances_per_eval_task=data_args.max_num_instances_per_eval_task,\n num_examples=data_args.num_examples,\n over_sampling=data_args.over_sampling\n )\n raw_datasets.cleanup_cache_files()\n\n # Load pretrained model and tokenizer\n #\n # Distributed training:\n # The .from_pretrained methods guarantee that only one local process can concurrently\n # download model & vocab.\n config = AutoConfig.from_pretrained(\n model_args.config_name if model_args.config_name else model_args.model_name_or_path,\n cache_dir=model_args.cache_dir,\n revision=model_args.model_revision,\n use_auth_token=True if model_args.use_auth_token else None,\n )\n tokenizer = AutoTokenizer.from_pretrained(\n model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path,\n cache_dir=model_args.cache_dir,\n use_fast=model_args.use_fast_tokenizer,\n revision=model_args.model_revision,\n use_auth_token=True if model_args.use_auth_token else None,\n )\n\n if 'bloom' in model_args.model_name_or_path.lower():\n model_class = BloomForCausalLM_WithLoss\n if tokenizer.pad_token is None:\n tokenizer.pad_token = tokenizer.eos_token\n tokenizer.padding_side = 'left'\n elif 'codegen' in model_args.model_name_or_path.lower():\n model_class = CodeGenForCausalLM_WithLoss\n tokenizer.pad_token = tokenizer.eos_token\n tokenizer.padding_side = 'left'\n\n elif 'neox' in model_args.model_name_or_path.lower(): # add neox\n model_class = GPTNeoXForCausalLM_WithLoss\n if tokenizer.pad_token is None:\n tokenizer.pad_token = tokenizer.eos_token\n tokenizer.padding_side = 'left'\n else:\n model_class = AutoModelForSeq2SeqLM\n model = model_class.from_pretrained(\n model_args.model_name_or_path,\n from_tf=bool(\".ckpt\" in model_args.model_name_or_path),\n config=config,\n cache_dir=model_args.cache_dir,\n revision=model_args.model_revision,\n use_auth_token=True if model_args.use_auth_token else None,\n )\n model.resize_token_embeddings(len(tokenizer))\n\n if (\n hasattr(model.config, \"max_position_embeddings\")\n and model.config.max_position_embeddings < data_args.max_source_length\n ):\n if model_args.resize_position_embeddings is None:\n logger.warning(\n f\"Increasing the model's number of position embedding vectors from {model.config.max_position_embeddings} \"\n f\"to {data_args.max_source_length}.\"\n )\n model.resize_position_embeddings(data_args.max_source_length)\n elif model_args.resize_position_embeddings:\n model.resize_position_embeddings(data_args.max_source_length)\n else:\n raise ValueError(\n f\"`--max_source_length` is set to {data_args.max_source_length}, but the model only has {model.config.max_position_embeddings}\"\n f\" position encodings. Consider either reducing `--max_source_length` to {model.config.max_position_embeddings} or to automatically \"\n \"resize the model's position encodings by passing `--resize_position_embeddings`.\"\n )\n\n if training_args.label_smoothing_factor > 0 and not hasattr(model, \"prepare_decoder_input_ids_from_labels\"):\n logger.warning(\n \"label_smoothing is enabled but the `prepare_decoder_input_ids_from_labels` method is not defined for\"\n f\"`{model.__class__.__name__}`. This will lead to loss being calculated twice and will take up more memory\"\n )\n\n if training_args.do_train:\n if \"train\" not in raw_datasets:\n raise ValueError(\"--do_train requires a train dataset\")\n train_dataset = raw_datasets[\"train\"]\n if data_args.max_train_samples is not None:\n train_dataset = train_dataset.select(range(data_args.max_train_samples))\n\n if training_args.do_eval:\n if \"validation\" not in raw_datasets:\n raise ValueError(\"--do_eval requires a validation dataset\")\n eval_dataset = raw_datasets[\"validation\"]\n if data_args.max_eval_samples is not None:\n eval_dataset = eval_dataset.select(range(data_args.max_eval_samples))\n\n if training_args.do_predict:\n if \"test\" not in raw_datasets:\n raise ValueError(\"--do_predict requires a test dataset\")\n predict_dataset = raw_datasets[\"test\"]\n if data_args.max_predict_samples is not None:\n predict_dataset = predict_dataset.select(range(data_args.max_predict_samples))\n\n # Data collator\n label_pad_token_id = -100 if data_args.ignore_pad_token_for_loss else tokenizer.pad_token_id\n data_collator = DataCollatorForUIE(\n tokenizer,\n model=model,\n padding=\"longest\",\n max_source_length=data_args.max_source_length,\n max_target_length=data_args.max_target_length,\n label_pad_token_id=label_pad_token_id,\n pad_to_multiple_of=8 if training_args.fp16 else None,\n add_task_name=data_args.add_task_name,\n add_dataset_name=data_args.add_dataset_name,\n common_dataset_name=data_args.common_dataset_name,\n num_examples=data_args.num_examples,\n input_record_file=data_args.input_record_file\n )\n # we don't want to remove unused columns because we will prepare each batch during training,\n # and some of the information will also be used in evaluation.\n training_args.remove_unused_columns = False\n\n # Metric\n def compute_rouge_metrics(dataset, preds, save_prefix=None):\n decoded_preds = skip_instructions(model, preds, tokenizer)\n references = [e[\"Instance\"][\"label\"] for e in dataset]\n result = compute_metrics(predictions=decoded_preds, references=references)\n result_per_task = compute_grouped_metrics(predictions=decoded_preds, references=references,\n groups=dataset[\"Task\"])\n result.update(result_per_task)\n categories = dataset[\"Dataset\"]\n result_per_category = compute_grouped_metrics(predictions=decoded_preds, references=references,\n groups=categories)\n result.update(result_per_category)\n prediction_lens = [np.count_nonzero(pred != tokenizer.pad_token_id) for pred in preds]\n result[\"gen_len\"] = np.mean(prediction_lens)\n result = {k: round(v, 4) for k, v in result.items()}\n if save_prefix is not None:\n with open(os.path.join(training_args.output_dir, f\"{save_prefix}_eval_predictions.jsonl\"), \"w\") as fout:\n for example, pred in zip(dataset, decoded_preds):\n fout.write(json.dumps({\n \"Task\": example[\"Task\"],\n \"Dataset\": example[\"Dataset\"],\n \"Instance\": example[\"Instance\"],\n \"Prediction\": pred\n }) + \"\\n\")\n return result\n\n print(f\"-----Gradient checkpointing: {training_args.gradient_checkpointing} -----\")\n if training_args.gradient_checkpointing:\n model.gradient_checkpointing_enable()\n\n trainer = UIETrainer(\n model=model,\n args=training_args,\n train_dataset=train_dataset if training_args.do_train else None,\n eval_dataset=eval_dataset if training_args.do_eval else None,\n tokenizer=tokenizer,\n data_collator=data_collator,\n compute_metrics=compute_rouge_metrics,\n callbacks=[DenserEvalCallback] if training_args.denser_evaluation else None\n )\n\n all_metrics = {\"run_name\": training_args.run_name}\n\n # Training\n # 训练epoch数,按照 num_train_epochs 传入,在trainer中解析\n # TODO, train debug, bloomz, flan-t5\n if training_args.do_train:\n checkpoint = None\n if training_args.resume_from_checkpoint is not None:\n checkpoint = training_args.resume_from_checkpoint\n elif last_checkpoint is not None:\n checkpoint = last_checkpoint\n train_result = trainer.train(resume_from_checkpoint=checkpoint)\n trainer.save_model() # Saves the tokenizer too for easy upload\n\n metrics = train_result.metrics\n max_train_samples = (\n data_args.max_train_samples if data_args.max_train_samples is not None else len(train_dataset)\n )\n metrics[\"train_samples\"] = min(max_train_samples, len(train_dataset))\n\n trainer.log_metrics(\"train\", metrics)\n trainer.save_metrics(\"train\", metrics)\n trainer.save_state()\n logger.info(f\"Metrics {metrics}\")\n all_metrics.update(metrics)\n\n # Evaluation\n results = {}\n # in case the batch is shorter than max length, the output should be padded\n max_new_tokens = (\n training_args.generation_max_length\n if training_args.generation_max_length is not None\n else data_args.max_target_length\n )\n\n num_beams = data_args.num_beams if data_args.num_beams is not None else training_args.generation_num_beams\n repetition_penalty = data_args.repetition_penalty\n\n if training_args.do_predict:\n logger.info(\"*** Prediction ***\")\n logger.info(\"*** Loading CheckPoint ***\")\n checkpoint = None\n if os.path.isdir(training_args.output_dir):\n checkpoint = get_last_checkpoint(training_args.output_dir)\n if training_args.resume_from_checkpoint is not None:\n checkpoint = training_args.resume_from_checkpoint\n # without last ckpt and resume ckpt, would predict with current model\n if checkpoint:\n model = model_class.from_pretrained(checkpoint)\n trainer = UIETrainer(\n model=model,\n args=training_args,\n train_dataset=train_dataset if training_args.do_train else None,\n eval_dataset=eval_dataset if training_args.do_eval else None,\n tokenizer=tokenizer,\n data_collator=data_collator,\n compute_metrics=compute_rouge_metrics,\n callbacks=[DenserEvalCallback] if training_args.denser_evaluation else None\n )\n\n if data_args.max_predict_samples is not None:\n predict_dataset = predict_dataset.select(range(data_args.max_predict_samples))\n\n predict_results = trainer.predict(\n predict_dataset,\n metric_key_prefix=\"predict\",\n max_new_tokens=max_new_tokens,\n num_beams=num_beams,\n repetition_penalty=repetition_penalty,\n pad_token_id=tokenizer.pad_token_id\n )\n metrics = predict_results.metrics\n max_predict_samples = (\n data_args.max_predict_samples if data_args.max_predict_samples is not None else len(predict_dataset)\n )\n metrics[\"predict_samples\"] = min(max_predict_samples, len(predict_dataset))\n\n trainer.log(metrics)\n trainer.log_metrics(\"predict\", metrics)\n trainer.save_metrics(\"predict\", metrics)\n all_metrics.update(metrics)\n\n if training_args.do_demo:\n logger.info(\"Serving the model as a demo...\")\n user_input = ''\n while True:\n user_input = input(\"Please enter your input to the model, or enter 'quit' to exit: \")\n if user_input.lower() == \"quit\":\n break\n inputs = tokenizer([user_input], return_tensors=\"pt\")\n _, preds, _ = trainer.prediction_step(model, inputs=inputs, prediction_loss_only=False)\n print(f\"Model generates: {tokenizer.decode(preds[0], skip_special_tokens=True)}\\n\\n\")\n\n return results\n\n\nif __name__ == \"__main__\":\n main()", "repo_name": "BeyonderXX/InstructUIE", "sub_path": "src/run_uie.py", "file_name": "run_uie.py", "file_ext": "py", "file_size_in_byte": 22942, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 253, "dataset": "github-code", "pt": "24", "api": [{"api_name": "os.environ", "line_number": 37, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "nltk.data.find", "line_number": 43, "usage_type": "call"}, {"api_name": "nltk.data", "line_number": 43, "usage_type": "attribute"}, {"api_name": "transformers.file_utils.is_offline_mode", "line_number": 45, "usage_type": "call"}, {"api_name": "filelock.FileLock", "line_number": 49, "usage_type": "call"}, {"api_name": "nltk.download", "line_number": 50, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 59, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 62, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 62, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 65, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 65, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 68, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 68, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 72, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 76, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 80, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 87, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 87, "usage_type": "call"}, {"api_name": "dataclasses.dataclass", "line_number": 53, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 101, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 102, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 105, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 108, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 111, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 111, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 116, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 119, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 122, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 122, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 126, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 126, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 134, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 134, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 141, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 141, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 147, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 147, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 154, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 157, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 161, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 161, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 168, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 168, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 175, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 175, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 182, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 182, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 186, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 192, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 192, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 196, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 196, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 200, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 200, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 204, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 204, "usage_type": "call"}, {"api_name": "dataclasses.dataclass", "line_number": 96, "usage_type": "name"}, {"api_name": "transformers.Seq2SeqTrainingArguments", "line_number": 211, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 212, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 212, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 216, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 216, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 220, "usage_type": "call"}, {"api_name": "dataclasses.dataclass", "line_number": 210, "usage_type": "name"}, {"api_name": "transformers.HfArgumentParser", "line_number": 228, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 229, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 232, "usage_type": "call"}, {"api_name": "os.path", "line_number": 232, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 232, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 237, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 240, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 240, "usage_type": "attribute"}, {"api_name": "datasets.utils.logging.set_verbosity", "line_number": 244, "usage_type": "call"}, {"api_name": "datasets.utils", "line_number": 244, "usage_type": "attribute"}, {"api_name": "transformers.utils.logging.set_verbosity", "line_number": 245, "usage_type": "call"}, {"api_name": "transformers.utils", "line_number": 245, "usage_type": "attribute"}, {"api_name": "transformers.utils.logging.enable_default_handler", "line_number": 246, "usage_type": "call"}, {"api_name": "transformers.utils", "line_number": 246, "usage_type": "attribute"}, {"api_name": "transformers.utils.logging.enable_explicit_format", "line_number": 247, "usage_type": "call"}, {"api_name": "transformers.utils", "line_number": 247, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 258, "usage_type": "call"}, {"api_name": "os.path", "line_number": 258, "usage_type": "attribute"}, {"api_name": "transformers.trainer_utils.get_last_checkpoint", "line_number": 259, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 260, "usage_type": "call"}, {"api_name": "transformers.set_seed", "line_number": 272, "usage_type": "call"}, {"api_name": "uie_dataset.gen_cache_path", "line_number": 273, "usage_type": "call"}, {"api_name": "datasets.load_dataset", "line_number": 276, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 277, "usage_type": "call"}, {"api_name": "os.path", "line_number": 277, "usage_type": "attribute"}, {"api_name": "transformers.AutoConfig.from_pretrained", "line_number": 295, "usage_type": "call"}, {"api_name": "transformers.AutoConfig", "line_number": 295, "usage_type": "name"}, {"api_name": "transformers.AutoTokenizer.from_pretrained", "line_number": 301, "usage_type": "call"}, {"api_name": "transformers.AutoTokenizer", "line_number": 301, "usage_type": "name"}, {"api_name": "model.bloom.BloomForCausalLM_WithLoss", "line_number": 310, "usage_type": "name"}, {"api_name": "model.codegen.CodeGenForCausalLM_WithLoss", "line_number": 315, "usage_type": "name"}, {"api_name": "model.gpt_neox.GPTNeoXForCausalLM_WithLoss", "line_number": 320, "usage_type": "name"}, {"api_name": "transformers.AutoModelForSeq2SeqLM", "line_number": 325, "usage_type": "name"}, {"api_name": "model.bloom", "line_number": 326, "usage_type": "name"}, {"api_name": "model.bloom.resize_token_embeddings", "line_number": 334, "usage_type": "call"}, {"api_name": "model.bloom", "line_number": 334, "usage_type": "name"}, {"api_name": "model.bloom.config", "line_number": 337, "usage_type": "attribute"}, {"api_name": "model.bloom", "line_number": 337, "usage_type": "name"}, {"api_name": "model.bloom.config", "line_number": 338, "usage_type": "attribute"}, {"api_name": "model.bloom", "line_number": 338, "usage_type": "name"}, {"api_name": "model.bloom.config", "line_number": 342, "usage_type": "attribute"}, {"api_name": "model.bloom", "line_number": 342, "usage_type": "name"}, {"api_name": "model.bloom.resize_position_embeddings", "line_number": 345, "usage_type": "call"}, {"api_name": "model.bloom", "line_number": 345, "usage_type": "name"}, {"api_name": "model.bloom.resize_position_embeddings", "line_number": 347, "usage_type": "call"}, {"api_name": "model.bloom", "line_number": 347, "usage_type": "name"}, {"api_name": "model.bloom.config", "line_number": 350, "usage_type": "attribute"}, {"api_name": "model.bloom", "line_number": 350, "usage_type": "name"}, {"api_name": "model.bloom.config", "line_number": 351, "usage_type": "attribute"}, {"api_name": "model.bloom", "line_number": 351, "usage_type": "name"}, {"api_name": "model.bloom", "line_number": 355, "usage_type": "argument"}, {"api_name": "model.bloom.__class__", "line_number": 358, "usage_type": "attribute"}, {"api_name": "model.bloom", "line_number": 358, "usage_type": "name"}, {"api_name": "uie_collator.DataCollatorForUIE", "line_number": 384, "usage_type": "call"}, {"api_name": "model.bloom", "line_number": 386, "usage_type": "name"}, {"api_name": "uie_trainer.skip_instructions", "line_number": 404, "usage_type": "call"}, {"api_name": "model.bloom", "line_number": 404, "usage_type": "argument"}, {"api_name": "compute_metrics.compute_metrics", "line_number": 406, "usage_type": "call"}, {"api_name": "compute_metrics.compute_grouped_metrics", "line_number": 407, "usage_type": "call"}, {"api_name": "compute_metrics.compute_grouped_metrics", "line_number": 411, "usage_type": "call"}, {"api_name": "numpy.count_nonzero", "line_number": 414, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 415, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 418, "usage_type": "call"}, {"api_name": "os.path", "line_number": 418, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 420, "usage_type": "call"}, {"api_name": "model.bloom.gradient_checkpointing_enable", "line_number": 430, "usage_type": "call"}, {"api_name": "model.bloom", "line_number": 430, "usage_type": "name"}, {"api_name": "uie_trainer.UIETrainer", "line_number": 432, "usage_type": "call"}, {"api_name": "model.bloom", "line_number": 433, "usage_type": "name"}, {"api_name": "uie_trainer.DenserEvalCallback", "line_number": 440, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 485, "usage_type": "call"}, {"api_name": "os.path", "line_number": 485, "usage_type": "attribute"}, {"api_name": "transformers.trainer_utils.get_last_checkpoint", "line_number": 486, "usage_type": "call"}, {"api_name": "model.bloom", "line_number": 491, "usage_type": "name"}, {"api_name": "uie_trainer.UIETrainer", "line_number": 492, "usage_type": "call"}, {"api_name": "model.bloom", "line_number": 493, "usage_type": "name"}, {"api_name": "uie_trainer.DenserEvalCallback", "line_number": 500, "usage_type": "name"}, {"api_name": "model.bloom", "line_number": 533, "usage_type": "argument"}]}
+{"seq_id": "35823515997", "text": "import os, sys\nfrom flask import Flask, request\nfrom utils import wit_response\nfrom pymessenger import Bot\n\napp = Flask(__name__)\n\nPAGE_ACCESS_TOKEN=\"EAAHLxMxj2OoBAC4gJIsp8fbtFKR3ICG3SRoV7UUZAq5cW8IsvA5TIlZBbTpLhe34cxGfxcgxT84gxsY324ZA12Yph3WF9Uaon51ZB0dLbqZAAc7q7K5wZAkPFGZCZCUlf5tcrSZBmYw7ZBp4LzCZAAZB5VCHcuh0Gi725kbKfJ61uGlw6AZDZD\"\nbot = Bot(PAGE_ACCESS_TOKEN)\n\n@app.route('/', methods=['GET'])\ndef verify():\n #Webhook verification\n #print(\"HELLO****************************\")\n if request.args.get(\"hub.mode\")==\"subscribe\" and request.args.get(\"hub.challenge\"):\n if not request.args.get(\"hub.verify_token\")==\"hello\":\n return \"Verification token mismatch\", 403\n return request.args[\"hub.challenge\"], 200\n return \"Hello World\", 200\n\n@app.route('/', methods=['POST'])\ndef webhook():\n #print(\"HI****************************\")\n data = request.get_json()\n print(\"data\")\n log(data)\n if data['object']=='page':\n for entry in data['entry']:\n for messaging_event in entry['messaging']:\n sender_id=messaging_event['sender']['id']\n recipient_id=messaging_event['recipient']['id']\n if messaging_event.get('message'):\n if 'text' in messaging_event['message']:\n messaging_text= messaging_event['message']['text']\n else:\n messaging_text='no text'\n response = None\n entity, value = wit_response(messaging_text)\n if entity == 'books':\n response = \"Ok. I will show you {} books.\".format(str(value))\n elif entity == \"greetings\":\n response = \"Hello, how can I help you?\"\n elif entity == \"intent\":\n response = \"Which subject books do you want?\"\n else:\n response = \"Sorry, I din't get you.\"\n bot.send_text_message(sender_id,response)\n return \"ok\", 200\n\ndef log(message):\n\tprint(message)\n\tsys.stdout.flush()\n\nif __name__=='__main__':\n\tapp.run(debug=True)\n", "repo_name": "Akshansh93/Education-Bot", "sub_path": "app_wit.py", "file_name": "app_wit.py", "file_ext": "py", "file_size_in_byte": 2161, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "pymessenger.Bot", "line_number": 9, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 15, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 16, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 18, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 18, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 24, "usage_type": "name"}, {"api_name": "utils.wit_response", "line_number": 38, "usage_type": "call"}, {"api_name": "sys.stdout.flush", "line_number": 52, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 52, "usage_type": "attribute"}]}
+{"seq_id": "72273549821", "text": "import concurrent.futures\nfrom concurrent.futures import ThreadPoolExecutor\n\nfrom tabulate import tabulate\n\nfrom .Config import Config\nfrom .Pull import Pull\nfrom ..model import (\n UpdateJson,\n TrackJson\n)\nfrom ..track import BaseTracks, LocalTracks, GithubTracks\nfrom ..utils import Log\n\n\nclass Sync:\n def __init__(self, root_folder, config, tracks=None):\n self._log = Log(\"Sync\", enable_log=config.enable_log, log_dir=config.log_dir)\n self._root_folder = root_folder\n self._pull = Pull(root_folder, config)\n\n self._json_folder = Config.get_json_folder(root_folder)\n self._modules_folder = Config.get_modules_folder(root_folder)\n self._config = config\n\n if tracks is None:\n self._tracks = BaseTracks()\n else:\n self._tracks = tracks\n\n self._updated_diff = list()\n\n def _update_jsons(self, track, force):\n module_folder = self._modules_folder.joinpath(track.id)\n\n if not track.enable:\n self._log.i(f\"_update_jsons: [{track.id}] -> update check has been disabled\")\n return None\n\n online_module, timestamp = self._pull.from_track(track)\n if online_module is None:\n return None\n\n update_json_file = module_folder.joinpath(UpdateJson.filename())\n track_json_file = module_folder.joinpath(TrackJson.filename())\n\n if force:\n for file in module_folder.glob(\"*\"):\n if file.name not in [\n TrackJson.filename(),\n online_module.zipfile_name,\n online_module.changelog_filename\n ]:\n file.unlink()\n\n if update_json_file.exists():\n update_json = UpdateJson.load(update_json_file)\n update_json.update(id=track.id)\n else:\n update_json = UpdateJson(\n id=track.id,\n timestamp=timestamp,\n versions=list()\n )\n\n version_item = online_module.to_VersionItem(timestamp)\n update_json.versions.append(version_item)\n\n max_num = self._config.max_num\n if track.max_num is not None:\n max_num = track.max_num\n\n if len(update_json.versions) > max_num:\n old_item = update_json.versions.pop(0)\n zipfile = module_folder.joinpath(old_item.zipfile_name)\n changelog = module_folder.joinpath(old_item.changelog_filename)\n\n for path in [zipfile, changelog]:\n if not (path.exists() and path.is_file()):\n continue\n\n self._log.d(f\"_update_jsons: [{track.id}] -> remove {path.name}\")\n path.unlink()\n\n track.last_update = timestamp\n track.versions = len(update_json.versions)\n\n update_json.write(update_json_file)\n track.write(track_json_file)\n\n if len(update_json.versions) >= 2:\n self._updated_diff.append(\n (update_json.versions[-2], online_module)\n )\n else:\n self._updated_diff.append(\n (None, online_module)\n )\n\n return online_module\n\n @staticmethod\n def _check_tracks(obj, cls):\n if type(obj) is BaseTracks:\n raise RuntimeError(\"tracks interface has not been created\")\n\n return isinstance(obj, cls)\n\n def create_github_tracks(self, api_token, after_date=None):\n self._tracks = GithubTracks(\n modules_folder=self._modules_folder,\n config=self._config,\n api_token=api_token,\n after_date=after_date\n )\n return self._tracks\n\n def create_local_tracks(self):\n self._tracks = LocalTracks(\n modules_folder=self._modules_folder,\n config=self._config\n )\n return self._tracks\n\n def update(self, module_ids=None, force=False, single=False, **kwargs):\n user_name = kwargs.get(\"user_name\")\n if user_name is not None:\n if self._check_tracks(self._tracks, GithubTracks):\n tracks = self._tracks.get_tracks(\n user_name=user_name,\n repo_names=module_ids,\n single=single,\n cover=kwargs.get(\"cover\", False),\n use_ssh=kwargs.get(\"use_ssh\", True)\n )\n else:\n msg = f\"unsupported tracks interface type [{type(self._tracks).__name__}]\"\n raise RuntimeError(msg)\n else:\n tracks = self._tracks.get_tracks(module_ids)\n\n with ThreadPoolExecutor(max_workers=1 if single else None) as executor:\n futures = []\n for track in tracks:\n futures.append(\n executor.submit(self._update_jsons, track=track, force=force)\n )\n\n for future in concurrent.futures.as_completed(futures):\n online_module = future.result()\n if online_module is not None:\n self._log.i(f\"update: [{online_module.id}] -> update to {online_module.version_display}\")\n\n def get_versions_diff(self):\n headers = [\"id\", \"name\", \"version\"]\n table = []\n\n if len(self._updated_diff) == 0:\n return None\n\n for last, new in self._updated_diff:\n version = new.version_display\n if last is not None:\n version = f\"{last.version_display} -> {version}\"\n\n name = new.name.replace(\"|\", \"_\")\n table.append(\n [new.id, name, version]\n )\n\n markdown_text = tabulate(table, headers, tablefmt=\"github\")\n return markdown_text\n", "repo_name": "ya0211/magisk-modules-repo-util", "sub_path": "sync/core/Sync.py", "file_name": "Sync.py", "file_ext": "py", "file_size_in_byte": 5698, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 34, "dataset": "github-code", "pt": "24", "api": [{"api_name": "utils.Log", "line_number": 18, "usage_type": "call"}, {"api_name": "Pull.Pull", "line_number": 20, "usage_type": "call"}, {"api_name": "Config.Config.get_json_folder", "line_number": 22, "usage_type": "call"}, {"api_name": "Config.Config", "line_number": 22, "usage_type": "name"}, {"api_name": "Config.Config.get_modules_folder", "line_number": 23, "usage_type": "call"}, {"api_name": "Config.Config", "line_number": 23, "usage_type": "name"}, {"api_name": "track.BaseTracks", "line_number": 27, "usage_type": "call"}, {"api_name": "track.id", "line_number": 34, "usage_type": "attribute"}, {"api_name": "track.enable", "line_number": 36, "usage_type": "attribute"}, {"api_name": "track.id", "line_number": 37, "usage_type": "attribute"}, {"api_name": "model.UpdateJson.filename", "line_number": 44, "usage_type": "call"}, {"api_name": "model.UpdateJson", "line_number": 44, "usage_type": "name"}, {"api_name": "model.TrackJson.filename", "line_number": 45, "usage_type": "call"}, {"api_name": "model.TrackJson", "line_number": 45, "usage_type": "name"}, {"api_name": "model.TrackJson.filename", "line_number": 50, "usage_type": "call"}, {"api_name": "model.TrackJson", "line_number": 50, "usage_type": "name"}, {"api_name": "model.UpdateJson.load", "line_number": 57, "usage_type": "call"}, {"api_name": "model.UpdateJson", "line_number": 57, "usage_type": "name"}, {"api_name": "track.id", "line_number": 58, "usage_type": "attribute"}, {"api_name": "model.UpdateJson", "line_number": 60, "usage_type": "call"}, {"api_name": "track.id", "line_number": 61, "usage_type": "attribute"}, {"api_name": "track.max_num", "line_number": 70, "usage_type": "attribute"}, {"api_name": "track.max_num", "line_number": 71, "usage_type": "attribute"}, {"api_name": "track.id", "line_number": 82, "usage_type": "attribute"}, {"api_name": "track.last_update", "line_number": 85, "usage_type": "attribute"}, {"api_name": "track.versions", "line_number": 86, "usage_type": "attribute"}, {"api_name": "track.write", "line_number": 89, "usage_type": "call"}, {"api_name": "track.BaseTracks", "line_number": 104, "usage_type": "name"}, {"api_name": "track.GithubTracks", "line_number": 110, "usage_type": "call"}, {"api_name": "track.LocalTracks", "line_number": 119, "usage_type": "call"}, {"api_name": "track.GithubTracks", "line_number": 128, "usage_type": "argument"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 142, "usage_type": "call"}, {"api_name": "concurrent.futures.futures.as_completed", "line_number": 149, "usage_type": "call"}, {"api_name": "concurrent.futures.futures", "line_number": 149, "usage_type": "attribute"}, {"api_name": "concurrent.futures", "line_number": 149, "usage_type": "name"}, {"api_name": "tabulate.tabulate", "line_number": 171, "usage_type": "call"}]}
+{"seq_id": "12872667250", "text": "from flask import Flask, render_template, request, redirect, flash\napp = Flask(__name__)\napp.secret_key = 'RobBoss'\n\n@app.route('/')\ndef index():\n\treturn render_template(\"index.html\")\n@app.route('/results', methods=['POST'])\ndef create():\n\tnamer = request.form['name']\n\tlocaler = request.form['locale']\n\tlingor = request.form['lingo']\n\tcommentr = request.form['comment']\n\n\t## validations yay ##\n\tif len(namer) < 1:\n\t\tflash('Yo dawg, gotta have a name! AMIRITE?')\n\t\treturn redirect('/')\n\telif len(commentr) > 120:\n\t\tflash('more register less comment young one')\n\t\treturn redirect('/')\n\telse:\n\t\tflash('thanks for registering broski!')\n\n\treturn render_template('result.html', named=namer, locale=localer, lingo=lingor, comment=commentr)\n\n\n\napp.run(debug=True)", "repo_name": "pinkshrub/numberGame", "sub_path": "server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 756, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "flask.Flask", "line_number": 2, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 7, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 10, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 11, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 11, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 12, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 12, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 13, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 13, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 25, "usage_type": "call"}]}
+{"seq_id": "38143770035", "text": "import matplotlib, sys\nmatplotlib.use('Agg')\nimport matplotlib.pyplot as m_plot\n\nif __name__=='__main__':\n data = []\n with open(sys.argv[1], 'r') as f:\n for line in f:\n data.append(map(float, line.strip().split(',')))\n\n fig = m_plot.figure(figsize=(14, 7))\n fig.suptitle('Seek time vs Hour', fontsize=20)\n\n colors = ('b','g','r','c','m','y','k')\n weekdays = ('Mon', 'Tues','Wed','Thurs','Fri','Sat','Sun')\n for i in range(0,7):\n hr_ave_by_wkday = [(h,s/c) for w,h,s,c in data if w==i]\n hr_ave_by_wkday.sort()\n values = zip(*hr_ave_by_wkday)\n\n ax = fig.add_axes([.1,.1,.7,.7])\n ax.plot(values[0], values[1], color=colors[i], label=weekdays[i])\n \n ax.legend(bbox_to_anchor=(1.05, 1), loc=2, borderaxespad=0.)\n ax.set_xlim([0,23])\n ax.spines['top'].set_visible(False)\n ax.spines['right'].set_visible(False)\n ax.get_xaxis().tick_bottom()\n ax.get_yaxis().tick_left()\n ax.set_ylabel('seek time')\n ax.set_xlabel('hour of day')\n fig.savefig(sys.argv[2])\n", "repo_name": "yinkelly/nyc-taxi", "sub_path": "1-seek-time/plot_results.py", "file_name": "plot_results.py", "file_ext": "py", "file_size_in_byte": 1054, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "matplotlib.use", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 7, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 32, "usage_type": "attribute"}]}
+{"seq_id": "2088839320", "text": "import sys\nfrom itertools import combinations\n\ninput = sys.stdin.readline\n\nN, M = map(int, input().split())\n\n\ndef calc_dist(home, chicken):\n # min dist 최댓값 초기화\n min_dist = 1e9\n\n # 치킨집을 순회하면서 계산\n for c in chicken:\n # abs를 이용한 최솟값 계산\n min_dist = min(min_dist, abs(c[0] - home[0]) + abs(c[1] - home[1]))\n # min dist 대입\n return min_dist\n\n\n# M : 치킨 집 개수\ngraph = [list(map(int, input().split())) for _ in range(N)]\n\n# 치킨과 집 집계\nchicken = [[c, j] for c in range(N) for j in range(N) if graph[c][j] == 2]\nhome = [[i, j] for i in range(N) for j in range(N) if graph[i][j] == 1]\n\n# 치킨에 대해서 많은 케이스 생성\nchicken_select = list(combinations(chicken, M))\n\n# 최소 치킨거리 계산\nmin_chicken_dist = int(10e9)\nfor chickens in chicken_select:\n dist = 0\n # 모든 집을 순회하며 치킨 거리 계산\n for h in home:\n dist += calc_dist(h, chickens)\n min_chicken_dist = min(dist, min_chicken_dist)\n\nprint(min_chicken_dist)\n", "repo_name": "hjun-park/Coding-test-self-study", "sub_path": "--2022backup/[02]구현/[G5]15686-치킨 배달.py", "file_name": "[G5]15686-치킨 배달.py", "file_ext": "py", "file_size_in_byte": 1065, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "sys.stdin", "line_number": 4, "usage_type": "attribute"}, {"api_name": "itertools.combinations", "line_number": 29, "usage_type": "call"}]}
+{"seq_id": "15188593159", "text": "# -*- coding: utf-8 -*-\r\n\r\n# __title__ = 'MyForms.py'\r\n# __author__ = 'YangYang'\r\n# __mtime__ = '2018.07.17'\r\n\r\nfrom django import forms\r\nfrom django.forms import widgets\r\n\r\nfrom blog.models import UserInfo\r\nfrom django.core.exceptions import ValidationError\r\n\r\nwid_01=widgets.TextInput(attrs={\"class\":\"form-control\"})\r\nwid_02=widgets.PasswordInput(attrs={\"class\":\"form-control\"})\r\nwid_03=widgets.EmailInput(attrs={\"class\":\"form-control\"})\r\n\r\n\r\nclass UserForm(forms.Form):\r\n\tuser=forms.CharField(max_length=32,\r\n\t error_messages={\"required\":\"该字段不能为空\"},\r\n\t label=\"用户名\", widget=wid_01)\r\n\tpwd=forms.CharField(max_length=32,\r\n\t error_messages={\"required\": \"该字段不能为空\"},\r\n\t label=\"密码\",\r\n\t widget=wid_02)\r\n\tre_pwd=forms.CharField(max_length=32,\r\n\t error_messages={\"required\": \"该字段不能为空\"},\r\n\t label=\"确认密码\",widget=wid_02)\r\n\temail=forms.EmailField(max_length=32,\r\n\t error_messages={\"required\": \"该字段不能为空\"},\r\n\t label=\"邮箱\",widget=wid_03)\r\n\r\n\r\n\tdef clean_user(self):\r\n\t\tuser = self.cleaned_data.get(\"user\")\r\n\t\tuser_obj = UserInfo.objects.filter(username=user).first()\r\n\t\tif not user_obj:\r\n\t\t\treturn user\r\n\t\telse:\r\n\t\t\traise ValidationError(\"该用户已注册\")\r\n\r\n\r\n\tdef clean(self):\r\n\t\tpwd=self.cleaned_data.get(\"pwd\")\r\n\t\tre_pwd=self.cleaned_data.get(\"re_pwd\")\r\n\r\n\t\tif pwd and re_pwd:\r\n\r\n\t\t\tif pwd==re_pwd:\r\n\t\t\t\treturn self.cleaned_data\r\n\t\t\telse:\r\n\t\t\t\traise ValidationError(\"两次密码不一致\")\r\n\t\telse:\r\n\t\t\treturn self.cleaned_data\r\n\r\n\r\n", "repo_name": "ryan-yang-2049/myblog", "sub_path": "myblog/blog/utils/MyForms.py", "file_name": "MyForms.py", "file_ext": "py", "file_size_in_byte": 1692, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.forms.widgets.TextInput", "line_number": 13, "usage_type": "call"}, {"api_name": "django.forms.widgets", "line_number": 13, "usage_type": "name"}, {"api_name": "django.forms.widgets.PasswordInput", "line_number": 14, "usage_type": "call"}, {"api_name": "django.forms.widgets", "line_number": 14, "usage_type": "name"}, {"api_name": "django.forms.widgets.EmailInput", "line_number": 15, "usage_type": "call"}, {"api_name": "django.forms.widgets", "line_number": 15, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 18, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 18, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 19, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 22, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 26, "usage_type": "name"}, {"api_name": "django.forms.EmailField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 29, "usage_type": "name"}, {"api_name": "blog.models.UserInfo.objects.filter", "line_number": 36, "usage_type": "call"}, {"api_name": "blog.models.UserInfo.objects", "line_number": 36, "usage_type": "attribute"}, {"api_name": "blog.models.UserInfo", "line_number": 36, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 40, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 52, "usage_type": "call"}]}
+{"seq_id": "74456429182", "text": "from datetime import datetime, timezone\nfrom dataclasses import dataclass\nfrom enum import Enum\n\nimport numpy as np\nfrom numpy import arcsin, pi, log10\nfrom numpy.polynomial import Polynomial\nfrom blocksim.satellite.Satellite import (\n CircleSatellite,\n generateWalkerDeltaConstellation,\n)\nfrom blocksim.utils import llavpa_to_itrf, itrf_to_azeld\nfrom blocksim.constants import Req, kb, c as clum\n\n\nclass EventType(Enum):\n RISE = \"rise\"\n SET = \"set\"\n\n\n@dataclass\nclass Event:\n satellite: int\n type: EventType\n date: float\n\n\ndef analyse_timeline(init, events, total_sim_time):\n tl = []\n for ksat, sat in enumerate(events):\n for e in sat:\n tl.append(Event(satellite=ksat, type=EventType.RISE, date=e[\"rise\"]))\n tl.append(Event(satellite=ksat, type=EventType.SET, date=e[\"set\"]))\n tl.sort(key=lambda x: x.date)\n\n nsat = len(np.where(init > 0)[0])\n blind_e = Event(satellite=-1, type=EventType.RISE, date=0)\n t_blind = 0\n e: Event\n nsat_max = nsat\n for e in tl:\n if e.type == EventType.RISE:\n if nsat == 0:\n t_blind += e.date - blind_e.date\n nsat += 1\n else:\n nsat -= 1\n if nsat == 0:\n blind_e = e\n elif nsat < 0:\n raise AssertionError\n if nsat_max < nsat:\n nsat_max = nsat\n print(e, nsat, t_blind)\n\n if nsat_max == 0:\n t_blind = total_sim_time\n\n return t_blind, nsat_max\n\n\ndef compute_elevation_mask(alt_km):\n # Parametres du probleme\n # ======================\n px = 36 # dBm\n NF = 3 # dB\n z1_J1 = 1.616339347\n eta = 0.7\n wl = clum / (2e9)\n cn0_lim = 46.46595211\n alpha = 0.0305\n\n # Calculs des constantes K et Q\n # =============================\n sma = Req + alt_km * 1e3\n K = (\n -36\n + px\n - NF\n + 10\n * log10((eta**2 * wl**2 * z1_J1**4 * sma**4) / (16 * pi**2 * kb * 290 * Req**4))\n )\n Q = 10 ** ((K - cn0_lim) / 20)\n\n # Calcul de l'élévation\n # =====================\n p = Polynomial(\n [\n alpha,\n 2,\n -(\n (2 * alpha**2 - 1) * sma**2\n + (1 - 2 * alpha**2) * Req**2\n + 2 * alpha * Q * Req\n + Q**2\n )\n / (alpha * sma**2 - alpha * Req**2),\n -(4 * alpha * sma**2 - 4 * alpha * Req**2 + 2 * Q * Req)\n / (alpha * sma**2 - alpha * Req**2),\n ((alpha**2 - 2) * sma**2 + (2 - alpha**2) * Req**2 + 2 * alpha * Q * Req)\n / (alpha * sma**2 - alpha * Req**2),\n (2 * alpha * sma**2 - 2 * alpha * Req**2 + 2 * Q * Req)\n / (alpha * sma**2 - alpha * Req**2),\n 1 / alpha,\n ]\n )\n\n rts = p.roots()\n r0 = rts[np.where((rts > arcsin(alpha)) & (rts < 1) & (np.abs(np.imag(rts)) < 1e-9))[0]]\n assert len(r0) == 1\n elev_mask = arcsin(r0[0])\n\n return elev_mask\n\n\ndef simulate(lat, inc, nsat, npla, pha, alt_km):\n elev_mask = compute_elevation_mask(alt_km)\n\n t0 = datetime(2023, 6, 27, 12, 0, 0, tzinfo=timezone.utc)\n firstraan = 0.0\n lon = 0.0\n # tps_max = 5 * 86400\n tps_max = 20000\n sma = Req + alt_km * 1e3\n\n satellites = generateWalkerDeltaConstellation(\n \"sim\", sma, inc, firstraan, nsat, npla, pha, t0, prop=CircleSatellite\n )\n obs = llavpa_to_itrf((lon, lat, 0, 0, 0, 0))\n events = list()\n init = list()\n sat: CircleSatellite\n for sat in satellites:\n pv_sat = sat.getGeocentricITRFPositionAt(0)\n _, el0, _, _, _, _ = itrf_to_azeld(obs, pv_sat)\n init.append(el0 - elev_mask)\n events.append(\n sat.find_events(obs, t0=0, t1=tps_max, elevation=elev_mask),\n )\n\n t_blind, nsat_max = analyse_timeline(np.array(init), events, tps_max)\n\n return t_blind, nsat_max, tps_max, elev_mask\n", "repo_name": "ydethe/constellation_design", "sub_path": "tests/simulate.py", "file_name": "simulate.py", "file_ext": "py", "file_size_in_byte": 3899, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "enum.Enum", "line_number": 16, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 21, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 36, "usage_type": "call"}, {"api_name": "blocksim.constants.c", "line_number": 69, "usage_type": "name"}, {"api_name": "blocksim.constants.Req", "line_number": 75, "usage_type": "name"}, {"api_name": "numpy.log10", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 81, "usage_type": "name"}, {"api_name": "blocksim.constants.kb", "line_number": 81, "usage_type": "name"}, {"api_name": "blocksim.constants.Req", "line_number": 81, "usage_type": "name"}, {"api_name": "numpy.polynomial.Polynomial", "line_number": 87, "usage_type": "call"}, {"api_name": "blocksim.constants.Req", "line_number": 93, "usage_type": "name"}, {"api_name": "blocksim.constants.Req", "line_number": 94, "usage_type": "name"}, {"api_name": "blocksim.constants.Req", "line_number": 97, "usage_type": "name"}, {"api_name": "blocksim.constants.Req", "line_number": 98, "usage_type": "name"}, {"api_name": "blocksim.constants.Req", "line_number": 99, "usage_type": "name"}, {"api_name": "blocksim.constants.Req", "line_number": 100, "usage_type": "name"}, {"api_name": "blocksim.constants.Req", "line_number": 101, "usage_type": "name"}, {"api_name": "blocksim.constants.Req", "line_number": 102, "usage_type": "name"}, {"api_name": "blocksim.constants.Req", "line_number": 103, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.arcsin", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.imag", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.arcsin", "line_number": 111, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 119, "usage_type": "call"}, {"api_name": "datetime.timezone.utc", "line_number": 119, "usage_type": "attribute"}, {"api_name": "datetime.timezone", "line_number": 119, "usage_type": "name"}, {"api_name": "blocksim.constants.Req", "line_number": 124, "usage_type": "name"}, {"api_name": "blocksim.satellite.Satellite.generateWalkerDeltaConstellation", "line_number": 126, "usage_type": "call"}, {"api_name": "blocksim.satellite.Satellite.CircleSatellite", "line_number": 127, "usage_type": "name"}, {"api_name": "blocksim.utils.llavpa_to_itrf", "line_number": 129, "usage_type": "call"}, {"api_name": "blocksim.satellite.Satellite.CircleSatellite", "line_number": 132, "usage_type": "name"}, {"api_name": "blocksim.utils.itrf_to_azeld", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 141, "usage_type": "call"}]}
+{"seq_id": "16791852081", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nThis module constructs and populates the deployment dictionary data\nstructure ``deploy_dict``. Below is an overview of the deployment dictionary:\n\n* Testing organization\n * Organization name, contact information\n* Testing Location\n * Site name, address, coordinates, and AQS site identifier\n* Deployment Information and Statistics\n * Unique deployment groups\n * Description of sensor uptime for each sensor unit\n * Evaluation parameter statistics\n * Precision\n * Error\n * Description of reference monitor, measured range during\n deployment period at 1-hour and 24-hour averages\n * Meteorological conditions\n * Description of temperature instrument, measured range during\n deployment period at 1-hour and 24-hour averages\n * Description of relative humidity instrument, measured range during\n deployment period at 1-hour and 24-hour averages\n\n================================================================================\n\n@Author:\n | Samuel Frederick, NSSC Contractor (ORAU)\n | U.S. EPA / ORD / CEMM / AMCD / SFSB\n\nCreated:\n Mon Nov 9 10:47:56 2020\nLast Updated:\n Tue Jul 12 13:38:00 2021\n\"\"\"\nimport pandas as pd\nimport numpy as np\nfrom datetime import datetime\nfrom sensortoolkit.calculate import uptime\nfrom sensortoolkit.lib_utils import _get_version\nfrom sensortoolkit.param import Parameter\nfrom sensortoolkit.datetime_utils import (deploy_timestamp_index,\n get_timestamp_interval)\n\ndef construct_deploy_dict(deploy_df, full_df_list, hourly_df_list,\n daily_df_list, sensor_name, testing_loc,\n testing_org, **kwargs):\n \"\"\"Create the deployment dictionary, initialize with sensor group info,\n time period of deployment, testing agency and location, and library version\n and time at which the dictionary were constructed.\n\n Determines which sensors match the beginning and end dates for deployment\n (provided a timedelta padding window of 1 day around the begin and end\n timestamps). Sensors measuring concurrently are grouped together as a\n `deployment group`. Sensors with beginning and end deployment dates that\n differ from the identified deployment group are assigned ``True`` for the\n ``deploy_dict`` sensor unit entry ``deploy_issues``.\n\n Args:\n deploy_df (pandas dataframe):\n A data frame containing the start time (`Begin`), end time (`End`),\n and total duration of evaluation period for each sensor in a\n deployment group.\n full_df_list (list):\n List of sensor data frames of length N (where N is the number of\n sensor units in a testing group). Data frames indexed by\n at recorded sampling frequency.\n hourly_df_list (list):\n List of sensor data frames of length N (where N is the number of\n sensor units in a testing group). Data frames indexed by\n DateTime at 1-hour averaged sampling frequency.\n daily_df_list (list):\n List of sensor data frames of length N (where N is the number of\n sensor units in a testing group). Data frames indexed by\n DateTime at 24-hour averaged sampling frequency.\n sensor_name (str):\n The make and model of the sensor being evaluated.\n testing_org (dict):\n A dictionary containing the information about the testing\n organization.\n testing_loc (dict):\n A dictionary containing information about the testing site. If the\n site is part of U.S. EPA’s Air Quality System (AQS), the AQS Site\n ID should be specified.\n\n Returns:\n deploy_dict (dict):\n Dictionary containing separate deployment group start and\n end times (based on the latest (max) start timestamp and earliest\n (min) end timestamp in group), deployment duration, and sensor\n serial IDs for devices within each deployment group.\n\n \"\"\"\n current_time = datetime.now().strftime('%Y-%m-%d %H:%M:%S %p')\n deploy_dict = {'sensortoolkit Version': _get_version(),\n 'Date of Analysis': current_time,\n 'Sensor Name': sensor_name,\n 'Sensor Firmware Version': kwargs.get('sensor_firmware', 'Unspecified'),\n 'Deployment Groups': {},\n 'Testing Organization': testing_org,\n 'Testing Location': testing_loc}\n\n deploy_grp_n = 1\n\n while deploy_df.empty is False:\n i = deploy_df.index[0]\n\n match_begin = abs(deploy_df.loc[i, 'Begin'] - deploy_df.loc[:, 'Begin']\n ) < pd.Timedelta('1 day')\n\n deploy = deploy_df[match_begin]\n\n # Date (YYYY-MM-DD) of deployment group end, calculate mode\n end_date = deploy.loc[:, \"End\"].dt.strftime(\"%Y-%m-%d\")\n end_date_mode = end_date.mode()[0]\n\n # Sensors that concluded deployment before end of majority of group\n deploy['Issues'] = end_date != end_date_mode\n\n serials = {str(i): serial for i, serial in zip(\n deploy.Sensor_Number, deploy.Sensor_Serial)}\n\n deployments = deploy_dict['Deployment Groups']\n\n deployments['Group ' + str(deploy_grp_n)] = {}\n deployments['Group ' + str(deploy_grp_n)]['sensors'] = {}\n\n sensor_info = {i: {'serial_id': j} for i, j in zip(serials.keys(),\n serials.values())}\n\n deployments['Group ' + str(deploy_grp_n)]['sensors'] = sensor_info\n\n deployments['Group ' + str(deploy_grp_n)]['eval_start'] = \\\n deploy.Begin.min().strftime(\"%Y-%m-%dT%H:%M:%S%z\")\n deployments['Group ' + str(deploy_grp_n)]['eval_end'] = \\\n deploy.End.max().strftime(\"%Y-%m-%dT%H:%M:%S%z\")\n deployments['Group ' + str(deploy_grp_n)]['eval_duration'] = \\\n str(abs(deploy.Begin.min() - deploy.End.max()))\n\n start = deployments['Group ' + str(deploy_grp_n)]['eval_start']\n end = deployments['Group ' + str(deploy_grp_n)]['eval_end']\n\n # round timestamp down to nearest hour\n start = pd.to_datetime(start).floor(freq='H')\n # round timestamp up to nearest hour\n end = pd.to_datetime(end).ceil(freq='H')\n\n for sensor_n in list(sensor_info.keys()):\n i = int(sensor_n) - 1\n\n full_df = full_df_list[i]\n hourly_df = hourly_df_list[i]\n daily_df = daily_df_list[i]\n\n # Record whether sensor encountered issues during deployment, ended\n # deployment early\n sensor_df = deploy[deploy.Sensor_Number == sensor_n]\n sensor_df = sensor_df.reset_index(drop=True)\n sensor_info[sensor_n]['deploy_issues'] = str(bool(\n sensor_df.Issues[0]))\n\n # Compute recording interval for data\n time_delta = get_timestamp_interval(full_df)\n sensor_info[sensor_n]['recording_interval'] = time_delta\n\n # 1-hr uptime\n sensor_h_uptime = uptime(hourly_df.loc[start:end, :], key=sensor_n)\n sensor_info[sensor_n]['uptime_1-hour'] = sensor_h_uptime[sensor_n]['Uptime']\n\n # 24-hr uptime\n sensor_d_uptime = uptime(daily_df.loc[start:end, :], key=sensor_n)\n sensor_info[sensor_n]['uptime_24-hour'] = sensor_d_uptime[sensor_n]['Uptime']\n\n deploy_df = deploy_df.drop(deploy.index, axis=0)\n deploy_grp_n += 1\n\n return deploy_dict\n\n\ndef deploy_ref_stats(deploy_dict, ref_df, cal_check_dict=None, param=None,\n ref_name=None):\n \"\"\"Add reference monitor statistics to the parameter statistics subfield in\n the deployment dictionary.\n\n Details added include:\n\n * The FRM/FEM monitor name\n * The minimum concentration recorded at the specified interval\n averaging.\n * The maximum concentration recorded at the specified interval\n averaging.\n * The number of intervals during which the FRM/FEM exceeds the goal\n concentration recommended by the performance targets testing report\n for elevated concentrations (goal :math:`\\\\geq`` three days).\n\n Args:\n deploy_dict (dict):\n Dictionary containing separate deployment group start and end times\n (based on the latest (max) start timestamp and earliest (min)\n end timestamp in group), deployment duration, and sensor serial IDs\n for devices within each deployment group.\n \tref_df (pandas dataframe):\n Dataframe for reference concentrations at either 1-hour or 24-hour\n averaging depending on the performance targets recommeneded\n averaging interval.\n \tcal_check_dict (dict):\n [Future feature] Dictionary for housing dates and descriptions of QC\n calibration checks as part of regularly scheduled and cataloged QC\n procedures.\n param_obj (str):\n The evaluation parameter\n ref_name (str):\n The name of the FRM/FEM monitor (make and model).\n\n Returns:\n deploy_dict:\n Dictionary containing separate deployment group start and end times\n (based on the latest (max) start timestamp and earliest (min)\n end timestamp in group), deployment duration, and sensor serial IDs\n for devices within each deployment group.\n\n \"\"\"\n param_obj = Parameter(param)\n param_name = param_obj.name\n\n date_index, avg_suffix = deploy_timestamp_index(ref_df,\n averaging_suffix=True)\n\n if param_name == 'PM25':\n conc_goal = 25 # Concentration goal: 25 ug/m^3 for at least one day\n elif param_name == 'O3':\n conc_goal = 60 # Concentration goal: 60 ppbv for at least one day\n ref_df[f'{param_name}_rolling_8-hour_Value'] = ref_df[f'{param_name}_Value'].rolling(window=8).mean()\n\n else:\n conc_goal = None\n\n for group in deploy_dict['Deployment Groups']:\n deploy = deploy_dict['Deployment Groups'][group]\n start = deploy['eval_start']\n end = deploy['eval_end']\n\n ref_data = ref_df.loc[start:end, param_name + '_Value']\n\n if param_name not in deploy:\n deploy[param_name] = {}\n deploy[param_name]['Reference'] = {}\n\n if 'Reference' not in deploy[param_name]:\n deploy[param_name]['Reference'] = {}\n\n stats_loc = deploy[param_name]['Reference']\n\n stats_loc['reference_name'] = ref_name\n stats_loc['conc_min' + avg_suffix] = \\\n float(\"{0:.3f}\".format(ref_data.min()))\n stats_loc['conc_max' + avg_suffix] = \\\n float(\"{0:.3f}\".format(ref_data.max()))\n stats_loc['conc_mean' + avg_suffix] = \\\n float(\"{0:.3f}\".format(ref_data.mean()))\n stats_loc['n_exceed_conc_goal' + avg_suffix] = \\\n int(ref_data.where(ref_data > conc_goal).count())\n\n if ref_data.dropna().empty:\n stats_loc['conc_min' + avg_suffix] = None\n stats_loc['conc_max' + avg_suffix] = None\n stats_loc['n_exceed_conc_goal' + avg_suffix] = None\n\n # add 8-hr rolling statistics\n if param_name =='O3':\n avg_suffix = '_rolling_8-hour'\n ref_data = ref_df.loc[start:end, f'{param_name}{avg_suffix}_Value']\n\n stats_loc['conc_min' + avg_suffix] = \\\n float(\"{0:.3f}\".format(ref_data.min()))\n stats_loc['conc_max' + avg_suffix] = \\\n float(\"{0:.3f}\".format(ref_data.max()))\n stats_loc['conc_mean' + avg_suffix] = \\\n float(\"{0:.3f}\".format(ref_data.mean()))\n\n return deploy_dict\n\n\ndef deploy_met_stats(deploy_dict, df_list, met_ref_df, operational_range):\n \"\"\"Add meteorological instrument statistics to the parameter statistics\n subfield in the deployment dictionary.\n\n Details added include:\n\n * The name of the instrument collocated nearby sensor deployment location.\n * The minimum value recorded at the specified interval averaging.\n * The maximum value recorded at the specified interval averaging.\n * The number of intervals during which the instrument exceeds the\n manufacturer's recommended target range for instrument performance.\n This is provisionally set for RH (exceedence when :math:`\\\\leq` 10% or\n :math:`\\\\geq` 90%) and Temp (exceedence when :math:`\\\\leq` -20 C or\n :math:`\\\\geq` 40 C).\n\n Args:\n \tdeploy_dict (dict):\n Dictionary containing separate deployment group start and end times\n (based on the latest (max) start timestamp and earliest (min)\n end timestamp in group), deployment duration, and sensor serial IDs\n for devices within each deployment group.\n \tdf_list (list):\n List of pandas dataframes for sensor measurements at either 1-hr or\n 24-hr averaging intervals.\n met_ref_df (pandas dataframe):\n A dataframe containing meteorological parameters recorded at the\n testing site during the evaluation period (either 1-hr or 24-hr\n averaging intervals).\n \toperational_range (dict):\n Dictionary for listing the operational range indicated by the\n sensor manufacturer for meteorological parameters, such as temp\n and RH.\n\n Returns:\n deploy_dict:\n Dictionary containing separate deployment group start and end times\n (based on the latest (max) start timestamp and earliest (min)\n end timestamp in group), deployment duration, and sensor serial IDs\n for devices within each deployment group.\n\n \"\"\"\n met_str = 'Meteorological Conditions'\n date_index, avg_suffix = deploy_timestamp_index(met_ref_df,\n averaging_suffix=True)\n\n #cal_check_dict = cal_check_dict['Met cal checks']\n for name in ['Temp', 'RH']:\n param_obj = Parameter(name)\n param_name = param_obj.name\n fmt_param = param_obj.format_name\n #fmt_param_units = param_obj.units\n\n no_data = False\n try:\n ref_name = met_ref_df.loc[:, param_name + '_Method'].dropna().apply(\n lambda x: str(x)).unique()[0]\n except IndexError:\n ref_name = 'Unknown Reference'\n except KeyError:\n # No met parameter data in passed reference dataframe\n no_data = True\n\n max_criterion = operational_range[param_name][1]\n min_criterion = operational_range[param_name][0]\n\n for group in deploy_dict['Deployment Groups']:\n deploy = deploy_dict['Deployment Groups'][group]\n start = deploy['eval_start']\n end = deploy['eval_end']\n\n if met_str not in deploy:\n deploy[met_str] = {}\n\n if fmt_param not in deploy[met_str]:\n deploy[met_str][fmt_param] = {}\n\n stats_loc = deploy[met_str][fmt_param]\n\n if not no_data:\n ref_data = met_ref_df.loc[start:end, param_name + '_Value']\n\n grp_idx = [int(i) - 1 for i in deploy['sensors'].keys()]\n data_pairs = []\n for idx in grp_idx:\n df = df_list[idx]\n start = df.index.min()\n end = df.index.max()\n data_pairs.append(\n met_ref_df.loc[start:end,\n param_name + '_Value'].dropna().size)\n\n stats_loc['instrument_name'] = ref_name\n stats_loc['min' + avg_suffix] = \\\n float(\"{0:.3f}\".format(ref_data.min()))\n stats_loc['max' + avg_suffix] = \\\n float(\"{0:.3f}\".format(ref_data.max()))\n\n if (max_criterion and min_criterion):\n value = int(ref_data.where((ref_data > max_criterion) |\n (ref_data < min_criterion)).count())\n else:\n value = None\n\n stats_loc['n_exceed_target_criteria' + avg_suffix] = value\n stats_loc['n_measurement_pairs' + avg_suffix] = np.mean(data_pairs)\n\n #deploy[met_str]['cal_check_dates'] = cal_check_dict\n else:\n stats_loc['instrument_name'] = ''\n stats_loc['min' + avg_suffix] = ''\n stats_loc['max' + avg_suffix] = ''\n stats_loc['n_exceed_target_criteria' + avg_suffix] = ''\n stats_loc['n_measurement_pairs' + avg_suffix] = ''\n\n return deploy_dict\n", "repo_name": "praful-dodda/sensortoolkit", "sub_path": "sensortoolkit/deploy/_create_deploy_dict.py", "file_name": "_create_deploy_dict.py", "file_ext": "py", "file_size_in_byte": 16818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "github-code", "pt": "24", "api": [{"api_name": "datetime.datetime.now", "line_number": 93, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 93, "usage_type": "name"}, {"api_name": "sensortoolkit.lib_utils._get_version", "line_number": 94, "usage_type": "call"}, {"api_name": "pandas.Timedelta", "line_number": 108, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 143, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 145, "usage_type": "call"}, {"api_name": "sensortoolkit.datetime_utils.get_timestamp_interval", "line_number": 162, "usage_type": "call"}, {"api_name": "sensortoolkit.calculate.uptime", "line_number": 166, "usage_type": "call"}, {"api_name": "sensortoolkit.calculate.uptime", "line_number": 170, "usage_type": "call"}, {"api_name": "sensortoolkit.param.Parameter", "line_number": 222, "usage_type": "call"}, {"api_name": "sensortoolkit.datetime_utils.deploy_timestamp_index", "line_number": 225, "usage_type": "call"}, {"api_name": "sensortoolkit.datetime_utils.deploy_timestamp_index", "line_number": 325, "usage_type": "call"}, {"api_name": "sensortoolkit.param.Parameter", "line_number": 330, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 387, "usage_type": "call"}]}
+{"seq_id": "26492379439", "text": "import logging\n\nclass PerfLogger:\n def __init__(self):\n logging.basicConfig(level=logging.DEBUG)\n \n def get_logger(self, name=None, log_file='./logs/output.log'):\n logger = logging.getLogger(name)\n formatter = logging.Formatter(u'%(asctime)s [%(levelname)8s] | %(message)s')\n file_handler = logging.FileHandler(log_file)\n file_handler.setFormatter(formatter)\n logger.addHandler(file_handler)\n\n return logger\n", "repo_name": "jinwonkim93/perf_stats", "sub_path": "perf_stats/logger.py", "file_name": "logger.py", "file_ext": "py", "file_size_in_byte": 466, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "logging.basicConfig", "line_number": 5, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 5, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 10, "usage_type": "call"}]}
+{"seq_id": "36786804815", "text": "import json\nfrom typing import Tuple, Callable, Any\n\nimport geopandas\nimport numpy as np\nimport pandas as pd\nimport requests\nfrom dotenv import dotenv_values\nfrom geopandas import GeoDataFrame, GeoSeries, points_from_xy\nfrom pandas import DataFrame\nfrom scipy.spatial import cKDTree\n\nfrom date_range import DateRange\n\nsecrets = dotenv_values('.env')\n\n\ndef get_waterlevel_flow_observations(ids: str, date_range: DateRange):\n parameters = {\n 'StationId': ids,\n 'Parameter': '1000,1001',\n 'ResolutionTime': 60,\n 'ReferenceTime': date_range\n }\n url = f\"https://hydapi.nve.no/api/v1/Observations\"\n request_headers = {\n \"Accept\": \"application/json\",\n \"X-API-Key\": \"JkbAM/hEkk+5Z7mJIlC3fQ==\",\n }\n nve_stations = requests.get(url, parameters, headers=request_headers)\n parsed_result = nve_stations.json()\n df = pd.json_normalize(parsed_result['data'])\n return df\n\n\ndef get_nearest_station_obesrvation(shape: GeoDataFrame, date_range: DateRange):\n url = f'https://hydapi.nve.no/api/v1/Stations?Active=1'\n request_headers = {\n \"Accept\": \"application/json\",\n \"X-API-Key\": \"JkbAM/hEkk+5Z7mJIlC3fQ==\",\n }\n nve_stations = requests.get(url, headers=request_headers)\n parsed_result = nve_stations.json()\n df = pd.json_normalize(parsed_result['data'])\n df['parameters'] = [[x['parameter'] for x in i] for i in df.seriesList]\n df = df[df['parameters'].apply(lambda x: 1000 in x or 1001 in x)]\n df['seriesList'] = [[x for x in i if x['parameter'] == 1000 or x['parameter'] == 1001] for i in df.seriesList]\n df['dateRange'] = [[[DateRange(f\"{y['dataFromTime'][:10]}/{y['dataToTime'][:10]}\") for y in x['resolutionList'] if\n y['resTime'] == 60] for x in i] for i in df.seriesList]\n df = df[df['dateRange'].apply(lambda x: bool(list(filter(None, x))))]\n gdf = geopandas.GeoDataFrame(\n df[[\"stationId\", \"stationName\", \"latitude\", \"longitude\", \"seriesList\"]],\n geometry=geopandas.points_from_xy(df.longitude, df.latitude), crs=4326)\n culvert = shape.sjoin_nearest(gdf).merge(gdf, left_on=\"index_right\", right_index=True)\n culvert_id = culvert.iloc[0]['stationId_x']\n observations = get_waterlevel_flow_observations(culvert_id, date_range)\n return observations\n", "repo_name": "Xevorius/7Analytics-API", "sub_path": "NVE/nve_api.py", "file_name": "nve_api.py", "file_ext": "py", "file_size_in_byte": 2305, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "dotenv.dotenv_values", "line_number": 15, "usage_type": "call"}, {"api_name": "date_range.DateRange", "line_number": 18, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 30, "usage_type": "call"}, {"api_name": "pandas.json_normalize", "line_number": 32, "usage_type": "call"}, {"api_name": "geopandas.GeoDataFrame", "line_number": 36, "usage_type": "name"}, {"api_name": "date_range.DateRange", "line_number": 36, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 42, "usage_type": "call"}, {"api_name": "pandas.json_normalize", "line_number": 44, "usage_type": "call"}, {"api_name": "date_range.DateRange", "line_number": 48, "usage_type": "call"}, {"api_name": "geopandas.GeoDataFrame", "line_number": 51, "usage_type": "call"}, {"api_name": "geopandas.points_from_xy", "line_number": 53, "usage_type": "call"}]}
+{"seq_id": "71720900542", "text": "from django.views.generic import ListView, DetailView\n\nfrom .models import Post, Tag\n\n\nclass PostList(ListView):\n model = Post\n paginate_by = 15\n\n def get_context_data(self, **kwargs):\n context = super().get_context_data(**kwargs)\n context['tags'] = Tag.objects.all()[:15]\n return context\n\n\nclass PostDetail(DetailView):\n model = Post\n\n def get_context_data(self, **kwargs):\n context = super().get_context_data(**kwargs)\n context['tags'] = Tag.objects.all()[:15]\n return context\n", "repo_name": "python-krasnodar/python-krasnodar.ru", "sub_path": "src/blog/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 536, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.views.generic.ListView", "line_number": 6, "usage_type": "name"}, {"api_name": "models.Post", "line_number": 7, "usage_type": "name"}, {"api_name": "models.Tag.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Tag.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "models.Tag", "line_number": 12, "usage_type": "name"}, {"api_name": "django.views.generic.DetailView", "line_number": 16, "usage_type": "name"}, {"api_name": "models.Post", "line_number": 17, "usage_type": "name"}, {"api_name": "models.Tag.objects.all", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Tag.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "models.Tag", "line_number": 21, "usage_type": "name"}]}
+{"seq_id": "19300503039", "text": "import keras.backend as K\nfrom keras.engine import Layer\nfrom keras.layers import Input, Embedding, Bidirectional, GRU, Convolution1D, GlobalMaxPooling1D, TimeDistributed, \\\n concatenate, Dense, Activation, Dropout\nfrom keras.models import Model\nfrom keras.optimizers import Adam\nfrom keras.utils.training_utils import multi_gpu_model\n\nfrom custom_layers import Squeeze, BetterTimeDistributed, RepeatToMatch, ElementAt, AttentionWeightedAverage\n\n__author__ = 'sjebbara'\n\n\ndef pairwise_ranking_metric(labels, diffs):\n return K.mean(diffs > 0)\n\n\ndef margin_ranking_loss(label, diff, margin=1.0):\n ranking_loss = K.maximum(0, margin - diff)\n\n return ranking_loss\n\n\ndef batch_margin_ranking_loss(dummy, scores, margin=1.0):\n pos_score = scores[:, :1]\n neg_scores = scores[:, 1:]\n\n ranking_losses = K.maximum(0.0, margin + neg_scores - pos_score) # pos score broadcasted\n\n ranking_loss = K.sum(ranking_losses, axis=-1) # todo: sum or mean?\n return ranking_loss\n\n\n# def batch_pairwise_margin_ranking_loss(labels, scores, margin=1.0):\n# pos_scores = K.expand_dims(scores[(labels >= 0.5).nonzero()], dim=-1)\n# neg_scores = K.expand_dims(scores[(labels <= 0.5).nonzero()], dim=-1)\n#\n# pos_tmp = K.variable(numpy.array([[margin]]))\n# pos_scores = K.concatenate((pos_scores, pos_tmp), axis=0)\n#\n# neg_tmp = K.variable(numpy.array([[0.]]))\n# neg_scores = K.concatenate((neg_scores, neg_tmp), axis=0)\n#\n# n_p = K.shape(pos_scores)[0]\n# n_n = K.shape(neg_scores)[0]\n# repeated_pos_scores = K.repeat(pos_scores, n_n, )\n# repeated_neg_scores = K.transpose(K.repeat(neg_scores, n_p))\n#\n# ranking_losses = K.maximum(0, margin - repeated_pos_scores + repeated_neg_scores)\n#\n# return K.sum(ranking_losses)\n\n\n### GENERAL COMPONENTS ###\ndef apply_sequence_model(layer_type, input_sequence, embedding_size, depth, kernel_size, dropout):\n # if pooling == \"last\":\n # return_sequences = False\n # else:\n # return_sequences = True\n\n embedding_sequence = input_sequence\n\n for d in range(depth):\n if layer_type == \"rnn\":\n embedding_sequence = Bidirectional(GRU(embedding_size, activation=\"selu\", kernel_initializer=\"he_normal\",\n return_sequences=True if d < depth - 1 else False))(\n embedding_sequence)\n elif layer_type == \"cnn\":\n embedding_sequence = Convolution1D(embedding_size, kernel_size=kernel_size, activation=\"selu\",\n kernel_initializer=\"he_normal\", padding=\"same\")(embedding_sequence)\n embedding_sequence = Dropout(dropout)(embedding_sequence)\n\n if layer_type == \"cnn\":\n embedding = GlobalMaxPooling1D()(embedding_sequence)\n else:\n embedding = embedding_sequence\n\n return embedding\n\n\n### GENERAL COMPONENTS ###\ndef apply_sequence_attention_model(layer_type, input_sequence, relative_position_embeddings, embedding_size, depth,\n kernel_size, dropout):\n # if pooling == \"last\":\n # return_sequences = False\n # else:\n # return_sequences = True\n\n embedding_sequence = input_sequence\n\n for d in range(depth):\n if layer_type == \"rnn\":\n forward_embedding_sequence, backward_embedding_sequence = Bidirectional(\n GRU(embedding_size, activation=\"selu\", kernel_initializer=\"he_normal\",\n return_sequences=True), merge_mode=None)(\n embedding_sequence)\n embedding_sequence = concatenate([forward_embedding_sequence, backward_embedding_sequence])\n elif layer_type == \"cnn\":\n embedding_sequence = Convolution1D(embedding_size, kernel_size=kernel_size, activation=\"selu\",\n kernel_initializer=\"he_normal\", padding=\"same\")(embedding_sequence)\n embedding_sequence = Dropout(dropout)(embedding_sequence)\n\n if layer_type == \"cnn\":\n summary_embedding = GlobalMaxPooling1D()(embedding_sequence)\n else:\n forward_embedding = ElementAt(-1)(forward_embedding_sequence)\n backward_embedding = ElementAt(0)(backward_embedding_sequence)\n\n summary_embedding = concatenate([forward_embedding, backward_embedding])\n\n repeated_summary_embedding = RepeatToMatch()([summary_embedding, input_sequence])\n concatenated_attention_inputs = concatenate(\n [embedding_sequence, relative_position_embeddings, repeated_summary_embedding])\n\n attention_embedding = AttentionWeightedAverage()([embedding_sequence, concatenated_attention_inputs])\n return attention_embedding\n\n\ndef get_text_encoder(vocab_size, embedding_size, kernel_size, depth, dropout, embedding_layer=None, layer_type=None):\n text_input = Input(shape=(None,), dtype='int32', name='text_input')\n\n if embedding_layer is None:\n embedding_layer = Embedding(input_dim=vocab_size, output_dim=embedding_size, name=\"word_embeddings\")\n\n embedding_sequence = embedding_layer(text_input)\n\n if layer_type is None:\n layer_type = \"cnn\"\n\n text_embedding = apply_sequence_model(layer_type, input_sequence=embedding_sequence, embedding_size=embedding_size,\n depth=depth, kernel_size=kernel_size, dropout=dropout)\n\n model = Model(inputs=[text_input], outputs=[text_embedding])\n\n return model\n\n\n### SPECIFIC COMPONENTS ###\n#\n# def get_question_encoder(word_vocab_size, word_embedding_size, question_embedding_size):\n# question_input = Input(shape=(None,), dtype='int32', name='question_input')\n#\n# word_embedding_sequence = Embedding(input_dim=word_vocab_size, output_dim=word_embedding_size,\n# name=\"word_embeddings\")(question_input)\n#\n# question_embedding = apply_sequence_model(\"cnn\", input_sequence=word_embedding_sequence,\n# embedding_size=question_embedding_size, depth=5, pooling=\"max\",\n# kernel_size=3)\n#\n# model = Model(inputs=[question_input], outputs=[question_embedding])\n# return model\n#\n#\n# def get_predicate_label_encoder(word_vocab_size, word_embedding_size, predicate_vocab_size, predicate_embedding_size):\n# predicate_label_input = Input(shape=(None,), dtype='int32', name='predicate_label_input')\n#\n# predicate_word_embedding_sequence = Embedding(input_dim=word_vocab_size, output_dim=word_embedding_size,\n# name=\"predicate_word_embeddings\")(predicate_label_input)\n#\n# predicate_label_embedding = apply_sequence_model(\"cnn\", input_sequence=predicate_word_embedding_sequence,\n# embedding_size=predicate_embedding_size, depth=2, pooling=\"max\",\n# kernel_size=3)\n#\n# model = Model(inputs=[predicate_label_input], outputs=[predicate_label_embedding])\n# return model\n#\n#\n# def get_entity_label_encoder(char_vocab_size, char_embedding_size, char_kernel_size, entity_embedding_size):\n# entity_label_input = Input(shape=(None,), dtype='int32', name='predicate_hierarchy_input')\n#\n# entity_label_embedding_sequence = Embedding(input_dim=char_vocab_size, output_dim=char_embedding_size,\n# name=\"entity_char_embeddings\")(entity_label_input)\n#\n# entity_label_embedding = apply_sequence_model(\"cnn\", input_sequence=entity_label_embedding_sequence,\n# embedding_size=entity_embedding_size, depth=2, pooling=\"max\",\n# kernel_size=3)\n#\n# model = Model(inputs=[entity_label_input], outputs=[entity_label_embedding])\n# return model\n#\n\ndef simple_joint_qa(conf, res, **kwargs):\n ### setup inputs ###\n # word_vocab_size = conf., word_embedding_size, word_embedding_weights, char_vocab_size, char_embedding_size,\n # match_embedding_size,\n\n ### INPUT LIST####\n inputs = []\n\n question_char_input = Input(shape=(None,), dtype='int32', name='question_char_input')\n question_token_input = Input(shape=(None,), dtype='int32', name='question_token_input')\n\n candidate_subject_labels_input = Input(shape=(None, None), dtype='int32', name='candidate_subject_labels_input')\n candidate_predicate_labels_input = Input(shape=(None, None), dtype='int32', name='candidate_predicate_labels_input')\n\n ## add to inputs\n inputs.append(question_char_input)\n inputs.append(question_token_input)\n inputs.append(candidate_subject_labels_input)\n inputs.append(candidate_predicate_labels_input)\n\n ### setup encoders ###\n word_embedding_layer = Embedding(input_dim=conf.word_vocab_size, output_dim=conf.word_embedding_size,\n weights=[res.word_embeddings.W], name=\"word_embeddings\")\n\n char_sequence_encoder = get_text_encoder(vocab_size=conf.char_vocab_size, embedding_size=conf.char_embedding_size,\n kernel_size=conf.question_char_kernel_size,\n depth=conf.question_char_cnn_depth, dropout=conf.dropout,\n layer_type=conf.layer_type)\n predicate_token_sequence_encoder = get_text_encoder(vocab_size=conf.word_vocab_size,\n embedding_size=conf.word_embedding_size,\n kernel_size=conf.predicate_token_kernel_size,\n depth=conf.predicate_token_cnn_depth,\n embedding_layer=word_embedding_layer, dropout=conf.dropout,\n layer_type=conf.layer_type)\n\n question_token_sequence_encoder = get_text_encoder(vocab_size=conf.word_vocab_size,\n embedding_size=conf.word_embedding_size,\n kernel_size=conf.question_token_kernel_size,\n depth=conf.question_token_cnn_depth,\n embedding_layer=word_embedding_layer, dropout=conf.dropout,\n layer_type=conf.layer_type)\n\n ### encode inputs ###\n question_char_embedding = char_sequence_encoder(question_char_input)\n question_token_embedding = question_token_sequence_encoder(question_token_input)\n\n subject_label_embeddings = BetterTimeDistributed(char_sequence_encoder)(candidate_subject_labels_input)\n predicate_label_embeddings = BetterTimeDistributed(predicate_token_sequence_encoder)(\n candidate_predicate_labels_input)\n\n repeated_question_char_embedding = RepeatToMatch()([question_char_embedding, subject_label_embeddings])\n repeated_question_token_embedding = RepeatToMatch()([question_token_embedding, predicate_label_embeddings])\n\n ### compute partial matches ###\n ## use graph embeddings\n if conf.use_graph_embeddings:\n candidate_subject_graph_embedding_input = Input(shape=(None,), dtype='int32',\n name='candidate_subject_graph_embeddings_input')\n\n candidate_predicate_graph_embedding_input = Input(shape=(None,), dtype='int32',\n name='candidate_predicate_graph_embeddings_input')\n\n ## add to inputs\n inputs.append(candidate_subject_graph_embedding_input)\n inputs.append(candidate_predicate_graph_embedding_input)\n\n graph_embedding_layer = Embedding(input_dim=conf.graph_vocab_size, output_dim=conf.graph_embedding_size,\n weights=[res.graph_embeddings.W], name=\"graph_embeddings\")\n\n subject_graph_embeddings = graph_embedding_layer(candidate_subject_graph_embedding_input)\n predicate_graph_embeddings = graph_embedding_layer(candidate_predicate_graph_embedding_input)\n\n subject_match_embeddings = concatenate(\n [repeated_question_char_embedding, subject_label_embeddings, subject_graph_embeddings])\n\n predicate_match_embeddings = concatenate(\n [repeated_question_token_embedding, predicate_label_embeddings, predicate_graph_embeddings])\n\n else:\n subject_match_embeddings = concatenate([repeated_question_char_embedding, subject_label_embeddings])\n predicate_match_embeddings = concatenate([repeated_question_token_embedding, predicate_label_embeddings])\n\n subject_match_embeddings = BetterTimeDistributed(\n Dense(conf.match_embedding_size, activation=\"selu\", kernel_initializer=\"he_normal\"))(subject_match_embeddings)\n subject_match_embeddings = Dropout(conf.dropout)(subject_match_embeddings)\n\n predicate_match_embeddings = BetterTimeDistributed(\n Dense(conf.match_embedding_size, activation=\"selu\", kernel_initializer=\"he_normal\"))(predicate_match_embeddings)\n predicate_match_embeddings = Dropout(conf.dropout)(predicate_match_embeddings)\n\n if conf.use_predicate_and_subject_outputs:\n ### SUBJECT output\n subject_match_scores = BetterTimeDistributed(Dense(1, activation=\"linear\"))(subject_match_embeddings)\n subject_match_scores = Squeeze()(subject_match_scores)\n subject_match_scores = Activation(\"sigmoid\")(subject_match_scores)\n subject_answer_scores = Layer(name=\"subject_answer_scores\")(subject_match_scores)\n\n ### PREDICATE output\n predicate_match_scores = BetterTimeDistributed(Dense(1, activation=\"linear\"))(predicate_match_embeddings)\n predicate_match_scores = Squeeze()(predicate_match_scores)\n predicate_match_scores = Activation(\"sigmoid\")(predicate_match_scores)\n predicate_answer_scores = Layer(name=\"predicate_answer_scores\")(predicate_match_scores)\n\n ### compute overall matches ###\n overall_match_embedding = concatenate([subject_match_embeddings, predicate_match_embeddings])\n overall_match_embedding = BetterTimeDistributed(\n Dense(conf.match_embedding_size, activation=\"selu\", kernel_initializer=\"he_normal\"))(\n overall_match_embedding)\n overall_match_embedding = Dropout(conf.dropout)(overall_match_embedding)\n\n overall_match_scores = BetterTimeDistributed(Dense(1, activation=\"linear\"))(overall_match_embedding)\n overall_match_scores = Squeeze()(overall_match_scores)\n\n overall_match_scores = Activation(\"softmax\")(overall_match_scores)\n\n answer_scores = Layer(name=\"answer_scores\")(overall_match_scores)\n\n #### 3 OUTPUTS -: pair, subject and predicate alone\n outputs = [answer_scores, subject_answer_scores, predicate_answer_scores]\n\n model = Model(inputs=inputs, outputs=outputs)\n model.compile(Adam(), {\"answer_scores\": \"categorical_crossentropy\",\n \"predicate_answer_scores\": \"binary_crossentropy\",\n \"subject_answer_scores\": \"binary_crossentropy\"},\n metrics=[\"accuracy\", \"crossentropy\"],\n loss_weights={\"answer_scores\": conf.answer_loss_weight,\n \"predicate_answer_scores\": conf.predicate_answer_loss_weight,\n \"subject_answer_scores\": conf.subject_answer_loss_weight})\n\n #### SINGLE OUTPUT : CandidatePAIR Scores\n else:\n ### compute overall matches ###\n overall_match_embedding = concatenate([subject_match_embeddings, predicate_match_embeddings])\n overall_match_embedding = BetterTimeDistributed(\n Dense(conf.match_embedding_size, activation=\"selu\", kernel_initializer=\"he_normal\"))(\n overall_match_embedding)\n overall_match_embedding = Dropout(conf.dropout)(overall_match_embedding)\n\n overall_match_scores = BetterTimeDistributed(Dense(1, activation=\"linear\"))(overall_match_embedding)\n overall_match_scores = Squeeze()(overall_match_scores)\n\n overall_match_scores = Activation(\"softmax\")(overall_match_scores)\n\n answer_scores = Layer(name=\"answer_scores\")(overall_match_scores)\n\n outputs = [answer_scores]\n\n model = Model(inputs=inputs, outputs=outputs)\n model.compile(Adam(), \"categorical_crossentropy\", metrics=[\"accuracy\", \"crossentropy\"])\n\n return model\n\n\ndef simple_joint_qa_predicate_model(conf, res, **kwargs):\n ### INPUT LIST####\n inputs = []\n\n ### setup encoders ###\n word_embedding_layer = Embedding(input_dim=conf.word_vocab_size, output_dim=conf.word_embedding_size,\n weights=[res.word_embeddings.W], name=\"word_embeddings\")\n\n\n\n if conf.predicate_encoder_embedding_type == \"word\":\n question_token_input = Input(shape=(None,), dtype='int32', name='question_token_input')\n inputs.append(question_token_input)\n token_embeddings = word_embedding_layer(question_token_input)\n\n resulting_embeddings = token_embeddings\n\n elif conf.predicate_encoder_embedding_type == \"char\":\n question_char_input = Input(shape=(None, None), dtype='int32', name='question_char_input')\n inputs.append(question_char_input)\n char_sequence_encoder = get_text_encoder(vocab_size=conf.char_vocab_size,\n embedding_size=conf.char_embedding_size,\n kernel_size=conf.question_char_kernel_size,\n depth=conf.question_char_cnn_depth, dropout=conf.dropout,\n layer_type=conf.layer_type)\n\n token_char_embeddings = BetterTimeDistributed(char_sequence_encoder)(question_char_input)\n\n resulting_embeddings = token_char_embeddings\n\n ### combination OF WORD and CHAR embeddings\n else:\n question_char_input = Input(shape=(None, None), dtype='int32', name='question_char_input')\n question_token_input = Input(shape=(None,), dtype='int32', name='question_token_input')\n\n ## add to inputs\n inputs.append(question_char_input)\n inputs.append(question_token_input)\n\n char_sequence_encoder = get_text_encoder(vocab_size=conf.char_vocab_size,\n embedding_size=conf.char_embedding_size,\n kernel_size=conf.question_char_kernel_size,\n depth=conf.question_char_cnn_depth, dropout=conf.dropout,\n layer_type=conf.layer_type)\n\n token_char_embeddings = BetterTimeDistributed(char_sequence_encoder)(question_char_input)\n\n token_embeddings = word_embedding_layer(question_token_input)\n\n ## concatenate\n resulting_embeddings = concatenate([token_embeddings, token_char_embeddings])\n\n if \"att\" in conf.predicate_encoder_embedding_type:\n relative_position_input = Input(shape=(None,), dtype='int32', name='relative_position_input')\n\n inputs.append(relative_position_input)\n\n relative_position_embeddings = Embedding(input_dim=conf.subject_position_max_distance * 2 + 1,\n output_dim=conf.distance_embedding_size, name=\"distance_embeddings\")(\n relative_position_input)\n\n text_embedding = apply_sequence_attention_model(conf.layer_type, input_sequence=resulting_embeddings,\n relative_position_embeddings=relative_position_embeddings,\n embedding_size=conf.predicate_embedding_size,\n depth=conf.predicate_embedding_depth,\n kernel_size=conf.predicate_embedding_kernel_size,\n dropout=conf.dropout)\n else:\n text_embedding = apply_sequence_model(conf.layer_type, input_sequence=resulting_embeddings,\n embedding_size=conf.predicate_embedding_size,\n depth=conf.predicate_embedding_depth,\n kernel_size=conf.predicate_embedding_kernel_size, dropout=conf.dropout)\n\n\n ### add DROPOUT\n text_embedding = Dropout(conf.dropout)(text_embedding)\n\n #### Different outputs based on predicate_model_type ####\n if conf.predicate_model_type == \"predict_all_predicates\":\n answer_scores = Dense(conf.predicate_vocab_size, activation=\"softmax\")(text_embedding)\n answer_scores = Layer(name=\"answer_scores\")(answer_scores)\n # loss_function = \"categorical_crossentropy\"\n # metrics = \"accuracy\"\n\n elif conf.predicate_model_type == \"predict_graph_embedding\":\n answer_scores = Dense(conf.graph_embedding_size, activation=\"selu\")(text_embedding)\n answer_scores = Dense(conf.graph_embedding_size, activation=\"linear\")(answer_scores)\n answer_scores = Layer(name=\"answer_scores\")(answer_scores)\n # loss_function = \"cosine_proximity\"\n # metrics = \"cosine_proximity\"\n else:\n #### BINARY classification for candidate predicate\n\n graph_embedding_layer = Embedding(input_dim=conf.graph_vocab_size, output_dim=conf.graph_embedding_size,\n weights=[res.graph_embeddings.W], name=\"graph_embeddings\")\n\n predicate_graph_embedding_input = Input(shape=(None,), dtype='int32', name='predicate_graph_embedding_input')\n\n inputs.append(predicate_graph_embedding_input)\n\n predicate_graph_embedding = graph_embedding_layer(predicate_graph_embedding_input)\n predicate_graph_embedding = ElementAt(0)(predicate_graph_embedding)\n\n concatenated = concatenate([predicate_graph_embedding, text_embedding])\n\n concatenated_layer_output = Dense(200, activation=\"selu\")(concatenated)\n answer_scores = Dense(1, activation=\"sigmoid\")(concatenated_layer_output)\n answer_scores = Layer(name=\"answer_scores\")(answer_scores)\n # loss_function = \"binary_crossentropy\"\n # metrics = \"accuracy\"\n\n outputs = [answer_scores]\n\n model = Model(inputs=inputs, outputs=outputs)\n\n # if conf.gpus > 1:\n # model = multi_gpu_model(model, gpus=conf.gpus)\n\n model.compile(Adam(), conf.predicate_model_loss_function, metrics=[conf.predicate_model_metrics])\n\n return model\n", "repo_name": "ag-sc/SimpleQA", "sub_path": "src/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 22709, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "keras.backend.mean", "line_number": 15, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 15, "usage_type": "name"}, {"api_name": "keras.backend.maximum", "line_number": 19, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 19, "usage_type": "name"}, {"api_name": "keras.backend.maximum", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 28, "usage_type": "name"}, {"api_name": "keras.backend.sum", "line_number": 30, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 30, "usage_type": "name"}, {"api_name": "keras.layers.Bidirectional", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers.GRU", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers.Convolution1D", "line_number": 69, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.layers.GlobalMaxPooling1D", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.layers.Bidirectional", "line_number": 93, "usage_type": "call"}, {"api_name": "keras.layers.GRU", "line_number": 94, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 97, "usage_type": "call"}, {"api_name": "keras.layers.Convolution1D", "line_number": 99, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 101, "usage_type": "call"}, {"api_name": "keras.layers.GlobalMaxPooling1D", "line_number": 104, "usage_type": "call"}, {"api_name": "custom_layers.ElementAt", "line_number": 106, "usage_type": "call"}, {"api_name": "custom_layers.ElementAt", "line_number": 107, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 109, "usage_type": "call"}, {"api_name": "custom_layers.RepeatToMatch", "line_number": 111, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 112, "usage_type": "call"}, {"api_name": "custom_layers.AttentionWeightedAverage", "line_number": 115, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 120, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 123, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 133, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 190, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 191, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 193, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 194, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 203, "usage_type": "call"}, {"api_name": "custom_layers.BetterTimeDistributed", "line_number": 228, "usage_type": "call"}, {"api_name": "custom_layers.BetterTimeDistributed", "line_number": 229, "usage_type": "call"}, {"api_name": "custom_layers.RepeatToMatch", "line_number": 232, "usage_type": "call"}, {"api_name": "custom_layers.RepeatToMatch", "line_number": 233, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 238, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 241, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 248, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 254, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 257, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 261, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 262, "usage_type": "call"}, {"api_name": "custom_layers.BetterTimeDistributed", "line_number": 264, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 265, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 266, "usage_type": "call"}, {"api_name": "custom_layers.BetterTimeDistributed", "line_number": 268, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 269, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 270, "usage_type": "call"}, {"api_name": "custom_layers.BetterTimeDistributed", "line_number": 274, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 274, "usage_type": "call"}, {"api_name": "custom_layers.Squeeze", "line_number": 275, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 276, "usage_type": "call"}, {"api_name": "keras.engine.Layer", "line_number": 277, "usage_type": "call"}, {"api_name": "custom_layers.BetterTimeDistributed", "line_number": 280, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 280, "usage_type": "call"}, {"api_name": "custom_layers.Squeeze", "line_number": 281, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 282, "usage_type": "call"}, {"api_name": "keras.engine.Layer", "line_number": 283, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 286, "usage_type": "call"}, {"api_name": "custom_layers.BetterTimeDistributed", "line_number": 287, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 288, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 290, "usage_type": "call"}, {"api_name": "custom_layers.BetterTimeDistributed", "line_number": 292, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 292, "usage_type": "call"}, {"api_name": "custom_layers.Squeeze", "line_number": 293, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 295, "usage_type": "call"}, {"api_name": "keras.engine.Layer", "line_number": 297, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 302, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 303, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 314, "usage_type": "call"}, {"api_name": "custom_layers.BetterTimeDistributed", "line_number": 315, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 316, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 318, "usage_type": "call"}, {"api_name": "custom_layers.BetterTimeDistributed", "line_number": 320, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 320, "usage_type": "call"}, {"api_name": "custom_layers.Squeeze", "line_number": 321, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 323, "usage_type": "call"}, {"api_name": "keras.engine.Layer", "line_number": 325, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 329, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 330, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 340, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 346, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 353, "usage_type": "call"}, {"api_name": "custom_layers.BetterTimeDistributed", "line_number": 361, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 367, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 368, "usage_type": "call"}, {"api_name": "custom_layers.BetterTimeDistributed", "line_number": 380, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 385, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 388, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 392, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 410, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 414, "usage_type": "call"}, {"api_name": "keras.engine.Layer", "line_number": 415, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 420, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 421, "usage_type": "call"}, {"api_name": "keras.engine.Layer", "line_number": 422, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 428, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 431, "usage_type": "call"}, {"api_name": "custom_layers.ElementAt", "line_number": 436, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 438, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 440, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 441, "usage_type": "call"}, {"api_name": "keras.engine.Layer", "line_number": 442, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 448, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 453, "usage_type": "call"}]}
+{"seq_id": "8941011676", "text": "import torch\n\nfrom megatron import get_args, print_rank_last\nfrom megatron import mpu\nfrom megatron.model.enums import AttnMaskType\nfrom megatron.model.bert_model import bert_extended_attention_mask, bert_position_ids\nfrom megatron.model.language_model import get_language_model\nfrom megatron.model.utils import get_linear_layer\nfrom megatron.model.utils import init_method_normal\nfrom megatron.model.utils import scaled_init_method_normal\nfrom .module import MegatronModule\n\n\nclass Classification(MegatronModule):\n\n def __init__(self,\n num_classes,\n num_tokentypes=2,\n pre_process=True,\n post_process=True):\n super(Classification, self).__init__(share_word_embeddings=False)\n args = get_args()\n\n self.num_classes = num_classes\n self.pre_process = pre_process\n self.post_process = post_process\n init_method = init_method_normal(args.init_method_std)\n\n self.language_model, self._language_model_key = get_language_model(\n num_tokentypes=num_tokentypes,\n add_pooler=True,\n encoder_attn_mask_type=AttnMaskType.padding,\n init_method=init_method,\n scaled_init_method=scaled_init_method_normal(args.init_method_std,\n args.num_layers),\n pre_process=self.pre_process,\n post_process=self.post_process)\n\n # Multi-choice head.\n if self.post_process:\n self.classification_dropout = torch.nn.Dropout(args.hidden_dropout)\n self.classification_head = get_linear_layer(args.hidden_size,\n self.num_classes,\n init_method)\n self._classification_head_key = 'classification_head'\n\n def set_input_tensor(self, input_tensor):\n \"\"\"See megatron.model.transformer.set_input_tensor()\"\"\"\n self.language_model.set_input_tensor(input_tensor)\n\n def forward(self, model_input, attention_mask, tokentype_ids=None):\n\n extended_attention_mask = bert_extended_attention_mask(attention_mask)\n input_ids = model_input\n position_ids = bert_position_ids(input_ids)\n\n lm_output = self.language_model(\n input_ids,\n position_ids,\n extended_attention_mask,\n tokentype_ids=tokentype_ids\n )\n\n if self.post_process:\n _, pooled_output = lm_output\n classification_output = self.classification_dropout(pooled_output)\n classification_logits = self.classification_head(classification_output)\n\n # Reshape back to separate choices.\n classification_logits = classification_logits.view(-1, self.num_classes)\n\n return classification_logits\n return lm_output\n\n def state_dict_for_save_checkpoint(self, destination=None, prefix='',\n keep_vars=False):\n \"\"\"For easy load when model is combined with other heads,\n add an extra key.\"\"\"\n\n state_dict_ = {}\n state_dict_[self._language_model_key] \\\n = self.language_model.state_dict_for_save_checkpoint(\n destination, prefix, keep_vars)\n if self.post_process:\n state_dict_[self._classification_head_key] \\\n = self.classification_head.state_dict(\n destination, prefix, keep_vars)\n return state_dict_\n\n def load_state_dict(self, state_dict, strict=True):\n \"\"\"Customized load.\"\"\"\n\n self.language_model.load_state_dict(\n state_dict[self._language_model_key], strict=strict)\n if self.post_process:\n if self._classification_head_key in state_dict:\n self.classification_head.load_state_dict(\n state_dict[self._classification_head_key], strict=strict)\n else:\n print_rank_last('***WARNING*** could not find {} in the checkpoint, '\n 'initializing to random'.format(\n self._classification_head_key))\n", "repo_name": "mlcommons/training", "sub_path": "large_language_model/megatron-lm/megatron/model/classification.py", "file_name": "classification.py", "file_ext": "py", "file_size_in_byte": 4184, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1501, "dataset": "github-code", "pt": "24", "api": [{"api_name": "module.MegatronModule", "line_number": 14, "usage_type": "name"}, {"api_name": "megatron.get_args", "line_number": 22, "usage_type": "call"}, {"api_name": "megatron.model.utils.init_method_normal", "line_number": 27, "usage_type": "call"}, {"api_name": "megatron.model.language_model.get_language_model", "line_number": 29, "usage_type": "call"}, {"api_name": "megatron.model.enums.AttnMaskType.padding", "line_number": 32, "usage_type": "attribute"}, {"api_name": "megatron.model.enums.AttnMaskType", "line_number": 32, "usage_type": "name"}, {"api_name": "megatron.model.utils.scaled_init_method_normal", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.Dropout", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "attribute"}, {"api_name": "megatron.model.utils.get_linear_layer", "line_number": 42, "usage_type": "call"}, {"api_name": "megatron.model.bert_model.bert_extended_attention_mask", "line_number": 53, "usage_type": "call"}, {"api_name": "megatron.model.bert_model.bert_position_ids", "line_number": 55, "usage_type": "call"}, {"api_name": "megatron.print_rank_last", "line_number": 100, "usage_type": "call"}]}
+{"seq_id": "71845732863", "text": "from rest_framework.authentication import TokenAuthentication\nfrom rest_framework import HTTP_HEADER_ENCODING, exceptions\n\n\n\nclass SyncClientAuthentication(TokenAuthentication):\n def __init__(self):\n from .client import Client\n self.client = Client()\n super(SyncClientAuthentication, self).__init__()\n\n def authenticate_credentials(self, key):\n if key:\n is_valid, client_errors = self.client.run_validation(key)\n if is_valid:\n return super(SyncClientAuthentication, self).authenticate_credentials(key)\n else:\n raise exceptions.AuthenticationFailed(client_errors)\n else:\n return exceptions.AuthenticationFailed('Authorization token not set')\n", "repo_name": "letsolexandr/django-single-auth-client", "sub_path": "single_sync/sync_client/authentication.py", "file_name": "authentication.py", "file_ext": "py", "file_size_in_byte": 757, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "rest_framework.authentication.TokenAuthentication", "line_number": 6, "usage_type": "name"}, {"api_name": "client.Client", "line_number": 9, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.AuthenticationFailed", "line_number": 18, "usage_type": "call"}, {"api_name": "rest_framework.exceptions", "line_number": 18, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.AuthenticationFailed", "line_number": 20, "usage_type": "call"}, {"api_name": "rest_framework.exceptions", "line_number": 20, "usage_type": "name"}]}
+{"seq_id": "36895911218", "text": "from django.contrib import admin\nfrom .models import Asset, Task, Worker, Allocation\nfrom uuid import UUID\n# Register your models here.\n# admin.site.register([Asset, Task, Worker, Allocation])\n@admin.register(Asset, Task, Worker)\nclass Admin(admin.ModelAdmin):\n list_display = ('name', 'id')\n\ndef allocation_details(obj):\n worker = Worker.objects.filter(id = UUID(str(obj.worker_id))).values()[0]['name']\n task = Task.objects.filter(id = UUID(str(obj.task_id))).values()[0]['name']\n asset = Asset.objects.filter(id = UUID(str(obj.asset_id))).values()[0]['name']\n return worker+' - '+task+' - '+asset\n@admin.register(Allocation)\nclass AllocAdmin(admin.ModelAdmin):\n list_display = (allocation_details,)\n", "repo_name": "mihirkawatra/housekeeping-portal", "sub_path": "household/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 720, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 7, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 6, "usage_type": "call"}, {"api_name": "models.Asset", "line_number": 6, "usage_type": "argument"}, {"api_name": "models.Task", "line_number": 6, "usage_type": "argument"}, {"api_name": "models.Worker", "line_number": 6, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "models.Worker.objects.filter", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Worker.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.Worker", "line_number": 11, "usage_type": "name"}, {"api_name": "uuid.UUID", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Task.objects.filter", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Task.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "models.Task", "line_number": 12, "usage_type": "name"}, {"api_name": "uuid.UUID", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Asset.objects.filter", "line_number": 13, "usage_type": "call"}, {"api_name": "models.Asset.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "models.Asset", "line_number": 13, "usage_type": "name"}, {"api_name": "uuid.UUID", "line_number": 13, "usage_type": "call"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 16, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 15, "usage_type": "call"}, {"api_name": "models.Allocation", "line_number": 15, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 15, "usage_type": "name"}]}
+{"seq_id": "44305486722", "text": "from __future__ import print_function # __future__模块可以引用未来版本python库中的函数,其实我的版本已经是3.x,所以没有必要走这一步\nfrom numpy import *\nimport operator\nfrom os import listdir\nfrom collections import Counter\nimport matplotlib\nimport matplotlib.pyplot as plt\n\nfr=open(\"E:\\MachineLearning\\KNN\\dataset\\datingTestSet2.txt\")\nnumberOfLines=len(fr.readlines())\nfr.seek(0)\nreturnMat=zeros((numberOfLines,3))\nclassLabelVector=[]\nindex=0\nfor line in fr.readlines():\n line=line.strip()\n listFromLine=line.split('\\t')\n returnMat[index,:]=listFromLine[0:3]\n classLabelVector.append(int(listFromLine[-1])) # classLabelVector 与 returnMat 一一对应,前者为所属标签,后者为具体数据\n index+=1\n\nlabels=classLabelVector\n\nplt.figure(figsize=(8, 5), dpi=80)\naxes = plt.subplot(111)\n# 将三类数据分别取出来\n# x轴代表飞行的里程数\n# y轴代表玩视频游戏的百分比\ntype1_x = []\ntype1_y = []\ntype2_x = []\ntype2_y = []\ntype3_x = []\ntype3_y = []\n\nfor i in range(len(labels)):\n if labels[i] == 1: # 不喜欢\n type1_x.append(returnMat[i][0])\n type1_y.append(returnMat[i][1])\n\n if labels[i] == 2: # 魅力一般\n type2_x.append(returnMat[i][0])\n type2_y.append(returnMat[i][1])\n\n if labels[i] == 3: # 极具魅力\n type3_x.append(returnMat[i][0])\n type3_y.append(returnMat[i][1])\n\ntype1 = axes.scatter(type1_x, type1_y, s=20, c='red')\ntype2 = axes.scatter(type2_x, type2_y, s=40, c='green')\ntype3 = axes.scatter(type3_x, type3_y, s=50, c='blue')\n\nplt.xlabel('Miles earned per year')\nplt.ylabel('Percentage of events spent playing video games')\naxes.legend((type1, type2, type3), (u'dislike', u'Charismatic', u'Very attractive'), loc=2,)\n\nplt.show()\n\n\"\"\"\n下面进行归一化操作\n公式:线性转换: newValue=(oldValue-min)/(max-min) max和min为该组数据集中的最大最小特征值\n 对数转换: y=log10(x)\n 反余切转换:y=arctan(x)*2/PI \n\"\"\"\n\nnormMat=zeros((numberOfLines,3))\nfor i in [0,1,2]:\n max=returnMat[:,i].max()\n min=returnMat[:,i].min()\n for j in range(1000):\n normMat[j,i]=(returnMat[j,i]-min)/(max-min)\n# done\n\n\"\"\"\n对于每一个在数据集中的数据点:\n 计算目标的数据点(需要分类的数据点)与该数据点的距离\n 将距离排序:从小到大\n 选取前K个最短距离\n 选取这K个中最多的分类类别\n 返回该类别来作为目标数据点的预测值\n\"\"\"\n\ninx=[0.40166314, 0.56719748, 0.52034602] # 作为输入的样本\ndataSetSize=normMat.shape[0]\ndiffMat = tile(inx, (dataSetSize,1))-normMat\nsqDiffMat = diffMat**2\ndistances=(sqDiffMat.sum(axis=1))**0.5\nsortedDistIndicies = distances.argsort() # argsort()函数是将x中的元素从小到大排列,提取其对应的index(索引号),所以并不会影响对应的label的值\n# k取5\nclassCount={}\nfor i in range(5):\n voteIlabel = labels[sortedDistIndicies[i]]\n classCount[voteIlabel] = classCount.get(voteIlabel,0) + 1\nsortedClassCount = sorted(classCount.items(), key=operator.itemgetter(1), reverse=True)\n", "repo_name": "xiuzheDorothy/DL_exercise", "sub_path": "KNN/test_for.py", "file_name": "test_for.py", "file_ext": "py", "file_size_in_byte": 3129, "program_lang": "python", "lang": "zh", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "operator.itemgetter", "line_number": 94, "usage_type": "call"}]}
+{"seq_id": "73188657983", "text": "#!/usr/bin/python\n#-*- encoding: utf-8 -*-\n\nfrom zope.i18n import translate\nfrom Products.Five.browser.pagetemplatefile import ViewPageTemplateFile\nfrom plone.app.layout.viewlets.common import TitleViewlet\nfrom plone.app.layout.viewlets.common import LogoViewlet\nfrom plone.app.layout.links.viewlets import FaviconViewlet\n\n\nclass TitleViewlet(TitleViewlet):\n \"\"\"Custom Title\n \"\"\"\n def update(self):\n super(TitleViewlet, self).update()\n self.site_title = u\"%s — %s\" % (\n translate('portal_title',\n domain='plone',\n context=self.request,\n default='TNCR GIS'),\n self.page_title)\n\n\nclass FaviconViewlet(FaviconViewlet):\n \"\"\"Custom Favicon\n \"\"\"\n _template = ViewPageTemplateFile('favicon.pt')\n\n\nclass LogoViewlet(LogoViewlet):\n \"\"\"Custom Logo\n \"\"\"\n def update(self):\n super(LogoViewlet, self).update()\n self.logo_tag = ''\n\n", "repo_name": "l34marr/tncr.theme", "sub_path": "src/tncr/theme/browser/viewlets.py", "file_name": "viewlets.py", "file_ext": "py", "file_size_in_byte": 1051, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "zope.i18n.translate", "line_number": 17, "usage_type": "call"}, {"api_name": "Products.Five.browser.pagetemplatefile.ViewPageTemplateFile", "line_number": 27, "usage_type": "call"}]}
+{"seq_id": "27830003279", "text": "from selenium import webdriver\nfrom bs4 import BeautifulSoup\nimport time\n\ndriver = webdriver.Chrome()\n\ndriver.get(\"https://localprod.pandateacher.com/python-manuscript/hello-spiderman/\")\nprint(\"页面请求成功\")\ntime.sleep(1)\n\nprint(\"点击提交按钮,进入禅页面\")\ndriver.find_element_by_css_selector(\"#div2 .sub\").click()\ntime.sleep(1)\n\npageSource = driver.page_source\nsoup = BeautifulSoup(pageSource, \"html.parser\")\n# print(soup)\ncontents = soup.find_all(class_=\"content\")\nfor c in contents:\n print(type(c))\n print(c.find(id=\"p\").text)\n\ntime.sleep(2)\ndriver.close()\nprint(\"浏览器关闭\")\n", "repo_name": "picktsh/python", "sub_path": "code2/day09/练习2_python禅2selenium&beautifulsoup.py", "file_name": "练习2_python禅2selenium&beautifulsoup.py", "file_ext": "py", "file_size_in_byte": 608, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 5, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 5, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 9, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 13, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 16, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 23, "usage_type": "call"}]}
+{"seq_id": "13970561162", "text": "from bearlibterminal import terminal\nfrom loguru import logger\n\nfrom utilities import configUtilities, armourManagement, common, display, input_handlers, mobileHelp\n\n\ndef shopkeeper_armour(gameworld, shopkeeper_id):\n game_config = configUtilities.load_config()\n selected_menu_option = 0\n flavour_column_text = []\n player_entity = mobileHelp.MobileUtilities.get_player_entity(gameworld=gameworld)\n player_names = mobileHelp.MobileUtilities.get_mobile_name_details(gameworld=gameworld, entity=player_entity)\n player_first_name = player_names[0]\n\n mobileHelp.MobileUtilities.clear_talk_to_me_flag(gameworld=gameworld, target_entity=shopkeeper_id)\n mobileHelp.MobileUtilities.set_spoken_to_before_flag_to_true(gameworld=gameworld, target_entity=shopkeeper_id)\n\n dialog_frame_start_x = configUtilities.get_config_value_as_integer(configfile=game_config,\n section='gui', parameter='DIALOG_FRAME_START_X')\n dialog_frame_start_y = configUtilities.get_config_value_as_integer(configfile=game_config,\n section='gui', parameter='DIALOG_FRAME_START_Y')\n dialog_frame_width = configUtilities.get_config_value_as_integer(configfile=game_config,\n section='gui', parameter='DIALOG_FRAME_WIDTH')\n\n common.CommonUtils.draw_dialog_ui(gameworld=gameworld, game_config=game_config, entity_speaking=shopkeeper_id)\n\n armour_details, as_prefix_list, px_att_bonus, px_att_name, px_flavour = armourManagement.ArmourUtilities.get_all_armour_modifiers()\n\n as_display_name = armour_details[0]\n as_material = armour_details[1]\n flavour_column_text.append(px_flavour)\n\n starter_text = \"Ahhh if it isn't $1\"\n return_text = common.CommonUtils.replace_value_in_event(event_string=starter_text, par1=player_first_name)\n intro_text = return_text + \", and, I see you're wearing some \" + as_material + ' ' + as_display_name + ' armour, ' + 'tell me, what kind of modifier would you like adding?'\n\n menu_options = as_prefix_list\n max_menu_option = len(menu_options) - 1\n\n # armour column titles\n flavour_colour_string = '[color=SHOPKEEPER_ARMOUR_COLUMN_FLAVOUR]'\n flavour_coloumn_one_title = 'Flavour'\n flavour_column_one_string = flavour_colour_string + flavour_coloumn_one_title\n\n valid_event = False\n while not valid_event:\n\n # intro text\n terminal.print_(x=dialog_frame_start_x + 2, y=dialog_frame_start_y + 2, width=dialog_frame_width - 5,\n s=intro_text)\n\n display.pointy_vertical_menu(header='', menu_options=menu_options, menu_start_x=dialog_frame_start_x + 3,\n menu_start_y=dialog_frame_start_y + 9, blank_line=True, selected_option=selected_menu_option)\n\n # display flavour columns\n terminal.printf(x=dialog_frame_start_x + 15, y=dialog_frame_start_y + 8, s=flavour_column_one_string)\n\n # display attribute to be modified\n fg = \"[color=SHOPKEEPER_ARMOUR_ATTRIBUTE_FLAVOUR]\"\n\n # display flavour text\n display.coloured_list(list_options=flavour_column_text[0],\n list_x=dialog_frame_start_x + 17, list_y=dialog_frame_start_y + 9,\n selected_option='nothing', blank_line=True, fg=fg)\n\n # blit the console\n terminal.refresh()\n\n event_to_be_processed, event_action = input_handlers.handle_game_keys()\n if event_action == 'quit':\n valid_event = True\n if event_action in ('up', 'down'):\n selected_menu_option = common.CommonUtils.move_menu_selection(event_action=event_action,\n selected_menu_option=selected_menu_option,\n max_menu_option=max_menu_option)\n if event_action == 'enter':\n valid_event = True\n # apply shopkeeper bonus\n logger.debug('Armour modifier chosen is {}', as_prefix_list[selected_menu_option])\n logger.debug('Attribute to be modified is {}', px_att_name[selected_menu_option])\n logger.debug('Attribute bonus is {}', px_att_bonus[selected_menu_option])\n\n armour_set = armourManagement.ArmourUtilities.create_full_armour_set(gameworld=gameworld, prefix=as_prefix_list[selected_menu_option],\n armourset='Embroided')\n armourManagement.ArmourUtilities.equip_full_set_of_armour(gameworld=gameworld, entity=player_entity, armourset=armour_set)\n\n armourManagement.ArmourUtilities.apply_major_attribute_bonus_to_full_armourset(gameworld=gameworld,\n player_entity=player_entity,\n attribute_name=as_prefix_list[selected_menu_option],\n attribute_bonus=px_att_bonus[\n selected_menu_option])\n\n armourManagement.ArmourUtilities.set_mobile_derived_armour_attribute(gameworld=gameworld, entity=player_entity)\n mobileHelp.MobileUtilities.set_mobile_derived_attributes(gameworld=gameworld, entity=player_entity)\n", "repo_name": "devonps/talesfromticronem", "sub_path": "ui/shopkeeper_armour.py", "file_name": "shopkeeper_armour.py", "file_ext": "py", "file_size_in_byte": 5482, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 4, "dataset": "github-code", "pt": "24", "api": [{"api_name": "utilities.configUtilities.load_config", "line_number": 8, "usage_type": "call"}, {"api_name": "utilities.configUtilities", "line_number": 8, "usage_type": "name"}, {"api_name": "utilities.mobileHelp.MobileUtilities.get_player_entity", "line_number": 11, "usage_type": "call"}, {"api_name": "utilities.mobileHelp.MobileUtilities", "line_number": 11, "usage_type": "attribute"}, {"api_name": "utilities.mobileHelp", "line_number": 11, "usage_type": "name"}, {"api_name": "utilities.mobileHelp.MobileUtilities.get_mobile_name_details", "line_number": 12, "usage_type": "call"}, {"api_name": "utilities.mobileHelp.MobileUtilities", "line_number": 12, "usage_type": "attribute"}, {"api_name": "utilities.mobileHelp", "line_number": 12, "usage_type": "name"}, {"api_name": "utilities.mobileHelp.MobileUtilities.clear_talk_to_me_flag", "line_number": 15, "usage_type": "call"}, {"api_name": "utilities.mobileHelp.MobileUtilities", "line_number": 15, "usage_type": "attribute"}, {"api_name": "utilities.mobileHelp", "line_number": 15, "usage_type": "name"}, {"api_name": "utilities.mobileHelp.MobileUtilities.set_spoken_to_before_flag_to_true", "line_number": 16, "usage_type": "call"}, {"api_name": "utilities.mobileHelp.MobileUtilities", "line_number": 16, "usage_type": "attribute"}, {"api_name": "utilities.mobileHelp", "line_number": 16, "usage_type": "name"}, {"api_name": "utilities.configUtilities.get_config_value_as_integer", "line_number": 18, "usage_type": "call"}, {"api_name": "utilities.configUtilities", "line_number": 18, "usage_type": "name"}, {"api_name": "utilities.configUtilities.get_config_value_as_integer", "line_number": 20, "usage_type": "call"}, {"api_name": "utilities.configUtilities", "line_number": 20, "usage_type": "name"}, {"api_name": "utilities.configUtilities.get_config_value_as_integer", "line_number": 22, "usage_type": "call"}, {"api_name": "utilities.configUtilities", "line_number": 22, "usage_type": "name"}, {"api_name": "utilities.common.CommonUtils.draw_dialog_ui", "line_number": 25, "usage_type": "call"}, {"api_name": "utilities.common.CommonUtils", "line_number": 25, "usage_type": "attribute"}, {"api_name": "utilities.common", "line_number": 25, "usage_type": "name"}, {"api_name": "utilities.armourManagement.ArmourUtilities.get_all_armour_modifiers", "line_number": 27, "usage_type": "call"}, {"api_name": "utilities.armourManagement.ArmourUtilities", "line_number": 27, "usage_type": "attribute"}, {"api_name": "utilities.armourManagement", "line_number": 27, "usage_type": "name"}, {"api_name": "utilities.common.CommonUtils.replace_value_in_event", "line_number": 34, "usage_type": "call"}, {"api_name": "utilities.common.CommonUtils", "line_number": 34, "usage_type": "attribute"}, {"api_name": "utilities.common", "line_number": 34, "usage_type": "name"}, {"api_name": "bearlibterminal.terminal.print_", "line_number": 49, "usage_type": "call"}, {"api_name": "bearlibterminal.terminal", "line_number": 49, "usage_type": "name"}, {"api_name": "utilities.display.pointy_vertical_menu", "line_number": 52, "usage_type": "call"}, {"api_name": "utilities.display", "line_number": 52, "usage_type": "name"}, {"api_name": "bearlibterminal.terminal.printf", "line_number": 56, "usage_type": "call"}, {"api_name": "bearlibterminal.terminal", "line_number": 56, "usage_type": "name"}, {"api_name": "utilities.display.coloured_list", "line_number": 62, "usage_type": "call"}, {"api_name": "utilities.display", "line_number": 62, "usage_type": "name"}, {"api_name": "bearlibterminal.terminal.refresh", "line_number": 67, "usage_type": "call"}, {"api_name": "bearlibterminal.terminal", "line_number": 67, "usage_type": "name"}, {"api_name": "utilities.input_handlers.handle_game_keys", "line_number": 69, "usage_type": "call"}, {"api_name": "utilities.input_handlers", "line_number": 69, "usage_type": "name"}, {"api_name": "utilities.common.CommonUtils.move_menu_selection", "line_number": 73, "usage_type": "call"}, {"api_name": "utilities.common.CommonUtils", "line_number": 73, "usage_type": "attribute"}, {"api_name": "utilities.common", "line_number": 73, "usage_type": "name"}, {"api_name": "loguru.logger.debug", "line_number": 79, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 79, "usage_type": "name"}, {"api_name": "loguru.logger.debug", "line_number": 80, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 80, "usage_type": "name"}, {"api_name": "loguru.logger.debug", "line_number": 81, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 81, "usage_type": "name"}, {"api_name": "utilities.armourManagement.ArmourUtilities.create_full_armour_set", "line_number": 83, "usage_type": "call"}, {"api_name": "utilities.armourManagement.ArmourUtilities", "line_number": 83, "usage_type": "attribute"}, {"api_name": "utilities.armourManagement", "line_number": 83, "usage_type": "name"}, {"api_name": "utilities.armourManagement.ArmourUtilities.equip_full_set_of_armour", "line_number": 85, "usage_type": "call"}, {"api_name": "utilities.armourManagement.ArmourUtilities", "line_number": 85, "usage_type": "attribute"}, {"api_name": "utilities.armourManagement", "line_number": 85, "usage_type": "name"}, {"api_name": "utilities.armourManagement.ArmourUtilities.apply_major_attribute_bonus_to_full_armourset", "line_number": 87, "usage_type": "call"}, {"api_name": "utilities.armourManagement.ArmourUtilities", "line_number": 87, "usage_type": "attribute"}, {"api_name": "utilities.armourManagement", "line_number": 87, "usage_type": "name"}, {"api_name": "utilities.armourManagement.ArmourUtilities.set_mobile_derived_armour_attribute", "line_number": 93, "usage_type": "call"}, {"api_name": "utilities.armourManagement.ArmourUtilities", "line_number": 93, "usage_type": "attribute"}, {"api_name": "utilities.armourManagement", "line_number": 93, "usage_type": "name"}, {"api_name": "utilities.mobileHelp.MobileUtilities.set_mobile_derived_attributes", "line_number": 94, "usage_type": "call"}, {"api_name": "utilities.mobileHelp.MobileUtilities", "line_number": 94, "usage_type": "attribute"}, {"api_name": "utilities.mobileHelp", "line_number": 94, "usage_type": "name"}]}
+{"seq_id": "33768748557", "text": "import logging\nimport argparse\n\nimport numpy as np\nimport onnx\n\nlogger = logging.getLogger(__name__)\nlogging.basicConfig(format = '%(asctime)s - %(levelname)s - %(name)s - %(message)s',\n datefmt = '%m/%d/%Y %H:%M:%S',\n level = logging.WARN)\n\nclass Dataloader:\n def __init__(self, batch_size):\n self.batch_size = batch_size\n shape = [[batch_size, 4, 64, 64], [batch_size], [batch_size, 77, 768]]\n dtype = ['float32', 'int64', 'float32']\n self.dataset = []\n for idx in range(0, len(shape)):\n tensor = np.random.uniform(size=shape[idx])\n tensor = tensor.astype(dtype[idx])\n self.dataset.append(tensor)\n\n def __iter__(self):\n yield self.dataset, 0\n\nif __name__ == \"__main__\":\n logger.info(\"Evaluating ONNXRuntime full precision accuracy and performance:\")\n parser = argparse.ArgumentParser(\n formatter_class=argparse.ArgumentDefaultsHelpFormatter\n )\n parser.add_argument(\n '--model_path',\n type=str,\n help=\"Pre-trained model on onnx file\"\n )\n parser.add_argument(\n '--benchmark',\n action='store_true', \\\n default=False\n )\n parser.add_argument(\n '--tune',\n action='store_true', \\\n default=False,\n help=\"whether quantize the model\"\n )\n parser.add_argument(\n '--output_model',\n type=str,\n help=\"output model path\"\n )\n parser.add_argument(\n '--mode',\n type=str,\n help=\"benchmark mode of performance or accuracy\"\n )\n parser.add_argument(\n '--quant_format',\n type=str,\n default='default', \n choices=['default', 'QDQ', 'QOperator'],\n help=\"quantization format\"\n )\n parser.add_argument(\n \"--batch_size\",\n default=1,\n type=int,\n )\n args = parser.parse_args()\n\n dataloader = Dataloader(args.batch_size)\n\n if args.benchmark and args.mode == 'performance':\n from neural_compressor.benchmark import fit\n from neural_compressor.config import BenchmarkConfig\n conf = BenchmarkConfig(warmup=10, iteration=1000, cores_per_instance=4, num_of_instance=1)\n fit(args.model_path, conf, b_dataloader=dataloader)\n if args.tune:\n from neural_compressor import quantization, PostTrainingQuantConfig\n config = PostTrainingQuantConfig(quant_format=args.quant_format, recipes={'graph_optimization_level':'ENABLE_EXTENDED'})\n q_model = quantization.fit(args.model_path, config, calib_dataloader=dataloader)\n\n q_model.save(args.output_model)\n", "repo_name": "intel/neural-compressor", "sub_path": "examples/onnxrt/image_recognition/unet/quantization/ptq_static/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2623, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1596, "dataset": "github-code", "pt": "24", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.WARN", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 19, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 28, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 29, "usage_type": "attribute"}, {"api_name": "neural_compressor.config.BenchmarkConfig", "line_number": 76, "usage_type": "call"}, {"api_name": "neural_compressor.benchmark.fit", "line_number": 77, "usage_type": "call"}, {"api_name": "neural_compressor.PostTrainingQuantConfig", "line_number": 80, "usage_type": "call"}, {"api_name": "neural_compressor.quantization.fit", "line_number": 81, "usage_type": "call"}, {"api_name": "neural_compressor.quantization", "line_number": 81, "usage_type": "name"}]}
+{"seq_id": "41108860274", "text": "import pandas as pd\r\nimport numpy as np\r\nimport matplotlib.pylab as plt\r\ndata = pd.read_csv('a.csv')# -*- coding: utf-8 -*-\r\ndateparse = lambda dates: pd.datetime.strptime(dates, '%d.%m.%Y')\r\ndata = pd.read_csv('a.csv', parse_dates=['Date'],index_col='Date', date_parser=dateparse)\r\nprint (data.head)\r\nprint (data.dtypes)\r\ndata.index\r\nts = data\r\nplt.plot(ts)\r\n \r\n\"\"\"\r\nРедактор Spyder\r\n\r\nЭто временный скриптовый файл.\r\n\"\"\"\r\n\r\n", "repo_name": "sergeyivanov01/PHBS_MLF_2018", "sub_path": "1987-2018 bp.py", "file_name": "1987-2018 bp.py", "file_ext": "py", "file_size_in_byte": 460, "program_lang": "python", "lang": "ru", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "24", "api": [{"api_name": "pandas.read_csv", "line_number": 4, "usage_type": "call"}, {"api_name": "pandas.datetime.strptime", "line_number": 5, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 5, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pylab.plot", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 11, "usage_type": "name"}]}
+{"seq_id": "17566722638", "text": "import random\nfrom django.shortcuts import get_object_or_404\nfrom rest_framework import status\nfrom rest_framework.decorators import (\n api_view,\n authentication_classes,\n permission_classes,\n)\nfrom rest_framework_jwt.authentication import JSONWebTokenAuthentication\nfrom rest_framework.permissions import IsAuthenticated\nfrom rest_framework.response import Response\nfrom .models import Genre, Movie, Review\nfrom .serializers import (\n GenreListSerializer,\n MovieListSerializer,\n MovieSerializer,\n MovieBackdropSerializer,\n ReviewSerializer\n)\n\n\n@api_view(['GET'])\ndef genre_list(request):\n genres = Genre.objects.all()\n serializer = GenreListSerializer(genres, many=True)\n return Response(serializer.data)\n\n\n@api_view(['GET'])\ndef movie_list(request, genre_pk):\n # print(genre_pk)\n movies = Movie.objects.all().filter(genres__id__icontains=genre_pk).order_by('-vote_average')[:10]\n serializer = MovieListSerializer(movies, many=True)\n # print('clear')\n return Response(serializer.data)\n\n\n@api_view(['GET'])\ndef movie(request, movie_pk):\n movie = get_object_or_404(Movie, pk=movie_pk)\n serializer = MovieSerializer(movie)\n return Response(serializer.data)\n\n\n@api_view(['GET'])\ndef movie_category(request, where):\n movies = Movie.objects.all().filter(where=where).order_by('-vote_average')[:10]\n serializer = MovieListSerializer(movies, many=True)\n # print(serializer.data)\n return Response(serializer.data)\n\n\n@api_view(['GET','POST'])\n@authentication_classes([JSONWebTokenAuthentication])\n@permission_classes([IsAuthenticated])\ndef like(request, movie_pk):\n movie = get_object_or_404(Movie, pk=movie_pk)\n if request.method == 'GET':\n if movie.like_users.filter(pk=request.user.pk).exists():\n return Response(True)\n else:\n return Response(False)\n else:\n if movie.like_users.filter(pk=request.user.pk).exists():\n # 좋아요 취소\n movie.like_users.remove(request.user)\n return Response({'like': False})\n else:\n # 좋아요\n movie.like_users.add(request.user)\n return Response({'like': True})\n\n\n# 리뷰 작성, 리뷰 리스트\n@api_view(['POST', 'GET'])\n@authentication_classes([JSONWebTokenAuthentication])\n@permission_classes([IsAuthenticated])\ndef review_list_create(request, movie_pk):\n if request.method == 'GET':\n reviews = Review.objects.all()\n serializer = ReviewSerializer(reviews, many=True)\n return Response(serializer.data)\n else:\n movie = get_object_or_404(Movie, pk=movie_pk)\n serializer = ReviewSerializer(data=request.data)\n if serializer.is_valid():\n serializer.save(movie=movie, user=request.user)\n return Response(serializer.data) \n\n\n# 내가 좋아요한 영화들\n@api_view(['GET'])\n@authentication_classes([JSONWebTokenAuthentication])\n@permission_classes([IsAuthenticated])\ndef my_movie(request):\n movies = request.user.like_movies\n # movies = Movie.objects.all()\n # print(request.method)\n serializer = MovieListSerializer(movies, many=True)\n return Response(serializer.data)\n\n\n@api_view(['GET'])\n@authentication_classes([JSONWebTokenAuthentication])\n@permission_classes([IsAuthenticated])\ndef recommend(request):\n movies = request.user.like_movies.all()\n genres = dict()\n for movie in movies:\n for g in movie.genres.all():\n # print(g)\n if g.name in genres:\n val = genres[g.id]\n genres[g.id] = val + 1\n else:\n genres[g.id] = 1\n # print(genres)\n max_v = max_g = 0\n for (key, val) in genres.items():\n if val > max_v:\n max_g = key\n max_v = val\n\n \n movies = Movie.objects.all().filter(genres__id__icontains=max_g).order_by('-vote_average')[:10]\n serializer = MovieListSerializer(movies, many=True)\n \n return Response(serializer.data)", "repo_name": "SaltCastle77/LuvLuvMovie", "sub_path": "final-pjt-back/movies/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3978, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "models.Genre.objects.all", "line_number": 24, "usage_type": "call"}, {"api_name": "models.Genre.objects", "line_number": 24, "usage_type": "attribute"}, {"api_name": "models.Genre", "line_number": 24, "usage_type": "name"}, {"api_name": "serializers.GenreListSerializer", "line_number": 25, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 26, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 22, "usage_type": "call"}, {"api_name": "models.Movie.objects.all", "line_number": 32, "usage_type": "call"}, {"api_name": "models.Movie.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "models.Movie", "line_number": 32, "usage_type": "name"}, {"api_name": "serializers.MovieListSerializer", "line_number": 33, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 35, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 29, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 40, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 40, "usage_type": "argument"}, {"api_name": "serializers.MovieSerializer", "line_number": 41, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 42, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 38, "usage_type": "call"}, {"api_name": "models.Movie.objects.all", "line_number": 47, "usage_type": "call"}, {"api_name": "models.Movie.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "models.Movie", "line_number": 47, "usage_type": "name"}, {"api_name": "serializers.MovieListSerializer", "line_number": 48, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 50, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 45, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 57, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 57, "usage_type": "argument"}, {"api_name": "rest_framework.response.Response", "line_number": 60, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 62, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 67, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 71, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 53, "usage_type": "call"}, {"api_name": "rest_framework.decorators.authentication_classes", "line_number": 54, "usage_type": "call"}, {"api_name": "rest_framework_jwt.authentication.JSONWebTokenAuthentication", "line_number": 54, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 55, "usage_type": "call"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 55, "usage_type": "name"}, {"api_name": "models.Review.objects.all", "line_number": 80, "usage_type": "call"}, {"api_name": "models.Review.objects", "line_number": 80, "usage_type": "attribute"}, {"api_name": "models.Review", "line_number": 80, "usage_type": "name"}, {"api_name": "serializers.ReviewSerializer", "line_number": 81, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 82, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 84, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 84, "usage_type": "argument"}, {"api_name": "serializers.ReviewSerializer", "line_number": 85, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 88, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 75, "usage_type": "call"}, {"api_name": "rest_framework.decorators.authentication_classes", "line_number": 76, "usage_type": "call"}, {"api_name": "rest_framework_jwt.authentication.JSONWebTokenAuthentication", "line_number": 76, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 77, "usage_type": "call"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 77, "usage_type": "name"}, {"api_name": "serializers.MovieListSerializer", "line_number": 99, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 100, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 92, "usage_type": "call"}, {"api_name": "rest_framework.decorators.authentication_classes", "line_number": 93, "usage_type": "call"}, {"api_name": "rest_framework_jwt.authentication.JSONWebTokenAuthentication", "line_number": 93, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 94, "usage_type": "call"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 94, "usage_type": "name"}, {"api_name": "models.Movie.objects.all", "line_number": 125, "usage_type": "call"}, {"api_name": "models.Movie.objects", "line_number": 125, "usage_type": "attribute"}, {"api_name": "models.Movie", "line_number": 125, "usage_type": "name"}, {"api_name": "serializers.MovieListSerializer", "line_number": 126, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 128, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 103, "usage_type": "call"}, {"api_name": "rest_framework.decorators.authentication_classes", "line_number": 104, "usage_type": "call"}, {"api_name": "rest_framework_jwt.authentication.JSONWebTokenAuthentication", "line_number": 104, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 105, "usage_type": "call"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 105, "usage_type": "name"}]}
+{"seq_id": "20021065187", "text": "import collections\nimport copy\nimport tempfile\nimport time\n\nfrom urllib.parse import urlparse\n\n## django imports\nfrom django.core.exceptions import ObjectDoesNotExist\n\n### apps imports\nfrom grabbers.utils import miners\n\nfrom grabbers.utils.confs import GrabberConf, MapperConf\n\nfrom information.models import PageData, make_control_key\nfrom urlocators.models import Page, Locator, make_url_id\nfrom workers.models import Job\nfrom drivers.models import Header\nfrom grabbers.models import Target, ElementAction, PostElementAction, Extractor, PageAction\nfrom workers.utils.tasksProducer import TaskFromUrls\n\nGrabis=miners.Grabis()\n\n# --- extract page & response data\n#########################\nclass Pythoness:\n '''\n '''\n\n def __init__(self):\n '''\n '''\n self._conf={}\n self._job=None\n self._data={\n 'page_data':[],\n }\n\n def set_job(self, job):\n '''\n '''\n self._job=job\n\n def set_task(self, task):\n '''\n '''\n self._task=task\n\n\n def set_grabber(self, grabber):\n '''\n '''\n self._conf=GrabberConf.toDict(grabber)\n print('[+] Setting Grabber Configuration')\n\n def map_targets(self, mapper, browser):\n '''\n action\n ------\n maps urls on target pages\n creates tasks with diferent taskconfigs\n save url with field name on information db\n\n obs\n ---\n this guy here is an hybric fellow\n so he is kind out place here...\n '''\n #set vars\n field_name=mapper.field_name\n print('[+] Starting Mapper [{}]'.format(field_name))\n selector=mapper.field_selector\n task_configs=TaskConfig.objects.filter(mapper=mapper)\n page_object=Grabis.load_page(self._job, browser, save_source=False)\n #mine target links\n try:\n gb=Grabis\n gb.set_selector(selector)\n gb.set_page_object(page_object)\n gb.grab()\n #mapper element attr is always generic link\n data=gb.get_data(field_name, ['generic_link',])\n except Exception as e:\n print('[-] Fail to extract link in mapper: [{}]'.format(e))\n return\n #create mapper tasks\n current_url = urlparse(browser.current_url)\n #-- build page obj\n page=self._build_urllocators_objs(current_url)\n urls=[]\n for element in data:\n element_index=element['index']\n for mined_url in element[field_name]:\n if not mined_url:continue\n full_url=self._map_url(mined_url, current_url)\n urls.append(full_url)\n self._save_field(field_name, full_url, element_index, page)\n #build tasks for next crawling\n for task_config in task_configs:\n TaskFromUrls._job=self._job\n TaskFromUrls._config=task_config\n TaskFromUrls.set_urls(urls)\n TaskFromUrls.build_tasks()\n print('[+] Done creating [{}] tasks for job [{}]'.format(\n task_configs*len(urls), self._job))\n\n #---- mapper helper\n ###################\n def _map_url(self, u, current_url):\n parsed_url = urlparse(u)\n if not parsed_url.netloc:\n parsed_url = parsed_url._replace(\n netloc=current_url.netloc.lower(),\n scheme=current_url.scheme)\n return parsed_url.geturl()\n\n def session(self, browser, element_index=-1):\n '''\n '''\n #get general vals\n browser_name=browser.__class__.__name__\n\n # set base values\n if 'target' in self._conf:\n selector=self._conf['target']['selector']\n print('[+] Start targeting selector [{0}]'.format(selector))\n page_object=Grabis.load_page(self._job, browser)\n try:\n gb=Grabis\n gb.set_selector(selector)\n gb.set_page_object(page_object)\n except Exception as e:\n print('[-] Fail to load conf in Grabis', e)\n return\n\n if 'extractors' in self._conf:\n field_name=self._conf['target']['name']\n attrs=self._conf['extractors']\n gb.grab()\n data=gb.get_data(field_name, attrs)\n self._data['page_data']=data\n\n #will deal with elements list to action and post action\n if 'element_action' in self._conf:\n #test browser type\n if browser_name == 'LeanRequests':\n raise TypeError('[-] Lean Requests has no action')\n #grab target elements\n toAction=browser.xpathToaction(selector)\n actionType=self._conf['element_action']\n #apply action\n for targetAction in toAction:\n if actionType == \"jsClick\":\n browser.clickByJS(targetAction)\n else:\n getattr(targetAction, actionType)()\n if not self._conf['post_action']:continue\n postAction=Pythoness()\n postAction.set_task(self._task)\n postAction.set_job(self._job)\n postAction.set_grabber(self._conf['post_action'])\n postAction.session(browser)\n postAction.save_data(browser)\n\n\n if 'page_action' in self._conf:\n #it's dirty - needs to improve\n page_action=self._conf['page_action']\n action_data={}\n\n #legacy - now is required\n if page_action == 'get_header_field':\n if browser_name != 'LeanRequests':\n raise TypeError('[-] Only Lean Requests has get header field')\n #target header field is passed as extractor\n action_data.update({'header_field':self._conf['extractors']})\n\n elif page_action == 'execute_script':\n if browser_name == 'LeanRequests':\n raise TypeError('[-] Lean Requests has no execute script')\n if not selector: #for now script is send by selector ---------\n raise TypeError('[-] Script in target is required')\n field_name=field_name=self._conf['target']['name']\n action_data={'field_name':field_name, 'script':selector}\n\n elif page_action == 'switch_to_frame':\n if browser_name == 'LeanRequests':\n raise TypeError('[-] Lean Requests has no execute script')\n if not selector: #for now script is send by selector ---------\n raise TypeError('[-] Frame set_selector is required')\n field_name=field_name=self._conf['target']['name']\n action_data={'field_name':field_name, 'xpath':selector}\n\n elif page_action == 'page_source':\n action_data={'job':self._job}\n\n elif page_action == 'take_screenshot':\n action_data={'job':self._job}\n\n print('[+] Start page action [{0}]'.format(page_action))\n getattr(browser, page_action)(page_data=self._data,\n action_data=action_data)\n\n #will call a post-action\n if self._conf['post_action'] and not 'element_action' in self._conf:\n postAction=Pythoness()\n postAction.set_task(self._task)\n postAction.set_job(self._job)\n postAction.set_grabber(self._conf['post_action'])\n postAction.session(browser)\n postAction.save_data(browser)\n\n\n def save_data(self, browser, target_url=None):\n '''\n '''\n url=target_url\n if not url:\n url=browser.current_url\n page=self._build_urllocators_objs(url)\n for dict_item in self._data['page_data']:\n element_index = dict_item.pop('index')\n for field_name, values in dict_item.items():\n for value in values:\n if not value:continue\n self._save_field(field_name, value, element_index, page)\n time.sleep(0.001)\n\n def save_condition(self, browser, condition):\n '''\n '''\n condition_type = condition.save_type\n condition_confs = condition.taskconfig_set.all()\n if condition_type == 'page_data':\n msg='[-] Condition::Page Data save was'' not implemented yet'\n raise NotImplemented(msg)\n if condition_type == 'silent':return\n headers=browser.get_headers()\n print(\"H======>{}\".format(headers))\n for k,v in headers.items():\n hea=Header()\n hea.field_name=k\n hea.field_value=v\n hea.header_name=browser._header_name\n hea.save()\n for conf in condition_confs:\n has_headers = conf.driver.headers.filter(\n field_name=hea.field_name)\n for h_header in has_headers:\n conf.driver.headers.remove(h_header)\n time.sleep(0.001)\n conf.driver.headers.add(hea)\n time.sleep(0.001)\n\n print('[+] Done saving condition::{}'.format(condition_type))\n\n\n\n # --- save data helpers\n ########################\n def _build_urllocators_objs(self, url):\n '''\n obs\n ---------\n this function is very similar\n to save page source in browser\n the only difference is html source persistence\n\n *** soon will merge both ****\n '''\n #build url relations\n url_id=make_url_id(url)\n try:\n locs=Locator.objects.get(url_id=url_id)\n except ObjectDoesNotExist:\n locs=Locator()\n locs.url=url\n locs.save()\n #build page relation\n try:\n page=Page.objects.get(addr=locs.id, job=self._job)\n except Page.DoesNotExist:\n page=Page()\n page.job=self._job\n page.task=self._task\n page.addr=locs\n page.save()\n return page\n\n def _save_field(self, field_name, value, element_index, page):\n '''\n '''\n control_key=make_control_key(field_name, value, page.id)\n is_duplicate=PageData.objects.filter(control_key=control_key)\n if is_duplicate.count():return\n pd=PageData()\n pd.field_name=field_name\n pd.field_value=value\n pd.element_index = element_index\n pd.page=page\n pd.job=self._job\n pd.task=self._task\n pd.save()\n\n @staticmethod\n def save_proxy_data(browser):\n '''\n '''\n print(browser.get_proxy_data())\n", "repo_name": "VulcanoAhab/delphi", "sub_path": "grabbers/utils/crushers.py", "file_name": "crushers.py", "file_ext": "py", "file_size_in_byte": 10796, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "grabbers.utils.miners.Grabis", "line_number": 23, "usage_type": "call"}, {"api_name": "grabbers.utils.miners", "line_number": 23, "usage_type": "name"}, {"api_name": "grabbers.utils.confs.GrabberConf.toDict", "line_number": 54, "usage_type": "call"}, {"api_name": "grabbers.utils.confs.GrabberConf", "line_number": 54, "usage_type": "name"}, {"api_name": "urllib.parse.urlparse", "line_number": 88, "usage_type": "call"}, {"api_name": "workers.utils.tasksProducer.TaskFromUrls._job", "line_number": 101, "usage_type": "attribute"}, {"api_name": "workers.utils.tasksProducer.TaskFromUrls", "line_number": 101, "usage_type": "name"}, {"api_name": "workers.utils.tasksProducer.TaskFromUrls._config", "line_number": 102, "usage_type": "attribute"}, {"api_name": "workers.utils.tasksProducer.TaskFromUrls", "line_number": 102, "usage_type": "name"}, {"api_name": "workers.utils.tasksProducer.TaskFromUrls.set_urls", "line_number": 103, "usage_type": "call"}, {"api_name": "workers.utils.tasksProducer.TaskFromUrls", "line_number": 103, "usage_type": "name"}, {"api_name": "workers.utils.tasksProducer.TaskFromUrls.build_tasks", "line_number": 104, "usage_type": "call"}, {"api_name": "workers.utils.tasksProducer.TaskFromUrls", "line_number": 104, "usage_type": "name"}, {"api_name": "urllib.parse.urlparse", "line_number": 111, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 228, "usage_type": "call"}, {"api_name": "drivers.models.Header", "line_number": 242, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 252, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 254, "usage_type": "call"}, {"api_name": "urlocators.models.make_url_id", "line_number": 273, "usage_type": "call"}, {"api_name": "urlocators.models.Locator.objects.get", "line_number": 275, "usage_type": "call"}, {"api_name": "urlocators.models.Locator.objects", "line_number": 275, "usage_type": "attribute"}, {"api_name": "urlocators.models.Locator", "line_number": 275, "usage_type": "name"}, {"api_name": "django.core.exceptions.ObjectDoesNotExist", "line_number": 276, "usage_type": "name"}, {"api_name": "urlocators.models.Locator", "line_number": 277, "usage_type": "call"}, {"api_name": "urlocators.models.Page.objects.get", "line_number": 282, "usage_type": "call"}, {"api_name": "urlocators.models.Page.objects", "line_number": 282, "usage_type": "attribute"}, {"api_name": "urlocators.models.Page", "line_number": 282, "usage_type": "name"}, {"api_name": "urlocators.models.Page.DoesNotExist", "line_number": 283, "usage_type": "attribute"}, {"api_name": "urlocators.models.Page", "line_number": 283, "usage_type": "name"}, {"api_name": "urlocators.models.Page", "line_number": 284, "usage_type": "call"}, {"api_name": "information.models.make_control_key", "line_number": 294, "usage_type": "call"}, {"api_name": "information.models.PageData.objects.filter", "line_number": 295, "usage_type": "call"}, {"api_name": "information.models.PageData.objects", "line_number": 295, "usage_type": "attribute"}, {"api_name": "information.models.PageData", "line_number": 295, "usage_type": "name"}, {"api_name": "information.models.PageData", "line_number": 297, "usage_type": "call"}]}
+{"seq_id": "26836613360", "text": "from cell import Cell\nfrom random import randint\nimport pygame\n\nclass Grid:\n\n\tdef __init__(self, length, height):\n\t\tself.board = [[Cell() for k in range(height)] for i in range(length)]\n\t\t\n\tdef generate(self, quota):\n\t\tmines = []\n\t\twhile quota > 0:\n\t\t\tfor i in range(len(self.board)):\n\t\t\t\tfor k in range(len(self.board[i])):\n\t\t\t\t\tif quota > 0:\n\t\t\t\t\t\tif self.board[i][k].value != \"×\":\n\t\t\t\t\t\t\trandom_value = randint(0, 100)\n\t\t\t\t\t\t\tif random_value > 99:\n\t\t\t\t\t\t\t\tquota -= 1\n\t\t\t\t\t\t\t\tself.board[i][k].value = \"×\"\n\t\t\t\t\t\t\t\tmines.append((i, k))\n\t\t\t\t\t\t\telse:\n\t\t\t\t\t\t\t\tif self.board[i][k] == \" \":\n\t\t\t\t\t\t\t\t\tself.board[i][k].value = \" \"\n\n\t\tfor coords in mines:\n\t\t\tfor i in range(max(0, coords[0]-1), min(len(self.board), coords[0]+2)):\n\t\t\t\tfor k in range(max(0, coords[1]-1), min(len(self.board[i]), coords[1]+2)):\n\t\t\t\t\tif self.board[i][k].value == \" \":\n\t\t\t\t\t\tself.board[i][k].value = \"1\"\n\t\t\t\t\telse:\n\t\t\t\t\t\tif self.board[i][k].value != \"×\":\n\t\t\t\t\t\t\tself.board[i][k].value = str(int(self.board[i][k].value)+1)\n\t\n\tdef display(self, window, font):\n\t\tcouleurs = { \"×\": (0, 0, 0), \"1\": (0, 0, 255), \"2\": (0, 128, 0), \"3\": (255, 0, 0), \"4\": (200, 0, 200), \"5\": (255, 255, 0), \"6\": (0, 255, 255), \"7\": (106, 230, 201), \"8\": (106, 230, 201), \" \": (255, 255, 255)}\n\t\tcell_false = pygame.image.load(\"./assets/cell_false.png\")\n\t\tcell_true = pygame.image.load(\"./assets/cell_true.png\")\n\t\tcell_flag = pygame.image.load(\"./assets/cell_flag.png\")\n\n\t\tfor i in range(len(self.board)):\n\t\t\tfor k in range(len(self.board[i])):\n\t\t\t\tif self.board[i][k].display == True:\n\t\t\t\t\tpygame.draw.rect(window, (192, 192, 192), (25*k, 25*i, 25, 25))\n\t\t\t\t\ttexte = font.render(self.board[i][k].value, True, couleurs[self.board[i][k].value])\n\t\t\t\t\twindow.blit(cell_true, (25*k, 25*i))\n\t\t\t\t\twindow.blit(texte, (25*k+8, 25*i+5))\n\t\t\t\telse:\n\t\t\t\t\tif self.board[i][k].flag == True:\n\t\t\t\t\t\twindow.blit(cell_flag, (25*k, 25*i))\n\t\t\t\t\telse:\n\t\t\t\t\t\twindow.blit(cell_false, (25*k, 25*i))\n\t\n\tdef game_end(self):\n\t\tfor i in range(len(self.board)):\n\t\t\tfor k in range(len(self.board[i])):\n\t\t\t\tif (self.board[i][k].value == \"×\" and self.board[i][k].flag == False) or (self.board[i][k].value != \"×\" and self.board[i][k].flag == True) or (self.board[i][k].value != \"×\" and self.board[i][k].display != True):\n\t\t\t\t\treturn False\n\t\treturn True\n\n\n\tdef game_over(self):\n\t\tfor i in range(len(self.board)):\n\t\t\tfor k in range(len(self.board[i])):\n\t\t\t\tself.board[i][k].display = True\n\n\tdef flood_fill(self, coords):\n\t\tmovements = [[0,-1], [0,1], [-1,0], [1,0], [-1,-1], [-1,1], [1,1], [1,-1]]\n\t\tremains_to_be_done = [coords]\n\t\tcell_edges = []\n\t\twhile len(remains_to_be_done) > 0:\n\t\t\tcurrent_cell = remains_to_be_done[0]\n\t\t\tdel remains_to_be_done[0]\n\t\t\tself.board[current_cell[0]][current_cell[1]].display = True\n\t\t\tcell_edges.extend(correct_cell_edges(self, current_cell))\n\t\t\tfor movement in movements:\n\t\t\t\tneighbour_cell = [current_cell[0]+movement[0], current_cell[1]+movement[1]]\n\t\t\t\tif correct_cell(self, neighbour_cell):\n\t\t\t\t\tremains_to_be_done.append(neighbour_cell)\n\t\tfor k in range(len(cell_edges)):\n\t\t\tself.board[cell_edges[k][0]][cell_edges[k][1]].display = True\n\ndef correct_cell(grid, cell):\n\treturn (0 <= cell[0] < len(grid.board)) and (0 <= cell[1] < len(grid.board[0])) and (grid.board[cell[0]][cell[1]].value == \" \") and (grid.board[cell[0]][cell[1]].display == False)\n\ndef correct_cell_edges(grid, cell):\n\tmovements = [[0,-1],[0,1],[-1,0],[1,0],[-1,-1],[-1,1],[1,1],[1,-1]]\n\tcell_edges = []\n\tcell_visited = cell\n\tfor movement in movements:\n\t\tcell = [cell_visited[0]+movement[0], cell_visited[1]+movement[1]]\n\t\tif 0 <= cell[0] < len(grid.board) and 0 <= cell[1] < len(grid.board[0]):\n\t\t\tif grid.board[cell[0]][cell[1]].value != \" \" and grid.board[cell[0]][cell[1]].value != \"X\":\n\t\t\t\tcell_edges.append(cell)\n\treturn cell_edges\n", "repo_name": "vosketalor/minesweeper", "sub_path": "grid.py", "file_name": "grid.py", "file_ext": "py", "file_size_in_byte": 3766, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "cell.Cell", "line_number": 8, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.image.load", "line_number": 37, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 38, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 44, "usage_type": "attribute"}]}
+{"seq_id": "71128056703", "text": "\n# Example Azure function which reacts to a file landing in a queue\n\n## Import Azure packages\nimport logging\nimport azure.functions as func\nfrom azure.identity import DefaultAzureCredential\nfrom azure.storage.blob import BlobServiceClient\n\n## Import Snowpark session module\nfrom snowflake.snowpark import Session\n\n## Import other packages\nimport os\nimport json\nfrom io import BytesIO\n\n## Define function to retrieve the desired\n## information from the input message\ndef parse_input_message(msg: func.QueueMessage):\n\n ### Retrieve message as JSON\n msg_json = msg.get_json()\n logging.info('Message JSON:')\n logging.info(msg_json)\n \n ### Retrieve message ID\n msg_id = msg_json[\"id\"]\n logging.info(f'Message ID: {msg_id}')\n\n ### Retrieve full file URL from input blob.\n ### The specific key varies depending on the type\n ### of storage container\n if \"url\" in msg_json[\"data\"] :\n file_path_url = msg_json[\"data\"][\"url\"]\n elif \"blobUrl\" in msg_json[\"data\"] :\n file_path_url = msg_json[\"data\"][\"blobUrl\"]\n else :\n logging.error(\"Function abort - Path URL does not match expected storage blob service URI\")\n raise ValueError(\"Function abort - Path URL does not match expected storage blob service URI\")\n \n logging.info(f'File path URL: {file_path_url}')\n \n '''\n Expected file URL format:\n https://.blob.core.windows.net//path/to/file.json\n \n Example expected file URL:\n https://my-storage-account.blob.core.windows.net/automated-function-trigger-demo/example_file.json\n '''\n\n ### Retrieve storage blob service uri\n storage_blob_service_uri = os.getenv(\"AZURE_STORAGE_IDENTITY__blobServiceUri\")\n\n ### Parse storage queue service URI from file path URL\n if file_path_url.startswith(storage_blob_service_uri) :\n file_path = file_path_url[1 + len(storage_blob_service_uri):]\n else :\n logging.info(f'Function abort - Path URL does not match expected storage blob service URI')\n return\n \n ### Split file path into container and relative file path\n container, relative_file_path = file_path.split('/', 1)\n\n return storage_blob_service_uri, container, relative_file_path\n\n## Define function to download full JSON file from blob\ndef azure_download_json_file(storage_blob_service_uri=None, container=None, relative_file_path=None):\n default_azure_credential = DefaultAzureCredential()\n blob_service_client = BlobServiceClient(storage_blob_service_uri, credential=default_azure_credential)\n blob_client = blob_service_client.get_blob_client(container=container, blob=relative_file_path)\n with BytesIO() as input_blob:\n blob_client.download_blob().download_to_stream(input_blob)\n input_blob.seek(0)\n json_input = json.load(input_blob)\n \n return json_input\n\n## Define function that retrieve the SQL statement\n## to execute from the JSON input\ndef retrieve_sql_statement_to_execute(json_input: dict):\n\n '''\n Expected format of JSON file:\n json_input = {\n \"sql_statement_to_execute\" : \"\"\n }\n '''\n\n ### Error if JSON file is not in expected format\n if \"sql_statement_to_execute\" not in json_input.keys() :\n logging.error(f\"Manual log - Downloaded file did not include the key 'sql_statement_to_execute'\")\n raise ValueError(\"Manual log - Downloaded file did not include the key 'sql_statement_to_execute'\")\n\n ### Retrieve the value as its own variable\n sql_statement_to_execute = json_input[\"sql_statement_to_execute\"]\n\n return sql_statement_to_execute\n \n## Function to create Snowpark session\ndef build_snowpark_session() :\n\n ### Retrieve connection parameters from app settings\n snowflake_connection_parameters = {\n \"account\": os.getenv(\"SNOWFLAKE_ACCOUNT\")\n , \"user\": os.getenv(\"SNOWFLAKE_USER\")\n , \"password\": os.getenv(\"SNOWFLAKE_PASSWORD\")\n , \"role\": os.getenv(\"SNOWFLAKE_ROLE\")\n , \"warehouse\": os.getenv(\"SNOWFLAKE_WAREHOUSE\")\n }\n\n ### Create Snowflake Snowpark session \n snowpark_session = Session.builder.configs(snowflake_connection_parameters).create()\n\n return snowpark_session\n\n## Define function that executes given SQL in Snowflake\ndef execute_sql_in_snowflake(sql_statement_to_execute: str):\n\n ### Create Snowflake Snowpark session \n snowpark_session = build_snowpark_session()\n\n ### Execute the SQL command in Snowflake\n ### and log the result\n sf_df_statement_result = snowpark_session.sql(sql_statement_to_execute).collect()\n \n logging.info(\"SQL statement result:\")\n logging.info(sf_df_statement_result)\n\n ### Close the Snowflake Snowpark Session\n snowpark_session.close()\n \n return\n\n## Define main function for Azure\ndef main(msg: func.QueueMessage):\n logging.info('Received new message from queue')\n logging.info(msg)\n\n ### Parse the input message for required information\n storage_blob_service_uri, container, relative_file_path = parse_input_message(msg)\n\n ### Retrieve JSON input from Azure storage\n json_input = azure_download_json_file(storage_blob_service_uri=storage_blob_service_uri, container=container, relative_file_path=relative_file_path)\n\n ### Retrieve the value as its own variable\n sql_statement_to_execute = retrieve_sql_statement_to_execute(json_input)\n \n ### Attempt to execute the SQL in Snowflake\n execute_sql_in_snowflake(sql_statement_to_execute)\n \n return", "repo_name": "interworks/Example-Snowpark-Azure-Functions", "sub_path": "azure_storage_trigger_leveraging_app_settings_directly/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 5268, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "azure.functions.QueueMessage", "line_number": 20, "usage_type": "attribute"}, {"api_name": "azure.functions", "line_number": 20, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 29, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 39, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 42, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 53, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 59, "usage_type": "call"}, {"api_name": "azure.identity.DefaultAzureCredential", "line_number": 69, "usage_type": "call"}, {"api_name": "azure.storage.blob.BlobServiceClient", "line_number": 70, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 72, "usage_type": "call"}, {"api_name": "json.load", "line_number": 75, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 92, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 105, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 106, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 107, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 108, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 109, "usage_type": "call"}, {"api_name": "snowflake.snowpark.Session.builder.configs", "line_number": 113, "usage_type": "call"}, {"api_name": "snowflake.snowpark.Session.builder", "line_number": 113, "usage_type": "attribute"}, {"api_name": "snowflake.snowpark.Session", "line_number": 113, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 127, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 128, "usage_type": "call"}, {"api_name": "azure.functions.QueueMessage", "line_number": 136, "usage_type": "attribute"}, {"api_name": "azure.functions", "line_number": 136, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 137, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 138, "usage_type": "call"}]}
+{"seq_id": "24661075680", "text": "from django.urls import path\n\nfrom . import views\nfrom contact_us.views import create_contact, faq\n\n\napp_name='adminpage'\n\nurlpatterns=[\n path('garbage_collectors_list', views.garbage_collectors_list, name='collectors_list'),\n path('accept_collector', views.create_garbage_collector, name=\"accept_collector\"),\n path(\"accept_list\", views.unaccepted_collectors_list, name=\"accept_list\"),\n path(\"accept_detail/\", views.unaccepted_collector_detail, name=\"accept_detail\"),\n path(\"accepted_collector/\", views.unaccepted_collector_changed, name=\"accepted_collector\"),\n path(\"contact\", create_contact, name=\"contact_create\"),\n path('faq', faq, name=\"faq\")\n\n]", "repo_name": "adebusola-prog/IN_OUT", "sub_path": "garbage_project/admin_page/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 680, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "contact_us.views.create_contact", "line_number": 15, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "contact_us.views.faq", "line_number": 16, "usage_type": "argument"}]}
+{"seq_id": "31483919047", "text": "import os \nimport datetime\nimport shutil #using copy2('src,'dst') cammand only\nass_list=[] #for associating the date to the image name\n\nos.chdir('G:\\\\DCIM\\\\100CANON')\nfor image in os.listdir(): #listing all files of the Card\n\t\n\ttimestamp=os.stat(image)[8] #gettingtimestamp\n\tdate_time=datetime.datetime.fromtimestamp(timestamp).strftime('%d-%b') #creating a date_time from the timestamp \n\t#fromtimestamp() returns a string\n\tloc_ass=[image,date_time] #list with image and date\n\tass_list.append(loc_ass) #appending to the ass_list\n\t\nos.chdir('F:\\\\LatestDSLR')\ncurrent_folders=[folder for folder in os.listdir() ] #listing current destination folders\nfor i in range(len(ass_list)):\n\tdate=ass_list[i][1] #getting the dates for each image\n\tsrc=f\"G:\\\\DCIM\\\\100CANON\\\\{ass_list[i][0]}\" #assigning the source to image location\n\n\n\tif os.path.exists(f\"F:\\\\LatestDSLR\\\\{date}\") ==True: #checking if there's already a date folder or not\n\t des=f\"F:\\\\LatestDSLR\\\\{date}\"\n\n\telse:\n\t\tos.mkdir(f\"{date}\")\n\t\tdes=f\"F:\\\\LatestDSLR\\\\{date}\" #creating the date foldee if not already present\n\t\n\n\tshutil.copy(src,des) #copying the file\n\n\nprint(\"DONE!!\")\n\t\n", "repo_name": "jitendrayt/DSLR-auto-photo-import", "sub_path": "main_code.py", "file_name": "main_code.py", "file_ext": "py", "file_size_in_byte": 1132, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "os.chdir", "line_number": 6, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 7, "usage_type": "call"}, {"api_name": "os.stat", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 15, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 26, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 30, "usage_type": "call"}]}
+{"seq_id": "40091510168", "text": "#! python3\n# pretty_stopwatch.py\n# Author: Michael Koundouros\n\"\"\"\nA stopwatch program that:\n1. Tracks the amount of time elapsed between presses of the ENTER key, with each key press starting a new “lap”\n on the timer.\n2. Prints the lap number, total time, and lap time.\n\"\"\"\n\n\nimport time\nimport pyperclip\n\n\n# Display the program's instructions.\nprint('Press ENTER to begin. Afterward, press ENTER to \"click\" the stopwatch. Press Ctrl-C to quit.')\ninput() # press Enter to begin\nprint('Started.')\nstartTime = time.time() # get the first lap's start time\nlastTime = startTime\nlapNum = 1\n\nclipboard = []\ntry:\n while True:\n input()\n lapTime = round(time.time() - lastTime, 2)\n totalTime = round(time.time() - startTime, 2)\n output_str = f'Lap #{lapNum:2}: {totalTime:5} ({lapTime:5})'\n print(output_str, end='')\n clipboard.append(output_str)\n lapNum += 1\n lastTime = time.time() # reset the last lap time\nexcept KeyboardInterrupt:\n # Handle the Ctrl-C exception to keep its error message from displaying.\n print('\\nDone.')\n pyperclip.copy('\\n'.join(clipboard))\n", "repo_name": "mkoundo/Automate_the_Boring_Stuff", "sub_path": "chapter_17_Time/pretty_stopwatch.py", "file_name": "pretty_stopwatch.py", "file_ext": "py", "file_size_in_byte": 1249, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "time.time", "line_number": 20, "usage_type": "call"}, {"api_name": "time.time", "line_number": 28, "usage_type": "call"}, {"api_name": "time.time", "line_number": 29, "usage_type": "call"}, {"api_name": "time.time", "line_number": 34, "usage_type": "call"}, {"api_name": "pyperclip.copy", "line_number": 38, "usage_type": "call"}]}
+{"seq_id": "37520989208", "text": "import datetime\nimport random\nimport secrets\nimport typing\n\nfrom google.protobuf.descriptor import EnumDescriptor, FieldDescriptor\nfrom google.protobuf.message import Message\nfrom google.protobuf.timestamp_pb2 import Timestamp\nfrom google.type.date_pb2 import Date\nfrom google.type.timeofday_pb2 import TimeOfDay\n\nfrom protarrow.common import M\nfrom protarrow.proto_to_arrow import is_map\n\nEPOCH_RATIO = 24 * 60 * 60\n\nUNIT_IN_NANOS = {\"s\": 1_000_000_000, \"ms\": 1_000_000, \"us\": 1_000, \"ns\": 1}\n\n\ndef random_string(count: int) -> str:\n return secrets.token_urlsafe(random.randint(0, count))\n\n\ndef random_bytes(count: int) -> bytes:\n return secrets.token_bytes(random.randint(0, count))\n\n\ndef random_timestamp() -> Timestamp:\n return Timestamp(\n seconds=random.randint(-9223372036, 9223372035),\n nanos=random.randint(0, 999_999_999),\n )\n\n\ndef random_date() -> Date:\n date = datetime.date.min + datetime.timedelta(days=random.randint(0, 3652058))\n return Date(year=date.year, month=date.month, day=date.day)\n\n\ndef random_time_of_day() -> TimeOfDay:\n return TimeOfDay(\n hours=random.randint(0, 23),\n minutes=random.randint(0, 59),\n seconds=random.randint(0, 59),\n nanos=random.randint(0, 999_999_999),\n )\n\n\nCPP_TYPE_GENERATOR = {\n FieldDescriptor.CPPTYPE_INT32: lambda: random.randint(-(2**31), 2**31 - 1),\n FieldDescriptor.CPPTYPE_INT64: lambda: random.randint(-(2**63), 2**63 - 1),\n FieldDescriptor.CPPTYPE_UINT32: lambda: random.randint(0, 2**32 - 1),\n FieldDescriptor.CPPTYPE_UINT64: lambda: random.randint(0, 2**64 - 1),\n FieldDescriptor.CPPTYPE_DOUBLE: lambda: random.uniform(-1, 1),\n FieldDescriptor.CPPTYPE_FLOAT: lambda: random.uniform(-1, 1),\n FieldDescriptor.CPPTYPE_BOOL: lambda: bool(random.getrandbits(1)),\n}\n\nTYPE_GENERATOR = {\n FieldDescriptor.TYPE_BYTES: random_bytes,\n FieldDescriptor.TYPE_STRING: random_string,\n}\n\nMESSAGE_GENERATORS = {\n Date.DESCRIPTOR: random_date,\n Timestamp.DESCRIPTOR: random_timestamp,\n TimeOfDay.DESCRIPTOR: random_time_of_day,\n}\n\n\ndef generate_message(message_type: typing.Type[M], repeated_count: int) -> M:\n message = message_type()\n for one_of in message_type.DESCRIPTOR.oneofs:\n one_of_index = random.randint(0, len(one_of.fields))\n if one_of_index < len(one_of.fields):\n field = one_of.fields[one_of_index]\n set_field(message, field, repeated_count)\n\n for field in message_type.DESCRIPTOR.fields:\n if field.containing_oneof is None:\n if (\n field.label == FieldDescriptor.LABEL_REPEATED\n or field.type != FieldDescriptor.TYPE_MESSAGE\n or random.getrandbits(1) == 1\n ):\n set_field(message, field, repeated_count)\n return message\n\n\ndef generate_messages(\n message_type: typing.Type[M], count: int, repeated_count: int = 10\n) -> typing.List[M]:\n return [generate_message(message_type, repeated_count) for _ in range(count)]\n\n\ndef set_field(message: Message, field: FieldDescriptor, count: int) -> None:\n data = generate_field_data(field, count)\n\n if field.label == FieldDescriptor.LABEL_REPEATED:\n field_value = getattr(message, field.name)\n if is_map(field):\n if (\n field.message_type.fields_by_name[\"value\"].type\n == FieldDescriptor.TYPE_MESSAGE\n ):\n for entry in data:\n field_value[entry.key].MergeFrom(entry.value)\n else:\n for entry in data:\n field_value[entry.key] = entry.value\n else:\n field_value.extend(data)\n elif field.type == FieldDescriptor.TYPE_MESSAGE:\n if random.getrandbits(1) == 1:\n getattr(message, field.name).CopyFrom(data)\n else:\n setattr(message, field.name, data)\n\n\ndef generate_field_data(field: FieldDescriptor, count: int):\n if field.label == FieldDescriptor.LABEL_REPEATED:\n size = random.randint(0, count)\n return [_generate_data(field, count) for _ in range(size)]\n else:\n return _generate_data(field, count)\n\n\ndef _generate_data(field: FieldDescriptor, count: int) -> typing.Any:\n if field.type == FieldDescriptor.TYPE_ENUM:\n return _generate_enum(field.enum_type)\n elif field.message_type in MESSAGE_GENERATORS:\n return MESSAGE_GENERATORS[field.message_type]()\n elif field.type == FieldDescriptor.TYPE_MESSAGE:\n return generate_message(field.message_type._concrete_class, count)\n elif field.type in TYPE_GENERATOR:\n return TYPE_GENERATOR[field.type](count)\n else:\n return CPP_TYPE_GENERATOR[field.cpp_type]()\n\n\ndef _generate_enum(enum: EnumDescriptor) -> int:\n return random.choice(enum.values).index\n\n\ndef truncate_nanos(message: Message, timestamp_unit: str, time_unit: str) -> Message:\n if message.DESCRIPTOR == Timestamp.DESCRIPTOR:\n message.nanos = (\n message.nanos // UNIT_IN_NANOS[timestamp_unit]\n ) * UNIT_IN_NANOS[timestamp_unit]\n elif message.DESCRIPTOR == TimeOfDay.DESCRIPTOR:\n message.nanos = (message.nanos // UNIT_IN_NANOS[time_unit]) * UNIT_IN_NANOS[\n time_unit\n ]\n else:\n for field in message.DESCRIPTOR.fields:\n if field.type == FieldDescriptor.TYPE_MESSAGE:\n if field.label == FieldDescriptor.LABEL_REPEATED:\n field_value = getattr(message, field.name)\n if (\n field.message_type is not None\n and field.message_type.GetOptions().map_entry\n ):\n if (\n field.message_type.fields_by_name[\"value\"].type\n == FieldDescriptor.TYPE_MESSAGE\n ):\n for key, value in field_value.items():\n truncate_nanos(value, timestamp_unit, time_unit)\n\n else:\n for item in field_value:\n truncate_nanos(item, timestamp_unit, time_unit)\n elif message.HasField(field.name):\n truncate_nanos(\n getattr(message, field.name), timestamp_unit, time_unit\n )\n return message\n", "repo_name": "tradewelltech/protarrow", "sub_path": "tests/random_generator.py", "file_name": "random_generator.py", "file_ext": "py", "file_size_in_byte": 6354, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 10, "dataset": "github-code", "pt": "24", "api": [{"api_name": "secrets.token_urlsafe", "line_number": 21, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 21, "usage_type": "call"}, {"api_name": "secrets.token_bytes", "line_number": 25, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 25, "usage_type": "call"}, {"api_name": "google.protobuf.timestamp_pb2.Timestamp", "line_number": 29, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 30, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 31, "usage_type": "call"}, {"api_name": "google.protobuf.timestamp_pb2.Timestamp", "line_number": 28, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 36, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 36, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 36, "usage_type": "call"}, {"api_name": "google.type.date_pb2.Date", "line_number": 37, "usage_type": "call"}, {"api_name": "google.type.date_pb2.Date", "line_number": 35, "usage_type": "name"}, {"api_name": "google.type.timeofday_pb2.TimeOfDay", "line_number": 41, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 42, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 43, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 44, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 45, "usage_type": "call"}, {"api_name": "google.type.timeofday_pb2.TimeOfDay", "line_number": 40, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.CPPTYPE_INT32", "line_number": 50, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 50, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.CPPTYPE_INT64", "line_number": 51, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 51, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.CPPTYPE_UINT32", "line_number": 52, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 52, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.CPPTYPE_UINT64", "line_number": 53, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 53, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.CPPTYPE_DOUBLE", "line_number": 54, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 54, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.CPPTYPE_FLOAT", "line_number": 55, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 55, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.CPPTYPE_BOOL", "line_number": 56, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 56, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 50, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 51, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 52, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 53, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 54, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 55, "usage_type": "call"}, {"api_name": "random.getrandbits", "line_number": 56, "usage_type": "call"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.TYPE_BYTES", "line_number": 60, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 60, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.TYPE_STRING", "line_number": 61, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 61, "usage_type": "name"}, {"api_name": "google.type.date_pb2.Date.DESCRIPTOR", "line_number": 65, "usage_type": "attribute"}, {"api_name": "google.type.date_pb2.Date", "line_number": 65, "usage_type": "name"}, {"api_name": "google.protobuf.timestamp_pb2.Timestamp.DESCRIPTOR", "line_number": 66, "usage_type": "attribute"}, {"api_name": "google.protobuf.timestamp_pb2.Timestamp", "line_number": 66, "usage_type": "name"}, {"api_name": "google.type.timeofday_pb2.TimeOfDay.DESCRIPTOR", "line_number": 67, "usage_type": "attribute"}, {"api_name": "google.type.timeofday_pb2.TimeOfDay", "line_number": 67, "usage_type": "name"}, {"api_name": "typing.Type", "line_number": 71, "usage_type": "attribute"}, {"api_name": "protarrow.common.M", "line_number": 71, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 74, "usage_type": "call"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.LABEL_REPEATED", "line_number": 82, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 82, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.TYPE_MESSAGE", "line_number": 83, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 83, "usage_type": "name"}, {"api_name": "random.getrandbits", "line_number": 84, "usage_type": "call"}, {"api_name": "typing.Type", "line_number": 91, "usage_type": "attribute"}, {"api_name": "protarrow.common.M", "line_number": 91, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 92, "usage_type": "attribute"}, {"api_name": "protarrow.common.M", "line_number": 92, "usage_type": "name"}, {"api_name": "google.protobuf.message.Message", "line_number": 96, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 96, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.LABEL_REPEATED", "line_number": 99, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 99, "usage_type": "name"}, {"api_name": "protarrow.proto_to_arrow.is_map", "line_number": 101, "usage_type": "call"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.TYPE_MESSAGE", "line_number": 104, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 104, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.TYPE_MESSAGE", "line_number": 113, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 113, "usage_type": "name"}, {"api_name": "random.getrandbits", "line_number": 114, "usage_type": "call"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 120, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.LABEL_REPEATED", "line_number": 121, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 121, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 122, "usage_type": "call"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 128, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.TYPE_ENUM", "line_number": 129, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 129, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.TYPE_MESSAGE", "line_number": 133, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 133, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 128, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.EnumDescriptor", "line_number": 141, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 142, "usage_type": "call"}, {"api_name": "google.protobuf.message.Message", "line_number": 145, "usage_type": "name"}, {"api_name": "google.protobuf.timestamp_pb2.Timestamp.DESCRIPTOR", "line_number": 146, "usage_type": "attribute"}, {"api_name": "google.protobuf.timestamp_pb2.Timestamp", "line_number": 146, "usage_type": "name"}, {"api_name": "google.type.timeofday_pb2.TimeOfDay.DESCRIPTOR", "line_number": 150, "usage_type": "attribute"}, {"api_name": "google.type.timeofday_pb2.TimeOfDay", "line_number": 150, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.TYPE_MESSAGE", "line_number": 156, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 156, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.LABEL_REPEATED", "line_number": 157, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 157, "usage_type": "name"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor.TYPE_MESSAGE", "line_number": 165, "usage_type": "attribute"}, {"api_name": "google.protobuf.descriptor.FieldDescriptor", "line_number": 165, "usage_type": "name"}]}
+{"seq_id": "12305424168", "text": "# Tamar Saad 207256991\r\n# Rachel Weinberger 208812628\r\nimport os.path\r\nimport random\r\nimport sys\r\nimport re\r\nimport math\r\nimport numpy as np\r\nfrom Bio.Seq import Seq\r\nfrom Bio import SeqIO\r\nimport pandas as pd\r\nimport matplotlib.pyplot as plt\r\nfrom multiprocessing import Pool\r\n\r\n\r\nclass SequenceClassifier:\r\n def __init__(self, pattern_file):\r\n self.regex_dict = self.__patterns_to_domains(pattern_file)\r\n\r\n def __prosite_to_python(self, pattern_dict):\r\n # dictionary that translate prosite format to python format\r\n dic = {\"-\": None, \"x\": \".\", \"(\": \"{\", \")\": \"}\", \"{\": \"[^\", \"}\": \"]\", \"<\": \"^\", \">\": \"$\", \",\": \",\"}\r\n RE_patterns = {}\r\n # going through each pattern and domain\r\n for pattern, domain in pattern_dict.items():\r\n # check if the format is valid\r\n if self.__check_prosite_format(pattern):\r\n # translate the pattern and pair it with its domain\r\n trans = str(pattern).maketrans(dic)\r\n re_pattern = str(pattern).translate(trans)\r\n RE_patterns[re_pattern] = domain\r\n return RE_patterns\r\n\r\n def __check_prosite_format(self, pattern):\r\n # create regexes that represent correct prosite format\r\n occurrences = \"(\\([\\d](,[\\d])?\\))?\"\r\n prefix = \"(<)?\"\r\n suffix = \"(>)?\"\r\n amino_group = \"(\\[[A-Z]*\\])\"\r\n amino_single = \"[A-Z]\"\r\n amino_any = \"x\"\r\n avoid = \"(\\{[A-Z]*\\})\"\r\n correct_format_once = \"((\" + amino_group + \"|\" + amino_single + \"|\" + amino_any + \"|\" + avoid + \")\" + occurrences + \")\"\r\n correct_format_repeat = \"((\" + amino_group + \"|\" + amino_single + \"|\" + amino_any + \"|\" + avoid + \")\" + occurrences + \"-)*\"\r\n correct_format = prefix + correct_format_repeat + correct_format_once + suffix\r\n\r\n # check if the pattern is a valid prosite format\r\n check = re.search(correct_format, pattern)\r\n # return boolean value\r\n return check.group() is pattern and re.compile(check.group())\r\n\r\n def __patterns_to_domains(self, pattern_file):\r\n # initialize empty dictionary\r\n dict_from_csv = {}\r\n # read file into dictionary\r\n try:\r\n dict_from_csv = pd.read_csv(pattern_file, header=0, index_col=0, squeeze=True).to_dict()\r\n except ValueError:\r\n print(\"File is not in wanted format\")\r\n # turn the prosite regex to python regex\r\n regex_dict = self.__prosite_to_python(dict_from_csv)\r\n return regex_dict\r\n\r\n def classify(self, seq_list, csv_file):\r\n protein_domains = dict()\r\n # for every protein in the list\r\n protein_domains[\"Sequence\"] = \"Domains\"\r\n for seq in seq_list:\r\n domains = \"\"\r\n flag = False\r\n # go through each regex\r\n for reg in self.regex_dict:\r\n # check if the pattern fits any of the proteins\r\n if re.search(reg, seq):\r\n domains += self.regex_dict[reg] + \";\"\r\n flag = True\r\n if not flag:\r\n domains = \"NA\"\r\n if flag:\r\n domains = domains[:-1]\r\n protein_domains[seq] = domains\r\n # turn dictionary into dataframe, and write it to csv file\r\n df = pd.DataFrame.from_dict(protein_domains, orient='index')\r\n df.to_csv(csv_file, index=True, header=False)\r\n\r\n\r\n# define DNA and RNA polymerase\r\nclass Polymerase:\r\n # constructor\r\n def __init__(self, type, error_rate=0):\r\n self.type = type\r\n self.error_rate = error_rate\r\n # define translation differently for RNA/DNA polymerase\r\n A_translation = \"\"\r\n if type == \"RNA\":\r\n A_translation = \"U\"\r\n elif type == \"DNA\":\r\n A_translation = \"T\"\r\n # initialize a dictionary for transcribing\r\n self.transcribing_dictionary = {\"T\": \"A\", \"t\": \"A\", \"C\": \"G\", \"c\": \"G\", \"G\": \"C\", \"g\": \"C\",\r\n \"A\": A_translation,\r\n \"a\": A_translation}\r\n\r\n # this function receive a DNA sequence and translate it to RNA\r\n def transcribe(self, dna_seq):\r\n\r\n # find the indices of the mutations\r\n num_of_mutants = math.ceil(self.error_rate * len(dna_seq))\r\n # locations = np.random.choice(range(1, len(dna_seq)), num_of_mutants)\r\n locations = random.sample(range(0, len(dna_seq)), num_of_mutants)\r\n # initialize an empty list\r\n rna = \"\"\r\n # translate the letter according to the dictionary\r\n for ind, nuc in enumerate(dna_seq):\r\n if nuc in self.transcribing_dictionary.keys():\r\n # if we need to insert mutation\r\n if ind in locations:\r\n # create a random mutation. if it's synonym- replace it\r\n nucleotides = set(self.transcribing_dictionary.values()) # unique trans dic values\r\n nucleotides = list(nucleotides)\r\n nucleotides.remove(self.transcribing_dictionary[nuc])\r\n mut = np.random.choice(nucleotides)\r\n # while mut == self.transcribing_dictionary[nuc]:\r\n # mut = np.random.choice(list(self.transcribing_dictionary.values()))\r\n # add the mutation to the sequence\r\n rna += mut\r\n else:\r\n rna += self.transcribing_dictionary[nuc]\r\n else:\r\n break\r\n # revers the sequence so it will be 5'-3', and cut the last character to fit the format\r\n if rna:\r\n return rna[::-1]\r\n else:\r\n return None\r\n\r\n\r\n# define ribosome\r\nclass Ribosome:\r\n # constructor\r\n def __init__(self, genetic_code, start_codons):\r\n self.genetic_code = genetic_code\r\n self.start_codons = start_codons\r\n\r\n # turn RNA sequence to the biggest protein available\r\n def synthesize(self, rna_seq):\r\n # calls to translate\r\n protein = self.translate(rna_seq)\r\n if protein:\r\n return protein\r\n else:\r\n return None\r\n\r\n # this function receives rna sequence and returns the codons of the longest reading frame\r\n def translate(self, rna_seq):\r\n # initialize the biggest protein with empty string\r\n max_protein = \"\"\r\n # go through every nucleotide and look for start codon (AUG)\r\n for i in range(len(rna_seq)):\r\n if rna_seq[i:i + 3] in self.start_codons:\r\n # initialize the protein we will go through\r\n protein = \"\"\r\n # go through each codon and add it to the protein. stop in stop codon or at the end of the sequence\r\n for j in range(i, len(rna_seq) - 2, 3):\r\n codon = rna_seq[j:j + 3]\r\n # check for stop codon\r\n if not self.genetic_code[codon]:\r\n i += 4\r\n break\r\n else:\r\n # add the codon to the protein\r\n protein += self.genetic_code[codon]\r\n # when the protein is done, check if it's bigger than the biggest protein, and replace it if so\r\n if len(protein) > len(max_protein):\r\n max_protein = protein\r\n return max_protein\r\n\r\n\r\nclass Cell:\r\n # constructor\r\n def __init__(self, name, genome, num_copies, genetic_code, start_codons, division_rate):\r\n self.name = name\r\n self.genome = []\r\n self.genome.append(genome)\r\n # input check\r\n if type(num_copies) is int and num_copies > 0:\r\n self.num_copies = num_copies\r\n self.genetic_code = genetic_code\r\n self.start_codons = start_codons\r\n # input check\r\n if type(division_rate) is int and division_rate > 1:\r\n self.division_rate = division_rate\r\n # initialize RNA+DNA polymerases\r\n self.RNA_Polymerase = Polymerase(\"RNA\", 0)\r\n self.DNA_Polymerase = Polymerase(\"DNA\", 0)\r\n # initialize Ribosome\r\n self.Ribosome = Ribosome(genetic_code, start_codons)\r\n\r\n # define a printing method\r\n def __str__(self):\r\n return \"<\" + str(self.name) + \", \" + str(self.num_copies) + \", \" + str(self.division_rate) + \">\"\r\n\r\n # returns a list of n identical cells, while n is the division rate\r\n def mitosis(self):\r\n return self * self.division_rate\r\n\r\n # define the * operator\r\n def __mul__(self, num):\r\n cells = [self]\r\n return cells * num\r\n\r\n # return 2 cells: one identical to the original, one with the complimentary genome. both with n/2 genome copies\r\n def meiosis(self):\r\n if self.num_copies % 2 != 0:\r\n return None\r\n new_cell = Cell(self.name, self.genome, (self.num_copies / 2), self.genetic_code, self.start_codons,\r\n self.division_rate)\r\n # find the complementary strands of the genome\r\n comp_genome = []\r\n for seq in self.genome:\r\n comp_genome.append(self.DNA_Polymerase.transcribe(seq))\r\n # define the complementary cell\r\n new_comp = Cell(self.name, comp_genome, self.num_copies / 2, self.genetic_code, self.start_codons,\r\n self.division_rate)\r\n return [new_cell, new_comp]\r\n\r\n # this function find microsatellites in repeats of 3-6\r\n def find_srr(self, dna_seq):\r\n # flag to know if there are satellites\r\n satellites = {}\r\n # looking for microsatellites in increasing sizes of nucleotides\r\n for size_of_match in range(1, 7):\r\n # looking for satellite in different reading frames\r\n for i in range(len(dna_seq) - (len(dna_seq) % size_of_match)):\r\n count = 1 # number of repeats\r\n # checking for repeats\r\n for j in range(i, len(dna_seq) - (len(dna_seq) % size_of_match), size_of_match):\r\n # if the sequences are the same- increase the counter\r\n if dna_seq[j: j + size_of_match] == dna_seq[j + size_of_match: j + 2 * size_of_match]:\r\n count += 1\r\n else:\r\n # if the sequences are different and there are more than 3 repeats- add it to the dictionary\r\n if count >= 3:\r\n satellite = dna_seq[j: j + size_of_match]\r\n # if the satellite exists already- check if the count is bigger\r\n if satellite in satellites.keys():\r\n if count <= satellites[satellite]:\r\n break\r\n # if the satellite didn't exist/the count is smaller- update the dictionary\r\n satellites[satellite] = count\r\n break\r\n size_of_match += 1\r\n if satellites:\r\n sat_list = \"\"\r\n for satellite, count in sorted(satellites.items(), key=lambda t: t[0]):\r\n sat_list += satellite + \",\" + str(count) + \";\"\r\n return sat_list[:-1]\r\n else:\r\n return None\r\n\r\n # returns tuple for every strand. each tuple contains: satellites, RNA transcribe, translated protein\r\n def repertoire(self):\r\n list_of_tuples = []\r\n for sequence in self.genome:\r\n # find the satellites\r\n satellites = self.find_srr(sequence)\r\n if not satellites:\r\n satellites = \"No simple repeats in DNA sequence\"\r\n # find RNA transcribe\r\n rna_seq = self.RNA_Polymerase.transcribe(sequence)\r\n # find the biggest protein available\r\n protein = self.Ribosome.synthesize(rna_seq)\r\n if not protein:\r\n protein = \"Non-coding RNA\"\r\n seq = (satellites, rna_seq, protein)\r\n # returns list of tuples\r\n list_of_tuples.append(seq)\r\n return list_of_tuples\r\n\r\n\r\n# inherit class from cell\r\nclass ProkaryoticCell(Cell):\r\n # constructor\r\n def __init__(self, genome):\r\n prokaryotic_genetic_code = {\r\n 'AUA': 'I', 'AUC': 'I', 'AUU': 'I', 'AUG': 'M',\r\n 'ACA': 'T', 'ACC': 'T', 'ACG': 'T', 'ACU': 'T',\r\n 'AAC': 'N', 'AAU': 'N', 'AAA': 'K', 'AAG': 'K',\r\n 'AGC': 'S', 'AGU': 'S', 'AGA': 'R', 'AGG': 'R',\r\n 'CUA': 'L', 'CUC': 'L', 'CUG': 'L', 'CUU': 'L',\r\n 'CCA': 'P', 'CCC': 'P', 'CCG': 'P', 'CCU': 'P',\r\n 'CAC': 'H', 'CAU': 'H', 'CAA': 'Q', 'CAG': 'Q',\r\n 'CGA': 'R', 'CGC': 'R', 'CGG': 'R', 'CGU': 'R',\r\n 'GUA': 'V', 'GUC': 'V', 'GUG': 'V', 'GUU': 'V',\r\n 'GCA': 'A', 'GCC': 'A', 'GCG': 'A', 'GCU': 'A',\r\n 'GAC': 'D', 'GAU': 'D', 'GAA': 'E', 'GAG': 'E',\r\n 'GGA': 'G', 'GGC': 'G', 'GGG': 'G', 'GGU': 'G',\r\n 'UCA': 'S', 'UCC': 'S', 'UCG': 'S', 'UCU': 'S',\r\n 'UUC': 'F', 'UUU': 'F', 'UUA': 'L', 'UUG': 'L',\r\n 'UAC': 'Y', 'UAU': 'Y', 'UAA': None, 'UAG': None,\r\n 'UGC': 'C', 'UGU': 'C', 'UGA': 'U', 'UGG': 'W'}\r\n start_codons = (\"AUG\", \"GUG\", \"UUG\")\r\n division_rate = 4\r\n num_copies = 1\r\n # calls parent's constructor\r\n super().__init__(\"ProKaryoticCell\", genome, num_copies, prokaryotic_genetic_code, start_codons, division_rate)\r\n\r\n\r\n# inherit class from cell\r\nclass EukaryoticCell(Cell):\r\n # constructor\r\n def __init__(self, name, genome, division_rate):\r\n standard_genetic_code = {\r\n 'AUA': 'I', 'AUC': 'I', 'AUU': 'I', 'AUG': 'M',\r\n 'ACA': 'T', 'ACC': 'T', 'ACG': 'T', 'ACU': 'T',\r\n 'AAC': 'N', 'AAU': 'N', 'AAA': 'K', 'AAG': 'K',\r\n 'AGC': 'S', 'AGU': 'S', 'AGA': 'R', 'AGG': 'R',\r\n 'CUA': 'L', 'CUC': 'L', 'CUG': 'L', 'CUU': 'L',\r\n 'CCA': 'P', 'CCC': 'P', 'CCG': 'P', 'CCU': 'P',\r\n 'CAC': 'H', 'CAU': 'H', 'CAA': 'Q', 'CAG': 'Q',\r\n 'CGA': 'R', 'CGC': 'R', 'CGG': 'R', 'CGU': 'R',\r\n 'GUA': 'V', 'GUC': 'V', 'GUG': 'V', 'GUU': 'V',\r\n 'GCA': 'A', 'GCC': 'A', 'GCG': 'A', 'GCU': 'A',\r\n 'GAC': 'D', 'GAU': 'D', 'GAA': 'E', 'GAG': 'E',\r\n 'GGA': 'G', 'GGC': 'G', 'GGG': 'G', 'GGU': 'G',\r\n 'UCA': 'S', 'UCC': 'S', 'UCG': 'S', 'UCU': 'S',\r\n 'UUC': 'F', 'UUU': 'F', 'UUA': 'L', 'UUG': 'L',\r\n 'UAC': 'Y', 'UAU': 'Y', 'UAA': None, 'UAG': None,\r\n 'UGC': 'C', 'UGU': 'C', 'UGA': None, 'UGG': 'W'}\r\n start_codons = \"AUG\"\r\n num_copies = 2\r\n # calls parent's constructor\r\n super().__init__(name, genome, num_copies, standard_genetic_code, start_codons, division_rate)\r\n\r\n\r\n# inherit from EukaryoticCell\r\nclass NeuronCell(EukaryoticCell):\r\n # constructor\r\n def __init__(self, genome):\r\n division_rate = 2\r\n # calls parent's constructor\r\n super().__init__(\"NeuronCell\", genome, division_rate)\r\n\r\n\r\n# inherit from EukaryoticCell\r\nclass StemCell(EukaryoticCell):\r\n def __init__(self, genome):\r\n division_rate = 3\r\n # calls parent's constructor\r\n super().__init__(\"StemCell\", genome, division_rate)\r\n\r\n\r\n# inherit from stem cell\r\nclass MutantCell(StemCell):\r\n def __init__(self, genome, num_mutations=0, error_rate=0.05):\r\n # call father's constructor\r\n super().__init__(genome)\r\n # override the name\r\n self.name = \"MutantCell\"\r\n # default mutation rate is 1:20\r\n self.DNA_Polymerase.error_rate = error_rate\r\n self.RNA_Polymerase.error_rate = 0\r\n # initialize num of muts to 0\r\n self.num_of_mutations = num_mutations\r\n self.num_of_new_mutations_per_generation = self.calculate_num_of_muts_per_generation()\r\n\r\n def calculate_num_of_muts_per_generation(self):\r\n num_of_new_mutations_per_generation = 0\r\n for g in self.genome:\r\n num_of_new_mutations_per_generation += math.ceil(len(g) * self.DNA_Polymerase.error_rate)\r\n return num_of_new_mutations_per_generation\r\n\r\n # returns a list of n identical cells, while n is the division rate\r\n def mitosis(self):\r\n # get a list of mutants\r\n mutants = self * self.division_rate\r\n mutants[0] = self\r\n return mutants\r\n\r\n # define the * operator\r\n def __mul__(self, num):\r\n # create mutant genome\r\n t_genome = self.get_mutant_genome()\r\n # check if the number of mutations is under 10\r\n if self.num_of_mutations + self.num_of_new_mutations_per_generation > 10:\r\n # create cancer cell\r\n mutant = CancerCell(t_genome,\r\n num_mutations=self.num_of_mutations + self.num_of_new_mutations_per_generation,\r\n error_rate=self.DNA_Polymerase.error_rate)\r\n else:\r\n # create mutant cell\r\n mutant = MutantCell(t_genome,\r\n num_mutations=self.num_of_mutations + self.num_of_new_mutations_per_generation,\r\n error_rate=self.DNA_Polymerase.error_rate)\r\n\r\n # def the number of mutations the cell has\r\n cells = [mutant]\r\n return cells * num\r\n\r\n def get_mutant_genome(self):\r\n # t_genome = []\r\n complement = None\r\n for dna in self.genome:\r\n complement = Seq(self.DNA_Polymerase.transcribe(dna))\r\n # t_genome.append(str(complement.reverse_complement()))\r\n complement = str(complement.reverse_complement())\r\n return complement\r\n\r\n\r\nclass CancerCell(MutantCell):\r\n def __init__(self, genome, num_mutations, error_rate=0.05):\r\n super().__init__(genome, num_mutations, error_rate=error_rate)\r\n self.division_rate = 10\r\n self.name = \"CancerCell\"\r\n\r\n\r\n# factory to initialize each cell\r\nclass CellFactory:\r\n def create_cell_object(self, name, genome):\r\n if name == \"NeuronCell\":\r\n return NeuronCell(genome)\r\n if name == \"StemCell\":\r\n return StemCell(genome)\r\n if name == \"ProkaryoticCell\":\r\n return ProkaryoticCell(genome)\r\n if name == \"MutantCell\":\r\n return MutantCell(genome)\r\n if name == \"CancerCell\":\r\n return CancerCell(genome, num_mutations=10)\r\n else:\r\n raise AssertionError(name)\r\n\r\n\r\n# check if the input sequences are valid as genome\r\ndef is_genome_valid(sequences):\r\n genome = (\"A\", \"T\", \"G\", \"C\",)\r\n for seq in sequences:\r\n matched_list = [characters in genome for characters in seq.upper()]\r\n assert all(matched_list), \"Invalid input \" + seq\r\n\r\n\r\ndef cells_divisions(cell, divisions_num, max_cell_num):\r\n divs = 0\r\n num_of_cells = 1\r\n cells = [cell]\r\n # while we didn't do maximum num of divisions:\r\n while divs < int(divisions_num):\r\n # go through each cell in the list\r\n for c in range(num_of_cells):\r\n # if we will not exceed from the max number of cells- do mitosis\r\n # the -1 is because we exclude the cell that actually is going through mitosis\r\n if len(cells) <= max_cell_num - (cells[c].division_rate - 1):\r\n # adding the new cells to the list, excluding the original cell that actually did mitosis\r\n cells += (cells[c].mitosis()[1:])\r\n else: # we have the max number of cells and can exit both loops\r\n break\r\n # after all the cells went through mitosis once- we increase the number of divisions\r\n divs += 1\r\n num_of_cells = len(cells)\r\n return cells\r\n\r\n\r\ndef get_different_proteins_from_cells(cells):\r\n proteins = []\r\n for cell in cells:\r\n for sequence in cell.genome:\r\n # rep = cell.repertoire()\r\n rna_seq = cell.RNA_Polymerase.transcribe(sequence)\r\n # find the biggest protein available\r\n cell_proteins = cell.Ribosome.synthesize(rna_seq)\r\n if not cell_proteins:\r\n cell_proteins = \"Non-coding RNA\"\r\n proteins += [cell_proteins]\r\n proteins = np.array(proteins)\r\n unique_proteins = np.unique(proteins)\r\n unique_proteins = np.delete(unique_proteins, np.where(unique_proteins == \"Non-coding RNA\"))\r\n return unique_proteins\r\n\r\n\r\ndef get_most_mutant_cell(mutant_cells):\r\n most_mutant = None\r\n muts_num = 0\r\n for cell in mutant_cells:\r\n if cell.num_of_mutations > muts_num:\r\n most_mutant = cell\r\n muts_num = cell.num_of_mutations\r\n return most_mutant\r\n\r\n\r\ndef get_genome_from_fasta(fastaFile):\r\n GenomicSequences = []\r\n # records = list(SeqIO.parse(fastaFile, \"fasta\"))\r\n # for r in records:\r\n # genome.append(r.seq)\r\n for record in SeqIO.parse(fastaFile, \"fasta\"):\r\n GenomicSequences.append(record.seq)\r\n return GenomicSequences\r\n\r\n\r\n\"\"\"\r\nthis function gets a sequence and returns a df that contains the number of cancer cells,\r\nmutant cells and different proteins, as a function of different error rates and divisions numbers\r\n\"\"\"\r\ndef get_results(seq):\r\n # initialize a list of error rates\r\n error_rates = np.arange(0.05, 0.5, 0.05)\r\n error_rates = np.concatenate(([0.01], error_rates, [0.5]))\r\n # initialize a list of divisions number\r\n division_numbers = range(1, 6)\r\n # initialize an empty dataframe\r\n df = pd.DataFrame(\r\n columns=['division_number', 'error_rate', 'number_of_cancer_cells', 'number_of_mutant_cells',\r\n 'number_of_proteins'])\r\n # go through each division number\r\n for div_num in division_numbers:\r\n # go through each error rate\r\n for error_rate in error_rates:\r\n # initialize a mutant cell\r\n mut_cell = MutantCell(str(seq), error_rate=error_rate)\r\n # get all the cells after mitosis\r\n cells = cells_divisions(mut_cell, div_num, float('inf'))\r\n # get number of cancer cells\r\n cancer_cells_num = len([c for c in cells if c.name == \"CancerCell\"])\r\n # number of mutant cells\r\n mutant_cell_num = len(cells) - cancer_cells_num\r\n # get a list of the different proteins\r\n proteins = get_different_proteins_from_cells(cells)\r\n df = df.append({'division_number': div_num, 'error_rate': error_rate,\r\n 'number_of_cancer_cells': cancer_cells_num, 'number_of_mutant_cells': mutant_cell_num,\r\n 'number_of_proteins': len(proteins)}, ignore_index=True)\r\n return df\r\n\r\n\r\n# option number 1 for the combined graphs- linear graphs\r\n\"\"\"\r\ndef create_combined_graphs(data_df, sequences):\r\n # the different options for x axis\r\n x_axes = [\"division_number\", \"error_rate\"]\r\n # go through each kind of x axis\r\n for ind, x in enumerate(x_axes):\r\n # initialize a new plot\r\n plt.figure()\r\n curves = []\r\n # create a curve for each sequence\r\n for num, seq in enumerate(sequences):\r\n # get the relevant data of the current sequence from the df\r\n data_df_seq = data_df[data_df['Sequence'].str.contains(str(seq))]\r\n y_axis = data_df_seq['number_of_proteins']\r\n x_axis = data_df_seq[x]\r\n curves.append(x_axis)\r\n plt.plot(x_axis, y_axis, label=(\"seq \" + str(num + 1)))\r\n plt.xlabel(x)\r\n plt.ylabel('number_of_proteins')\r\n title = \"number of proteins as a function of \" + x\r\n plt.title(title)\r\n plt.legend()\r\n plt.savefig(\"exercise4_207256991_208812628_\" + title + \".png\")\r\n\"\"\"\r\n\r\n\r\n\"\"\"\r\nthis function creates a plot for each sequence:\r\nx axis: error rate\r\ny axis: divisions number\r\nsize+color of points: number of proteins\r\n\"\"\"\r\n# option number 2 to combined graphs: scatter plots\r\ndef create_combined_graphs_2(data_df, sequences):\r\n # go through each kind of x axis\r\n for ind, seq in enumerate(sequences):\r\n # initialize a new plot\r\n plt.figure()\r\n # get the data of one sequence\r\n data_df_seq = data_df[data_df['Sequence'].str.contains(str(seq))]\r\n # set the axes\r\n x_axis = data_df_seq[\"error_rate\"]\r\n y_axis = data_df_seq[\"division_number\"]\r\n # number of proteins\r\n prot_num = data_df_seq['number_of_proteins']\r\n # create scatter plot:\r\n # define the sizes and colors to represent the number of protein\r\n plt.scatter(x_axis, y_axis, c=prot_num, cmap='viridis', s=prot_num, alpha=0.7)\r\n # labels and titles\r\n plt.xlabel(\"error rate\")\r\n plt.ylabel(\"division number\")\r\n title = f\"sequence {ind+1} protein number ~ division number and error rate\"\r\n plt.suptitle(f\"sequence {seq}\")\r\n plt.title(\"number of proteins is represented by size and color\")\r\n plt.colorbar()\r\n plt.savefig(\"exercise4_207256991_208812628_\" + title + \".png\")\r\n\r\n\r\n# gets the required plots: a plot for number of proteins, number of mutant cells\r\n# and number of cancer cells as a function of error rate\r\ndef get_graphs_per_seq(sequences):\r\n # read the csv file we created\r\n data_df = pd.read_csv('exercise4_207256991_208812628.csv', sep=',')\r\n # keep only the rows of division_number = 5\r\n data_df_max_divs = data_df[data_df['division_number'].astype(str).str.contains('5')]\r\n plot_num = 1\r\n # go through each sequence\r\n for ind, seq in enumerate(sequences):\r\n # keep only the data for this sequence\r\n data_df_seq = data_df_max_divs[data_df_max_divs['Sequence'].str.contains(str(seq))]\r\n # define the x axis as error rate\r\n x_axis = data_df_seq['error_rate']\r\n # define the different options for y axis\r\n y_axes = [\"number_of_cancer_cells\", \"number_of_mutant_cells\", \"number_of_proteins\"]\r\n # go through each option\r\n for y in y_axes:\r\n # define a specific plot and clear it from previous plots\r\n plt.figure(plot_num)\r\n # define the y axis\r\n y_axis = data_df_seq[y]\r\n plt.plot(x_axis, y_axis)\r\n # create labels\r\n plt.xlabel('error_rate')\r\n plt.ylabel(y)\r\n plt.title(str(seq))\r\n # save the plot\r\n plt.savefig(\"exercise4_207256991_208812628_\" + f\"seq_{ind+1}_\" + y + \".png\")\r\n plot_num += 1\r\n plt.clf()\r\n # create_combined_graphs(data_df, sequences)\r\n # create the combined graphs\r\n create_combined_graphs_2(data_df, sequences)\r\n\r\n\r\ndef main():\r\n random.seed(1)\r\n # input checkups\r\n assert len(sys.argv) == 2, \"Wrong number of inputs\"\r\n fastaFile = sys.argv[1]\r\n assert os.path.isfile(fastaFile), \"input is not a file \" + fastaFile\r\n name, extension = os.path.splitext(fastaFile)\r\n assert extension == \".fa\", \"input must be fasta file \" + fastaFile\r\n # get list of the sequences from the input fasta file\r\n # get the sequences names from the file\r\n seqs = []\r\n seq_names = []\r\n for seq_record in SeqIO.parse(fastaFile, \"fasta\"):\r\n seqs.append(str(seq_record.seq))\r\n seq_names.append(seq_record.id)\r\n sequences = pd.DataFrame(list(zip(seq_names, seqs)), columns=['Name', 'Seq'])\r\n\r\n # check input sequences\r\n is_genome_valid(seqs)\r\n\r\n # initialize the results dataframe\r\n df = pd.DataFrame(\r\n columns=['Sequence', 'division_number', 'error_rate', 'number_of_cancer_cells', 'number_of_mutant_cells',\r\n 'number_of_proteins'])\r\n # go through each sequence\r\n for index, seq in sequences.iterrows():\r\n # create an iterable list of the sequence for the map function\r\n # it's just a list with the same sequence 3 times\r\n itr_seq = [seq[\"Seq\"]] * 3\r\n # create 3 processes\r\n with Pool(3) as p:\r\n # the output is a list of 3 dataframes, one from each process\r\n results = p.map(get_results, itr_seq)\r\n # calculate the average result\r\n results = (results[0] + results[1] + results[2]) / 3\r\n # create a list with the sequence name, for the csv file\r\n seq_list = [seq['Name']] * len(results.index)\r\n # insert the list to the dataframe in the first column\r\n results.insert(0, \"Sequence\", seq_list)\r\n # concatenate the results dataframe to the df of all the sequences\r\n df = pd.concat([df, results])\r\n # save results to csv\r\n df.to_csv('exercise4_207256991_208812628.csv', index=False)\r\n # create the required graphs\r\n get_graphs_per_seq(sequences['Name'])\r\n\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n", "repo_name": "TamarSaad/University-Assignments", "sub_path": "Lamut/ex4/exercise4_207256991_208812628.py", "file_name": "exercise4_207256991_208812628.py", "file_ext": "py", "file_size_in_byte": 28627, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "re.search", "line_number": 48, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 50, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 57, "usage_type": "call"}, {"api_name": "re.search", "line_number": 74, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 83, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 83, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 108, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 122, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 370, "usage_type": "call"}, {"api_name": "Bio.Seq.Seq", "line_number": 404, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 474, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 475, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 476, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 476, "usage_type": "call"}, {"api_name": "Bio.SeqIO.parse", "line_number": 495, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 495, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 506, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 507, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 511, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 572, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 572, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 582, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 582, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 584, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 584, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 585, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 585, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.suptitle", "line_number": 587, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 587, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 588, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 588, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 589, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 589, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 590, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 590, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 597, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 612, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 612, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 615, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 615, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 617, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 617, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 618, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 618, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 619, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 619, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 621, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 621, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 623, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 623, "usage_type": "name"}, {"api_name": "random.seed", "line_number": 630, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 632, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 633, "usage_type": "attribute"}, {"api_name": "os.path.path.isfile", "line_number": 634, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 634, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 634, "usage_type": "name"}, {"api_name": "os.path.path.splitext", "line_number": 635, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 635, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 635, "usage_type": "name"}, {"api_name": "Bio.SeqIO.parse", "line_number": 641, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 641, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 644, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 650, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 659, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 669, "usage_type": "call"}]}
+{"seq_id": "39124116231", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# # COMPUTER VISION AND IOT INTERN\n# \n# # SPARKS FOUNDATION\n# \n# ## TASK 1 : Object Detection\n# \n# ## Detect the object in a photo/video\n# \n# ## By: Swati Namdev\n\n# In[1]:\n\n\nimport numpy as np\nimport cv2\n\n\n# In[2]:\n\n\nnet=cv2.dnn.readNet(\"yolov3.weights\",\"yolov3.cfg\")\n\n\n# In[3]:\n\n\nclasses=[]\nwith open(\"coco.names\",\"r\") as f:\n classes=[line.strip()for line in f.readlines()]\nprint(classes)\n\n\n# In[4]:\n\n\nlayer_names=net.getLayerNames()\noutput_layers=[layer_names[i[0]-1] for i in net.getUnconnectedOutLayers()]\n\n\n# In[5]:\n\n\ncolors=np.random.uniform(0,255, size=(len(classes),3))\n\n\n# In[6]:\n\n\nimg=cv2.imread(\"room_ser.jpg\")\nimg=cv2.resize(img,None,fx=0.4,fy=0.4)\nheight,width,channels=img.shape\nblob=cv2.dnn.blobFromImage(img,0.00392,(416,416),(0,0,0),True,crop=False)\n\n\n# In[7]:\n\n\n''''for b in blob:\n for n,img_blob in enumerate(b):\n cv2.imshow(str(n),img_blob)'''\n\n\n# In[8]:\n\n\nnet.setInput(blob)\nouts=net.forward(output_layers)\n\n\n# In[9]:\n\n\nclass_ids=[]\nconfidences=[]\nboxes=[]\n\n\n# In[10]:\n\n\nfor out in outs:\n for detection in out:\n scores=detection[5:]\n class_id=np.argmax(scores)\n confidence=scores[class_id]\n if confidence>0.5:\n center_x=int(detection[0]*width)\n center_y=int(detection[1]*height)\n w=int(detection[2]*width)\n h=int(detection[1]*height)\n \n x=int(center_x-w/2)\n y=int(center_y-h/2)\n \n boxes.append([x,y,w,h])\n confidences.append(float(confidence))\n class_ids.append(class_id)\n\n\n# In[11]:\n\n\nprint(len(boxes))\n\n\n# In[12]:\n\n\n#number_objects_detected=len(boxes)\nindexes=cv2.dnn.NMSBoxes(boxes,confidences,0.5,0.4)\n#print(indexes)\nfont=cv2.FONT_HERSHEY_PLAIN\nfor i in range(len(boxes)):\n if i in indexes:\n x,y,w,h=boxes[i]\n label=str(classes[class_ids[i]])\n color=colors[i]\n cv2.rectangle(img,(x,y),(x+w,y+h),color,2)\n cv2.putText(img,label,(x,y+30),font,3,color,3)\n #print(label)\n\n\n# In[ ]:\n\n\n\n\n\n# In[13]:\n\n\ncv2.imshow(\"Image\",img)\ncv2.waitKey(0)\n\n\n# In[14]:\n\n\ncv2.destroyAllWindows()\n\n\n# In[ ]:\n\n\n\n\n", "repo_name": "swati0806/OBJECT_DETECTION", "sub_path": "photo_Object_detection1.py", "file_name": "photo_Object_detection1.py", "file_ext": "py", "file_size_in_byte": 2162, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "cv2.dnn.readNet", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 46, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 53, "usage_type": "call"}, {"api_name": "cv2.dnn.blobFromImage", "line_number": 55, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 55, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 87, "usage_type": "call"}, {"api_name": "cv2.dnn.NMSBoxes", "line_number": 113, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 113, "usage_type": "attribute"}, {"api_name": "cv2.FONT_HERSHEY_PLAIN", "line_number": 115, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 121, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 122, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 135, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 136, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 142, "usage_type": "call"}]}
+{"seq_id": "71540964543", "text": "import unittest\nfrom jettools import validation\nimport typing\n\n\ndef _verify_dict_list(d) -> bool:\n for key in d:\n if not isinstance(key, str):\n return False\n if not isinstance(d[key], list):\n return False\n for item in d[key]:\n if not isinstance(item, int):\n return False\n return True\n\n\nclass TestAreValidArgs(unittest.TestCase):\n def test_should_raise_exception_if_not_all_args_are_validators(self):\n with self.assertRaises(ValueError):\n validation.are_valid_args([validation.Validator('test', 2, int), 3])\n\n def test_should_return_true_if_none(self):\n self.assertTrue(validation.are_valid_args([]))\n\n def test_should_return_false_if_an_arg_doesnt_validate(self):\n args = [\n validation.Validator('name', 4, str),\n validation.Validator('test3', 6.2, [str, int])\n ]\n self.assertFalse(validation.are_valid_args(args))\n\n def test_should_return_true_if_valid(self):\n args = [\n validation.Validator('name', 4, int),\n validation.Validator('test2', 'tst', str),\n validation.Validator('test3', 6.2, [str, int, float]),\n validation.Validator('test4', (2,), typing.Tuple),\n ]\n result = validation.are_valid_args(args)\n print(result.message)\n self.assertTrue(result)\n\n def test_should_raise_exception_if_parameterized_type(self):\n args = [\n validation.Validator('test4', (2,), typing.Tuple[str]),\n ]\n with self.assertRaises(ValueError):\n result = validation.are_valid_args(args)\n\n def test_should_validate_with_functions(self):\n d = {'test': [1, 2, 3, 4]}\n args = [\n validation.Validator('name', d, _verify_dict_list),\n validation.Validator('test2', d['test'], lambda x: isinstance(x, typing.List))\n ]\n result = validation.are_valid_args(args)\n print(result.message)\n self.assertTrue(result)\n", "repo_name": "jettdc/jetts-tools", "sub_path": "test/test_validation.py", "file_name": "test_validation.py", "file_ext": "py", "file_size_in_byte": 2027, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "unittest.TestCase", "line_number": 18, "usage_type": "attribute"}, {"api_name": "jettools.validation.are_valid_args", "line_number": 21, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 21, "usage_type": "name"}, {"api_name": "jettools.validation.Validator", "line_number": 21, "usage_type": "call"}, {"api_name": "jettools.validation.are_valid_args", "line_number": 24, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 24, "usage_type": "name"}, {"api_name": "jettools.validation.Validator", "line_number": 28, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 28, "usage_type": "name"}, {"api_name": "jettools.validation.Validator", "line_number": 29, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 29, "usage_type": "name"}, {"api_name": "jettools.validation.are_valid_args", "line_number": 31, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 31, "usage_type": "name"}, {"api_name": "jettools.validation.Validator", "line_number": 35, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 35, "usage_type": "name"}, {"api_name": "jettools.validation.Validator", "line_number": 36, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 36, "usage_type": "name"}, {"api_name": "jettools.validation.Validator", "line_number": 37, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 37, "usage_type": "name"}, {"api_name": "jettools.validation.Validator", "line_number": 38, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 38, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 38, "usage_type": "attribute"}, {"api_name": "jettools.validation.are_valid_args", "line_number": 40, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 40, "usage_type": "name"}, {"api_name": "jettools.validation.Validator", "line_number": 46, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 46, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 46, "usage_type": "attribute"}, {"api_name": "jettools.validation.are_valid_args", "line_number": 49, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 49, "usage_type": "name"}, {"api_name": "jettools.validation.Validator", "line_number": 54, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 54, "usage_type": "name"}, {"api_name": "jettools.validation.Validator", "line_number": 55, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 55, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 55, "usage_type": "attribute"}, {"api_name": "jettools.validation.are_valid_args", "line_number": 57, "usage_type": "call"}, {"api_name": "jettools.validation", "line_number": 57, "usage_type": "name"}]}
+{"seq_id": "17842867359", "text": "import os\nfrom unittest.mock import patch\n\nimport pytest\nimport shutil\nimport unittest\nfrom typing import Optional\n\nimport ray.air\nfrom sklearn.datasets import load_breast_cancer\nfrom sklearn.utils import shuffle\n\nfrom ray import tune\nfrom ray.air.config import RunConfig, ScalingConfig\nfrom ray.air.examples.pytorch.torch_linear_example import (\n train_func as linear_train_func,\n)\nfrom ray.data import Dataset, Datasource, ReadTask, from_pandas, read_datasource\nfrom ray.data.block import BlockMetadata\nfrom ray.train.torch import TorchTrainer\nfrom ray.train.trainer import BaseTrainer\nfrom ray.train.xgboost import XGBoostTrainer\nfrom ray.tune import Callback, TuneError, CLIReporter\nfrom ray.tune.result import DEFAULT_RESULTS_DIR\nfrom ray.tune.tune_config import TuneConfig\nfrom ray.tune.tuner import Tuner\n\n\nclass DummyTrainer(BaseTrainer):\n _scaling_config_allowed_keys = BaseTrainer._scaling_config_allowed_keys + [\n \"num_workers\",\n \"use_gpu\",\n \"resources_per_worker\",\n \"placement_strategy\",\n ]\n\n def training_loop(self) -> None:\n for i in range(5):\n with tune.checkpoint_dir(step=i) as checkpoint_dir:\n path = os.path.join(checkpoint_dir, \"checkpoint\")\n with open(path, \"w\") as f:\n f.write(str(i))\n tune.report(step=i)\n\n\nclass FailingTrainer(DummyTrainer):\n def training_loop(self) -> None:\n raise RuntimeError(\"There is an error in trainer!\")\n\n\nclass TestDatasource(Datasource):\n def __init__(self, do_shuffle: bool):\n self._shuffle = do_shuffle\n\n def prepare_read(self, parallelism: int, **read_args):\n import pyarrow as pa\n\n def load_data():\n data_raw = load_breast_cancer(as_frame=True)\n dataset_df = data_raw[\"data\"]\n dataset_df[\"target\"] = data_raw[\"target\"]\n if self._shuffle:\n dataset_df = shuffle(dataset_df)\n return [pa.Table.from_pandas(dataset_df)]\n\n meta = BlockMetadata(\n num_rows=None,\n size_bytes=None,\n schema=None,\n input_files=None,\n exec_stats=None,\n )\n return [ReadTask(load_data, meta)]\n\n\ndef gen_dataset_func(do_shuffle: Optional[bool] = False) -> Dataset:\n test_datasource = TestDatasource(do_shuffle)\n return read_datasource(test_datasource)\n\n\ndef gen_dataset_func_eager():\n data_raw = load_breast_cancer(as_frame=True)\n dataset_df = data_raw[\"data\"]\n dataset_df[\"target\"] = data_raw[\"target\"]\n dataset = from_pandas(dataset_df)\n return dataset\n\n\nclass TunerTest(unittest.TestCase):\n \"\"\"The e2e test for hparam tuning using Tuner API.\"\"\"\n\n def test_tuner_with_xgboost_trainer(self):\n \"\"\"Test a successful run.\"\"\"\n shutil.rmtree(\n os.path.join(DEFAULT_RESULTS_DIR, \"test_tuner\"), ignore_errors=True\n )\n trainer = XGBoostTrainer(\n label_column=\"target\",\n params={},\n datasets={\"train\": gen_dataset_func_eager()},\n )\n # prep_v1 = StandardScaler([\"worst radius\", \"worst area\"])\n # prep_v2 = StandardScaler([\"worst concavity\", \"worst smoothness\"])\n param_space = {\n \"scaling_config\": ScalingConfig(num_workers=tune.grid_search([1, 2])),\n # \"preprocessor\": tune.grid_search([prep_v1, prep_v2]),\n \"datasets\": {\n \"train\": tune.grid_search(\n [gen_dataset_func(), gen_dataset_func(do_shuffle=True)]\n ),\n },\n \"params\": {\n \"objective\": \"binary:logistic\",\n \"tree_method\": \"approx\",\n \"eval_metric\": [\"logloss\", \"error\"],\n \"eta\": tune.loguniform(1e-4, 1e-1),\n \"subsample\": tune.uniform(0.5, 1.0),\n \"max_depth\": tune.randint(1, 9),\n },\n }\n tuner = Tuner(\n trainable=trainer,\n run_config=RunConfig(name=\"test_tuner\"),\n param_space=param_space,\n tune_config=TuneConfig(mode=\"min\", metric=\"train-error\"),\n # limiting the number of trials running at one time.\n # As the unit test only has access to 4 CPUs on Buildkite.\n _tuner_kwargs={\"max_concurrent_trials\": 1},\n )\n results = tuner.fit()\n assert len(results) == 4\n\n def test_tuner_with_xgboost_trainer_driver_fail_and_resume(self):\n # So that we have some global checkpointing happening.\n os.environ[\"TUNE_GLOBAL_CHECKPOINT_S\"] = \"1\"\n shutil.rmtree(\n os.path.join(DEFAULT_RESULTS_DIR, \"test_tuner_driver_fail\"),\n ignore_errors=True,\n )\n trainer = XGBoostTrainer(\n label_column=\"target\",\n params={},\n datasets={\"train\": gen_dataset_func_eager()},\n )\n # prep_v1 = StandardScaler([\"worst radius\", \"worst area\"])\n # prep_v2 = StandardScaler([\"worst concavity\", \"worst smoothness\"])\n param_space = {\n \"scaling_config\": ScalingConfig(num_workers=tune.grid_search([1, 2])),\n # \"preprocessor\": tune.grid_search([prep_v1, prep_v2]),\n \"datasets\": {\n \"train\": tune.grid_search(\n [gen_dataset_func(), gen_dataset_func(do_shuffle=True)]\n ),\n },\n \"params\": {\n \"objective\": \"binary:logistic\",\n \"tree_method\": \"approx\",\n \"eval_metric\": [\"logloss\", \"error\"],\n \"eta\": tune.loguniform(1e-4, 1e-1),\n \"subsample\": tune.uniform(0.5, 1.0),\n \"max_depth\": tune.randint(1, 9),\n },\n }\n\n class FailureInjectionCallback(Callback):\n \"\"\"Inject failure at the configured iteration number.\"\"\"\n\n def __init__(self, num_iters=10):\n self.num_iters = num_iters\n\n def on_step_end(self, iteration, trials, **kwargs):\n if iteration == self.num_iters:\n print(f\"Failing after {self.num_iters} iters.\")\n raise RuntimeError\n\n tuner = Tuner(\n trainable=trainer,\n run_config=RunConfig(\n name=\"test_tuner_driver_fail\", callbacks=[FailureInjectionCallback()]\n ),\n param_space=param_space,\n tune_config=TuneConfig(mode=\"min\", metric=\"train-error\"),\n # limiting the number of trials running at one time.\n # As the unit test only has access to 4 CPUs on Buildkite.\n _tuner_kwargs={\"max_concurrent_trials\": 1},\n )\n with self.assertRaises(TuneError):\n tuner.fit()\n\n # Test resume\n restore_path = os.path.join(DEFAULT_RESULTS_DIR, \"test_tuner_driver_fail\")\n tuner = Tuner.restore(restore_path)\n # A hack before we figure out RunConfig semantics across resumes.\n tuner._local_tuner._run_config.callbacks = None\n results = tuner.fit()\n assert len(results) == 4\n\n def test_tuner_trainer_fail(self):\n trainer = FailingTrainer()\n param_space = {\n \"scaling_config\": ScalingConfig(num_workers=tune.grid_search([1, 2]))\n }\n tuner = Tuner(\n trainable=trainer,\n run_config=RunConfig(name=\"test_tuner_trainer_fail\"),\n param_space=param_space,\n tune_config=TuneConfig(mode=\"max\", metric=\"iteration\"),\n )\n results = tuner.fit()\n assert len(results) == 2\n for i in range(2):\n assert results[i].error\n\n def test_tuner_with_torch_trainer(self):\n \"\"\"Test a successful run using torch trainer.\"\"\"\n shutil.rmtree(\n os.path.join(DEFAULT_RESULTS_DIR, \"test_tuner_torch\"), ignore_errors=True\n )\n # The following two should be tunable.\n config = {\"lr\": 1e-2, \"hidden_size\": 1, \"batch_size\": 4, \"epochs\": 10}\n scaling_config = ScalingConfig(num_workers=1, use_gpu=False)\n trainer = TorchTrainer(\n train_loop_per_worker=linear_train_func,\n train_loop_config=config,\n scaling_config=scaling_config,\n )\n param_space = {\n \"scaling_config\": ScalingConfig(num_workers=tune.grid_search([1, 2])),\n \"train_loop_config\": {\n \"batch_size\": tune.grid_search([4, 8]),\n \"epochs\": tune.grid_search([5, 10]),\n },\n }\n tuner = Tuner(\n trainable=trainer,\n run_config=RunConfig(name=\"test_tuner\"),\n param_space=param_space,\n tune_config=TuneConfig(mode=\"min\", metric=\"loss\"),\n )\n results = tuner.fit()\n assert len(results) == 8\n\n def test_tuner_run_config_override(self):\n trainer = DummyTrainer(run_config=RunConfig(stop={\"metric\": 4}))\n tuner = Tuner(trainer)\n\n assert tuner._local_tuner._run_config.stop == {\"metric\": 4}\n\n\n@pytest.mark.parametrize(\n \"params_expected\",\n [\n (\n {\"run_config\": RunConfig(progress_reporter=CLIReporter())},\n lambda kw: isinstance(kw[\"progress_reporter\"], CLIReporter),\n ),\n (\n {\"tune_config\": TuneConfig(reuse_actors=True)},\n lambda kw: kw[\"reuse_actors\"] is True,\n ),\n (\n {\"run_config\": RunConfig(log_to_file=\"some_file\")},\n lambda kw: kw[\"log_to_file\"] == \"some_file\",\n ),\n (\n {\"tune_config\": TuneConfig(max_concurrent_trials=3)},\n lambda kw: kw[\"max_concurrent_trials\"] == 3,\n ),\n (\n {\"tune_config\": TuneConfig(time_budget_s=60)},\n lambda kw: kw[\"time_budget_s\"] == 60,\n ),\n ],\n)\ndef test_tuner_api_kwargs(params_expected):\n tuner_params, assertion = params_expected\n\n tuner = Tuner(lambda config: 1, **tuner_params)\n\n caught_kwargs = {}\n\n def catch_kwargs(**kwargs):\n caught_kwargs.update(kwargs)\n\n with patch(\"ray.tune.impl.tuner_internal.run\", catch_kwargs):\n tuner.fit()\n\n assert assertion(caught_kwargs)\n\n\ndef test_tuner_fn_trainable_checkpoint_at_end_true():\n tuner = Tuner(\n lambda config, checkpoint_dir: 1,\n run_config=ray.air.RunConfig(\n checkpoint_config=ray.air.CheckpointConfig(checkpoint_at_end=True)\n ),\n )\n with pytest.raises(TuneError):\n tuner.fit()\n\n\ndef test_tuner_fn_trainable_checkpoint_at_end_false():\n tuner = Tuner(\n lambda config, checkpoint_dir: 1,\n run_config=ray.air.RunConfig(\n checkpoint_config=ray.air.CheckpointConfig(checkpoint_at_end=False)\n ),\n )\n tuner.fit()\n\n\ndef test_tuner_fn_trainable_checkpoint_at_end_none():\n tuner = Tuner(\n lambda config, checkpoint_dir: 1,\n run_config=ray.air.RunConfig(\n checkpoint_config=ray.air.CheckpointConfig(checkpoint_at_end=None)\n ),\n )\n tuner.fit()\n\n\nif __name__ == \"__main__\":\n import sys\n\n sys.exit(pytest.main([\"-v\", __file__] + sys.argv[1:]))\n", "repo_name": "spacegoing/myray_2.0.0", "sub_path": "tune/tests/test_tuner.py", "file_name": "test_tuner.py", "file_ext": "py", "file_size_in_byte": 11050, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "ray.train.trainer.BaseTrainer", "line_number": 29, "usage_type": "name"}, {"api_name": "ray.train.trainer.BaseTrainer._scaling_config_allowed_keys", "line_number": 30, "usage_type": "attribute"}, {"api_name": "ray.train.trainer.BaseTrainer", "line_number": 30, "usage_type": "name"}, {"api_name": "ray.tune.checkpoint_dir", "line_number": 39, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 39, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "ray.tune.report", "line_number": 43, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 43, "usage_type": "name"}, {"api_name": "ray.data.Datasource", "line_number": 51, "usage_type": "name"}, {"api_name": "sklearn.datasets.load_breast_cancer", "line_number": 59, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 63, "usage_type": "call"}, {"api_name": "pyarrow.Table.from_pandas", "line_number": 64, "usage_type": "call"}, {"api_name": "pyarrow.Table", "line_number": 64, "usage_type": "attribute"}, {"api_name": "ray.data.block.BlockMetadata", "line_number": 66, "usage_type": "call"}, {"api_name": "ray.data.ReadTask", "line_number": 73, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 76, "usage_type": "name"}, {"api_name": "ray.data.read_datasource", "line_number": 78, "usage_type": "call"}, {"api_name": "ray.data.Dataset", "line_number": 76, "usage_type": "name"}, {"api_name": "sklearn.datasets.load_breast_cancer", "line_number": 82, "usage_type": "call"}, {"api_name": "ray.data.from_pandas", "line_number": 85, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 89, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 95, "usage_type": "call"}, {"api_name": "ray.tune.result.DEFAULT_RESULTS_DIR", "line_number": 95, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "ray.train.xgboost.XGBoostTrainer", "line_number": 97, "usage_type": "call"}, {"api_name": "ray.air.config.ScalingConfig", "line_number": 105, "usage_type": "call"}, {"api_name": "ray.tune.grid_search", "line_number": 105, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 105, "usage_type": "name"}, {"api_name": "ray.tune.grid_search", "line_number": 108, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 108, "usage_type": "name"}, {"api_name": "ray.tune.loguniform", "line_number": 116, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 116, "usage_type": "name"}, {"api_name": "ray.tune.uniform", "line_number": 117, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 117, "usage_type": "name"}, {"api_name": "ray.tune.randint", "line_number": 118, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 118, "usage_type": "name"}, {"api_name": "ray.tune.tuner.Tuner", "line_number": 121, "usage_type": "call"}, {"api_name": "ray.air.config.RunConfig", "line_number": 123, "usage_type": "call"}, {"api_name": "ray.tune.tune_config.TuneConfig", "line_number": 125, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 135, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 137, "usage_type": "call"}, {"api_name": "ray.tune.result.DEFAULT_RESULTS_DIR", "line_number": 137, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 137, "usage_type": "attribute"}, {"api_name": "ray.train.xgboost.XGBoostTrainer", "line_number": 140, "usage_type": "call"}, {"api_name": "ray.air.config.ScalingConfig", "line_number": 148, "usage_type": "call"}, {"api_name": "ray.tune.grid_search", "line_number": 148, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 148, "usage_type": "name"}, {"api_name": "ray.tune.grid_search", "line_number": 151, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 151, "usage_type": "name"}, {"api_name": "ray.tune.loguniform", "line_number": 159, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 159, "usage_type": "name"}, {"api_name": "ray.tune.uniform", "line_number": 160, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 160, "usage_type": "name"}, {"api_name": "ray.tune.randint", "line_number": 161, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 161, "usage_type": "name"}, {"api_name": "ray.tune.Callback", "line_number": 165, "usage_type": "name"}, {"api_name": "ray.tune.tuner.Tuner", "line_number": 176, "usage_type": "call"}, {"api_name": "ray.air.config.RunConfig", "line_number": 178, "usage_type": "call"}, {"api_name": "ray.tune.tune_config.TuneConfig", "line_number": 182, "usage_type": "call"}, {"api_name": "ray.tune.TuneError", "line_number": 187, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 191, "usage_type": "call"}, {"api_name": "ray.tune.result.DEFAULT_RESULTS_DIR", "line_number": 191, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 191, "usage_type": "attribute"}, {"api_name": "ray.tune.tuner.Tuner.restore", "line_number": 192, "usage_type": "call"}, {"api_name": "ray.tune.tuner.Tuner", "line_number": 192, "usage_type": "name"}, {"api_name": "ray.air.config.ScalingConfig", "line_number": 201, "usage_type": "call"}, {"api_name": "ray.tune.grid_search", "line_number": 201, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 201, "usage_type": "name"}, {"api_name": "ray.tune.tuner.Tuner", "line_number": 203, "usage_type": "call"}, {"api_name": "ray.air.config.RunConfig", "line_number": 205, "usage_type": "call"}, {"api_name": "ray.tune.tune_config.TuneConfig", "line_number": 207, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 216, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 217, "usage_type": "call"}, {"api_name": "ray.tune.result.DEFAULT_RESULTS_DIR", "line_number": 217, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 217, "usage_type": "attribute"}, {"api_name": "ray.air.config.ScalingConfig", "line_number": 221, "usage_type": "call"}, {"api_name": "ray.train.torch.TorchTrainer", "line_number": 222, "usage_type": "call"}, {"api_name": "ray.air.examples.pytorch.torch_linear_example.train_func", "line_number": 223, "usage_type": "name"}, {"api_name": "ray.air.config.ScalingConfig", "line_number": 228, "usage_type": "call"}, {"api_name": "ray.tune.grid_search", "line_number": 228, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 228, "usage_type": "name"}, {"api_name": "ray.tune.grid_search", "line_number": 230, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 230, "usage_type": "name"}, {"api_name": "ray.tune.grid_search", "line_number": 231, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 231, "usage_type": "name"}, {"api_name": "ray.tune.tuner.Tuner", "line_number": 234, "usage_type": "call"}, {"api_name": "ray.air.config.RunConfig", "line_number": 236, "usage_type": "call"}, {"api_name": "ray.tune.tune_config.TuneConfig", "line_number": 238, "usage_type": "call"}, {"api_name": "ray.air.config.RunConfig", "line_number": 244, "usage_type": "call"}, {"api_name": "ray.tune.tuner.Tuner", "line_number": 245, "usage_type": "call"}, {"api_name": "ray.tune.tuner.Tuner", "line_number": 278, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 285, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 250, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 250, "usage_type": "attribute"}, {"api_name": "ray.air.config.RunConfig", "line_number": 254, "usage_type": "call"}, {"api_name": "ray.tune.CLIReporter", "line_number": 254, "usage_type": "call"}, {"api_name": "ray.tune.CLIReporter", "line_number": 255, "usage_type": "argument"}, {"api_name": "ray.tune.tune_config.TuneConfig", "line_number": 258, "usage_type": "call"}, {"api_name": "ray.air.config.RunConfig", "line_number": 262, "usage_type": "call"}, {"api_name": "ray.tune.tune_config.TuneConfig", "line_number": 266, "usage_type": "call"}, {"api_name": "ray.tune.tune_config.TuneConfig", "line_number": 270, "usage_type": "call"}, {"api_name": "ray.tune.tuner.Tuner", "line_number": 292, "usage_type": "call"}, {"api_name": "ray.air.air.RunConfig", "line_number": 294, "usage_type": "call"}, {"api_name": "ray.air.air", "line_number": 294, "usage_type": "attribute"}, {"api_name": "ray.air", "line_number": 294, "usage_type": "name"}, {"api_name": "ray.air.air.CheckpointConfig", "line_number": 295, "usage_type": "call"}, {"api_name": "ray.air.air", "line_number": 295, "usage_type": "attribute"}, {"api_name": "ray.air", "line_number": 295, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 298, "usage_type": "call"}, {"api_name": "ray.tune.TuneError", "line_number": 298, "usage_type": "argument"}, {"api_name": "ray.tune.tuner.Tuner", "line_number": 303, "usage_type": "call"}, {"api_name": "ray.air.air.RunConfig", "line_number": 305, "usage_type": "call"}, {"api_name": "ray.air.air", "line_number": 305, "usage_type": "attribute"}, {"api_name": "ray.air", "line_number": 305, "usage_type": "name"}, {"api_name": "ray.air.air.CheckpointConfig", "line_number": 306, "usage_type": "call"}, {"api_name": "ray.air.air", "line_number": 306, "usage_type": "attribute"}, {"api_name": "ray.air", "line_number": 306, "usage_type": "name"}, {"api_name": "ray.tune.tuner.Tuner", "line_number": 313, "usage_type": "call"}, {"api_name": "ray.air.air.RunConfig", "line_number": 315, "usage_type": "call"}, {"api_name": "ray.air.air", "line_number": 315, "usage_type": "attribute"}, {"api_name": "ray.air", "line_number": 315, "usage_type": "name"}, {"api_name": "ray.air.air.CheckpointConfig", "line_number": 316, "usage_type": "call"}, {"api_name": "ray.air.air", "line_number": 316, "usage_type": "attribute"}, {"api_name": "ray.air", "line_number": 316, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 325, "usage_type": "call"}, {"api_name": "pytest.main", "line_number": 325, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 325, "usage_type": "attribute"}]}
+{"seq_id": "43225683062", "text": "from torch.utils.data import Dataset, DataLoader\nimport pytorch_lightning as pl\nimport torch\nimport math\nimport numpy as np\nimport json\nfrom .chunking_data_util import ChunkingDataset\nimport nltk\n\ngiga_path = './dataset/ggw_data/org_data/'\ngiga_train_src_path = giga_path + 'train.src.txt'\ngiga_train_tgt_path = giga_path + 'train.tgt.txt'\ngiga_dev_src_path = giga_path + 'selected_dev.src.txt'\ngiga_dev_tgt_path = giga_path + 'selected_dev.tgt.txt'\ngiga_test_src_path = giga_path + 'test.src.txt'\ngiga_test_tgt_path = giga_path + 'test.tgt.txt'\n\nconll_train_src_path = \"./dataset/conll_2000/conll_pretrain_src.txt\" \nconll_train_tgt_path = \"./dataset/conll_2000/conll_pretrain_tgt.txt\" # word level\nconll_test_src_path = \"./dataset/conll_2000/test_conll_src.txt\"\nconll_test_tgt_path = \"./dataset/conll_2000/test_conll_tgt.txt\"\nconll_dev_src_path = \"./dataset/conll_2000/dev_conll_src.txt\"\n\nsnli_path = './dataset/snli_1.0/'\nsnli_train_path = snli_path + 'snli_1.0_train.jsonl'\nsnli_dev_path = snli_path + 'snli_1.0_dev.jsonl'\nsnli_test_path = snli_path + 'snli_1.0_test.jsonl'\n\nmnli_path = './dataset/multinli_1.0/'\nmnli_train_path = mnli_path + 'multinli_1.0_train.jsonl'\nmnli_dev_path = mnli_path + 'multinli_1.0_dev_mismatched.jsonl'\nmnli_test_path = mnli_path + 'multinli_1.0_dev_matched.jsonl'\n\nwmt_train_src_path = './translation/WMT14-en-de/train.en'\nwmt_train_tgt_path = './translation/WMT14-en-de/train.de'\nwmt_test_src_path = './translation/WMT14-en-de/newstest2014.en'\nwmt_test_tgt_path = './translation/WMT14-en-de/newstest2014.de'\n\ndef create_wmt_dict(src_path, tgt_path):\n data = []\n with open(src_path) as sf:\n with open(tgt_path) as tf:\n for id, line in enumerate(zip(sf, tf)):\n dict = {}\n src, tgt = line[0], line[1]\n dict['src'] = src.strip().replace(\" ##AT##-##AT## \", \"-\")\n dict['tgt'] = tgt.strip().replace(\" ##AT##-##AT## \", \"-\")\n data.append(dict)\n return data\n\ndef create_giga_dict(src_path, tgt_path, replace_unk=True):\n data = []\n with open(src_path) as sf:\n with open(tgt_path) as tf:\n for id, line in enumerate(zip(sf, tf)):\n dict = {}\n src, tgt = line[0], line[1]\n if replace_unk:\n src = src.strip().replace(\"UNK\", \"\")\n tgt = tgt.strip().replace(\"UNK\", \"\")\n else:\n src = src.strip()\n tgt = tgt.strip()\n dict['src'] = src\n dict['tgt'] = tgt\n data.append(dict)\n return data\n\ndef read_snli_entailment_data(data_path):\n data = []\n with open(data_path) as f:\n lines = f.readlines()\n for line in lines:\n this_data = {}\n tmp = json.loads(line)\n if tmp['gold_label'] == 'entailment':\n this_data['src'] = ' '.join(nltk.word_tokenize(tmp['sentence1'].strip()))\n this_data['tgt'] = ' '.join(nltk.word_tokenize(tmp['sentence2'].strip()))\n data.append(this_data)\n return data\n\nclass SummarizationDataset(Dataset):\n def __init__(self, data, tokenizer, max_src_len, max_tgt_len, cut_rate):\n self.data = data\n self.tokenizer = tokenizer\n self.max_src_len = max_src_len\n self.max_tgt_len = max_tgt_len\n self.cut_rate = cut_rate\n \n def __len__(self):\n return len(self.data)\n \n def __getitem__(self, index: int):\n this_data = self.data[index]\n source = this_data['src']\n target = this_data['tgt']\n encoded_src = self.tokenizer(source, max_length=self.max_src_len, padding=\"max_length\", truncation=True, return_tensors=\"pt\")\n encoded_tgt = self.tokenizer(target, max_length=self.max_tgt_len, padding=\"max_length\", truncation=True, return_tensors=\"pt\")\n\n # cut_rate = np.clip(np.random.normal(0.6, 0.1), 0, 1)\n a = math.ceil((torch.sum(encoded_src[\"attention_mask\"].flatten(), dim=-1) * self.cut_rate).item())\n mes_label = torch.zeros(encoded_src[\"attention_mask\"].flatten().shape)\n mes_label[0:a] = 1\n\n return dict(\n src_text=source, \n tgt_text=target, \n src_input_ids=encoded_src[\"input_ids\"].flatten(),\n src_attention_mask=encoded_src[\"attention_mask\"].flatten(),\n tgt_input_ids=encoded_tgt[\"input_ids\"].flatten(),\n tgt_attention_mask=encoded_tgt[\"attention_mask\"].flatten(),\n mse_label=mes_label,\n )\n\nclass DataModule(pl.LightningDataModule):\n def __init__(self, tokenizer, hparams, max_src_len=128, max_tgt_len=64):\n super().__init__()\n self.tokenizer = tokenizer\n self.batch_size = hparams.batch_size\n self.max_src_len = max_src_len\n self.max_tgt_len = max_tgt_len\n self.train_dataset = hparams.train_dataset\n self.test_dataset = hparams.test_dataset\n self.predict_dataset = hparams.predict_dataset\n self.cut_rate = hparams.cut_rate\n self.predict_mode = None\n\n if self.train_dataset == 'giga':\n self.train_data = create_giga_dict(giga_train_src_path, giga_train_tgt_path)\n elif self.train_dataset == 'mnli':\n self.train_data = read_snli_entailment_data(mnli_train_path)\n else:\n exit('No such train dataset')\n \n if self.test_dataset == 'giga':\n self.test_data = create_giga_dict(giga_test_src_path, giga_test_tgt_path)\n elif self.test_dataset == 'mnli':\n self.test_data = read_snli_entailment_data(mnli_test_path)\n else:\n exit('No such test dataset')\n \n if self.predict_dataset == 'wmt' :\n self.predcit_data = create_wmt_dict(wmt_test_src_path, wmt_test_tgt_path) \n elif self.predict_dataset == 'giga' :\n self.predcit_data = create_giga_dict(giga_test_src_path, giga_test_tgt_path)\n elif self.predict_dataset == 'mnli':\n self.predcit_data= read_snli_entailment_data(mnli_test_path)\n else:\n exit('No such predict dataset')\n \n self.conll_test_data = []\n with open(conll_test_src_path) as f:\n # with open(conll_dev_src_path) as f:\n data = f.readlines()\n for d in data:\n this_data = {}\n this_data['src'] = d.strip()\n self.conll_test_data.append(this_data)\n print('Done!')\n \n def setup(self, stage):\n if stage == \"fit\" or stage is None:\n self.train_dataset = SummarizationDataset(self.train_data, self.tokenizer, self.max_src_len, self.max_tgt_len, self.cut_rate)\n # self.dev_dataset = SummarizationDataset(self.dev_data, self.tokenizer, self.max_src_len, self.max_tgt_len, self.cut_rate)\n if stage == \"test\" or stage is None:\n self.test_dataset = SummarizationDataset(self.test_data, self.tokenizer, self.max_src_len, self.max_tgt_len, self.cut_rate)\n if stage == \"predict\" or stage is None:\n self.predict_dataset = ChunkingDataset(self.predcit_data, self.tokenizer, self.max_src_len)\n self.predict_conll_dataset = ChunkingDataset(self.conll_test_data, self.tokenizer, self.max_src_len)\n\n def train_dataloader(self):\n # torch.manual_seed(42)\n torch.manual_seed(59)\n return DataLoader(self.train_dataset, batch_size=self.batch_size, shuffle=True, num_workers=6)\n \n # def val_dataloader(self):\n # return DataLoader(self.dev_dataset, batch_size=self.batch_size)\n\n def test_dataloader(self):\n return DataLoader(self.test_dataset, batch_size=self.batch_size)\n \n def predict_dataloader(self):\n if self.predict_mode == 'test':\n return DataLoader(self.predict_dataset, batch_size=1)\n elif self.predict_mode == 'conll':\n return DataLoader(self.predict_conll_dataset, batch_size=1)\n # elif self.predict_mode == 'dev':\n # return DataLoader(self.predict_val_dataset, batch_size=1)\n else:\n print('predict: No such option!')\n", "repo_name": "MANGA-UOFA/UCHRNN", "sub_path": "data_util/generation_data_util.py", "file_name": "generation_data_util.py", "file_ext": "py", "file_size_in_byte": 8182, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "json.loads", "line_number": 75, "usage_type": "call"}, {"api_name": "nltk.word_tokenize", "line_number": 77, "usage_type": "call"}, {"api_name": "nltk.word_tokenize", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 82, "usage_type": "name"}, {"api_name": "math.ceil", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 102, "usage_type": "call"}, {"api_name": "pytorch_lightning.LightningDataModule", "line_number": 115, "usage_type": "attribute"}, {"api_name": "chunking_data_util.ChunkingDataset", "line_number": 168, "usage_type": "call"}, {"api_name": "chunking_data_util.ChunkingDataset", "line_number": 169, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 174, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 180, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 184, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 186, "usage_type": "call"}]}
+{"seq_id": "24565120784", "text": "from datetime import datetime\nimport webob\nimport webob.dec\nfrom xml.dom import minidom\n\nimport nova.api.openstack.views.versions\nfrom nova.api.openstack import wsgi\n\n\nVERSIONS = {\n \"v1.0\": {\n \"id\": \"v1.0\",\n \"status\": \"DEPRECATED\",\n \"updated\": \"2011-01-21T11:33:21Z\",\n \"links\": [\n {\n \"rel\": \"describedby\",\n \"type\": \"application/pdf\",\n \"href\": \"http://docs.rackspacecloud.com/\"\n \"servers/api/v1.0/cs-devguide-20110125.pdf\",\n },\n {\n \"rel\": \"describedby\",\n \"type\": \"application/vnd.sun.wadl+xml\",\n \"href\": \"http://docs.rackspacecloud.com/\"\n \"servers/api/v1.0/application.wadl\",\n },\n ],\n \"media-types\": [\n {\n \"base\": \"application/xml\",\n \"type\": \"application/vnd.openstack.compute-v1.0+xml\",\n },\n {\n \"base\": \"application/json\",\n \"type\": \"application/vnd.openstack.compute-v1.0+json\",\n }\n ],\n },\n \"v1.1\": {\n \"id\": \"v1.1\",\n \"status\": \"CURRENT\",\n \"updated\": \"2011-01-21T11:33:21Z\",\n \"links\": [\n {\n \"rel\": \"describedby\",\n \"type\": \"application/pdf\",\n \"href\": \"http://docs.rackspacecloud.com/\"\n \"servers/api/v1.1/cs-devguide-20110125.pdf\",\n },\n {\n \"rel\": \"describedby\",\n \"type\": \"application/vnd.sun.wadl+xml\",\n \"href\": \"http://docs.rackspacecloud.com/\"\n \"servers/api/v1.1/application.wadl\",\n },\n ],\n \"media-types\": [\n {\n \"base\": \"application/xml\",\n \"type\": \"application/vnd.openstack.compute-v1.1+xml\",\n },\n {\n \"base\": \"application/json\",\n \"type\": \"application/vnd.openstack.compute-v1.1+json\",\n }\n ],\n },\n}\n\n\nclass Versions(wsgi.Resource):\n def __init__(self):\n metadata = {\n \"attributes\": {\n \"version\": [\"status\", \"id\"],\n \"link\": [\"rel\", \"href\"],\n }\n }\n\n headers_serializer = VersionsHeadersSerializer()\n\n body_serializers = {\n 'application/atom+xml': VersionsAtomSerializer(metadata=metadata),\n 'application/xml': VersionsXMLSerializer(metadata=metadata),\n }\n serializer = wsgi.ResponseSerializer(\n body_serializers=body_serializers,\n headers_serializer=headers_serializer)\n\n supported_content_types = ('application/json',\n 'application/xml',\n 'application/atom+xml')\n deserializer = VersionsRequestDeserializer(\n supported_content_types=supported_content_types)\n\n wsgi.Resource.__init__(self, None, serializer=serializer,\n deserializer=deserializer)\n\n def dispatch(self, request, *args):\n \"\"\"Respond to a request for all OpenStack API versions.\"\"\"\n builder = nova.api.openstack.views.versions.get_view_builder(request)\n if request.path == '/':\n # List Versions\n return builder.build_versions(VERSIONS)\n else:\n # Versions Multiple Choice\n return builder.build_choices(VERSIONS, request)\n\n\nclass VersionV10(object):\n def show(self, req):\n builder = nova.api.openstack.views.versions.get_view_builder(req)\n return builder.build_version(VERSIONS['v1.0'])\n\n\nclass VersionV11(object):\n def show(self, req):\n builder = nova.api.openstack.views.versions.get_view_builder(req)\n return builder.build_version(VERSIONS['v1.1'])\n\n\nclass VersionsRequestDeserializer(wsgi.RequestDeserializer):\n def get_expected_content_type(self, request):\n supported_content_types = list(self.supported_content_types)\n if request.path != '/':\n # Remove atom+xml accept type for 300 responses\n if 'application/atom+xml' in supported_content_types:\n supported_content_types.remove('application/atom+xml')\n\n return request.best_match_content_type(supported_content_types)\n\n def get_action_args(self, request_environment):\n \"\"\"Parse dictionary created by routes library.\"\"\"\n args = {}\n if request_environment['PATH_INFO'] == '/':\n args['action'] = 'index'\n else:\n args['action'] = 'multi'\n\n return args\n\n\nclass VersionsXMLSerializer(wsgi.XMLDictSerializer):\n #TODO(wwolf): this is temporary until we get rid of toprettyxml\n # in the base class (XMLDictSerializer), which I plan to do in\n # another branch\n def to_xml_string(self, node, has_atom=False):\n self._add_xmlns(node, has_atom)\n return node.toxml(encoding='UTF-8')\n\n def _versions_to_xml(self, versions, name=\"versions\", xmlns=None):\n root = self._xml_doc.createElement(name)\n root.setAttribute(\"xmlns\", wsgi.XMLNS_V11)\n root.setAttribute(\"xmlns:atom\", wsgi.XMLNS_ATOM)\n\n for version in versions:\n root.appendChild(self._create_version_node(version))\n\n return root\n\n def _create_media_types(self, media_types):\n base = self._xml_doc.createElement('media-types')\n for type in media_types:\n node = self._xml_doc.createElement('media-type')\n node.setAttribute('base', type['base'])\n node.setAttribute('type', type['type'])\n base.appendChild(node)\n\n return base\n\n def _create_version_node(self, version, create_ns=False):\n version_node = self._xml_doc.createElement('version')\n if create_ns:\n xmlns = wsgi.XMLNS_V11\n xmlns_atom = wsgi.XMLNS_ATOM\n version_node.setAttribute('xmlns', xmlns)\n version_node.setAttribute('xmlns:atom', xmlns_atom)\n\n version_node.setAttribute('id', version['id'])\n version_node.setAttribute('status', version['status'])\n if 'updated' in version:\n version_node.setAttribute('updated', version['updated'])\n\n if 'media-types' in version:\n media_types = self._create_media_types(version['media-types'])\n version_node.appendChild(media_types)\n\n link_nodes = self._create_link_nodes(self._xml_doc, version['links'])\n for link in link_nodes:\n version_node.appendChild(link)\n\n return version_node\n\n def index(self, data):\n self._xml_doc = minidom.Document()\n node = self._versions_to_xml(data['versions'])\n\n return self.to_xml_string(node)\n\n def show(self, data):\n self._xml_doc = minidom.Document()\n node = self._create_version_node(data['version'], True)\n\n return self.to_xml_string(node)\n\n def multi(self, data):\n self._xml_doc = minidom.Document()\n node = self._versions_to_xml(data['choices'], 'choices',\n xmlns=wsgi.XMLNS_V11)\n\n return self.to_xml_string(node)\n\n\nclass VersionsAtomSerializer(wsgi.XMLDictSerializer):\n #TODO(wwolf): this is temporary until we get rid of toprettyxml\n # in the base class (XMLDictSerializer), which I plan to do in\n # another branch\n def to_xml_string(self, node, has_atom=False):\n self._add_xmlns(node, has_atom)\n return node.toxml(encoding='UTF-8')\n\n def __init__(self, metadata=None, xmlns=None):\n self.metadata = metadata or {}\n if not xmlns:\n self.xmlns = wsgi.XMLNS_ATOM\n else:\n self.xmlns = xmlns\n\n def _create_text_elem(self, name, text, type=None):\n elem = self._xml_doc.createElement(name)\n if type:\n elem.setAttribute('type', type)\n elem_text = self._xml_doc.createTextNode(text)\n elem.appendChild(elem_text)\n return elem\n\n def _get_most_recent_update(self, versions):\n recent = None\n for version in versions:\n updated = datetime.strptime(version['updated'],\n '%Y-%m-%dT%H:%M:%SZ')\n if not recent:\n recent = updated\n elif updated > recent:\n recent = updated\n\n return recent.strftime('%Y-%m-%dT%H:%M:%SZ')\n\n def _get_base_url(self, link_href):\n # Make sure no trailing /\n link_href = link_href.rstrip('/')\n return link_href.rsplit('/', 1)[0] + '/'\n\n def _create_detail_meta(self, root, version):\n title = self._create_text_elem('title', \"About This Version\",\n type='text')\n\n updated = self._create_text_elem('updated', version['updated'])\n\n uri = version['links'][0]['href']\n id = self._create_text_elem('id', uri)\n\n link = self._xml_doc.createElement('link')\n link.setAttribute('rel', 'self')\n link.setAttribute('href', uri)\n\n author = self._xml_doc.createElement('author')\n author_name = self._create_text_elem('name', 'Rackspace')\n author_uri = self._create_text_elem('uri', 'http://www.rackspace.com/')\n author.appendChild(author_name)\n author.appendChild(author_uri)\n\n root.appendChild(title)\n root.appendChild(updated)\n root.appendChild(id)\n root.appendChild(author)\n root.appendChild(link)\n\n def _create_list_meta(self, root, versions):\n title = self._create_text_elem('title', \"Available API Versions\",\n type='text')\n # Set this updated to the most recently updated version\n recent = self._get_most_recent_update(versions)\n updated = self._create_text_elem('updated', recent)\n\n base_url = self._get_base_url(versions[0]['links'][0]['href'])\n id = self._create_text_elem('id', base_url)\n\n link = self._xml_doc.createElement('link')\n link.setAttribute('rel', 'self')\n link.setAttribute('href', base_url)\n\n author = self._xml_doc.createElement('author')\n author_name = self._create_text_elem('name', 'Rackspace')\n author_uri = self._create_text_elem('uri', 'http://www.rackspace.com/')\n author.appendChild(author_name)\n author.appendChild(author_uri)\n\n root.appendChild(title)\n root.appendChild(updated)\n root.appendChild(id)\n root.appendChild(author)\n root.appendChild(link)\n\n def _create_version_entries(self, root, versions):\n for version in versions:\n entry = self._xml_doc.createElement('entry')\n\n id = self._create_text_elem('id', version['links'][0]['href'])\n title = self._create_text_elem('title',\n 'Version %s' % version['id'],\n type='text')\n updated = self._create_text_elem('updated', version['updated'])\n\n entry.appendChild(id)\n entry.appendChild(title)\n entry.appendChild(updated)\n\n for link in version['links']:\n link_node = self._xml_doc.createElement('link')\n link_node.setAttribute('rel', link['rel'])\n link_node.setAttribute('href', link['href'])\n if 'type' in link:\n link_node.setAttribute('type', link['type'])\n\n entry.appendChild(link_node)\n\n content = self._create_text_elem('content',\n 'Version %s %s (%s)' %\n (version['id'],\n version['status'],\n version['updated']),\n type='text')\n\n entry.appendChild(content)\n root.appendChild(entry)\n\n def index(self, data):\n self._xml_doc = minidom.Document()\n node = self._xml_doc.createElementNS(self.xmlns, 'feed')\n self._create_list_meta(node, data['versions'])\n self._create_version_entries(node, data['versions'])\n\n return self.to_xml_string(node)\n\n def show(self, data):\n self._xml_doc = minidom.Document()\n node = self._xml_doc.createElementNS(self.xmlns, 'feed')\n self._create_detail_meta(node, data['version'])\n self._create_version_entries(node, [data['version']])\n\n return self.to_xml_string(node)\n\n\nclass VersionsHeadersSerializer(wsgi.ResponseHeadersSerializer):\n def multi(self, response, data):\n response.status_int = 300\n\n\ndef create_resource(version='1.0'):\n controller = {\n '1.0': VersionV10,\n '1.1': VersionV11,\n }[version]()\n\n body_serializers = {\n 'application/xml': VersionsXMLSerializer(),\n 'application/atom+xml': VersionsAtomSerializer(),\n }\n serializer = wsgi.ResponseSerializer(body_serializers)\n\n supported_content_types = ('application/json',\n 'application/xml',\n 'application/atom+xml')\n deserializer = wsgi.RequestDeserializer(\n supported_content_types=supported_content_types)\n\n return wsgi.Resource(controller, serializer=serializer,\n deserializer=deserializer)\n", "repo_name": "nii-cloud/dodai-compute", "sub_path": "nova/api/openstack/versions.py", "file_name": "versions.py", "file_ext": "py", "file_size_in_byte": 13246, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 12, "dataset": "github-code", "pt": "24", "api": [{"api_name": "nova.api.openstack.wsgi.Resource", "line_number": 72, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 72, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.ResponseSerializer", "line_number": 87, "usage_type": "call"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 87, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.Resource.__init__", "line_number": 97, "usage_type": "call"}, {"api_name": "nova.api.openstack.wsgi.Resource", "line_number": 97, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 97, "usage_type": "name"}, {"api_name": "nova.api.openstack.views.versions.api.openstack.views.versions.get_view_builder", "line_number": 102, "usage_type": "call"}, {"api_name": "nova.api.openstack.views.versions.api", "line_number": 102, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.views.versions", "line_number": 102, "usage_type": "name"}, {"api_name": "nova.api.openstack.views.versions.api.openstack.views.versions.get_view_builder", "line_number": 113, "usage_type": "call"}, {"api_name": "nova.api.openstack.views.versions.api", "line_number": 113, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.views.versions", "line_number": 113, "usage_type": "name"}, {"api_name": "nova.api.openstack.views.versions.api.openstack.views.versions.get_view_builder", "line_number": 119, "usage_type": "call"}, {"api_name": "nova.api.openstack.views.versions.api", "line_number": 119, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.views.versions", "line_number": 119, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.RequestDeserializer", "line_number": 123, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 123, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.XMLDictSerializer", "line_number": 144, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 144, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.XMLNS_V11", "line_number": 154, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 154, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.XMLNS_ATOM", "line_number": 155, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 155, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.XMLNS_V11", "line_number": 175, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 175, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.XMLNS_ATOM", "line_number": 176, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 176, "usage_type": "name"}, {"api_name": "xml.dom.minidom.Document", "line_number": 196, "usage_type": "call"}, {"api_name": "xml.dom.minidom", "line_number": 196, "usage_type": "name"}, {"api_name": "xml.dom.minidom.Document", "line_number": 202, "usage_type": "call"}, {"api_name": "xml.dom.minidom", "line_number": 202, "usage_type": "name"}, {"api_name": "xml.dom.minidom.Document", "line_number": 208, "usage_type": "call"}, {"api_name": "xml.dom.minidom", "line_number": 208, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.XMLNS_V11", "line_number": 210, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 210, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.XMLDictSerializer", "line_number": 215, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 215, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.XMLNS_ATOM", "line_number": 226, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 226, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 241, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 241, "usage_type": "name"}, {"api_name": "xml.dom.minidom.Document", "line_number": 340, "usage_type": "call"}, {"api_name": "xml.dom.minidom", "line_number": 340, "usage_type": "name"}, {"api_name": "xml.dom.minidom.Document", "line_number": 348, "usage_type": "call"}, {"api_name": "xml.dom.minidom", "line_number": 348, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.ResponseHeadersSerializer", "line_number": 356, "usage_type": "attribute"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 356, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.ResponseSerializer", "line_number": 371, "usage_type": "call"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 371, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.RequestDeserializer", "line_number": 376, "usage_type": "call"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 376, "usage_type": "name"}, {"api_name": "nova.api.openstack.wsgi.Resource", "line_number": 379, "usage_type": "call"}, {"api_name": "nova.api.openstack.wsgi", "line_number": 379, "usage_type": "name"}]}
+{"seq_id": "13159953774", "text": "from flask import Flask, render_template, request\nimport hashlib\n\napp = Flask(__name__)\n\ndef get_gravatar_url(email):\n # Convert the email address to lowercase and encode it\n email_bytes = email.lower().encode('utf-8')\n # Generate the MD5 hash of the email\n digest = hashlib.md5(email_bytes).hexdigest()\n # Return the Gravatar URL\n return f'https://www.gravatar.com/avatar/{digest}'\n\n@app.route('/', methods=['GET', 'POST'])\ndef index():\n gravatar_url = None\n if request.method == 'POST':\n email = request.form.get('email')\n if email:\n gravatar_url = get_gravatar_url(email)\n return render_template('index.html', gravatar_url=gravatar_url)\n\nif __name__ == '__main__':\n app.run(debug=True)\n", "repo_name": "sivori/chatgpt-projects", "sub_path": "gravatar-retrieval/gravatar-retrieve.py", "file_name": "gravatar-retrieve.py", "file_ext": "py", "file_size_in_byte": 744, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "flask.Flask", "line_number": 4, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 17, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 18, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 18, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 21, "usage_type": "call"}]}
+{"seq_id": "36174619998", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri May 26 22:01:16 2017\n\n@author: Jonas Stein\n\"\"\"\n\nimport datetime\n\ndef CreateLogfile():\n MyLogTime = datetime.datetime.now()\n MyFileName = MyLogTime.strftime(\"protokoll_%Y-%m-%dT%H_%M_%S.csv\")\n f = open(MyFileName, 'w')\n f.write('\"Nr.\",\"Datum Uhrzeit\",\"Dauer (s)\",\"Geschwindigkeit (km/h)\"\\n')\n f.close\n return(MyFileName)\n\ndef LogThis(MyFileName, Text):\n f = open(MyFileName, 'a')\n f.write(Text + '\\n')\n f.close\n\n\ndef measure(Distance_meter, FzID, MyFileName):\n Tstart = datetime.datetime.now()\n input(\"# (Messung läuft...) Stopp mit Return\")\n Tstop = datetime.datetime.now()\n \n TdifferenceSeconds = (Tstop-Tstart).total_seconds()\n \n Speed_m_s = Distance_meter / TdifferenceSeconds\n Speed_km_h = Speed_m_s /1000 * 3600\n StrTimeStamp = Tstart.strftime(\"%Y-%m-%d %H:%M:%S\")\n \n print('%03d, %s, %.6f, %.1f'%(FzID,StrTimeStamp,TdifferenceSeconds,Speed_km_h))\n LogThis(MyFileName,'%03d,%s,%.6f,%.1f'%(FzID,StrTimeStamp,TdifferenceSeconds,Speed_km_h))\n\ndef main():\n FzID = 0\n\n MyLogFile = CreateLogfile()\n\n Distance_meter = 86.5\n print('Geschwindigkeitsmessung.\\n Distanz zwischen Start- und Stoppsignal: %f (m)\\n\\n'%Distance_meter)\n print('Nr., Datum Uhrzeit, Dauer (s), Geschwindigkeit (km/h)')\n while True:\n KeyByUser = input(\"# (Standby) Start mit Return, Quit mit q\")\n if KeyByUser != '':\n print('Ende der Messserie.\\n')\n print('Protokoll öffnen mit')\n print('libreoffice %s'%(MyLogFile))\n break\n elif KeyByUser == '':\n FzID = FzID + 1\n measure(Distance_meter, FzID, MyLogFile)\n \n\nif __name__ == '__main__':\n main()\n", "repo_name": "jonasstein/speed", "sub_path": "speed.py", "file_name": "speed.py", "file_ext": "py", "file_size_in_byte": 1831, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "datetime.datetime.now", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 12, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 26, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "attribute"}]}
+{"seq_id": "42568766138", "text": "from django.contrib import admin\nimport api.models\n\n\n@admin.register(api.models.Reading)\nclass MeterReadingAdmin(admin.ModelAdmin):\n list_display = (\n \"mpan_core\",\n \"register_id\",\n \"meter_id\",\n \"reading_value\",\n \"reading_taken_at\",\n \"reading_flag\",\n \"reading_method\",\n \"file_name\",\n )\n search_fields = (\n \"mpan_core__mpan_core__startswith\",\n \"meter_id__id__startswith\",\n )\n", "repo_name": "ryjinh/meter_reading", "sub_path": "api/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 455, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 5, "usage_type": "call"}, {"api_name": "django.contrib.admin", "line_number": 5, "usage_type": "name"}, {"api_name": "api.models.models", "line_number": 5, "usage_type": "attribute"}, {"api_name": "api.models", "line_number": 5, "usage_type": "name"}]}
+{"seq_id": "34550182891", "text": "import sys\n\nfrom PyQt6.QtCore import (\n QSize, Qt,\n QSortFilterProxyModel,\n\n)\nfrom PyQt6.QtWidgets import (\n QApplication, QMainWindow, \n QWidget, QTabWidget, \n QTableWidget, QTableWidgetItem,\n QPushButton, QLineEdit, QCheckBox, QLabel,\n QGridLayout, QVBoxLayout, QHBoxLayout,\n QFileDialog, QHeaderView, \n QInputDialog, QFormLayout, QDialog, QMessageBox, QErrorMessage,\n \n)\n\nfrom PyQt6.QtGui import (\n QAction, QIcon,\n)\n\n# from pyroGamer.DataManager.Configs import (\n# LocalConfig, CloudConfig,\n# )\n\nimport subprocess\nimport multiprocessing\nfrom pathlib import Path\nimport json\nimport argparse\nimport pprint\nimport textwrap\nfrom colorama import Fore, Back, Style, init\ninit(autoreset=True)\n\n\nfrom pyroGamer.GUI.Hub.Elements.Tables import ProjectTable\n\nclass LocalTab(QWidget):\n def __init__(self):\n super().__init__()\n\n self.layout = QVBoxLayout()\n\n buttonsLayout = QHBoxLayout()\n buttonsLayout.addWidget(QLineEdit())\n openProjectButton = QPushButton(\"Open\")\n buttonsLayout.addWidget(openProjectButton)\n openProjectButton.clicked.connect(self.open_project)\n newProjectButton = QPushButton(\"New\")\n buttonsLayout.addWidget(newProjectButton)\n newProjectButton.clicked.connect(self.new_project)\n self.layout.addLayout(buttonsLayout)\n\n self.layout.addWidget(ProjectTable.Local())\n \n self.setLayout(self.layout)\n\n\n \n\n def open_project(self):\n fname = QFileDialog.getOpenFileName(self, 'Open file', '' ,\"Json files (*.json)\")\n if fname[0] == \"\":\n return\n path = Path(fname[0])\n\n # LocalConfig.AddProject(path)\n print(Fore.BLUE + \"Calling \" + \"pyroGamer.GUI.Hub.Configs --AddExistingProject\" + \" to open project...\")\n result = subprocess.run(['python', '-m', 'pyroGamer.GUI.Hub.Configs',\n '--AddExistingProject', str(path.as_posix())],\n capture_output=True, text=True)\n \n if result.returncode != 0:\n print(Fore.LIGHTBLACK_EX + textwrap.indent(pprint.pformat(result.stdout, width=100), '>'))\n print(Fore.RED + Style.BRIGHT + \"Error: \" + str(result.stderr))\n sys.exit(1)\n\n self.RefreshTable()\n\n\n def new_project(self):\n inputDialog = QDialog()\n inputDialog.setWindowTitle(\"Create new project\")\n inputDialog.setFixedSize(200, 200)\n mainLayout = QVBoxLayout()\n \n nameLayout = QHBoxLayout()\n # label = QLabel(\"Project Name:\")\n # nameLayout.addWidget(label)\n NameInput = QLineEdit()\n NameInput.setPlaceholderText(\"Project Name\")\n nameLayout.addWidget(NameInput)\n mainLayout.addLayout(nameLayout)\n\n pathLayout = QHBoxLayout()\n # label = QLabel(\"Parent Folder Path:\")\n # label.setAlignment(Qt.AlignmentFlag.AlignCenter)\n # pathLayout.addWidget(label)\n PathInput = QLineEdit()\n PathInput.setPlaceholderText(\"Parent Folder Path\")\n PathInput.setReadOnly(True)\n pathLayout.addWidget(PathInput)\n\n def browse():\n return PathInput.setText(QFileDialog.getExistingDirectory(self, \"Select Directory\"))\n \n browseAction = QAction(\"Browse\")\n browseAction.setIcon(QIcon(\"pyroGamer/Hub/icons/blue-document-search-result.png\"))\n browseAction.triggered.connect(browse)\n PathInput.addAction(browseAction, QLineEdit.ActionPosition.TrailingPosition)\n\n mainLayout.addLayout(pathLayout)\n\n buttonsLayout = QVBoxLayout()\n CreateButton = QPushButton(\"Create\")\n buttonsLayout.addWidget(CreateButton)\n def fields_filled():\n return NameInput.text() != \"\" and PathInput.text() != \"\"\n \n def createProject():\n if not fields_filled():\n QMessageBox.critical(None, \"Error\", \"Please fill in all fields\")\n return\n \n print(Fore.BLUE + \"Calling \" + \"pyroGamer.GUI.Hub.Configs AddNewProject --name --path\" + \" to add new project...\")\n result = subprocess.run(['python', '-m', 'pyroGamer.GUI.Hub.Configs',\n 'AddNewProject', '--name', NameInput.text(), '--path', Path(PathInput.text()).as_posix()],\n capture_output=True, text=True)\n \n if(result.returncode != 0):\n print(Fore.LIGHTBLACK_EX + textwrap.indent(pprint.pformat(result.stdout, width=100), '>'))\n print(Fore.RED + Style.BRIGHT + \"Error: \" + str(result.stderr))\n sys.exit(1)\n \n self.RefreshTable()\n inputDialog.close()\n \n CreateButton.clicked.connect(createProject)\n\n CancelButton = QPushButton(\"Cancel\")\n buttonsLayout.addWidget(CancelButton)\n CancelButton.clicked.connect(lambda: inputDialog.close())\n mainLayout.addLayout(buttonsLayout)\n\n inputDialog.setLayout(mainLayout)\n\n inputDialog.exec()\n\n def RefreshTable(self):\n for i in range(self.layout.count()):\n widget = self.layout.itemAt(i).widget()\n if widget is not None:\n widget.deleteLater()\n self.layout.removeWidget(widget)\n\n self.layout.addWidget(ProjectTable.Local())\n self.setLayout(self.layout)\n \n\n\n\n# class Local(QWidget):\n# def __init__(self, ProjectListPath):\n# super().__init__()\n# layout = QVBoxLayout()\n# buttonsLayout = QHBoxLayout()\n# buttonsLayout.addWidget(QLineEdit())\n \n# openProjectButton = QPushButton(\"Open\")\n# buttonsLayout.addWidget(openProjectButton)\n# openProjectButton.clicked.connect(self.open_project)\n\n# newProjectButton = QPushButton(\"New\")\n# buttonsLayout.addWidget(newProjectButton)\n# newProjectButton.clicked.connect(self.new_project)\n# layout.addLayout(buttonsLayout)\n\n# layout.addWidget(Projects_Table.Local(ProjectListPath))\n\n# self.setLayout(layout)\n\n# def open_project(self):\n# fname = QFileDialog.getOpenFileName(self, 'Open file', '' ,\"Json files (*.json)\")\n# if fname[0] == \"\":\n# return\n# path = Path(fname[0])\n\n# LocalConfig.AddProject(path)\n# Local.RefreshTable()\n\n# def new_project(self): \n# inputDialog = QDialog()\n\n# mainLayout = QVBoxLayout()\n# mainLayout.addWidget(QLabel(\"Create new project\"))\n \n# nameLayout = QHBoxLayout()\n# nameLayout.addWidget(QLabel(\"Project Name:\"))\n# NameInput = QLineEdit()\n# nameLayout.addWidget(NameInput)\n# mainLayout.addLayout(nameLayout)\n\n# pathLayout = QHBoxLayout()\n# pathLayout.addWidget(QLabel(\"Parent Folder Path:\"))\n# PathInput = QLineEdit()\n# PathInput.setReadOnly(True)\n# pathLayout.addWidget(PathInput)\n# BrowsePathButton = QPushButton(\"Browse\")\n# pathLayout.addWidget(BrowsePathButton)\n# def browse():\n# return PathInput.setText(QFileDialog.getExistingDirectory(self, \"Select Directory\"))\n# BrowsePathButton.clicked.connect(browse)\n# mainLayout.addLayout(pathLayout)\n\n# buttonsLayout = QHBoxLayout()\n# CreateButton = QPushButton(\"Create\")\n# buttonsLayout.addWidget(CreateButton)\n# def fields_filled():\n# return NameInput.text() != \"\" and PathInput.text() != \"\"\n# def createProject():\n# if not fields_filled():\n# QMessageBox.critical(None, \"Error\", \"Please fill in all fields\")\n# return\n# LocalConfig.AddEmptyProject(name = NameInput.text(), path = Path(PathInput.text())) \n# Local.RefreshTable() \n# inputDialog.close()\n \n# CreateButton.clicked.connect(createProject)\n\n# CancelButton = QPushButton(\"Cancel\")\n# buttonsLayout.addWidget(CancelButton)\n# CancelButton.clicked.connect(lambda: inputDialog.close())\n# mainLayout.addLayout(buttonsLayout)\n\n# inputDialog.setLayout(mainLayout)\n\n# inputDialog.exec()\n\n \n# def RefreshTable():\n# Local.clearTable()\n# Projects_Table.populateTable(Projects_Table.table, LocalConfig.GetProjectList())\n\n# def clearTable():\n# Projects_Table.table.clearContents()\n# Projects_Table.table.setRowCount(0)\n\n\n# class Cloud(QWidget):\n# def __init__(self):\n# super().__init__()\n# layout = QGridLayout()\n\n# layout.addWidget(Projects_Table.Cloud())\n\n# self.setLayout(layout)\n\n# class Projects_Table(QTableWidget):\n# LocalTable = QTableWidget()\n\n# def Cloud():\n# return Projects_Table.From(CloudConfig)\n\n \n# def Local(ProjectListPath):\n# Projects_Table.initTable(Projects_Table.LocalTable)\n\n# headers = []\n# for col in range(Projects_Table.table.columnCount()):\n# header_item = Projects_Table.table.horizontalHeaderItem(col)\n# if header_item is not None:\n# headers.append(header_item.text())\n\n\n# def item_changed(item):\n# project_list = configClass.GetProjectList()\n\n# if (item.column() == headers.index(\"Name\")):\n# configClass.SetName(project_list[item.row()][\"ID\"], item.text())\n# # elif (item.column() == headers.index(\"Path\")):\n# # configClass.SetPath(project_list[item.row()][\"ID\"], item.text())\n# # elif (item.column() == headers.index(\"Created\")):\n# # configClass.SetCreated(project_list[item.row()][\"ID\"], item.text())\n# else:\n# pass\n\n# Projects_Table.table.itemChanged.connect(item_changed)\n\n# Projects_Table.populateTable(Projects_Table.table, configClass.GetProjectList())\n\n# return Projects_Table.table\n \n\n\n\n# def populateTable(table, project_list):\n# if(table is None):\n# table = Projects_Table.table\n# table.setRowCount(len(project_list))\n\n# for row, project in enumerate(project_list):\n# table.setCellWidget(row, 0, Buttons.Star(project))\n\n# project_name = QTableWidgetItem(project[\"Project Name\"])\n# project_name.setFlags(Qt.ItemFlag.ItemIsEditable | Qt.ItemFlag.ItemIsEnabled)\n# table.setItem(row, 1, project_name)\n\n# project_path = QTableWidgetItem(project[\"Project Path\"])\n# project_path.setFlags(Qt.ItemFlag.ItemIsEnabled)\n# table.setItem(row, 2, project_path)\n\n# created_date = QTableWidgetItem(project[\"Created\"])\n# created_date.setFlags(Qt.ItemFlag.ItemIsEnabled)\n# table.setItem(row, 3, created_date)\n\n# table.setCellWidget(row, 4, Buttons.Play(project))\n# table.setCellWidget(row, 5, Buttons.Options(project))\n\n# return table\n \n# def initTable(table):\n# # headers = tableConfig['TABLE_HEADERS']\n# # table.setColumnCount(len(headers))\n# # table.setHorizontalHeaderLabels(headers)\n\n# table.setSortingEnabled(True)\n# table.setStyleSheet(\n# \"QTableWidget::item:selected {\"\n# \" background-color: transparent;\"\n# \" color: black;\"\n# \"}\"\n \n# \"QTableWidget {\"\n# \" gridline-color: transparent;\"\n# \" border: none;\" # Remove the border around the table\n# \"}\"\n# \"QHeaderView::section {\"\n# \" background-color: lightgray;\"\n# \" border: none;\" # Remove the header borders\n# \" padding: 4px;\" # Add padding to headers\n# \"}\"\n# ) \n# # table_unit = tableConfig['TABLE_UNIT']\n# table.setColumnWidth(0, 25)\n# table.setColumnWidth(1, 25 * 5)\n# table.setColumnWidth(2, 25 * 7)\n# table.setColumnWidth(3, 25 * 3)\n# table.setColumnWidth(4, 25)\n# table.setColumnWidth(5, 25)\n\n# return table\n\n\n# # Only Local for now\n# class Buttons(QPushButton):\n# class Star(QPushButton):\n# def __init__(self, project):\n# super().__init__()\n# self.setCheckable(True)\n# self.toggled.connect(self.buttonClicked(project))\n# self.setStyleSheet(\n# \"QPushButton {\"\n# \" border: none;\" # Remove the button border\n# \"}\"\n# \"QPushButton:checked {\"\n# \" color: black;\"\n# \"}\"\n# \"QPushButton:hover {\"\n# \" color: grey;\"\n# \"}\"\n# )\n# if(project[\"Star\"]):\n# self.setChecked(True)\n# self.setText(\"★\")\n# else:\n# self.setChecked(False)\n# self.setText(\"☆\")\n\n# font = self.font()\n# font.setPointSize(15)\n# self.setFont(font)\n\n# def buttonClicked(self, project):\n# def toggle():\n# if(self.isChecked()):\n# # print(f\"Starred + {project['Project Name']}\")\n# # project[\"Star\"] = True\n# self.setText(\"★\")\n# LocalConfig.SetStar(project[\"ID\"], True)\n# else:\n# # print(f\"Unstarred + {project['Project Name']}\")\n# # project[\"Star\"] = False\n# self.setText(\"☆\")\n# LocalConfig.SetStar(project[\"ID\"], False)\n# return toggle\n \n# class Play(QPushButton):\n# def __init__(self, project):\n# super().__init__()\n# self.setText(\"▶\")\n# self.activeProjectPath = project[\"Project Path\"]\n# self.clicked.connect(self.buttonClicked)\n# self.setStyleSheet(\n# \"QPushButton {\"\n# \" border: none;\"\n# \" text-align: center;\"\n# \" vertical-align: middle;\"\n# \"}\"\n# \"QPushButton:hover {\"\n# \" color: green;\"\n# \"}\"\n# )\n# font = self.font()\n# font.setPointSize(25)\n# self.setFont(font)\n\n# def buttonClicked(self, project):\n# subprocess.Popen([\"python\", \"-m\", \"pyroGamer.Editor\"] + [\"--projectPath\", self.activeProjectPath])\n\n# class Options(QPushButton):\n# def __init__(self, project):\n# super().__init__()\n# self.setText(\"⋮\")\n# self.clicked.connect(lambda: print(\"Options\"))\n# # hide the button's border\n# self.setStyleSheet(\n# \"QPushButton {\"\n# \" border: none;\"\n# \" text-align: center;\"\n# \" vertical-align: middle;\"\n# \"}\"\n# \"QPushButton:hover {\"\n# \" color: green;\"\n# \"}\"\n# )\n# font = self.font()\n# font.setPointSize(25)\n# self.setFont(font)\n", "repo_name": "BobuDragos/pyroGamer", "sub_path": "GUI/Hub/Elements/Tabs.py", "file_name": "Tabs.py", "file_ext": "py", "file_size_in_byte": 15247, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "colorama.init", "line_number": 35, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QWidget", "line_number": 40, "usage_type": "name"}, {"api_name": "PyQt6.QtWidgets.QVBoxLayout", "line_number": 44, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QHBoxLayout", "line_number": 46, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLineEdit", "line_number": 47, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QPushButton", "line_number": 48, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QPushButton", "line_number": 51, "usage_type": "call"}, {"api_name": "pyroGamer.GUI.Hub.Elements.Tables.ProjectTable.Local", "line_number": 56, "usage_type": "call"}, {"api_name": "pyroGamer.GUI.Hub.Elements.Tables.ProjectTable", "line_number": 56, "usage_type": "name"}, {"api_name": "PyQt6.QtWidgets.QFileDialog.getOpenFileName", "line_number": 64, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QFileDialog", "line_number": 64, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 67, "usage_type": "call"}, {"api_name": "colorama.Fore.BLUE", "line_number": 70, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 70, "usage_type": "name"}, {"api_name": "subprocess.run", "line_number": 71, "usage_type": "call"}, {"api_name": "colorama.Fore.LIGHTBLACK_EX", "line_number": 76, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 76, "usage_type": "name"}, {"api_name": "textwrap.indent", "line_number": 76, "usage_type": "call"}, {"api_name": "pprint.pformat", "line_number": 76, "usage_type": "call"}, {"api_name": "colorama.Fore.RED", "line_number": 77, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 77, "usage_type": "name"}, {"api_name": "colorama.Style.BRIGHT", "line_number": 77, "usage_type": "attribute"}, {"api_name": "colorama.Style", "line_number": 77, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 78, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QDialog", "line_number": 84, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QVBoxLayout", "line_number": 87, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QHBoxLayout", "line_number": 89, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLineEdit", "line_number": 92, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QHBoxLayout", "line_number": 97, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLineEdit", "line_number": 101, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QFileDialog.getExistingDirectory", "line_number": 107, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QFileDialog", "line_number": 107, "usage_type": "name"}, {"api_name": "PyQt6.QtGui.QAction", "line_number": 109, "usage_type": "call"}, {"api_name": "PyQt6.QtGui.QIcon", "line_number": 110, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLineEdit.ActionPosition", "line_number": 112, "usage_type": "attribute"}, {"api_name": "PyQt6.QtWidgets.QLineEdit", "line_number": 112, "usage_type": "name"}, {"api_name": "PyQt6.QtWidgets.QVBoxLayout", "line_number": 116, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QPushButton", "line_number": 117, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QMessageBox.critical", "line_number": 124, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QMessageBox", "line_number": 124, "usage_type": "name"}, {"api_name": "colorama.Fore.BLUE", "line_number": 127, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 127, "usage_type": "name"}, {"api_name": "subprocess.run", "line_number": 128, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 129, "usage_type": "call"}, {"api_name": "colorama.Fore.LIGHTBLACK_EX", "line_number": 133, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 133, "usage_type": "name"}, {"api_name": "textwrap.indent", "line_number": 133, "usage_type": "call"}, {"api_name": "pprint.pformat", "line_number": 133, "usage_type": "call"}, {"api_name": "colorama.Fore.RED", "line_number": 134, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 134, "usage_type": "name"}, {"api_name": "colorama.Style.BRIGHT", "line_number": 134, "usage_type": "attribute"}, {"api_name": "colorama.Style", "line_number": 134, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 135, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QPushButton", "line_number": 142, "usage_type": "call"}, {"api_name": "pyroGamer.GUI.Hub.Elements.Tables.ProjectTable.Local", "line_number": 158, "usage_type": "call"}, {"api_name": "pyroGamer.GUI.Hub.Elements.Tables.ProjectTable", "line_number": 158, "usage_type": "name"}]}
+{"seq_id": "40812497598", "text": "from django.conf import settings\nfrom rest_framework.response import Response\nfrom rest_framework.views import APIView\n\n\nclass ConfigView(APIView):\n def get(self, request):\n return Response(\n {\n key.lower(): value\n for key, value in settings.LICO.items()\n }\n )\n\n\nclass VersionView(APIView):\n def get(self, request):\n from ._version import build_date, version\n from .subapp import iter_sub_apps\n\n return Response(\n {\n 'version': version,\n 'build_date': build_date,\n 'subapps': [\n {\n 'name': app.dist.project_name,\n 'version': app.dist.version\n }\n for app in iter_sub_apps()\n ]\n }\n )\n\n\nclass ApplicationConfigView(APIView):\n app = None\n\n def get(self, request):\n return Response(\n self.app.on_show_config(settings)\n if self.app is not None else {}\n )\n", "repo_name": "lenovo/openlico", "sub_path": "core/base/lico/core/base/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1076, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 10, "dataset": "github-code", "pt": "24", "api": [{"api_name": "rest_framework.views.APIView", "line_number": 6, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.settings.LICO.items", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.settings.LICO", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 11, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 16, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 21, "usage_type": "call"}, {"api_name": "_version.version", "line_number": 23, "usage_type": "name"}, {"api_name": "_version.build_date", "line_number": 24, "usage_type": "name"}, {"api_name": "subapp.iter_sub_apps", "line_number": 30, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 36, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 40, "usage_type": "call"}, {"api_name": "django.conf.settings", "line_number": 41, "usage_type": "argument"}]}
+{"seq_id": "10243213772", "text": "import pygame\nfrom pygame import Rect, Color\nfrom pygame.constants import K_LEFT, K_RIGHT\n\nfrom pymunk import Body, Poly\nimport pymunk\n\nfrom config import GRAVITY, WIDTH, HEIGHT\n\n\nclass Player:\n\n def __init__ (self, space):\n \n # size\n self.w = 25\n self.h = 25\n\n # position\n self.x = WIDTH // 2 - self.w // 2\n self.y = HEIGHT // 2 - self.h // 2\n\n # pygame rectangle\n self.rect = Rect (self.x, self.y, self.w, self.h)\n self.color = Color (209, 87, 0)\n\n # physics\n self.rigidbody = Body (0, 5, body_type=Body.DYNAMIC)\n self.rigidbody.position = self.x, self.y\n \n self.hitbox = pymunk.Circle (self.rigidbody, self.w / 2)\n self.hitbox.mass = 10\n self.hitbox.elasticity = 0\n self.hitbox.friction = 0\n\n space.add (self.rigidbody, self.hitbox)\n\n \n\n \n def update (self, dt):\n\n keys = pygame.key.get_pressed()\n\n if keys [K_LEFT] or keys [K_RIGHT]:\n\n if keys [K_LEFT]:\n self.rigidbody.apply_force_at_world_point ( (-1000,0), (0,0) )\n\n if keys [K_RIGHT]:\n self.rigidbody.apply_force_at_world_point ( (1000,0), (0,0) )\n\n\n\n\n x = int (self.rigidbody.position.x)\n y = int (self.rigidbody.position.y)\n self.x = x\n self.y = y\n self.rect.update(x, y, self.w, self.h)\n\n return\n\n\n\n def draw (self, window):\n \n pygame.draw.rect (window, self.color, self.rect)\n\n return", "repo_name": "CSnackerman/pymunk_test", "sub_path": "player.py", "file_name": "player.py", "file_ext": "py", "file_size_in_byte": 1523, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "config.WIDTH", "line_number": 20, "usage_type": "name"}, {"api_name": "config.HEIGHT", "line_number": 21, "usage_type": "name"}, {"api_name": "pygame.Rect", "line_number": 24, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 25, "usage_type": "call"}, {"api_name": "pymunk.Body", "line_number": 28, "usage_type": "call"}, {"api_name": "pymunk.Body.DYNAMIC", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pymunk.Circle", "line_number": 31, "usage_type": "call"}, {"api_name": "pygame.key.get_pressed", "line_number": 43, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pygame.constants.K_LEFT", "line_number": 45, "usage_type": "name"}, {"api_name": "pygame.constants.K_RIGHT", "line_number": 45, "usage_type": "name"}, {"api_name": "pygame.constants.K_LEFT", "line_number": 47, "usage_type": "name"}, {"api_name": "pygame.constants.K_RIGHT", "line_number": 50, "usage_type": "name"}, {"api_name": "pygame.draw.rect", "line_number": 68, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 68, "usage_type": "attribute"}]}
+{"seq_id": "21436718121", "text": "import subprocess\nimport inquirer\nimport re\n\n# commands\ncreate_list = \"adb shell pm list packages > ~/package_list.txt\"\npackage_list = \"adb shell pm list packages\"\nremove = \"adb shell pm uninstall --user 0 \" # Space needed at the end. -k: Keep data and cache directories after package removal.\nadb_shell = \"adb shell\"\nsearch_command = \"adb shell pm list packages \" # Space needed at the end.\ntext_append = \" > ~/package_list.txt\" # Space needed in front.\n\n\ndef remove_packs(list_file): # CREATE TEXT FILE:\n with open('package_list.txt', 'w+'):\n subprocess.call(list_file, shell=True)\n\n with open('package_list.txt', 'r') as f: # READ TEXT FILE:\n lines = f.readlines()\n\n items = [] # CLEANUP STRINGS AND GET LIST OF PACKAGES\n for i in lines:\n the_list = i.split(\":\")\n items.append(the_list[1].replace(\"\\n\", \"\"))\n\n if len(items) == 0:\n print(\"\\n***Nothing found\")\n else:\n print(\"### TOTAL PACKAGES:\", len(items))\n\n package_object = [\n inquirer.Checkbox('packages_name',\n message=\"Select the packages you would like to remove with the space bar\",\n choices=items,\n ),\n ]\n package_selections = inquirer.prompt(package_object)\n packs_for_removal = package_selections.get('packages_name')\n i = 0\n\n # This is just confirmation that at least one package has been selected:\n if len(packs_for_removal) == 0:\n print(\"\\n***No selection: back to menu...\\n\")\n elif packs_for_removal:\n selection = input(\"\\n***Are you sure you want to remove these package(s)? y/N: \").lower()\n if selection == \"y\":\n while i < len(packs_for_removal):\n print(\"Removing package: \", i, \":\", packs_for_removal[i], \"...\")\n subprocess.call(remove + packs_for_removal[i], shell=True)\n i += 1\n elif selection == \"n\":\n print(\"\\n***Nothing to remove, back to menu...\\n\")\n elif selection != \"y\" or selection != \"n\":\n print(\"\\n***Bad input, only 'y' or 'n'.\\n\")\n\n\nprint(\"### Package Remover Interface for ADB ###\")\nresults = 0\nwhile results != 4:\n choose = [\n inquirer.List('menu',\n message=\"Please, choose an option\",\n choices=[('List all packages and select for removal', '1'), ('Search packages by keyword', '2'),\n ('Spawn ADB shell', '3'), ('Exit', '4')],\n ),\n ]\n choice = inquirer.prompt(choose)\n results = choice['menu'][0] # Option selected by user\n\n if results == \"1\":\n remove_packs(create_list)\n\n elif results == \"2\":\n search_filter = input(\"Search any part of the package hierarchy:\").lower()\n while True:\n if not re.match(\"^[A-Za-z0-9_]*$\", search_filter):\n print(\"\\n***Only alphanumeric and underscore characters!\\n\")\n break\n elif not search_filter:\n print(\"\\n***Please search for something.\\n\")\n break\n else:\n triumvirate = search_command + search_filter + text_append\n remove_packs(triumvirate)\n break\n\n elif results == \"3\":\n subprocess.call(adb_shell, shell=True)\n\n elif results == \"4\":\n print(\"Bye!\")\n break\n", "repo_name": "fercastt/Android-Package-Remover-Interface", "sub_path": "pack_remover.py", "file_name": "pack_remover.py", "file_ext": "py", "file_size_in_byte": 3368, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "24", "api": [{"api_name": "subprocess.call", "line_number": 16, "usage_type": "call"}, {"api_name": "inquirer.Checkbox", "line_number": 32, "usage_type": "call"}, {"api_name": "inquirer.prompt", "line_number": 37, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 49, "usage_type": "call"}, {"api_name": "inquirer.List", "line_number": 61, "usage_type": "call"}, {"api_name": "inquirer.prompt", "line_number": 67, "usage_type": "call"}, {"api_name": "re.match", "line_number": 76, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 88, "usage_type": "call"}]}
+{"seq_id": "70969524541", "text": "import logging\nfrom aiogram.utils.exceptions import BotBlocked\nfrom aiogram import Bot, Dispatcher, types\nfrom aiogram.contrib.fsm_storage.memory import MemoryStorage\nimport asyncio\nimport config as conf\nfrom navigate import register_navigate\nfrom aiogram.types import BotCommand\n\nlogger = logging.getLogger(__name__)\n\nasync def set_commands(bot: Bot):\n commands = [\n BotCommand(command='/menu', description=\"Меню\")\n ]\n await bot.set_my_commands(commands)\n\nasync def main():\n # Настройка логирования в stdout\n logging.basicConfig(\n level=logging.INFO,\n format=\"%(asctime)s - %(levelname)s - %(name)s - %(message)s\",\n )\n logger.error(\"Starting bot\")\n\n # Парсинг файла конфигурации\n config = conf.load_config(\"bot.ini\")\n\n # Объявление и инициализация объектов бота и диспетчера\n bot = Bot(token=config.tg_bot.token)\n dp = Dispatcher(bot, storage=MemoryStorage())\n register_navigate(dp)\n\n @dp.errors_handler(exception=BotBlocked)\n async def error_bot_blocked(update: types.Update, exception: BotBlocked):\n # Update: объект события от Telegram. Exception: объект исключения\n # Здесь можно как-то обработать блокировку, например, удалить пользователя из БД\n print(f\"Error: {exception}\")\n return True\n\n await set_commands(bot)\n await dp.start_polling()\n\n\nif __name__ == \"__main__\":\n asyncio.run(main())\n", "repo_name": "zhakos/tgbot_for_ali", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1593, "program_lang": "python", "lang": "ru", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "aiogram.Bot", "line_number": 12, "usage_type": "name"}, {"api_name": "aiogram.types.BotCommand", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 21, "usage_type": "attribute"}, {"api_name": "config.load_config", "line_number": 27, "usage_type": "call"}, {"api_name": "aiogram.Bot", "line_number": 30, "usage_type": "call"}, {"api_name": "config.tg_bot", "line_number": 30, "usage_type": "attribute"}, {"api_name": "aiogram.Dispatcher", "line_number": 31, "usage_type": "call"}, {"api_name": "aiogram.contrib.fsm_storage.memory.MemoryStorage", "line_number": 31, "usage_type": "call"}, {"api_name": "navigate.register_navigate", "line_number": 32, "usage_type": "call"}, {"api_name": "aiogram.types.Update", "line_number": 35, "usage_type": "attribute"}, {"api_name": "aiogram.types", "line_number": 35, "usage_type": "name"}, {"api_name": "aiogram.utils.exceptions.BotBlocked", "line_number": 35, "usage_type": "name"}, {"api_name": "aiogram.utils.exceptions.BotBlocked", "line_number": 34, "usage_type": "name"}, {"api_name": "asyncio.run", "line_number": 46, "usage_type": "call"}]}
+{"seq_id": "37079197239", "text": "import time\nimport tempfile\nimport shutil\nimport simplejson as json\nimport subprocess\nfrom os import remove\nfrom oio.common.http import get_pool_manager\n\nfrom tests.utils import BaseTestCase, random_str, random_id\n\n\ndef _key(rec):\n return '|'.join((rec['container_id'], rec['content_id'], rec['chunk_id']))\n\n\nclass TestRdirServer(BaseTestCase):\n def setUp(self):\n super(TestRdirServer, self).setUp()\n self.http_pool = get_pool_manager(max_retries=10)\n self.num, self.db_path, self.host, self.port = self.get_service('rdir')\n self.vol = self._volume()\n\n def tearDown(self):\n super(TestRdirServer, self).tearDown()\n self.http_pool.clear()\n\n def _volume(self):\n return random_id(8)\n\n def _record(self):\n return {\"container_id\": random_id(64),\n \"content_id\": random_id(32),\n \"chunk_id\": random_id(64),\n \"mtime\": 17}\n\n def _rdir_url(self, tail):\n return 'http://{0}:{1}{2}'.format(self.host, self.port, tail)\n\n def _get(self, url, **kwargs):\n return self.request('GET', self._rdir_url(url), **kwargs)\n\n def _post(self, url, **kwargs):\n return self.request('POST', self._rdir_url(url), **kwargs)\n\n def _delete(self, url, **kwargs):\n return self.request('DELETE', self._rdir_url(url), **kwargs)\n\n def test_explicit_create(self):\n rec = self._record()\n\n # try to push on unknown volume\n resp = self._post(\n \"/v1/rdir/push\", params={'vol': self.vol},\n data=json.dumps(rec))\n self.assertEqual(resp.status, 404)\n\n # The fetch fails\n resp = self._post(\"/v1/rdir/fetch\", params={'vol': self.vol})\n self.assertEqual(resp.status, 404)\n\n # create volume\n resp = self._post(\"/v1/rdir/create\", params={'vol': self.vol})\n self.assertEqual(resp.status, 201)\n\n # the fetch returns an empty array\n resp = self._post(\"/v1/rdir/fetch\", params={'vol': self.vol})\n self.assertEqual(resp.status, 200)\n self.assertEqual(self.json_loads(resp.data), [])\n\n # now the push must succeed\n resp = self._post(\n \"/v1/rdir/push\", params={'vol': self.vol},\n data=json.dumps(rec))\n self.assertEqual(resp.status, 204)\n\n # we must fetch the same data\n resp = self._post(\"/v1/rdir/fetch\", params={'vol': self.vol})\n self.assertEqual(resp.status, 200)\n reference = [\n [_key(rec), {'mtime': rec['mtime'], 'rtime': 0}]\n ]\n self.assertListEqual(self.json_loads(resp.data), reference)\n\n # deleting must succeed\n resp = self._delete(\n \"/v1/rdir/delete\", params={'vol': self.vol},\n data=json.dumps(rec))\n self.assertEqual(resp.status, 204)\n\n # fetching must return an empty array\n resp = self._post(\"/v1/rdir/fetch\", params={'vol': self.vol})\n self.assertEqual(resp.status, 200)\n self.assertEqual(self.json_loads(resp.data), [])\n\n def test_implicit_create(self):\n rec = self._record()\n\n # try to push on unknown volume\n resp = self._post(\n \"/v1/rdir/push\", params={'vol': self.vol},\n data=json.dumps(rec))\n self.assertEqual(resp.status, 404)\n\n # try to push on unknown volume WITH create flag\n resp = self._post(\n \"/v1/rdir/push\", params={'vol': self.vol, 'create': True},\n data=json.dumps(rec))\n self.assertEqual(resp.status, 204)\n\n # We must fetch the same data\n resp = self._post(\"/v1/rdir/fetch\", params={'vol': self.vol})\n self.assertEqual(resp.status, 200)\n self.assertEqual(self.json_loads(resp.data), [\n [_key(rec), {'mtime': rec['mtime'], 'rtime': 0}]\n ])\n\n def test_push_missing_fields(self):\n rec = self._record()\n\n # DB creation\n resp = self._post(\"/v1/rdir/create\", params={'vol': self.vol})\n self.assertEqual(resp.status, 201)\n\n for k in ['container_id', 'content_id', 'chunk_id']:\n save = rec.pop(k)\n # push an incomplete record\n resp = self._post(\n \"/v1/rdir/push\", params={'vol': self.vol},\n data=json.dumps(rec))\n self.assertEqual(resp.status, 400)\n # check we list nothing\n resp = self._post(\"/v1/rdir/fetch\", params={'vol': self.vol})\n self.assertEqual(resp.status, 200)\n self.assertListEqual(self.json_loads(resp.data), [])\n rec[k] = save\n\n def test_lock_unlock(self):\n who = random_str(64)\n\n # lock without who, DB not created\n resp = self._post(\n \"/v1/rdir/admin/lock\", params={'vol': self.vol},\n data=json.dumps({}))\n self.assertEqual(resp.status, 400)\n\n # lock with who, DB not created\n resp = self._post(\n \"/v1/rdir/admin/lock\", params={'vol': self.vol},\n data=json.dumps({'who': who}))\n self.assertEqual(resp.status, 404)\n\n # DB creation\n resp = self._post(\"/v1/rdir/create\", params={'vol': self.vol})\n self.assertEqual(resp.status, 201)\n\n # lock without who\n resp = self._post(\n \"/v1/rdir/admin/lock\", params={'vol': self.vol},\n data=json.dumps({}))\n self.assertEqual(resp.status, 400)\n\n # lock\n resp = self._post(\n \"/v1/rdir/admin/lock\", params={'vol': self.vol},\n data=json.dumps({'who': who}))\n self.assertEqual(resp.status, 204)\n\n # double lock, different who\n resp = self._post(\n \"/v1/rdir/admin/lock\", params={'vol': self.vol},\n data=json.dumps({'who': random_str(64)}))\n self.assertEqual(resp.status, 403)\n body = self.json_loads(resp.data)\n self.assertEqual(body['message'], \"Already locked by %s\" % who)\n\n # unlock\n resp = self._post(\"/v1/rdir/admin/unlock\", params={'vol': self.vol})\n self.assertEqual(resp.status, 204)\n\n def test_rdir_clear_and_lock(self):\n rec = self._record()\n who = random_id(32)\n\n # push with autocreate\n resp = self._post(\n \"/v1/rdir/push\", params={'vol': self.vol, 'create': True},\n data=json.dumps(rec))\n self.assertEqual(resp.status, 204)\n\n # lock\n resp = self._post(\n \"/v1/rdir/admin/lock\", params={'vol': self.vol},\n data=json.dumps({'who': who}))\n self.assertEqual(resp.status, 204)\n\n # try to clear while the lock is held\n resp = self._post(\"/v1/rdir/admin/clear\", params={'vol': self.vol})\n self.assertEqual(resp.status, 403)\n\n # unlock\n resp = self._post(\"/v1/rdir/admin/unlock\", params={'vol': self.vol})\n self.assertEqual(resp.status, 204)\n\n # clear all entries\n resp = self._post(\n \"/v1/rdir/admin/clear\", params={'vol': self.vol},\n data=json.dumps({'all': True}))\n self.assertEqual(resp.status, 200)\n self.assertEqual(self.json_loads(resp.data), {'removed': 1})\n\n def test_vol_status(self):\n # Status on inexistant DB\n resp = self._post(\"/v1/rdir/status\", params={'vol': self.vol})\n self.assertEqual(resp.status, 404)\n\n # DB creation\n resp = self._post(\"/v1/rdir/create\", params={'vol': self.vol})\n self.assertEqual(resp.status, 201)\n\n # Status on an empty DB\n resp = self._get(\"/v1/rdir/status\", params={'vol': self.vol})\n self.assertEqual(resp.status, 200)\n self.assertEqual(self.json_loads(resp.data),\n {'chunk': {'total': 0}, 'container': {}})\n\n\nclass TestRdirServer2(TestRdirServer):\n def setUp(self):\n super(TestRdirServer2, self).setUp()\n self.host = '127.0.0.1'\n self.port = 5999\n self.db_path = tempfile.mkdtemp()\n self.cfg_path = tempfile.mktemp()\n with open(self.cfg_path, 'w') as f:\n f.write(\"[rdir-server]\\n\")\n f.write(\"bind_addr = {0}\\n\".format(self.host))\n f.write(\"bind_port = {0}\\n\".format(self.port))\n f.write(\"namespace = {0}\\n\".format(self.ns))\n f.write(\"db_path = {0}\\n\".format(self.db_path))\n f.write(\"syslog_prefix = OIO,OPENIO,rdir,1\\n\")\n\n self.child = subprocess.Popen(['oio-rdir-server', self.cfg_path],\n close_fds=True)\n if not self._wait_for_that_fucking_slow_startup_on_travis():\n self.child.kill()\n raise Exception(\"The RDIR server is too long to start\")\n\n def tearDown(self):\n super(TestRdirServer2, self).tearDown()\n self.http_pool.clear()\n self._kill_and_watch_it_die()\n shutil.rmtree(self.db_path)\n remove(self.cfg_path)\n\n def _kill_and_watch_it_die(self):\n self.child.terminate()\n self.child.wait()\n\n def _wait_for_that_fucking_slow_startup_on_travis(self):\n for i in range(5):\n if self._check_for_server():\n return True\n time.sleep(i * 0.2)\n return False\n\n def _check_for_server(self):\n hexport = \"%04X\" % self.port\n with open(\"/proc/net/tcp\", \"r\") as f:\n for line in f:\n tokens = line.strip().split()\n port = tokens[1][9:13]\n if port == hexport:\n return True\n return False\n\n def test_status(self):\n\n # check the service has no opened DB\n resp = self._get('/status')\n self.assertEqual(resp.status, 200)\n self.assertEqual(self.json_loads(resp.data), {'opened_db_count': 0})\n\n # DB creation\n resp = self._post(\"/v1/rdir/create\", params={'vol': self.vol})\n self.assertEqual(resp.status, 201)\n\n # The base remains open after it has been created\n resp = self._get('/status')\n self.assertEqual(resp.status, 200)\n self.assertEqual(self.json_loads(resp.data), {'opened_db_count': 1})\n", "repo_name": "kamel-rahim/oio-sds", "sub_path": "tests/functional/rdir/test_server.py", "file_name": "test_server.py", "file_ext": "py", "file_size_in_byte": 10148, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "github-code", "pt": "24", "api": [{"api_name": "tests.utils.BaseTestCase", "line_number": 16, "usage_type": "name"}, {"api_name": "oio.common.http.get_pool_manager", "line_number": 19, "usage_type": "call"}, {"api_name": "tests.utils.random_id", "line_number": 28, "usage_type": "call"}, {"api_name": "tests.utils.random_id", "line_number": 31, "usage_type": "call"}, {"api_name": "tests.utils.random_id", "line_number": 32, "usage_type": "call"}, {"api_name": "tests.utils.random_id", "line_number": 33, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 54, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 73, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 87, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 101, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 107, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 129, "usage_type": "call"}, {"api_name": "tests.utils.random_str", "line_number": 138, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 143, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 149, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 159, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 165, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 171, "usage_type": "call"}, {"api_name": "tests.utils.random_str", "line_number": 171, "usage_type": "call"}, {"api_name": "tests.utils.random_id", "line_number": 182, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 187, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 193, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 207, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 232, "usage_type": "call"}, {"api_name": "tempfile.mktemp", "line_number": 233, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 242, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 252, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 253, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 263, "usage_type": "call"}]}
+{"seq_id": "71450409982", "text": "from flask import Flask, render_template, request\nfrom scrape import scrape_page\n\n\napp = Flask(__name__) \n\n\n@app.route('/')\ndef home():\n\treturn render_template('home.html')\n\n\n@app.route('/search', methods=['POST'])\ndef search():\n\tquery = str(request.form['query'])\n\tdata = scrape_page(query)\n\treturn render_template('result.html', data=data)\n\n\nif __name__ == '__main__':\n\tapp.run(host=\"0.0.0.0\", port=3167)", "repo_name": "AlexMathew/people-search", "sub_path": "routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 406, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 15, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 15, "usage_type": "name"}, {"api_name": "scrape.scrape_page", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 17, "usage_type": "call"}]}
+{"seq_id": "2303183809", "text": "import torch\nimport numpy as np\n\nKERNEL_SIZE = 1\n\n\ndef create_conv_1x1(weight, bias, in_c, out_c):\n assert weight.shape == (out_c, in_c, 1, 1)\n assert bias.shape == (out_c,)\n torch_conv_float = torch.nn.Conv2d(in_c, out_c, KERNEL_SIZE, 1, 0)\n torch_conv_float.weight.data = torch.from_numpy(weight)\n torch_conv_float.bias.data = torch.from_numpy(bias)\n return torch_conv_float\n\n\ndef do_np_conv_1x1(weight, bias, input):\n \"\"\"Do simplist 1x1 conv with out dilation and padding\n\n Args:\n weight(np.array): conv weight, shape: (out_c, in_c, kernel_size, kernel_size)\n bias(np.array): conv bias, shape: (out_c, )\n input(np.array): input, shape: (n, c, h, w)\n \"\"\"\n out_c = weight.shape[0]\n in_c = weight.shape[1]\n input_h = input.shape[2]\n input_w = input.shape[3]\n\n assert weight.shape[2] == KERNEL_SIZE\n assert weight.shape[3] == KERNEL_SIZE\n assert out_c == bias.shape[0]\n assert in_c == input.shape[1]\n reshape_input = input.reshape((in_c, -1))\n multi_result = np.matmul(weight.reshape((out_c, in_c)), reshape_input)\n return (\n multi_result + np.tile(bias.reshape(-1, 1), (1, input_h * input_w))\n ).reshape((1, out_c, input_h, input_w))\n", "repo_name": "GouMinghao/naive_ptq", "sub_path": "py/naive_ptq/float_op/conv.py", "file_name": "conv.py", "file_ext": "py", "file_size_in_byte": 1225, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "torch.nn.Conv2d", "line_number": 10, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 10, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 11, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 36, "usage_type": "call"}]}
+{"seq_id": "3970609811", "text": "from django import forms\nfrom django.forms import ModelForm\nfrom .models import Task, Course\n\n\n# Creates a widget to be used for inputting dates\nclass DateInput(forms.DateInput):\n input_type = 'date'\n\n\n# Uses the Course model so the entered fields can be registered to the Course table\nclass CourseForm(ModelForm):\n\n class Meta:\n model = Course\n fields = ['course_name']\n\n\n# Uses the Task model so the entered fields can be registered to the Task table\nclass TaskForm(ModelForm):\n\n # Displays only the courses registered by the logged in user\n def __init__(self, user, *args, **kwargs):\n super(TaskForm, self).__init__(*args, **kwargs)\n self.fields['course'].queryset = Course.objects.filter(user=user)\n\n class Meta:\n model = Task\n fields = ['title', 'type', 'date_assigned',\n 'date_due', 'course', 'description']\n # Date input widgets used for 'date assigned' and 'date due'\n widgets = {\n 'date_assigned': DateInput(),\n 'date_due': DateInput(),\n }\n", "repo_name": "karsteneugene/Django_Planner", "sub_path": "Django Planner/planner/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 1066, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.forms.DateInput", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 7, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 12, "usage_type": "name"}, {"api_name": "models.Course", "line_number": 15, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 20, "usage_type": "name"}, {"api_name": "models.Course.objects.filter", "line_number": 25, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 25, "usage_type": "name"}, {"api_name": "models.Task", "line_number": 28, "usage_type": "name"}]}
+{"seq_id": "1168571491", "text": "\"\"\"Countdown Calendar\n\nHacks and Tweaks\nsort it!\n sort by the date list.sort(key=lambda x: x[1]\n lambda function is a small anonymous function\n lambda arguments: expression\n x = lambda a : a + 10\n print(x(5)) _> OUTPUT 15\nrestyle the text\n make it look pretty\nset reminders\n\n\"\"\"\n\nfrom tkinter import Tk, Canvas\nfrom datetime import date,datetime\n\n\"\"\"make sure students use the proper format for %d/%m/%y' there can be no spaces and year must be 2 digits not 4\"\"\"\ndef get_events():\n list_events = []\n with open('events.txt') as file: #opens text file\n for line in file: #runs loop for each line in the text file\n line = line.rstrip('\\n') #you must remove \\n or else line will look like this ['Halloween','31/10/22\\n']\n current_event =line.split(',') #turn string line into an array with two string items\n event_date = datetime.strptime(current_event[1],'%d/%m/%y').date() #(string content, format)\n # print(current_event[1])\n current_event[1]=event_date #second item in list is now an actual date (not a string anymore)\n print(current_event[1])\n list_events.append(current_event) #this is will add the formated date of event to our list\n return list_events #returns a 2d list [['Halloween,date],[christmas, date],...]\n\ndef days_between_dates(date1,date2): #function that counts the number of days between two dates\n time_between = str(date1-date2) #variable stores difference of dates as a string\n #if a Hallowen is 27 days away, the string in time_between -> '27 days, 0:00:00 h:m:s all we need is the 27\n number_of_days = time_between.split(' ')\n return number_of_days[0]\n\n\n\nroot = Tk() #Tkinter window\nc = Canvas(root,width=800,height=800, bg='green')\nc.pack() #Tkinter window\nc.create_text(100,50, anchor='w', fill='pink', font='Courier 36 bold underline', text='Ms.T\\'s Calendar')\n\"\"\" \nThis line adds text onto the c canvas. The text starts at x=100 , y=50. \nThe starting coordinate is at the left (west) of the text \nnow we want to loop through every special event in our txt list and calculate how many days away we are\n\"\"\"\nevents = get_events()\ntoday = date.today()\n\nevents.sort(key=lambda x: x[1]) #use the second item in the list (date) to sort function\n\nvertical_space = 100 #moves the y coordinate so eevery date is on its own line\nfor event in events:\n event_name = event[0]\n days_until = days_between_dates(event[1],today)\n\n if int(days_until) <= 7:\n text_color = 'red'\n else:\n text_color = 'black'\n\n display = 'It is %s days until %s' % (days_until, event_name)\n c.create_text(100,vertical_space,anchor='w',fill=text_color, font='Arial 28 bold', text=display)\n vertical_space +=30\n\nroot.mainloop()\n\n\n\n\n\n\n", "repo_name": "thturin/countdown_Calendar", "sub_path": "countdown_calendar_update.py", "file_name": "countdown_calendar_update.py", "file_ext": "py", "file_size_in_byte": 2776, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 26, "usage_type": "name"}, {"api_name": "tkinter.Tk", "line_number": 41, "usage_type": "call"}, {"api_name": "tkinter.Canvas", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 51, "usage_type": "name"}]}
+{"seq_id": "8265130339", "text": "from django.urls import path\n\nfrom rest_framework.authtoken.views import obtain_auth_token\n\nfrom .views import AddEmployeeAPI, CreateProfileAPI, ProfileAPI, ListOfPayrollsAPI, Logout, ListOfProfilesAPI, \\\n PayrollsAPI\n\nurlpatterns = [\n path(\"login/\", obtain_auth_token),\n path(\"add_employee/\", AddEmployeeAPI.as_view()),\n path(\"create_profile//\", CreateProfileAPI.as_view(), name=\"create_profile\"),\n path(\"profile/\", ProfileAPI.as_view()),\n path(\"list_of_profiles/\", ListOfProfilesAPI.as_view()),\n path(\"salaries/\", PayrollsAPI.as_view()),\n path(\"list_of_salaries/\", ListOfPayrollsAPI.as_view()),\n path(\"logout/\", Logout.as_view())\n]\n", "repo_name": "hmpsl99/hr_solution", "sub_path": "hr_solution_back/Hr/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 672, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "rest_framework.authtoken.views.obtain_auth_token", "line_number": 9, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "views.AddEmployeeAPI.as_view", "line_number": 10, "usage_type": "call"}, {"api_name": "views.AddEmployeeAPI", "line_number": 10, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "views.CreateProfileAPI.as_view", "line_number": 11, "usage_type": "call"}, {"api_name": "views.CreateProfileAPI", "line_number": 11, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "views.ProfileAPI.as_view", "line_number": 12, "usage_type": "call"}, {"api_name": "views.ProfileAPI", "line_number": 12, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "views.ListOfProfilesAPI.as_view", "line_number": 13, "usage_type": "call"}, {"api_name": "views.ListOfProfilesAPI", "line_number": 13, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "views.PayrollsAPI.as_view", "line_number": 14, "usage_type": "call"}, {"api_name": "views.PayrollsAPI", "line_number": 14, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "views.ListOfPayrollsAPI.as_view", "line_number": 15, "usage_type": "call"}, {"api_name": "views.ListOfPayrollsAPI", "line_number": 15, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "views.Logout.as_view", "line_number": 16, "usage_type": "call"}, {"api_name": "views.Logout", "line_number": 16, "usage_type": "name"}]}
+{"seq_id": "22241722053", "text": "import warnings\n\n# Dependency imports\nimport numpy as np\nimport tensorflow.compat.v2 as tf\n\nfrom tensorflow_probability.python.internal import broadcast_util as bu\nfrom tensorflow_probability.python.internal import distribution_util as dist_util\nfrom tensorflow_probability.python.internal import dtype_util\nfrom tensorflow_probability.python.internal import prefer_static as ps\nfrom tensorflow_probability.python.internal import tensorshape_util\nfrom tensorflow_probability.python.math.gradient import value_and_gradient as tfp_math_value_and_gradients\nfrom tensorflow.python.util import deprecation # pylint: disable=g-direct-tensorflow-import\n\n\n__all__ = [\n 'choose',\n 'choose_from',\n 'enable_store_parameters_in_results',\n 'index_remapping_gather',\n 'is_list_like',\n 'is_namedtuple_like',\n 'make_name',\n 'maybe_call_fn_and_grads',\n 'prepare_state_parts',\n 'PrettyNamedTupleMixin',\n 'safe_sum',\n 'SEED_CTOR_ARG_DEPRECATION_MSG',\n 'set_doc',\n 'strip_seeds',\n 'warn_if_parameters_are_not_simple_tensors',\n]\n\n\nJAX_MODE = False\n\nSEED_CTOR_ARG_DEPRECATION_MSG = (\n 'Seeding `tfp.mcmc.TransitionKernel` instances by constructor argument is '\n 'deprecated. Use the `seed` argument to `tfp.mcmc.sample_chain` or '\n 'directly on `one_step`. The legacy behavior is still supported and should '\n 'be through 2020-09-20.')\n\n\nclass PrettyNamedTupleMixin(object):\n \"\"\"Mixin adding a nicer `__repr__` for `namedtuple`s.\"\"\"\n __slots__ = ()\n\n def __repr__(self):\n return '{}(\\n{}\\n)'.format(\n type(self).__name__,\n ',\\n'.join(' {}={}'.format(k, repr(v).replace('\\n', '\\n '))\n for (k, v) in self._asdict().items()))\n\n\ndef prepare_state_parts(state_or_state_part, dtype=None, name=None):\n \"\"\"Calls c2t on each element or the entirety if not iterable; returns list.\"\"\"\n # Don't use tf.name_scope since this function has ct2-like semantics.\n is_multipart = is_list_like(state_or_state_part)\n state_parts = state_or_state_part if is_multipart else [state_or_state_part]\n state_parts = [tf.convert_to_tensor(x, dtype=dtype, name=name)\n for x in state_parts]\n return state_parts, is_multipart\n\n\ndef is_list_like(x):\n \"\"\"Helper which returns `True` if input is `list`-like.\"\"\"\n return isinstance(x, (tuple, list))\n\n\ndef is_namedtuple_like(x):\n \"\"\"Helper which returns `True` if input is `collections.namedtuple`-like.\"\"\"\n try:\n for fn in x._fields:\n _ = getattr(x, fn)\n return True\n except AttributeError:\n return False\n\n\ndef make_name(super_name, default_super_name, sub_name):\n \"\"\"Helper which makes a `str` name; useful for tf.name_scope.\"\"\"\n name = super_name if super_name is not None else default_super_name\n if sub_name is not None:\n name += '_' + sub_name\n return name\n\n\ndef _choose_base_case(is_accepted,\n proposed,\n current,\n name=None,\n addr=None,):\n \"\"\"Helper to `choose` which expand_dims `is_accepted` and applies tf.where.\"\"\"\n def _where(proposed, current):\n \"\"\"Wraps `tf.where`.\"\"\"\n if proposed is current:\n return proposed\n\n # Handle CompositeTensor types at the leafmost `addr`.\n flat_p = tf.nest.flatten(proposed, expand_composites=True)\n flat_c = tf.nest.flatten(current, expand_composites=True)\n\n res = []\n for p, c in zip(flat_p, flat_c):\n # Preserve the name from `current` so names can propagate from\n # `bootstrap_results`.\n name = getattr(c, 'name', None)\n if name is not None:\n name = name.rpartition('/')[2].rsplit(':', 1)[0]\n # Since this is an internal utility it is ok to assume\n # tf.shape(proposed) == tf.shape(current).\n res.append(\n tf.where(bu.left_justified_expand_dims_like(is_accepted, p), p, c,\n name=name))\n return tf.nest.pack_sequence_as(current, res, expand_composites=True)\n\n with tf.name_scope(name or 'choose'):\n if not is_list_like(proposed):\n return _where(proposed, current)\n return tf.nest.pack_sequence_as(\n current,\n [(_choose_recursive(is_accepted, p, c, name=name, addr=f'{addr}[i]')\n if is_namedtuple_like(p) else\n _where(p, c)) for i, (p, c) in enumerate(zip(proposed, current))])\n\n\ndef _choose_recursive(is_accepted, proposed, current, name=None, addr=''):\n \"\"\"Recursion helper which also reports the address of any failures.\"\"\"\n with tf.name_scope(name or 'choose'):\n if not is_namedtuple_like(proposed):\n return _choose_base_case(is_accepted, proposed, current, name=name,\n addr=addr)\n if not isinstance(proposed, type(current)):\n raise TypeError(\n f'Type of `proposed` ({type(proposed).__name__}) must be identical '\n f'to type of `current` ({type(current).__name__}). (At \"{addr}\".)')\n items = {}\n for fn in proposed._fields:\n items[fn] = _choose_recursive(is_accepted,\n getattr(proposed, fn),\n getattr(current, fn),\n name=name,\n addr=f'{addr}/{fn}')\n return type(proposed)(**items)\n\n\ndef choose(is_accepted, proposed, current, name=None):\n \"\"\"Helper which expand_dims `is_accepted` then applies tf.where.\"\"\"\n return _choose_recursive(is_accepted, proposed, current, name=name)\n\n\ndef _nest_choose(is_accepted, proposed, current):\n \"\"\"Like `choose` but not limited to list, tuple, namedtuple.\"\"\"\n result_parts = choose(is_accepted,\n tf.nest.flatten(proposed, expand_composites=True),\n tf.nest.flatten(current, expand_composites=True))\n return tf.nest.pack_sequence_as(\n proposed, result_parts, expand_composites=True)\n\n\ndef choose_from(n, options):\n \"\"\"Helper to select the n-th option from a list of options.\n\n This is useful when `n` is not a concrete value. Also note that\n the value of `n` will be clipped to the edges of the interval\n `[0, len(options) - 1]`.\n\n Args:\n n: Scalar `int` `Tensor` option.\n options: List of options to choose from. All the options should have the\n same nested structure.\n\n Returns:\n The n-th option among `options`.\n \"\"\"\n if len(options) == 1:\n return options[0]\n m = len(options) // 2\n return _nest_choose(n < m, choose_from(n, options[:m]),\n choose_from(n - m, options[m:]))\n\n\ndef strip_seeds(obj):\n if not is_namedtuple_like(obj):\n return obj\n return type(obj)(**{fn: strip_seeds(fv) if fn != 'seed' else []\n for fn, fv in obj._asdict().items()})\n\n\ndef safe_sum(x, alt_value=-np.inf, name=None):\n \"\"\"Elementwise adds list members, replacing non-finite results with alt_value.\n\n Typically the `alt_value` is chosen so the `MetropolisHastings`\n `TransitionKernel` always rejects the proposal.\n\n Args:\n x: Python `list` of `Tensors` to elementwise add.\n alt_value: Python scalar used to replace any elementwise sums which would\n otherwise be non-finite.\n name: Python `str` name prefixed to Ops created by this function.\n Default value: `None` (i.e., \"safe_sum\").\n\n Returns:\n safe_sum: `Tensor` representing the elementwise sum of list of `Tensor`s\n `x` or `alt_value` where sums are non-finite.\n\n Raises:\n TypeError: if `x` is not list-like.\n ValueError: if `x` is empty.\n \"\"\"\n with tf.name_scope(name or 'safe_sum'):\n if not is_list_like(x):\n raise TypeError('Expected list input.')\n if not x:\n raise ValueError('Input should not be empty.')\n in_shape = x[0].shape\n x = tf.add_n(x)\n x = tf.where(tf.math.is_finite(x), x, tf.constant(alt_value, dtype=x.dtype))\n tensorshape_util.set_shape(x, in_shape)\n return x\n\n\ndef set_doc(value):\n \"\"\"Decorator to programmatically set a function docstring.\"\"\"\n def _doc(func):\n func.__doc__ = value\n return func\n return _doc\n\n\ndef _value_and_gradients(fn, fn_arg_list, result=None, grads=None, name=None):\n \"\"\"Helper to `maybe_call_fn_and_grads`.\"\"\"\n with tf.name_scope(name or 'value_and_gradients'):\n\n def _convert_to_tensor(x, name):\n ctt = lambda x_: None if x_ is None else tf.convert_to_tensor( # pylint: disable=g-long-lambda\n x_, name=name)\n return [ctt(x_) for x_ in x] if is_list_like(x) else ctt(x)\n\n fn_arg_list = (list(fn_arg_list) if is_list_like(fn_arg_list)\n else [fn_arg_list])\n fn_arg_list = _convert_to_tensor(fn_arg_list, 'fn_arg')\n\n if result is None and grads is None and (JAX_MODE or\n not tf.executing_eagerly()):\n # Currently, computing gradient is not working well with caching in\n # tensorflow eager mode (see below), so we will handle that case\n # separately.\n return tfp_math_value_and_gradients(fn, fn_arg_list)\n\n if result is None:\n result = fn(*fn_arg_list)\n if grads is None:\n assert tf.executing_eagerly()\n # Ensure we disable bijector cacheing in eager mode.\n # TODO(b/72831017): Remove this once bijector cacheing is fixed for\n # eager mode.\n fn_arg_list = [0 + x for x in fn_arg_list]\n\n result = _convert_to_tensor(result, 'fn_result')\n\n if grads is not None:\n grads = _convert_to_tensor(grads, 'fn_grad')\n return result, grads\n\n _, grads = tfp_math_value_and_gradients(fn, fn_arg_list)\n\n return result, grads\n\n\ndef maybe_call_fn_and_grads(fn,\n fn_arg_list,\n result=None,\n grads=None,\n check_non_none_grads=True,\n name=None):\n \"\"\"Calls `fn` and computes the gradient of the result wrt `args_list`.\"\"\"\n with tf.name_scope(name or 'maybe_call_fn_and_grads'):\n fn_arg_list = (list(fn_arg_list) if is_list_like(fn_arg_list)\n else [fn_arg_list])\n result, grads = _value_and_gradients(fn, fn_arg_list, result, grads)\n if not all(dtype_util.is_floating(r.dtype)\n for r in (result if is_list_like(result) else [result])): # pylint: disable=superfluous-parens\n raise TypeError('Function result must be a `Tensor` with `float` '\n '`dtype`.')\n if len(fn_arg_list) != len(grads):\n raise ValueError('Function args must be in one-to-one correspondence '\n 'with grads.')\n if check_non_none_grads and any(g is None for g in grads):\n raise ValueError('Encountered `None` gradient.\\n'\n ' fn_arg_list: {}\\n'\n ' grads: {}'.format(fn_arg_list, grads))\n return result, grads\n\n\ndef enable_store_parameters_in_results(kernel):\n \"\"\"Enables the `store_parameters_in_results` parameter in a chain of kernels.\n\n This is a temporary utility for use during the transition period of the\n parameter storage methods.\n\n Args:\n kernel: A TransitionKernel.\n\n Returns:\n kernel: The same kernel, but recreated with `store_parameters_in_results`\n recursively set to `True` in its parameters and its inner kernels (as\n appropriate).\n \"\"\"\n kernel_stack = []\n while hasattr(kernel, 'parameters') and 'inner_kernel' in kernel.parameters:\n kernel_stack.append(kernel)\n kernel = kernel.parameters['inner_kernel']\n\n def _recreate_kernel(kernel, parameters):\n new_parameters = kernel.parameters.copy()\n new_parameters.update(parameters)\n if 'store_parameters_in_results' in new_parameters:\n new_parameters['store_parameters_in_results'] = True\n with deprecation.silence():\n return type(kernel)(**new_parameters)\n\n if hasattr(kernel, 'parameters'):\n kernel = _recreate_kernel(kernel, {})\n\n for outer_kernel in reversed(kernel_stack):\n outer_kernel = _recreate_kernel(outer_kernel, {'inner_kernel': kernel})\n kernel = outer_kernel\n\n return kernel\n\n\ndef _is_tensor_like(param):\n if is_list_like(param):\n return all([_is_tensor_like(p) for p in param])\n if isinstance(param, tf.Tensor):\n return True\n elif isinstance(param, tf.Variable):\n return False\n else:\n return np.array(param).dtype != np.object_\n\n\ndef warn_if_parameters_are_not_simple_tensors(params_dict):\n for param_name, param in params_dict.items():\n if not _is_tensor_like(param):\n warnings.warn(\n '`{}` is not a `tf.Tensor`, Python number, or Numpy array. If this '\n 'parameter is mutable (e.g., a `tf.Variable`), then the '\n 'behavior implied by `store_parameters_in_results` will silently '\n 'change on 2019-08-01. Please consult the docstring for '\n '`store_parameters_in_results` details and use '\n '`store_parameters_in_results=True` to silence this warning.'.format(\n param_name))\n\n\ndef index_remapping_gather(params,\n indices,\n axis=0,\n indices_axis=0,\n name='index_remapping_gather'):\n \"\"\"Gather values from `axis` of `params` using `indices_axis` of `indices`.\n\n The shape of `indices` must broadcast to that of `params` when\n their `indices_axis` and `axis` (respectively) are aligned:\n\n ```python\n # params.shape:\n [p[0], ..., ..., p[axis], ..., ..., p[rank(params)] - 1])\n # indices.shape:\n [i[0], ..., i[indices_axis], ..., i[rank(indices)] - 1])\n ```\n\n In particular, `params` must have at least as many\n leading dimensions as `indices` (`axis >= indices_axis`), and at least as many\n trailing dimensions (`rank(params) - axis >= rank(indices) - indices_axis`).\n\n The `result` has the same shape as `params`, except that the dimension\n of size `p[axis]` is replaced by one of size `i[indices_axis]`:\n\n ```python\n # result.shape:\n [p[0], ..., ..., i[indices_axis], ..., ..., p[rank(params) - 1]]\n ```\n\n In the case where `rank(params) == 5`, `rank(indices) == 3`, `axis = 2`, and\n `indices_axis = 1`, the result is given by\n\n ```python\n # alignment is: v axis\n # params.shape == [p[0], p[1], p[2], p[3], p[4]]\n # indices.shape == [i[0], i[1], i[2]]\n # ^ indices_axis\n result[i, j, k, l, m] = params[i, j, indices[j, k, l], l, m]\n ```\n\n Args:\n params: `N-D` `Tensor` (`N > 0`) from which to gather values.\n Number of dimensions must be known statically.\n indices: `Tensor` with values in `{0, ..., params.shape[axis] - 1}`, whose\n shape broadcasts to that of `params` as described above.\n axis: Python `int` axis of `params` from which to gather.\n indices_axis: Python `int` axis of `indices` to align with the `axis`\n over which `params` is gathered.\n name: String name for scoping created ops.\n\n Returns:\n `Tensor` composed of elements of `params`.\n\n Raises:\n ValueError: If shape/rank requirements are not met.\n \"\"\"\n with tf.name_scope(name):\n params = tf.convert_to_tensor(params, name='params')\n indices = tf.convert_to_tensor(indices, name='indices')\n\n params_ndims = tensorshape_util.rank(params.shape)\n indices_ndims = tensorshape_util.rank(indices.shape)\n # `axis` dtype must match ndims, which are 64-bit Python ints.\n axis = tf.get_static_value(ps.convert_to_shape_tensor(axis, dtype=tf.int64))\n indices_axis = tf.get_static_value(\n ps.convert_to_shape_tensor(indices_axis, dtype=tf.int64))\n\n if params_ndims is None:\n raise ValueError(\n 'Rank of `params`, must be known statically. This is due to '\n 'tf.gather not accepting a `Tensor` for `batch_dims`.')\n\n if axis is None:\n raise ValueError(\n '`axis` must be known statically. This is due to '\n 'tf.gather not accepting a `Tensor` for `batch_dims`.')\n\n if indices_axis is None:\n raise ValueError(\n '`indices_axis` must be known statically. This is due to '\n 'tf.gather not accepting a `Tensor` for `batch_dims`.')\n\n if indices_axis > axis:\n raise ValueError(\n '`indices_axis` should be <= `axis`, but was {} > {}'.format(\n indices_axis, axis))\n\n if params_ndims < 1:\n raise ValueError(\n 'Rank of params should be `> 0`, but was {}'.format(params_ndims))\n\n if indices_ndims is not None and indices_ndims < 1:\n raise ValueError(\n 'Rank of indices should be `> 0`, but was {}'.format(indices_ndims))\n\n if (indices_ndims is not None and\n (indices_ndims - indices_axis > params_ndims - axis)):\n raise ValueError(\n '`rank(params) - axis` ({} - {}) must be >= `rank(indices) - '\n 'indices_axis` ({} - {}), but was not.'.format(\n params_ndims, axis, indices_ndims, indices_axis))\n\n # `tf.gather` requires the axis to be the rightmost batch ndim. So, we\n # transpose `indices_axis` to be the rightmost dimension of `indices`...\n transposed_indices = dist_util.move_dimension(indices,\n source_idx=indices_axis,\n dest_idx=-1)\n\n # ... and `axis` to be the corresponding (aligned as in the docstring)\n # dimension of `params`.\n broadcast_indices_ndims = indices_ndims + (axis - indices_axis)\n transposed_params = dist_util.move_dimension(\n params,\n source_idx=axis,\n dest_idx=broadcast_indices_ndims - 1)\n\n # Next we broadcast `indices` so that its shape has the same prefix as\n # `params.shape`.\n transposed_params_shape = ps.shape(transposed_params)\n result_shape = ps.concat([\n transposed_params_shape[:broadcast_indices_ndims - 1],\n ps.shape(indices)[indices_axis:indices_axis + 1],\n transposed_params_shape[broadcast_indices_ndims:]], axis=0)\n broadcast_indices = ps.broadcast_to(\n transposed_indices,\n result_shape[:broadcast_indices_ndims])\n\n result_t = tf.gather(transposed_params,\n broadcast_indices,\n batch_dims=broadcast_indices_ndims - 1,\n axis=broadcast_indices_ndims - 1)\n return dist_util.move_dimension(result_t,\n source_idx=broadcast_indices_ndims - 1,\n dest_idx=axis)\n", "repo_name": "tensorflow/probability", "sub_path": "tensorflow_probability/python/mcmc/internal/util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 18283, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3997, "dataset": "github-code", "pt": "24", "api": [{"api_name": "tensorflow.compat.v2.convert_to_tensor", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 60, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.nest.flatten", "line_number": 100, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.nest", "line_number": 100, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 100, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.nest.flatten", "line_number": 101, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.nest", "line_number": 101, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 101, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.where", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 113, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.broadcast_util.left_justified_expand_dims_like", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.broadcast_util", "line_number": 113, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.nest.pack_sequence_as", "line_number": 115, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.nest", "line_number": 115, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 115, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.name_scope", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 117, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.nest.pack_sequence_as", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.nest", "line_number": 120, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 120, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.name_scope", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 129, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.nest.flatten", "line_number": 155, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.nest", "line_number": 155, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 155, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.nest.flatten", "line_number": 156, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.nest", "line_number": 156, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 156, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.nest.pack_sequence_as", "line_number": 157, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.nest", "line_number": 157, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 157, "usage_type": "name"}, {"api_name": "numpy.inf", "line_number": 190, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2.name_scope", "line_number": 211, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 211, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.add_n", "line_number": 217, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 217, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.where", "line_number": 218, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 218, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.is_finite", "line_number": 218, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 218, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2.constant", "line_number": 218, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.tensorshape_util.set_shape", "line_number": 219, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.tensorshape_util", "line_number": 219, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.name_scope", "line_number": 233, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 233, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.convert_to_tensor", "line_number": 236, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 236, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.executing_eagerly", "line_number": 245, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 245, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.math.gradient.value_and_gradient", "line_number": 249, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.executing_eagerly", "line_number": 254, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 254, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.math.gradient.value_and_gradient", "line_number": 266, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.name_scope", "line_number": 278, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 278, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.dtype_util.is_floating", "line_number": 282, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.dtype_util", "line_number": 282, "usage_type": "name"}, {"api_name": "tensorflow.python.util.deprecation.silence", "line_number": 320, "usage_type": "call"}, {"api_name": "tensorflow.python.util.deprecation", "line_number": 320, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.Tensor", "line_number": 336, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 336, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.Variable", "line_number": 338, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 338, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 341, "usage_type": "call"}, {"api_name": "numpy.object_", "line_number": 341, "usage_type": "attribute"}, {"api_name": "warnings.warn", "line_number": 347, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.name_scope", "line_number": 413, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 413, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.convert_to_tensor", "line_number": 414, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 414, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.convert_to_tensor", "line_number": 415, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 415, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.tensorshape_util.rank", "line_number": 417, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.tensorshape_util", "line_number": 417, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.tensorshape_util.rank", "line_number": 418, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.tensorshape_util", "line_number": 418, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.get_static_value", "line_number": 420, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 420, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.prefer_static.convert_to_shape_tensor", "line_number": 420, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.prefer_static", "line_number": 420, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.int64", "line_number": 420, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2.get_static_value", "line_number": 421, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 421, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.prefer_static.convert_to_shape_tensor", "line_number": 422, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.prefer_static", "line_number": 422, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.int64", "line_number": 422, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 422, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.distribution_util.move_dimension", "line_number": 461, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.distribution_util", "line_number": 461, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.distribution_util.move_dimension", "line_number": 468, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.distribution_util", "line_number": 468, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.prefer_static.shape", "line_number": 475, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.prefer_static", "line_number": 475, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.prefer_static.concat", "line_number": 476, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.prefer_static", "line_number": 476, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.prefer_static.shape", "line_number": 478, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.prefer_static", "line_number": 478, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.prefer_static.broadcast_to", "line_number": 480, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.prefer_static", "line_number": 480, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.gather", "line_number": 484, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 484, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.distribution_util.move_dimension", "line_number": 488, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.distribution_util", "line_number": 488, "usage_type": "name"}]}
+{"seq_id": "6134198425", "text": "#using CNN with edge detection + contours\n\n\nimport tensorflow as tf\nimport numpy as np\nimport cv2\nimport time\nfrom tensorflow import keras\nfrom tensorflow.keras import layers\nfrom tensorflow.keras.models import Sequential\nfrom b_serial import sendColor\nimport json\nimport pathlib\nprint(tf.__version__)\n# colors 0 red, 1 green, 2, yellow, 3, emergency\nfrom keras.models import load_model\nmodel = load_model(r'models/vehicles_light_new.h5')\nSerialDump=dict()\nimport threading\n\nminArea=10000\nfilepath=r\"D:\\Python\\SmartVehicleGuidance\\testImages\\test_new\\test1.jpg\"\nimg_height = 120\nimg_width = 180\nmodel.summary()\n\ndef nothing(value):\n pass\n \nclass_names = ['car', 'otherStuff', 'truck']\n\nlane1= [0 , 250]\n\nlaneColors=[1,1,1]\ncolorsChanged=False\nready=True\nimg=cv2.imread(filepath)\n\ndef colorUpdate(SerialDump):\n y=json.dumps(SerialDump)\n sendColor(y)\n time.sleep(1)\n colorsChanged=False\n ready=True\n\n#Uncomment to use Live Video Feed\ncap= cv2.VideoCapture(1)\nOnce=True\nt1 = threading.Thread(target=colorUpdate, args=(SerialDump,))\niterations_Dilate=0\niterations_Erode=0\n\n\n\ncv2.namedWindow(\"Settings\", cv2.WINDOW_AUTOSIZE)\ncv2.createTrackbar('MinArea', \"Settings\", 5, 100, nothing)\ncv2.createTrackbar('Dilation', \"Settings\", 0, 5, nothing)\ncv2.createTrackbar('Erosion', \"Settings\", 0, 5, nothing)\ncv2.createTrackbar('Brightness', \"Settings\", 0, 200, nothing)\ncv2.createTrackbar('Blur', \"Settings\", 1, 9, nothing)\ncv2.createTrackbar('TopLine', \"Settings\", 0, 50, nothing)\ncv2.createTrackbar('FirstLane', \"Settings\", 100, 300, nothing)\n\n\nprevColors=[0,0,0]\nwhile Once:\n start=time.time()\n #Uncomment to use a single file\n #img=cv2.imread(filepath)\n\n #Uncomment to use live video Feed\n _,img=cap.read()\n \n img= cv2.resize(img, (600,600),interpolation=cv2.INTER_AREA)\n gray=cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)\n matrix = np.ones(img.shape, dtype = \"uint8\") * cv2.getTrackbarPos('Brightness','Settings')\n blurIndex=cv2.getTrackbarPos('Blur','Settings')\n if blurIndex>0:\n img = cv2.blur(img,(blurIndex,blurIndex))\n l1Endpoint=cv2.getTrackbarPos('FirstLane','Settings')\n lane1[1]=l1Endpoint \n y=cv2.getTrackbarPos('TopLine','Settings')\n img= cv2.line(img, (l1Endpoint,500),(l1Endpoint,y), (255,255,255), 5)\n img= cv2.line(img, (0,y),(500,y), (255,255,255), 5)\n kernel = np.array([[-1,-1,-1], \n [-1,9,-1], \n [-1,-1,-1]])\n #img = cv2.filter2D(img, -1, kernel)\n \n img = cv2.add(img, matrix)\n canny= cv2.Canny(gray, 50, 100)\n kernel = np.ones((3,3), np.uint8)\n kernel2 = np.ones((3,3), np.uint8)\n \n \n dilated = cv2.dilate(canny, kernel2, iterations=cv2.getTrackbarPos('Dilation','Settings'))\n eroded = cv2.erode(dilated, kernel, iterations=cv2.getTrackbarPos('Erosion','Settings'))\n eroded[y+1]=0\n eroded[y]=0\n eroded[y-1]=0\n \n cv2.imshow(\"binary\", eroded)\n contours, hierarchies= cv2.findContours(eroded, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE)\n \n for i,contour in enumerate(contours):\n if (hierarchies[0][i][3])<0:\n (x,y,w,h) = cv2.boundingRect(contour)\n if w*h > (cv2.getTrackbarPos('MinArea','Settings')*1000):\n d_object=img[y:y+h,x:x+w]\n d_object=cv2.resize(d_object, (img_width,img_height),interpolation=cv2.INTER_AREA)\n d_object=cv2.cvtColor(d_object, cv2.COLOR_BGR2RGB)\n img_array = tf.expand_dims(d_object, 0)\n predictions = model.predict(img_array)\n score = tf.nn.softmax(predictions[0])\n cv2.rectangle(img, (x,y), (x+w,y+h), (255, 0, 0), 2)\n objectDetected= class_names[np.argmax(score)]\n \n img = cv2.putText(img, \"{} {:.2f}\".format(class_names[np.argmax(score)], 100 * np.max(score)), (x,y+15), cv2.FONT_HERSHEY_SIMPLEX, 1, (0,0,255),2, cv2.LINE_AA)\n if objectDetected==\"truck\":\n if ((x+w)/2)>lane1[0] and ((x+w)/2)lane1[0] and ((x+w)/2) threshold else 0\n return avg_loss, score\n\n\n def compute_L2_loss(self, predict_points, label_points, threshold = 0.):\n # label = rect.rectify(label_points)\n label = label_points\n predicted_label = rect.rectify(predict_points)\n\n diff = label - predicted_label\n \n square_diff = diff**2\n dists = square_diff.sum(axis=1, dtype=np.int32)\n dists = np.sqrt(dists)\n avg_loss = np.sum(dists)/4\n\n score = 1 if avg_loss > threshold else 0\n return avg_loss, score\n\n def compute_overlap(self, predict_points, label_points, threshold = 0.):\n # label_points = rect.rectify(label_points)\n # predict_points = rect.rectify(predict_points)\n\n predicted_polygon = Polygon(predict_points)\n labeled_polygon = Polygon(label_points)\n # print(predicted_polygon)\n # print(labeled_polygon)\n intersection = predicted_polygon.intersection(labeled_polygon)\n union = predicted_polygon.union(labeled_polygon)\n \n overlap = intersection.area / union.area\n\n score = 1 if overlap > threshold else 0\n return overlap, score\n \n\nif __name__ == '__main__':\n fake_labeled_points = np.array([[0, 0], [0, 2], [2, 0], [2, 2]])\n fake_predicted_points = np.array([[1, 1], [1, 3], [3, 1], [3, 3]])\n\n evaluator = Evaluator(sys.argv[1])\n\n print( 'L1_loss = {}'.format(evaluator.error_func['L1'](fake_labeled_points,\n fake_predicted_points)) )\n\n print( 'L2_loss = {}'.format(evaluator.error_func['L2'](fake_labeled_points,\n fake_predicted_points)) )\n\n print( 'overlap = {}'.format(evaluator.error_func['overlap'](fake_labeled_points,\n fake_predicted_points)) )\n\n\n", "repo_name": "ZhengPhoenix/ObjectDetect", "sub_path": "Evaluation/Evaluator.py", "file_name": "Evaluator.py", "file_ext": "py", "file_size_in_byte": 3209, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "rect.rectify", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 41, "usage_type": "call"}, {"api_name": "rect.rectify", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 55, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 57, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 66, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 81, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 83, "usage_type": "attribute"}]}
+{"seq_id": "22239995073", "text": "import itertools\n# Dependency imports\nfrom absl.testing import parameterized\nimport numpy as np\nimport tensorflow.compat.v2 as tf\n\nfrom tensorflow_probability.python.bijectors import ascending as tfb\nfrom tensorflow_probability.python.distributions import categorical\nfrom tensorflow_probability.python.distributions import kullback_leibler\nfrom tensorflow_probability.python.distributions import logistic\nfrom tensorflow_probability.python.distributions import ordered_logistic as ol\nfrom tensorflow_probability.python.internal import test_util\n\n\n@test_util.test_all_tf_execution_regimes\nclass OrderedLogisticTest(test_util.TestCase):\n\n def _random_cutpoints(self, shape):\n return self._ordered.forward(self._rng.randn(*shape))\n\n def _random_location(self, shape):\n return self._rng.randn(*shape)\n\n def setUp(self):\n self._ordered = tfb.Ascending()\n self._rng = test_util.test_np_rng()\n super(OrderedLogisticTest, self).setUp()\n\n @parameterized.parameters(\n itertools.product(['cutpoints', 'loc', 'both'], [[], [1], [1, 2, 3]])\n )\n def testBatchShapes(self, test, batch_shape):\n\n if test == 'cutpoints':\n cutpoints = self._random_cutpoints(batch_shape + [2])\n loc = tf.constant(0., dtype=tf.float64)\n elif test == 'loc':\n cutpoints = tf.constant([1., 2.], dtype=tf.float64)\n loc = self._random_location(batch_shape)\n elif test == 'both':\n cutpoints = self._random_cutpoints(batch_shape + [2])\n loc = self._random_location(batch_shape)\n\n dist = ol.OrderedLogistic(cutpoints=cutpoints, loc=loc)\n\n self.assertAllEqual(dist.batch_shape, batch_shape)\n self.assertAllEqual(\n self.evaluate(dist.batch_shape_tensor()), batch_shape)\n\n self.assertAllEqual(dist.event_shape, [])\n self.assertAllEqual(self.evaluate(dist.event_shape_tensor()), [])\n\n categorical_probs = dist.categorical_probs()\n categorical_probs_shape = tf.shape(categorical_probs)\n self.assertAllEqual(\n self.evaluate(categorical_probs_shape), batch_shape + [3])\n\n sample = dist.sample(seed=test_util.test_seed())\n sample_shape = tf.shape(sample)\n self.assertAllEqual(self.evaluate(sample_shape), batch_shape)\n\n prob_sample_shape = tf.shape(dist.prob(sample))\n survival_prob_sample_shape = tf.shape(dist.survival_function(sample))\n self.assertAllEqual(self.evaluate(prob_sample_shape), batch_shape)\n self.assertAllEqual(self.evaluate(survival_prob_sample_shape), batch_shape)\n\n n = [4, 5]\n sample_n = dist.sample(n, seed=test_util.test_seed())\n sample_n_shape = tf.shape(sample_n)\n self.assertAllEqual(self.evaluate(sample_n_shape), n + batch_shape)\n\n prob_sample_n_shape = tf.shape(dist.prob(sample_n))\n survival_prob_sample_n_shape = tf.shape(dist.survival_function(sample_n))\n self.assertAllEqual(self.evaluate(prob_sample_n_shape), n + batch_shape)\n self.assertAllEqual(\n self.evaluate(survival_prob_sample_n_shape), n + batch_shape)\n\n def testProbs(self):\n # survival functions\n # P(Y > 0) = sigmoid(1) = 0.7310586\n # P(Y > 1) = sigmoid(0) = 0.5\n # P(Y > 2) = sigmoid(-1) = 0.26894143\n\n # probs\n # P(Y = 0) = 1. - sigmoid(1) = 0.2689414\n # P(Y = 1) = sigmoid(1) - sigmoid(0) = 0.2310586\n # P(Y = 2) = sigmoid(0) - sigmoid(-1) = 0.23105857\n # P(Y = 3) = sigmoid(-1) = 0.26894143\n expected_probs = [0.2689414, 0.2310586, 0.23105857, 0.26894143]\n expected_survival_probs = 1. - np.cumsum(expected_probs)\n dist = ol.OrderedLogistic(cutpoints=[-1., 0., 1.], loc=0.)\n\n categorical_probs = self.evaluate(dist.categorical_probs())\n self.assertAllClose(expected_probs, categorical_probs, atol=1e-6)\n\n probs = np.flip(self.evaluate(dist.prob([3, 2, 1, 0])))\n self.assertAllClose(expected_probs, probs, atol=1e-6)\n\n survival_probs = self.evaluate(dist.survival_function([0, 1, 2, 3]))\n self.assertAllClose(expected_survival_probs, survival_probs, atol=1e-6)\n\n zero_probs = self.evaluate(dist.prob([-1, 4]))\n self.assertAllClose([0., 0.], zero_probs, atol=1e-6)\n\n out_of_bounds_survival_probs = self.evaluate(\n dist.survival_function([-2, -1, 4, 5]))\n self.assertAllClose(\n [1., 1., 0., 0.], out_of_bounds_survival_probs, atol=1e-6)\n\n def testMode(self):\n # 2 cutpoints i.e. 3 possible outcomes. 3 \"batched\" distributions with the\n # logistic distribution location well within the large cutpoint spacing so\n # mode is obvious\n dist = ol.OrderedLogistic(cutpoints=[-10., 10.], loc=[-20., 0., 20.])\n mode = self.evaluate(dist.mode())\n self.assertAllEqual([0, 1, 2], mode)\n\n def testSample(self):\n # as per `testProbs`\n dist = ol.OrderedLogistic(cutpoints=[-1., 0., 1.], loc=0.)\n samples = self.evaluate(dist.sample(int(1e5), seed=test_util.test_seed()))\n expected_probs = [0.2689414, 0.2310586, 0.23105857, 0.26894143]\n for k, p in enumerate(expected_probs):\n self.assertAllClose(np.mean(samples == k), p, atol=0.01)\n\n def testEntropyAgainstCategoricalDistribution(self):\n cutpoints = self._random_cutpoints([3])\n loc = self._random_location([2])\n dist = ol.OrderedLogistic(cutpoints=cutpoints, loc=loc)\n categorical_dist = categorical.Categorical(dist.categorical_log_probs())\n expected_entropy = self.evaluate(categorical_dist.entropy())\n entropy = self.evaluate(dist.entropy())\n self.assertAllClose(expected_entropy, entropy)\n\n def testEntropyAgainstSampling(self):\n cutpoints = self._random_cutpoints([4])\n loc = self._random_location([])\n dist = ol.OrderedLogistic(cutpoints=cutpoints, loc=loc)\n samples = dist.sample(int(1e5), seed=test_util.test_seed())\n entropy_samples = self.evaluate(-dist.log_prob(samples))\n entropy = self.evaluate(dist.entropy())\n self.assertAllMeansClose(entropy_samples, entropy, axis=0, atol=0.01)\n\n @parameterized.parameters(1, 10, 25)\n def testKLAgainstCategoricalDistribution(self, batch_size):\n cutpoints = self._random_cutpoints([100])\n a_loc = self._random_location([batch_size])\n b_loc = self._random_location([batch_size])\n\n a = ol.OrderedLogistic(cutpoints=cutpoints, loc=a_loc, validate_args=True)\n b = ol.OrderedLogistic(cutpoints=cutpoints, loc=b_loc, validate_args=True)\n\n a_cat = categorical.Categorical(\n logits=a.categorical_log_probs(), validate_args=True)\n b_cat = categorical.Categorical(\n logits=b.categorical_log_probs(), validate_args=True)\n\n kl = self.evaluate(kullback_leibler.kl_divergence(a, b))\n self.assertEqual(kl.shape, (batch_size,))\n\n kl_expected = self.evaluate(kullback_leibler.kl_divergence(a_cat, b_cat))\n self.assertAllClose(kl, kl_expected)\n\n kl_same = self.evaluate(kullback_leibler.kl_divergence(a, a))\n self.assertAllClose(kl_same, np.zeros_like(kl_expected))\n\n def testKLAgainstSampling(self):\n a_cutpoints = self._random_cutpoints([4])\n b_cutpoints = self._random_cutpoints([4])\n loc = self._random_location([])\n\n a = ol.OrderedLogistic(cutpoints=a_cutpoints, loc=loc)\n b = ol.OrderedLogistic(cutpoints=b_cutpoints, loc=loc)\n\n samples = a.sample(int(1e5), seed=test_util.test_seed())\n kl_samples = self.evaluate(a.log_prob(samples) - b.log_prob(samples))\n kl = self.evaluate(kullback_leibler.kl_divergence(a, b))\n\n self.assertAllMeansClose(kl_samples, kl, axis=0, atol=2e-2)\n\n def testLatentLogistic(self):\n loc = self._random_location([2])\n cutpoints = self._random_cutpoints([2])\n latent = logistic.Logistic(loc=loc, scale=1.)\n ordered = ol.OrderedLogistic(cutpoints=cutpoints, loc=loc)\n ordered_cdf = self.evaluate(ordered.cdf([0, 1]))\n latent_cdf = self.evaluate(latent.cdf(cutpoints))\n self.assertAllClose(ordered_cdf, latent_cdf)\n\n def testUnorderedCutpointsFails(self):\n with self.assertRaisesRegexp(\n ValueError, 'Argument `cutpoints` must be non-decreasing.'):\n dist = ol.OrderedLogistic(\n cutpoints=[1., 0.9], loc=0.0, validate_args=True)\n self.evaluate(dist.mode())\n\nif __name__ == '__main__':\n test_util.main()\n", "repo_name": "tensorflow/probability", "sub_path": "tensorflow_probability/python/distributions/ordered_logistic_test.py", "file_name": "ordered_logistic_test.py", "file_ext": "py", "file_size_in_byte": 8003, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3997, "dataset": "github-code", "pt": "24", "api": [{"api_name": "tensorflow_probability.python.internal.test_util.TestCase", "line_number": 16, "usage_type": "attribute"}, {"api_name": "tensorflow_probability.python.internal.test_util", "line_number": 16, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.bijectors.ascending.Ascending", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.bijectors.ascending", "line_number": 25, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.test_util.test_np_rng", "line_number": 26, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.test_util", "line_number": 26, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.constant", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 36, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.float64", "line_number": 36, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2.constant", "line_number": 38, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 38, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.float64", "line_number": 38, "usage_type": "attribute"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic.OrderedLogistic", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic", "line_number": 44, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.shape", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 54, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.test_util.test_seed", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.test_util", "line_number": 58, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.shape", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 59, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.shape", "line_number": 62, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 62, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.shape", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 63, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.test_util.test_seed", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.test_util", "line_number": 68, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.shape", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 69, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.shape", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 72, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.shape", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 73, "usage_type": "name"}, {"api_name": "absl.testing.parameterized.parameters", "line_number": 29, "usage_type": "call"}, {"api_name": "absl.testing.parameterized", "line_number": 29, "usage_type": "name"}, {"api_name": "itertools.product", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic.OrderedLogistic", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic", "line_number": 91, "usage_type": "name"}, {"api_name": "numpy.flip", "line_number": 96, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic.OrderedLogistic", "line_number": 114, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic", "line_number": 114, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic.OrderedLogistic", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic", "line_number": 120, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.test_util.test_seed", "line_number": 121, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.test_util", "line_number": 121, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 124, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic.OrderedLogistic", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic", "line_number": 129, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.categorical.Categorical", "line_number": 130, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.categorical", "line_number": 130, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic.OrderedLogistic", "line_number": 138, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic", "line_number": 138, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.test_util.test_seed", "line_number": 139, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.test_util", "line_number": 139, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic.OrderedLogistic", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic", "line_number": 150, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic.OrderedLogistic", "line_number": 151, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic", "line_number": 151, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.categorical.Categorical", "line_number": 153, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.categorical", "line_number": 153, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.categorical.Categorical", "line_number": 155, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.categorical", "line_number": 155, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.kullback_leibler.kl_divergence", "line_number": 158, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.kullback_leibler", "line_number": 158, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.kullback_leibler.kl_divergence", "line_number": 161, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.kullback_leibler", "line_number": 161, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.kullback_leibler.kl_divergence", "line_number": 164, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.kullback_leibler", "line_number": 164, "usage_type": "name"}, {"api_name": "numpy.zeros_like", "line_number": 165, "usage_type": "call"}, {"api_name": "absl.testing.parameterized.parameters", "line_number": 144, "usage_type": "call"}, {"api_name": "absl.testing.parameterized", "line_number": 144, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic.OrderedLogistic", "line_number": 172, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic", "line_number": 172, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic.OrderedLogistic", "line_number": 173, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic", "line_number": 173, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.test_util.test_seed", "line_number": 175, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.test_util", "line_number": 175, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.kullback_leibler.kl_divergence", "line_number": 177, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.kullback_leibler", "line_number": 177, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.logistic.Logistic", "line_number": 184, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.logistic", "line_number": 184, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic.OrderedLogistic", "line_number": 185, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic", "line_number": 185, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic.OrderedLogistic", "line_number": 193, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.ordered_logistic", "line_number": 193, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.test_util.test_all_tf_execution_regimes", "line_number": 15, "usage_type": "attribute"}, {"api_name": "tensorflow_probability.python.internal.test_util", "line_number": 15, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.test_util.main", "line_number": 198, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.test_util", "line_number": 198, "usage_type": "name"}]}
+{"seq_id": "26639298985", "text": "import sys\nfrom PyQt5.QtWidgets import QWidget, QProgressBar, QVBoxLayout, QPushButton,QApplication, QLabel\nfrom PyQt5.QtCore import Qt\n\n\n\nclass ProgressWindow(QWidget):\n def __init__(self):\n super().__init__()\n self.setWindowFlag(Qt.FramelessWindowHint)\n self.setWindowModality(Qt.ApplicationModal)\n self.setWindowFlag(Qt.Tool) #hides window from taskbar\n\n layout = QVBoxLayout()\n\n self.label = QLabel(\"Please wait until the computation is completed.\")\n self.label.setStyleSheet('font-size: 24px; height: 24px;')\n layout.addWidget(self.label)\n\n self.label2 = QLabel()\n self.label2.setStyleSheet('font-size: 12px; height: 12px;')\n layout.addWidget(self.label2)\n\n self.progressBar = QProgressBar()\n self.progressBar.setMinimum(0)\n self.progressBar.setValue(0)\n layout.addWidget(self.progressBar)\n\n self.setLayout(layout)\n\n\n def increasebar_by(self, value):\n self.progressBar.setValue(self.progressBar.value()+value)\n if self.progressBar.value() >= self.progressBar.maximum(): self.close()\n\n def updatetext(self, text):\n self.label2.setText(self.label2.text() + '\\n' + text)\n \n def closeself(self):\n self.close()\n \n\n#if __name__ == \"__main__\":\n# app = QApplication(sys.argv)\n# MainWindow = ProgressWindow()\n# MainWindow.show()\n # sys.exit(app.exec_())", "repo_name": "HSPR-Software/HSPR", "sub_path": "src/ui/computation_progressBar.py", "file_name": "computation_progressBar.py", "file_ext": "py", "file_size_in_byte": 1426, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3, "dataset": "github-code", "pt": "24", "api": [{"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 7, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.FramelessWindowHint", "line_number": 10, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 10, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.ApplicationModal", "line_number": 11, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 11, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.Tool", "line_number": 12, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 12, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 14, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 16, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 20, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QProgressBar", "line_number": 24, "usage_type": "call"}]}
+{"seq_id": "30718987960", "text": "import argparse\nimport pickle\n\nimport torch\nfrom datasets import load_dataset\nfrom tqdm import tqdm\n\nfrom AttentionsAnalysis.AnalysisGenerator import AnalysisGenerator\nfrom DataModels.DataType import DataType\nfrom DataModels.ModelMetadata import ModelMetadata\nfrom DataModels.Sample import Sample\n\n\nclass AttentionsDataCreator:\n \"\"\"\n This class is used to create pipeline of comparisons with standard deviation between attentions of two models.\n \"\"\"\n\n def __init__(self, model1_metadata: ModelMetadata, model2_metadata: ModelMetadata, use_dummy_dataset: bool = False,\n start_example=None, end_example=None,\n metric: str = \"Cosine\"):\n self.model1_metadata = model1_metadata\n self.model2_metadata = model2_metadata\n self.metric = metric\n self.device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n self.analysis_generator = AnalysisGenerator(model1_metadata, model2_metadata, metric=metric)\n self.start_example = start_example\n self.end_example = end_example\n self.dataset = self.load_dummy_dataset() if use_dummy_dataset else self.load_dataset(self.start_example,\n self.end_example)\n\n def get_correlations_attentions_comparisons(self, sample1: Sample, sample2: Sample):\n attention_model1, attention_model2 = self.analysis_generator.get_attentions(sample1, sample2)\n correlations_attentions_comparisons = self.analysis_generator.get_correlations_of_attentions(attention_model1,\n attention_model2)\n return correlations_attentions_comparisons\n\n def run(self):\n correlations = {}\n for i in tqdm(range(len(self.dataset))):\n sample1 = Sample(id=self.dataset[i][\"id\"], text=self.dataset[i][\"text\"], audio=self.dataset[i][\"audio\"])\n sample2 = sample1\n try:\n correlations_attentions_comparisons = self.get_correlations_attentions_comparisons(\n sample1, sample2)\n correlations[sample1.id] = correlations_attentions_comparisons\n except AssertionError as e:\n example = f\"{sample1.text}\" if sample1.text == sample2.text else f\"{sample1.text} and {sample2.text}\"\n print(f\"Failed to calculate for sample {example}\")\n # save results to pickle file\n self.save_correlations(correlations)\n\n def load_dataset(self, start_example=None, end_example=None):\n if start_example is not None and end_example is not None:\n dataset = load_dataset(\"librispeech_asr\", 'clean', split=f'validation[{start_example}:{end_example}]')\n else:\n dataset = load_dataset(\"librispeech_asr\", 'clean', split='validation')\n return dataset\n\n def load_dummy_dataset(self):\n dataset = load_dataset(\"patrickvonplaten/librispeech_asr_dummy\", 'clean', split='validation')\n return dataset\n\n def save_correlations(self, correlations):\n model_name1 = self.model1_metadata.model_name\n model_name2 = self.model2_metadata.model_name\n # if the model name is 'facebook/wav2vec2-base-960h' we need to remove the '/' from the name\n if '/' in model_name1:\n model_name1 = model_name1.replace('/', '')\n if '/' in model_name2:\n model_name2 = model_name2.replace('/', '')\n path = f'correlations_for_{model_name1}_and_{model_name2}_{self.start_example}_{self.end_example}_{self.metric}.pkl'\n with open(path, 'wb') as handle:\n pickle.dump(correlations, handle)\n\n\nif __name__ == \"__main__\":\n # add arguments from command line with argparse\n parser = argparse.ArgumentParser()\n parser.add_argument(\"--experiment_name\", type=str, choices=[\"text_to_text\", \"audio_to_audio\", \"text_to_audio\"],\n default=\"text_to_audio\", help=\"The name of the experiment to run\")\n parser.add_argument(\"--use_dummy_dataset\", type=bool, default=False,\n help=\"Whether to use a dummy dataset for the experiment\")\n parser.add_argument(\"--start_example\", type=int, default=0)\n parser.add_argument(\"--end_example\", type=int, default=2703)\n parser.add_argument(\"--metric\", type=str, default='jaccard',\n help=\"Which metric to use\")\n args = parser.parse_args()\n\n if args.experiment_name == \"text_to_text\":\n model1_metadata = ModelMetadata(model_name=\"bert-base-uncased\", data_type=DataType.Text,\n align_tokens_to_bert_tokens=False, use_cls_and_sep=True)\n model2_metadata = ModelMetadata(model_name=\"roberta-base\", data_type=DataType.Text,\n align_tokens_to_bert_tokens=False, use_cls_and_sep=True)\n\n elif args.experiment_name == \"text_to_audio\":\n model1_metadata = ModelMetadata(model_name=\"bert-base-uncased\", data_type=DataType.Text,\n align_tokens_to_bert_tokens=False, use_cls_and_sep=True)\n model2_metadata = ModelMetadata(model_name=\"facebook/wav2vec2-base-960h\", data_type=DataType.Audio,\n align_tokens_to_bert_tokens=True, use_cls_and_sep=True)\n\n elif args.experiment_name == \"audio_to_audio\":\n raise Exception(\"Experiment audio_to_audio currently not supported\")\n\n else:\n raise Exception(\"Experiment name is not valid\")\n\n attention_similarity = AttentionsDataCreator(model1_metadata, model2_metadata,\n use_dummy_dataset=args.use_dummy_dataset,\n start_example=args.start_example, end_example=args.end_example,\n metric=args.metric)\n\n attention_similarity.run()\n", "repo_name": "eliyahabba/SoundOfAttention", "sub_path": "AttentionsStatisticsExperiments/CorrelationsAttentionsDataCreator.py", "file_name": "CorrelationsAttentionsDataCreator.py", "file_ext": "py", "file_size_in_byte": 5948, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "DataModels.ModelMetadata.ModelMetadata", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 25, "usage_type": "attribute"}, {"api_name": "AttentionsAnalysis.AnalysisGenerator.AnalysisGenerator", "line_number": 26, "usage_type": "call"}, {"api_name": "DataModels.Sample.Sample", "line_number": 32, "usage_type": "name"}, {"api_name": "tqdm.tqdm", "line_number": 40, "usage_type": "call"}, {"api_name": "DataModels.Sample.Sample", "line_number": 41, "usage_type": "call"}, {"api_name": "datasets.load_dataset", "line_number": 55, "usage_type": "call"}, {"api_name": "datasets.load_dataset", "line_number": 57, "usage_type": "call"}, {"api_name": "datasets.load_dataset", "line_number": 61, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 74, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 79, "usage_type": "call"}, {"api_name": "DataModels.ModelMetadata.ModelMetadata", "line_number": 91, "usage_type": "call"}, {"api_name": "DataModels.DataType.DataType.Text", "line_number": 91, "usage_type": "attribute"}, {"api_name": "DataModels.DataType.DataType", "line_number": 91, "usage_type": "name"}, {"api_name": "DataModels.ModelMetadata.ModelMetadata", "line_number": 93, "usage_type": "call"}, {"api_name": "DataModels.DataType.DataType.Text", "line_number": 93, "usage_type": "attribute"}, {"api_name": "DataModels.DataType.DataType", "line_number": 93, "usage_type": "name"}, {"api_name": "DataModels.ModelMetadata.ModelMetadata", "line_number": 97, "usage_type": "call"}, {"api_name": "DataModels.DataType.DataType.Text", "line_number": 97, "usage_type": "attribute"}, {"api_name": "DataModels.DataType.DataType", "line_number": 97, "usage_type": "name"}, {"api_name": "DataModels.ModelMetadata.ModelMetadata", "line_number": 99, "usage_type": "call"}, {"api_name": "DataModels.DataType.DataType.Audio", "line_number": 99, "usage_type": "attribute"}, {"api_name": "DataModels.DataType.DataType", "line_number": 99, "usage_type": "name"}]}
+{"seq_id": "25088368821", "text": "########### Kismet to Wigle CSV Conversion Scipt ###########\n\nimport os \nimport pathlib\nimport requests\n\n# Define path of .kismet files\npath = pathlib.Path(__file__).parent.resolve()\nfiles = os.listdir(path)\n\n# Define .kismet file list\nKismet = []\n\n# Append all .kismet files to list \nfor file in files:\n length = len(file)\n len1 = length - 7\n ext = file[len1:length]\n if ext == '.kismet':\n Kismet.append(file)\n\ntotal = len(Kismet)\ni = 0 \n\n# Convert Kismet to Wigle CSV\nfor kis in Kismet:\n i = i + 1\n command = 'kismetdb_to_wiglecsv --in ' + kis + ' --out ' + kis[0:len(kis) - 7] + '.csv'\n print(str(i) + '/' + str(total))\n print(\"Converting *\" + kis + '*')\n os.system(command)\n print('Done!' + '\\n')\n\n# Upload CSV Files to Wigle\ndef upload_files():\n url = 'https://api.wigle.net/api/v2/file/upload' \n path = pathlib.Path(__file__).parent.resolve()\n files = os.listdir(path)\n\n x = 0\n csv = []\n for file in files:\n length = len(file)\n len1 = length - 4\n ext = file[len1:length]\n if ext == '.csv':\n csv.append(file)\n\n Total = len(csv)\n\n for file_csv in csv:\n x = x + 1\n data = {'file': (file_csv, open(file_csv, 'rb'), 'multipart/form-data', {'Expires': '0'})}\n print(str(x) + '/' + str(Total))\n print('Uploading *' + file_csv + '*')\n r = requests.post(url, files = data, headers = {'Authorization': 'Basic YOUR API TOKEN HERE'})\n status = r.status_code\n if status == 200:\n print('Done!' + '\\n')\n elif status != 200:\n print('Failed...')\n print('Status Code: ' + str(status) + '\\n')\n\ndef ask_upload():\n upload = input('Upload CSVs to Wigle?(y/n)')\n\n if upload == 'Y' or upload == 'N':\n print('\\n' + 'Please enter y or n')\n ask_upload()\n elif upload == 'y':\n upload_files()\n return upload\n\nask_upload()\n", "repo_name": "TWinston-66/Kismet-to-Wigle-CSV", "sub_path": "Kismet_To_CSV.py", "file_name": "Kismet_To_CSV.py", "file_ext": "py", "file_size_in_byte": 1921, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "pathlib.Path", "line_number": 8, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 9, "usage_type": "call"}, {"api_name": "os.system", "line_number": 31, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 37, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 38, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 56, "usage_type": "call"}]}
+{"seq_id": "30954817750", "text": "import pytest\n\nfrom .solution import frequency_sort\n\n\ndef checkio_main(items, solution):\n assert frequency_sort(items) == solution\n\n\n@pytest.mark.parametrize(\n \"items, solution\",\n [\n ([4, 6, 2, 2, 6, 4, 4, 4], [4, 4, 4, 4, 6, 6, 2, 2]),\n ([\"bob\", \"bob\", \"carl\", \"alex\", \"bob\"], [\"bob\", \"bob\", \"bob\", \"carl\", \"alex\"]),\n ([17, 99, 42], [17, 99, 42]),\n ([], []),\n ([1], [1]),\n ],\n)\ndef test(items, solution):\n checkio_main(items, solution)\n", "repo_name": "LachlanMarnham/competitive_programming", "sub_path": "checkio/home/sort_array_by_element_frequency/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 487, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "solution.frequency_sort", "line_number": 7, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 10, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 10, "usage_type": "attribute"}]}
+{"seq_id": "21052505242", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[17]:\n\n\nimport pandas as pd\nimport matplotlib.pyplot as plt\nget_ipython().run_line_magic('matplotlib', 'inline')\nimport plotly.offline as pyo\nimport plotly.graph_objs as go\nimport plotly.express as px\nimport plotly.io as pio\nimport chart_studio\nimport chart_studio.plotly as py\nimport chart_studio.tools as tls\n\nusername ='yashkaransingh'\napi_key='sTvu3qg24uYodX2dC47M'\n#set credentials to import to plotly\nchart_studio.tools.set_credentials_file(username= username, api_key= api_key)\nchart_studio.tools.set_config_file(sharing='public')\n\n\n# In[18]:\n\n\ncg= pd.read_csv(\"cgl2-edited.csv\")\ncg.columns\ncg.head\n\n\n# In[19]:\n\n\nlength =cg.LENGTH\ntcg=cg.T_CG_COAT\nbcg=cg.B_CG_COAT\n\n\n# In[20]:\n\n\ntrace0 = go.Box(\n y= length,\n name= 'length'\n)\ntrace1 =go.Box(\n y= tcg,\n name= 'coating weight-top'\n)\ntrace2 =go.Box(\n y= bcg,\n name= 'Coating weight-bottom'\n)\n\n\n# In[21]:\n\n\ndata = [trace0,trace1,trace2]\nlayout = go.Layout(title = 'COATING GAUGE')\n\n\n# In[24]:\n\n\nfig = go.Figure(data = data,layout = layout)\n\n\n# In[ ]:\n\n\n\n\n\n# In[25]:\n\n\npy.plot(fig, filename='boxplot2')\n\n\n# In[ ]:\n\n\n\n\n", "repo_name": "yashkaransingh/Exploring-Google-Data-Studio", "sub_path": "coating_gauge.py", "file_name": "coating_gauge.py", "file_ext": "py", "file_size_in_byte": 1131, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "chart_studio.tools.set_credentials_file", "line_number": 21, "usage_type": "call"}, {"api_name": "chart_studio.tools", "line_number": 21, "usage_type": "attribute"}, {"api_name": "chart_studio.tools.set_config_file", "line_number": 22, "usage_type": "call"}, {"api_name": "chart_studio.tools", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 28, "usage_type": "call"}, {"api_name": "plotly.graph_objs.Box", "line_number": 44, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 44, "usage_type": "name"}, {"api_name": "plotly.graph_objs.Box", "line_number": 48, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 48, "usage_type": "name"}, {"api_name": "plotly.graph_objs.Box", "line_number": 52, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 52, "usage_type": "name"}, {"api_name": "plotly.graph_objs.Layout", "line_number": 62, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 62, "usage_type": "name"}, {"api_name": "plotly.graph_objs.Figure", "line_number": 68, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 68, "usage_type": "name"}, {"api_name": "chart_studio.plotly.plot", "line_number": 80, "usage_type": "call"}, {"api_name": "chart_studio.plotly", "line_number": 80, "usage_type": "name"}]}
+{"seq_id": "3554084927", "text": "# coding: utf-8\n# pylint: disable= invalid-name, unused-import\n\"\"\"For compatibility.\"\"\"\nfrom pykrige.ok import OrdinaryKriging\nfrom pykrige.ok3d import OrdinaryKriging3D\nfrom pykrige.uk import UniversalKriging\nfrom pykrige.uk3d import UniversalKriging3D\n\n# sklearn\ntry:\n # keep train_test_split here for backward compatibility\n from sklearn.base import BaseEstimator, ClassifierMixin, RegressorMixin\n from sklearn.model_selection import train_test_split\n\n SKLEARN_INSTALLED = True\n\nexcept ImportError:\n SKLEARN_INSTALLED = False\n\n train_test_split = None\n\n class RegressorMixin:\n \"\"\"Mock RegressorMixin.\"\"\"\n\n class ClassifierMixin:\n \"\"\"Mock ClassifierMixin.\"\"\"\n\n class BaseEstimator:\n \"\"\"Mock BaseEstimator.\"\"\"\n\n\nkrige_methods = {\n \"ordinary\": OrdinaryKriging,\n \"universal\": UniversalKriging,\n \"ordinary3d\": OrdinaryKriging3D,\n \"universal3d\": UniversalKriging3D,\n}\n\nthreed_krige = (\"ordinary3d\", \"universal3d\")\n\nkrige_methods_kws = {\n \"ordinary\": [\n \"anisotropy_scaling\",\n \"anisotropy_angle\",\n \"enable_statistics\",\n \"coordinates_type\",\n ],\n \"universal\": [\n \"anisotropy_scaling\",\n \"anisotropy_angle\",\n \"drift_terms\",\n \"point_drift\",\n \"external_drift\",\n \"external_drift_x\",\n \"external_drift_y\",\n \"functional_drift\",\n ],\n \"ordinary3d\": [\n \"anisotropy_scaling_y\",\n \"anisotropy_scaling_z\",\n \"anisotropy_angle_x\",\n \"anisotropy_angle_y\",\n \"anisotropy_angle_z\",\n ],\n \"universal3d\": [\n \"anisotropy_scaling_y\",\n \"anisotropy_scaling_z\",\n \"anisotropy_angle_x\",\n \"anisotropy_angle_y\",\n \"anisotropy_angle_z\",\n \"drift_terms\",\n \"functional_drift\",\n ],\n}\n\n\nclass SklearnException(Exception):\n \"\"\"Exception for missing scikit-learn.\"\"\"\n\n\ndef validate_method(method):\n \"\"\"Validate the kriging method in use.\"\"\"\n if method not in krige_methods.keys():\n raise ValueError(\n \"Kriging method must be one of {}\".format(krige_methods.keys())\n )\n\n\ndef validate_sklearn():\n \"\"\"Validate presence of scikit-learn.\"\"\"\n if not SKLEARN_INSTALLED:\n raise SklearnException(\n \"sklearn needs to be installed in order to use this module\"\n )\n\n\nclass Krige(RegressorMixin, BaseEstimator):\n \"\"\"\n A scikit-learn wrapper class for Ordinary and Universal Kriging.\n\n This works with both Grid/RandomSearchCv for finding the best\n Krige parameters combination for a problem.\n\n Parameters\n ----------\n method: str, optional\n type of kriging to be performed\n variogram_model: str, optional\n variogram model to be used during Kriging\n nlags: int\n see OK/UK class description\n weight: bool\n see OK/UK class description\n n_closest_points: int\n number of closest points to be used during Ordinary Kriging\n verbose: bool\n see OK/UK class description\n exact_values : bool\n see OK/UK class description\n variogram_parameters : list or dict\n see OK/UK class description\n variogram_function : callable\n see OK/UK class description\n anisotropy_scaling : tuple\n single value for 2D (UK/OK) and two values in 3D (UK3D/OK3D)\n anisotropy_angle : tuple\n single value for 2D (UK/OK) and three values in 3D (UK3D/OK3D)\n enable_statistics : bool\n see OK class description\n coordinates_type : str\n see OK/UK class description\n drift_terms : list of strings\n see UK/UK3D class description\n point_drift : array_like\n see UK class description\n ext_drift_grid : tuple\n Holding the three values external_drift, external_drift_x and\n external_drift_z for the UK class\n functional_drift : list of callable\n see UK/UK3D class description\n \"\"\"\n\n def __init__(\n self,\n method=\"ordinary\",\n variogram_model=\"linear\",\n nlags=6,\n weight=False,\n n_closest_points=10,\n verbose=False,\n exact_values=True,\n pseudo_inv=False,\n pseudo_inv_type=\"pinv\",\n variogram_parameters=None,\n variogram_function=None,\n anisotropy_scaling=(1.0, 1.0),\n anisotropy_angle=(0.0, 0.0, 0.0),\n enable_statistics=False,\n coordinates_type=\"euclidean\",\n drift_terms=None,\n point_drift=None,\n ext_drift_grid=(None, None, None),\n functional_drift=None,\n ):\n validate_method(method)\n self.variogram_model = variogram_model\n self.variogram_parameters = variogram_parameters\n self.variogram_function = variogram_function\n self.nlags = nlags\n self.weight = weight\n self.verbose = verbose\n self.exact_values = exact_values\n self.pseudo_inv = pseudo_inv\n self.pseudo_inv_type = pseudo_inv_type\n self.anisotropy_scaling = anisotropy_scaling\n self.anisotropy_angle = anisotropy_angle\n self.enable_statistics = enable_statistics\n self.coordinates_type = coordinates_type\n self.drift_terms = drift_terms\n self.point_drift = point_drift\n self.ext_drift_grid = ext_drift_grid\n self.functional_drift = functional_drift\n self.model = None # not trained\n self.n_closest_points = n_closest_points\n self.method = method\n\n def fit(self, x, y, *args, **kwargs):\n \"\"\"\n Fit the current model.\n\n Parameters\n ----------\n x: ndarray\n array of Points, (x, y) pairs of shape (N, 2) for 2d kriging\n array of Points, (x, y, z) pairs of shape (N, 3) for 3d kriging\n y: ndarray\n array of targets (N, )\n \"\"\"\n val_kw = \"val\" if self.method in threed_krige else \"z\"\n setup = dict(\n variogram_model=self.variogram_model,\n variogram_parameters=self.variogram_parameters,\n variogram_function=self.variogram_function,\n nlags=self.nlags,\n weight=self.weight,\n verbose=self.verbose,\n exact_values=self.exact_values,\n pseudo_inv=self.pseudo_inv,\n pseudo_inv_type=self.pseudo_inv_type,\n )\n add_setup = dict(\n anisotropy_scaling=self.anisotropy_scaling[0],\n anisotropy_angle=self.anisotropy_angle[0],\n enable_statistics=self.enable_statistics,\n coordinates_type=self.coordinates_type,\n anisotropy_scaling_y=self.anisotropy_scaling[0],\n anisotropy_scaling_z=self.anisotropy_scaling[1],\n anisotropy_angle_x=self.anisotropy_angle[0],\n anisotropy_angle_y=self.anisotropy_angle[1],\n anisotropy_angle_z=self.anisotropy_angle[2],\n drift_terms=self.drift_terms,\n point_drift=self.point_drift,\n external_drift=self.ext_drift_grid[0],\n external_drift_x=self.ext_drift_grid[1],\n external_drift_y=self.ext_drift_grid[2],\n functional_drift=self.functional_drift,\n )\n for kw in krige_methods_kws[self.method]:\n setup[kw] = add_setup[kw]\n input_kw = self._dimensionality_check(x)\n input_kw.update(setup)\n input_kw[val_kw] = y\n self.model = krige_methods[self.method](**input_kw)\n\n def _dimensionality_check(self, x, ext=\"\"):\n if self.method in (\"ordinary\", \"universal\"):\n if x.shape[1] != 2:\n raise ValueError(\"2d krige can use only 2d points\")\n else:\n return {\"x\" + ext: x[:, 0], \"y\" + ext: x[:, 1]}\n if self.method in (\"ordinary3d\", \"universal3d\"):\n if x.shape[1] != 3:\n raise ValueError(\"3d krige can use only 3d points\")\n else:\n return {\n \"x\" + ext: x[:, 0],\n \"y\" + ext: x[:, 1],\n \"z\" + ext: x[:, 2],\n }\n\n def predict(self, x, *args, **kwargs):\n \"\"\"\n Predict.\n\n Parameters\n ----------\n x: ndarray\n array of Points, (x, y) pairs of shape (N, 2) for 2d kriging\n array of Points, (x, y, z) pairs of shape (N, 3) for 3d kriging\n Returns\n -------\n Prediction array\n \"\"\"\n if not self.model:\n raise Exception(\"Not trained. Train first\")\n points = self._dimensionality_check(x, ext=\"points\")\n return self.execute(points, *args, **kwargs)[0]\n\n def execute(self, points, *args, **kwargs):\n # TODO array of Points, (x, y) pairs of shape (N, 2)\n \"\"\"\n Execute.\n\n Parameters\n ----------\n points: dict\n\n Returns\n -------\n Prediction array\n Variance array\n \"\"\"\n default_kw = dict(style=\"points\", backend=\"loop\")\n default_kw.update(kwargs)\n points.update(default_kw)\n if isinstance(self.model, (OrdinaryKriging, OrdinaryKriging3D)):\n points.update(dict(n_closest_points=self.n_closest_points))\n else:\n print(\"n_closest_points will be ignored for UniversalKriging\")\n prediction, variance = self.model.execute(**points)\n return prediction, variance\n\n\ndef check_sklearn_model(model, task=\"regression\"):\n \"\"\"Check the sklearn method in use.\"\"\"\n if task == \"regression\":\n if not (isinstance(model, BaseEstimator) and isinstance(model, RegressorMixin)):\n raise RuntimeError(\n \"Needs to supply an instance of a scikit-learn regression class.\"\n )\n elif task == \"classification\":\n if not (\n isinstance(model, BaseEstimator) and isinstance(model, ClassifierMixin)\n ):\n raise RuntimeError(\n \"Needs to supply an instance of a scikit-learn classification class.\"\n )\n", "repo_name": "GeoStat-Framework/PyKrige", "sub_path": "src/pykrige/compat.py", "file_name": "compat.py", "file_ext": "py", "file_size_in_byte": 9889, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 682, "dataset": "github-code", "pt": "24", "api": [{"api_name": "sklearn.model_selection.train_test_split", "line_number": 20, "usage_type": "name"}, {"api_name": "pykrige.ok.OrdinaryKriging", "line_number": 33, "usage_type": "name"}, {"api_name": "pykrige.uk.UniversalKriging", "line_number": 34, "usage_type": "name"}, {"api_name": "pykrige.ok3d.OrdinaryKriging3D", "line_number": 35, "usage_type": "name"}, {"api_name": "pykrige.uk3d.UniversalKriging3D", "line_number": 36, "usage_type": "name"}, {"api_name": "pykrige.ok.OrdinaryKriging", "line_number": 286, "usage_type": "name"}, {"api_name": "pykrige.ok3d.OrdinaryKriging3D", "line_number": 286, "usage_type": "name"}]}
+{"seq_id": "26875790435", "text": "from torchstudio.modules import Analyzer\r\nfrom typing import List\r\nimport numpy as np\r\nfrom random import randint\r\nimport zlib\r\nimport matplotlib as mpl\r\nimport matplotlib.pyplot as plt\r\nfrom matplotlib.patches import Circle\r\nimport PIL\r\nimport sys\r\n\r\nclass Multiclass(Analyzer):\r\n \"\"\"Analyze the distribution of multiclass datasets\r\n (single integer output, single label prediction)\r\n https://en.wikipedia.org/wiki/Multiclass_classification\r\n\r\n Args:\r\n train: If True, analyze the training set.\r\n If False, analyze the validation set.\r\n If None, analyze the entire dataset.\r\n \"\"\"\r\n def __init__(self, train=True):\r\n super().__init__(train)\r\n\r\n def start_analysis(self, num_samples: int, input_tensors_id: List[int], output_tensors_id: List[int], labels: List[str]):\r\n self.num_samples=num_samples\r\n self.input_tensors_id=input_tensors_id\r\n self.output_tensors_id=output_tensors_id\r\n self.labels=labels\r\n\r\n self.classes_weight=[]\r\n self.classes_label=[]\r\n self.classes_randomness=0\r\n\r\n self.tensor_id=None\r\n self.classes={}\r\n self.classes_sequence=bytearray()\r\n\r\n\r\n def analyze_sample(self, sample: List[np.array], training_sample: bool):\r\n if self.tensor_id is None:\r\n for id in self.output_tensors_id:\r\n if \"int\" in str(sample[id].dtype) and len(sample[id].shape)==0:\r\n self.tensor_id=id\r\n break\r\n if self.tensor_id is None:\r\n raise ValueError('Multiclass analysis requires a single integer output tensor')\r\n\r\n class_id=sample[self.tensor_id].item()\r\n self.classes_sequence.extend(class_id.to_bytes(2,'little'))\r\n if class_id not in self.classes:\r\n self.classes[class_id]=1\r\n else:\r\n self.classes[class_id]+=1\r\n\r\n def finish_analysis(self):\r\n num_registered_classes=len(self.classes)\r\n random_sequence=bytearray()\r\n for i in range(self.num_samples):\r\n random_sequence.extend(randint(0, num_registered_classes).to_bytes(2,'little'))\r\n self.classes_randomness=len(zlib.compress(self.classes_sequence))/len(zlib.compress(random_sequence))\r\n\r\n #prepare weights and labels\r\n last_class_id=0\r\n for class_id in self.classes:\r\n last_class_id=max(last_class_id,class_id)\r\n self.classes_weight=[]\r\n self.classes_label=[]\r\n for class_id in range(last_class_id+1):\r\n if class_id not in self.classes:\r\n self.classes_weight.append(0)\r\n else:\r\n self.classes_weight.append(self.classes[class_id])\r\n if class_id>=len(self.labels):\r\n self.classes_label.append(str(class_id))\r\n else:\r\n self.classes_label.append(self.labels[class_id])\r\n\r\n return self.classes_weight\r\n\r\n def generate_report(self, size, dpi):\r\n if not self.classes_weight:\r\n raise ValueError(\"Nothing to report\")\r\n\r\n #set up matplotlib renderer, style, figure and axis\r\n mpl.use('agg') #https://www.namingcrisis.net/post/2019/03/11/interactive-matplotlib-ipython/\r\n plt.style.use('dark_background')\r\n plt.rcParams.update({'font.size': 8})\r\n fig, [ax1, ax2] = plt.subplots(1 if size[0]>size[1] else 2, 2 if size[0]>size[1] else 1, figsize=(size[0]/dpi, size[1]/dpi), dpi=dpi)\r\n\r\n ax1.set_title(\"Class Distribution\")\r\n ax1.pie(self.classes_weight, labels=self.classes_label, autopct='%1.1f%%', colors=plt.cm.tab10.colors, startangle=90, counterclock=False)\r\n\r\n ax2.set_title(\"Class Randomness\")\r\n _, _, autopct = ax2.pie([self.classes_randomness,max(0,1-self.classes_randomness)], autopct='%1.1f%%', textprops={'fontsize': 16}, pctdistance=0, colors=[\"#b03070\",\"black\"], startangle=90, counterclock=False)\r\n autopct[1].set_visible(False)\r\n ax2.add_patch(Circle( (0,0), 0.6, color='black'))\r\n\r\n plt.tight_layout(pad=0)\r\n\r\n canvas = plt.get_current_fig_manager().canvas\r\n canvas.draw()\r\n img = PIL.Image.frombytes('RGB',canvas.get_width_height(),canvas.tostring_rgb())\r\n plt.close()\r\n return img\r\n\r\n", "repo_name": "TorchStudio/torchstudio", "sub_path": "torchstudio/analyzers/multiclass.py", "file_name": "multiclass.py", "file_ext": "py", "file_size_in_byte": 4272, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 362, "dataset": "github-code", "pt": "24", "api": [{"api_name": "torchstudio.modules.Analyzer", "line_number": 12, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 25, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 40, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 60, "usage_type": "call"}, {"api_name": "zlib.compress", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.use", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 87, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 88, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 92, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.patches.Circle", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.get_current_fig_manager", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "PIL.Image.frombytes", "line_number": 103, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 103, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.close", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}]}
+{"seq_id": "30307436787", "text": "# from datetime import date # released min date(1, 1, 1)\nfrom mimetypes import MimeTypes\nfrom os.path import basename, getsize # posixpath\nfrom random import randrange\nfrom re import IGNORECASE, sub\nfrom subprocess import call, check_output\n# extremely fast C++ JSON parser and serialization library\nfrom rapidjson import loads\nfrom sxtools.logging import get_basic_logger\nlogger = get_basic_logger() # sXtools.log\nfrom sxtools.utils import cache # keep thumbnails in one place\n\n\nclass Scene:\n\n def __init__(self, path: str) -> None:\n\n self.path, self.ext = path, path[-3:].lower()\n self.size = getsize(path) # os.stat()\n self.performers = list()\n self.title, self.paysite = '', ''\n self.released = None # not set yet | date.min\n self.ffprobed = False # get metadata using ffprobe, e.g. bitrate, resolution, duration\n\n def __str__(self) -> str:\n\n return self.sanitize()\n\n def is_valid(self) -> bool:\n '''\n check if scene well-parsed is\n '''\n return bool(self.released and self.performers and self.paysite)\n\n def basename(self) -> str:\n '''\n return the final component of a pathname\n '''\n return basename(self.path)\n\n def viewname(self) -> str:\n '''\n return the viewname of a scene\n '''\n return f'd=\"{self.released or \"\"}\" p=\"{self.perfs_as_string()}\" s=\"{self.paysite}\" t=\"{self.title}\"'\n\n def set_title(self, title: str = '') -> None:\n '''\n clean title, remove multiple spaces and assign it to the title\n '''\n t = ' '.join((title or '').split())\n t = sub(' part (\\d+)', ' (Part \\g<1>)', t, flags=IGNORECASE)\n # t = sub('\\'[A-Z]', lambda x: x.group(0).lower(), m.group('title'))\n self.title = t # assign a fmt'd string\n\n def resolution(self) -> str:\n '''\n return scene's resolution\n '''\n return f'{self.w}x{self.h}' if self.ffprobed else 'not yet scanned'\n\n def perfs_as_string(self) -> str:\n '''\n return performers as a well-formed string\n '''\n return ', '.join(self.performers[:-2] + [' & '.join(self.performers[-2:])])\n\n def mimetype(self) -> str:\n '''\n return scene's mimetype, e.g. video-mp4\n '''\n mt = MimeTypes().guess_type(self.path)[0]\n if not mt:\n mt = 'video'\n return mt.replace(chr(47), chr(45)) # return mimetype\n\n def scan(self) -> None:\n '''\n read out video file's metadata using \"ffprobe\" (part of \"ffmpeg\")\n '''\n logger.debug(f'Scanning {self.viewname()}')\n try:\n meta = loads(check_output(['ffprobe',\n '-v', 'fatal',\n '-select_streams',\n 'v:0',\n '-show_entries',\n 'format=duration,bit_rate:stream=width,height',\n '-of', 'json', self.path]))\n # get width & height of the scene\n self.h = meta['streams'][0]['height']\n self.w = meta['streams'][0]['width'] # v:0\n # get scene's bitrate\n self.bitrate = meta['format']['bit_rate']\n # get duration in seconds as float\n self.duration = meta['format']['duration']\n # check for suspicious resolutions\n if self.w != 1920: # 1080p\n logger.debug(\n f'Suspicious Resolution {self.w}x{self.h}')\n if not cache(self.basename()):\n call(['ffmpeg',\n '-ss', str(randrange(int(float(self.duration)))),\n '-v',\n 'error',\n '-i', self.path,\n '-vframes', '1',\n cache(self.basename(), True), '-y'])\n side = min(self.w, self.h)\n call(['convert',\n cache(self.basename()),\n '-gravity', 'Center',\n '-crop', f'{side}x{side}+0+0',\n #'-unsharp', '0.25x0.25+8+0.065',\n '-resize', '49',\n '-quality', '90', cache(self.basename())])\n self.ffprobed = True\n except (FileNotFoundError) as e:\n logger.warning(f'No such file or directory: {e.filename}')\n\n # be sure \"ffprobe\" is installed, run \"apt install ffmpeg\" otherwise\n\n def sanitize(self) -> str:\n '''\n return sanitized scene's name as a string\n '''\n name = basename(self.path)[:-4] # .lower()\n sep = '.' if name.count('.') > name.count(' ') else ' '\n for ss in list(['[pt]', '[', ']']):\n name = name.replace(ss, sep)\n droplist = ['720p','1080p','1920p','2160p','bj','int','mp4','mp4-gush','mp4-ktr','mp4-nbq','mp4-wrb','repack','xxx']\n return sep.join(\n x for x in name.split(sep) if x and x.lower() not in droplist)\n", "repo_name": "nschepsen/sxtools-manager", "sub_path": "src/sxtools/core/videoscene.py", "file_name": "videoscene.py", "file_ext": "py", "file_size_in_byte": 4913, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "24", "api": [{"api_name": "sxtools.logging.get_basic_logger", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.getsize", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 39, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 52, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 52, "usage_type": "name"}, {"api_name": "mimetypes.MimeTypes", "line_number": 72, "usage_type": "call"}, {"api_name": "rapidjson.loads", "line_number": 83, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 83, "usage_type": "call"}, {"api_name": "sxtools.utils.cache", "line_number": 101, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 102, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 103, "usage_type": "call"}, {"api_name": "sxtools.utils.cache", "line_number": 108, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 110, "usage_type": "call"}, {"api_name": "sxtools.utils.cache", "line_number": 111, "usage_type": "call"}, {"api_name": "sxtools.utils.cache", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 127, "usage_type": "call"}]}
+{"seq_id": "8715247946", "text": "import numpy as np\r\nimport pandas as pd\r\nimport scipy.stats\r\n\r\nfrom scipy.optimize import curve_fit \r\n\r\nimport plot_asymptotes_library as pal\r\n\r\ndef fit_to_line(xarray, yarray, alpha = 0.05):\r\n \"\"\"\r\n Purpose: Fit (x,y) data using linear regression; return intercept, slope,\r\n and their confidence intervals. Confidence intervals are based on desired \r\n type 1 error. \r\n\r\n Parameters\r\n ----------\r\n xarray : Collection (list, numpy array, ...) of Floats\r\n x-values\r\n yarray : Collection (list, numpy array, ...) of Floats\r\n y-values\r\n alpha : Float\r\n Desired Type 1 Error for the confidence intervals, the default is 0.05.\r\n\r\n Returns\r\n -------\r\n list\r\n [[y0, y0_lower, y0_upper],[slope, slope_lower, slope_upper], p_value]\r\n\r\n \"\"\"\r\n\r\n #For confidence intervals, calculate Z-score based on desired T1E\r\n z_score = scipy.stats.norm.ppf(1 - alpha/2)\r\n\r\n #Linear regression; returns slope, intercept, and their standard errors\r\n r = scipy.stats.linregress(xarray,yarray)\r\n\r\n #Get estimated parameters\r\n y0 = r[1]\r\n slope = r[0]\r\n p_value = r.pvalue\r\n\r\n #Calculate the ends of the confidence intervals\r\n y0_upper = y0 + z_score*r.intercept_stderr\r\n y0_lower = y0 - z_score*r.intercept_stderr\r\n slope_lower = slope - 1.96*r.stderr\r\n slope_upper = slope + 1.96*r.stderr\r\n\r\n #Return information on the [intercept], [slope], and p value \r\n return [[y0, y0_lower, y0_upper],\r\n [slope, slope_lower, slope_upper], \r\n p_value]\r\n \r\n\r\ndef calc_asympt_fx(x, lamb, gamma):\r\n \"\"\"\r\n Purpose: Calculate f(x) = lambda*(1- exp(-x/gamma))\r\n\r\n Parameters\r\n ----------\r\n x : Float\r\n x-value\r\n lamb : Float\r\n \"Lambda\": Asymptote value as x approaches infinity\r\n gamma : Float\r\n Decay parameter; same units as x\r\n\r\n Returns\r\n -------\r\n Float\r\n f(x)\r\n\r\n \"\"\"\r\n \r\n return lamb*(1 - np.exp(-x/gamma))\r\n\r\n\r\ndef fit_to_asymp(xarray, yarray):\r\n \"\"\"\r\n Purpose: Fit (x,y) values to an asymptote function. \r\n Return lambda and gamma. \r\n\r\n Parameters\r\n ----------\r\n xarray : Collection (list, numpy array, ...) of Floats\r\n x values\r\n yarray : Collection (list, numpy array, ...) of Floats\r\n y values\r\n\r\n Returns\r\n -------\r\n list\r\n [lambda, gamma]\r\n\r\n \"\"\"\r\n \r\n parameters, covariance = curve_fit(calc_asympt_fx, xarray, yarray)\r\n \r\n return [parameters[0], parameters[1]]\r\n\r\n\r\ndef fit_to_asymp_get_CI(xarray, yarray, alpha = 0.05):\r\n \"\"\"\r\n Purpose: Purpose: Fit (x,y) values to an asymptote function. \r\n Return lambda, gamma, and their confidence intervals.\r\n\r\n Parameters\r\n ----------\r\n xarray : Collection (list, numpy array, ...) of Floats\r\n x-values\r\n yarray : Collection (list, numpy array, ...) of Floats\r\n y-values\r\n alpha : Float, optional\r\n Desired Type 1 Error for the confidence intervals. The default is 0.05.\r\n\r\n Returns\r\n -------\r\n list of lists\r\n [lambda, lambda lower, lambda upper], \r\n [gamma, gamma lower, gamma upper]\r\n\r\n \"\"\"\r\n \r\n #For confidence intervals, calculate Z-score based on desired T1E\r\n z_score = scipy.stats.norm.ppf(1 - alpha/2)\r\n\r\n #Curve fit outputs parameters and covariance \r\n popt, pcov = curve_fit(calc_asympt_fx, xarray, yarray) \r\n \r\n #Get relative error from covariance\r\n sigma = np.sqrt(np.diag(pcov))\r\n \r\n #Calculate confidence intervals for lambda and gamma\r\n lambCI = popt[0], popt[0] - z_score*sigma[0], popt[0] + z_score*sigma[0]\r\n gammaCI = popt[1], popt[1] - z_score*sigma[1], popt[1] + z_score*sigma[1]\r\n \r\n return [lambCI, gammaCI]\r\n\r\n\r\ndef fit_asympt_bootstrap(xarray, yarray, simuls, plot_each = False):\r\n \"\"\"\r\n Purpose: Perform bootstrap simulations on (x,y) data; return the \r\n asymptote parameters for every simulation.\r\n\r\n Parameters\r\n ----------\r\n xarray : Collection (list, numpy array, ...) of Floats\r\n x-values \r\n yarray : Collection (list, numpy array, ...) of Floats\r\n y-values \r\n simuls : Integer \r\n Number of bootstrap simulations\r\n plot_each : Boolean, optional\r\n Option whether to plot each bootstrap and fit. The default is False.\r\n\r\n Returns\r\n -------\r\n list of lists\r\n Lists of each estimand from each bootstrap simulation \r\n [lambda results], [gamma results]\r\n\r\n \"\"\"\r\n \r\n #N: Number of data points\r\n N = len(xarray)\r\n \r\n #Create a Pandas dataframe to take advantage of pandas' sample-function.\r\n xarray = np.array(xarray)\r\n yarray = np.array(yarray)\r\n df = pd.DataFrame({'x': xarray, 'y':yarray})\r\n \r\n #Collection to keep the results from each simulation\r\n lambda_c = []\r\n gamma_c = []\r\n \r\n #For each bootstrap simulation\r\n for s in range(0,simuls):\r\n \r\n #Randomly draw (with replacement) N data points\r\n sampled = df.sample(n = N, replace = True)\r\n \r\n #Fit to asymptote\r\n r = fit_to_asymp(sampled[\"x\"], sampled[\"y\"])\r\n \r\n #Update the collections of results\r\n lambda_c.append(r[0])\r\n gamma_c.append(r[1])\r\n \r\n #Plot option\r\n if plot_each:\r\n pal.plot_asymptote(sampled[\"x\"], sampled[\"y\"], \r\n limits = [[0,18],[0,1]], \r\n ax_labels = [\"Area (mm$^{2}$)\", \"Accuracy\"],\r\n fit=\"none\",\r\n graph_filename = \"bootstrap.jpg\")\r\n pal.plot_asymptote(sampled[\"x\"], sampled[\"y\"], \r\n limits = [[0,18],[0,1]], \r\n ax_labels = [\"Area (mm$^{2}$)\", \"Accuracy\"],\r\n fit=\"asymptote\",\r\n graph_filename = \"bootstrap_fitted.jpg\")\r\n \r\n return [lambda_c, gamma_c]\r\n\r\n\r\ndef get_percentiles_from_list(rlist, alpha = 0.05):\r\n \"\"\"\r\n Purpose: Take a list of values, sort, and get percentile values.\r\n\r\n Parameters\r\n ----------\r\n rlist : List of Floats\r\n \r\n alpha: Float, optional\r\n Desired Type 1 Error for the confidence intervals. The default is 0.05.\r\n \r\n\r\n Returns\r\n -------\r\n list of Floats\r\n [median, lowerbound, upperbound]\r\n\r\n \"\"\"\r\n \r\n #Number of values in the list\r\n n = len(rlist)\r\n \r\n #Sort the list\r\n rlist = np.sort(np.array(rlist))\r\n \r\n #Calculate percentiles\r\n median = rlist[int(0.5*n)]\r\n lowerbound = rlist[int(alpha/2*n)]\r\n upperbound = rlist[int((1-alpha/2)*n)]\r\n \r\n return [median, lowerbound, upperbound]\r\n \r\n\r\ndef simulate_binomial_error(prob, n):\r\n \"\"\"\r\n Purpose: Take an observed probability, from n observations -> \r\n simulate a new probability with error from the binomial distribution.\r\n\r\n Parameters\r\n ----------\r\n prob : Float\r\n Probability (0 to 1)\r\n n : Integer\r\n Number of observations from which prob was calculated.\r\n\r\n Returns\r\n -------\r\n prob_new : Float\r\n Probability with error simulated from the binomial distribution.\r\n\r\n \"\"\"\r\n prob_new = np.random.normal(loc = prob, scale = (prob*(1-prob)/n)**0.5)\r\n \r\n return prob_new \r\n \r\n\r\ndef monteCarlo_asymptotes_with_error(xarray, lamb, gamma, n, \r\n simuls, plot_each=False):\r\n \"\"\"\r\n Purpose: Obtain a collection of asymptote fit parameters from \r\n Monte Carlo simulations. In each simulation, error from the binomial\r\n distribution is added to a perfect asympote at each of the specified \r\n x-values. Return fit parameters from each simulation.\r\n\r\n Parameters\r\n ----------\r\n xarray : Collection (list, numpy array, ...) of Floats\r\n x-values at which you want to simulate each point of the asymptote.\r\n lamb : Float\r\n \"Lambda\": Asymptote value as x approaches infinity\r\n gamma : Float\r\n Decay parameter; same units as x\r\n n : Integer\r\n Number of observations from which the probabilities were calculated.\r\n simuls : Integer \r\n Number of bootstrap simulations\r\n plot_each : Boolean, optional\r\n Option whether to plot each bootstrap and fit. The default is False.\r\n\r\n Returns\r\n -------\r\n list of lists (Floats)\r\n [simulation results for lambda], [simulations results for gamma]\r\n\r\n \"\"\"\r\n #Collections for the results of each MC simulation\r\n lamb_c = []\r\n gamma_c = []\r\n \r\n #Calculate the errorless asymptote, at each x, given the lamb and gamma.\r\n vectorized = np.vectorize(calc_asympt_fx) \r\n yarray = vectorized(xarray, lamb, gamma)\r\n\r\n #Vectorize the function that adds binomial error.\r\n vectorized_binomial = np.vectorize(simulate_binomial_error)\r\n\r\n #For each MC simulation...\r\n for s in range(0,simuls):\r\n #Simulate binomial-distribution error for each y point.\r\n yMC = vectorized_binomial(yarray, n)\r\n \r\n #Fit the (x,y) data to an asymptote\r\n r = fit_to_asymp(xarray, yMC)\r\n \r\n #Plot option \r\n if plot_each:\r\n pal.plot_asymptote(xarray, yMC, \r\n limits = [[0,18],[0,1]], \r\n ax_labels = [\"Area (mm$^{2}$)\", \"Accuracy\"],\r\n fit=\"none\",\r\n graph_filename = \"output_perfect.jpg\")\r\n pal.plot_asymptote(xarray, yMC, \r\n limits = [[0,18],[0,1]], \r\n ax_labels = [\"Area (mm$^{2}$)\", \"Accuracy\"],\r\n fit=\"asymptote\",\r\n graph_filename = \"output_perfect.jpg\") \r\n \r\n #Update collections of results.\r\n lamb_c.append(r[0])\r\n gamma_c.append(r[1])\r\n \r\n return [lamb_c, gamma_c]\r\n\r\n\r\ndef simulate_T1E(simuls, xarray, yarray = [],\r\n n = 1, simulation = \"binomialMC\", \r\n fit = \"linear\", plot_each = False):\r\n \"\"\"\r\n Purpose: Simulate type 1 error. Data can be simulated by either\r\n using the binomial distribution (\"binomialMC\") or random drawing \r\n with replacement from experimental data (bootstrap).\r\n\r\n Parameters\r\n ----------\r\n simuls : Integer\r\n Number of simulations to run\r\n xarray : Collection (list, numpy array, ...) of Floats\r\n x-values at which to simulate\r\n yarray : Collection (list, numpy array, ...) of Floats, optional\r\n If \"bootstrap\" simulation chosen, y values from which to draw\r\n randomly. The default is [].\r\n n : Integer, optional\r\n If \"binomialMC\" is chosen, the number of observations to use\r\n in the simulation of binomial error. The default is 1.\r\n simulation : String, optional\r\n Two options: \"binomialMC\" and \"bootstrap\". The default is \"binomialMC\".\r\n fit : String, optional\r\n Two options: \"linear\" and \"asymptote\" fit. The default is \"linear\".\r\n plot_each : Boolean, optional\r\n If true, plots every simulation and its fit. The default is False.\r\n\r\n Returns\r\n -------\r\n Float\r\n Type 1 Error\r\n\r\n \"\"\"\r\n \r\n #If an invalid option was selected for simulation type...\r\n if simulation not in [\"binomialMC\", \"bootstrap\"]:\r\n print(\"Error in simulate_T1E. For simulation, \\\r\n choose either: binomialMC or bootstrap\")\r\n return\r\n \r\n #If an invalid option was selected for fit...\r\n if fit not in [\"linear\", \"asymptote\"]:\r\n print(\"Error in simulate_T1E. For fit, \\\r\n choose either: linear or asymptote\")\r\n return\r\n \r\n #Number of data points and average y-value\r\n n_points = len(xarray) \r\n av_y = np.mean(yarray)\r\n \r\n #Keep track of the number of positive results.\r\n pos = 0\r\n \r\n #For each simulation...\r\n for s in range(0,simuls):\r\n #----------------------------------------------------------------------\r\n #Simulate the data using specified method\r\n #----------------------------------------------------------------------\r\n if simulation == \"binomialMC\":\r\n #Make y values array, all equal to the average-y of the input\r\n simulated = np.repeat(av_y, n_points)\r\n \r\n #Simulate the error for each using the binomial distribution\r\n vectorized = np.vectorize(simulate_binomial_error) \r\n simulated = vectorized(simulated, n)\r\n \r\n if simulation == \"bootstrap\":\r\n if len(yarray) == 0:\r\n print(\"Error in simulate_T1E. yarray must not be empty\")\r\n return\r\n \r\n #Make a dataframe to take advantadge of sample\r\n df = pd.DataFrame({'x': xarray, 'y':yarray})\r\n simulated = df.sample(n = n_points, replace = True)\r\n \r\n #Control the type by turning it into a numpy array\r\n simulated = np.array(simulated.y)\r\n\r\n #----------------------------------------------------------------------\r\n #Fit using specified method\r\n #----------------------------------------------------------------------\r\n if fit == \"linear\":\r\n\r\n fline = fit_to_line(xarray, simulated)\r\n\r\n #A positive result if both lower and upper bound are the same sign\r\n if fline[1][1]*fline[1][2] > 0:\r\n pos = pos + 1\r\n \r\n if fit == \"asymptote\":\r\n fasymp = fit_to_asymp_get_CI(xarray, simulated)\r\n \r\n #Positive result if the lower bound is above zero\r\n if fasymp[1][1] > 0:\r\n pos = pos + 1\r\n \r\n #----------------------------------------------------------------------\r\n #If plotting option was chosen\r\n #----------------------------------------------------------------------\r\n if plot_each:\r\n if fit == \"linear\": \r\n pal.plot_asymptote(xarray, simulated, \r\n limits = [[0,18],[0,1]], \r\n ax_labels = [\"Area (mm$^{2}$)\", \"Accuracy\"],\r\n fit=\"none\",\r\n graph_filename = \"output_none.jpg\")\r\n\r\n pal.plot_asymptote(xarray, simulated, \r\n limits = [[0,18],[0,1]], \r\n ax_labels = [\"Area (mm$^{2}$)\", \"Accuracy\"],\r\n fit=\"linear\",\r\n graph_filename = \"output_linear.jpg\")\r\n \r\n if fit == \"asymptote\":\r\n pal.plot_asymptote(xarray, simulated, \r\n limits = [[0,18],[0,1]], \r\n ax_labels = [\"Area (mm$^{2}$)\", \"Accuracy\"],\r\n fit=\"none\",\r\n graph_filename = \"output_none.jpg\")\r\n\r\n pal.plot_asymptote(xarray, simulated, \r\n limits = [[0,18],[0,1]], \r\n ax_labels = [\"Area (mm$^{2}$)\", \"Accuracy\"],\r\n fit=\"asymptote\",\r\n graph_filename = \"output_linear.jpg\") \r\n #Done with simulations\r\n \r\n #Return the ratio of positive results to total simulations\r\n return pos/simuls", "repo_name": "djarenas/accuracy_asymptote", "sub_path": "asymptote_functions_library.py", "file_name": "asymptote_functions_library.py", "file_ext": "py", "file_size_in_byte": 15385, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "scipy.stats.stats.norm.ppf", "line_number": 32, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 32, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 32, "usage_type": "name"}, {"api_name": "scipy.stats.stats.linregress", "line_number": 35, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 35, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 35, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 74, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 96, "usage_type": "call"}, {"api_name": "scipy.stats.stats.norm.ppf", "line_number": 124, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 124, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 124, "usage_type": "name"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 168, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 169, "usage_type": "call"}, {"api_name": "plot_asymptotes_library.plot_asymptote", "line_number": 190, "usage_type": "call"}, {"api_name": "plot_asymptotes_library.plot_asymptote", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 255, "usage_type": "attribute"}, {"api_name": "numpy.vectorize", "line_number": 294, "usage_type": "call"}, {"api_name": "numpy.vectorize", "line_number": 298, "usage_type": "call"}, {"api_name": "plot_asymptotes_library.plot_asymptote", "line_number": 310, "usage_type": "call"}, {"api_name": "plot_asymptotes_library.plot_asymptote", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 388, "usage_type": "call"}, {"api_name": "numpy.vectorize", "line_number": 391, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 400, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 404, "usage_type": "call"}, {"api_name": "plot_asymptotes_library.plot_asymptote", "line_number": 429, "usage_type": "call"}, {"api_name": "plot_asymptotes_library.plot_asymptote", "line_number": 435, "usage_type": "call"}, {"api_name": "plot_asymptotes_library.plot_asymptote", "line_number": 442, "usage_type": "call"}, {"api_name": "plot_asymptotes_library.plot_asymptote", "line_number": 448, "usage_type": "call"}]}
+{"seq_id": "21046191987", "text": "from setuptools import setup, find_packages\n\nwith open(\"requirements.txt\") as f:\n\tinstall_requires = f.read().strip().split(\"\\n\")\n\n# get version from __version__ variable in antoryum_gayrimenkul/__init__.py\nfrom antoryum_gayrimenkul import __version__ as version\n\nsetup(\n\tname=\"antoryum_gayrimenkul\",\n\tversion=version,\n\tdescription=\"Antoryum Website & Uygulama\",\n\tauthor=\"Harpiya Software Technologies\",\n\tauthor_email=\"info@harpiya.cloud\",\n\tpackages=find_packages(),\n\tzip_safe=False,\n\tinclude_package_data=True,\n\tinstall_requires=install_requires\n)\n", "repo_name": "harpiyacloud/antoryum_gayrimenkul", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 549, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "setuptools.setup", "line_number": 9, "usage_type": "call"}, {"api_name": "antoryum_gayrimenkul.__version__", "line_number": 11, "usage_type": "name"}, {"api_name": "setuptools.find_packages", "line_number": 15, "usage_type": "call"}]}
+{"seq_id": "33450231135", "text": "from aiogram import Dispatcher\nfrom aiogram.types import BotCommand\nfrom loguru import logger\n\ncommands = {\n \"/help\": \"Получить справочник.\",\n \"/settings\": \"Настройки аккаунта\",\n \"/start\": \"Получить клавиатуру|Начать диалог.\",\n}\n\n\nasync def set_bot_commands(dispatcher: Dispatcher):\n await dispatcher.bot.set_my_commands(\n commands=[\n BotCommand(\n command, description\n ) for command, description in commands.items()\n ]\n )\n logger.info(\"Installed bot commands\")\n", "repo_name": "ExissBrr/Telegram-Bot-Template", "sub_path": "utils/misc/set_bot_commands.py", "file_name": "set_bot_commands.py", "file_ext": "py", "file_size_in_byte": 596, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "aiogram.Dispatcher", "line_number": 12, "usage_type": "name"}, {"api_name": "aiogram.types.BotCommand", "line_number": 15, "usage_type": "call"}, {"api_name": "loguru.logger.info", "line_number": 20, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 20, "usage_type": "name"}]}
+{"seq_id": "3005241536", "text": "import argparse\r\nimport pathlib\r\nimport os\r\n\r\nparser = argparse.ArgumentParser(description='Training resusable components')\r\nparser.add_argument('--input_path', type=str, help='The text to reverse.')\r\nparser.add_argument('--output_path', type=str, help='Path of the local file where the Output data should be written')\r\nargs = parser.parse_args()\r\n\r\nprint(args.input_path)\r\n\r\nwith open(args.input_path + '.txt', 'r') as input_file:\r\n input_string = input_file.read()\r\n\r\nreversed_string = input_string[::-1]\r\n\r\npath = pathlib.PurePath(args.output_path)\r\nos.makedirs(path.parent)\r\n\r\nwith open(args.output_path + '.txt', 'w') as ouput_file:\r\n ouput_file.write(reversed_string)\r\n", "repo_name": "juansebashr/VertexPipelinesCICD", "sub_path": "src/reverse.py", "file_name": "reverse.py", "file_ext": "py", "file_size_in_byte": 679, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 5, "usage_type": "call"}, {"api_name": "pathlib.PurePath", "line_number": 17, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 18, "usage_type": "call"}]}
+{"seq_id": "4631505243", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# # Forest Fire analysis and prediction\n\n# - Data Collection\n# - Data Pre-Processing\n# - Exploratory Data Analysis\n# - Feature Engineering\n# - Feature Selection\n# - Model Building\n# - Model Selection\n# - Hyperparameter Tuning\n# - Regression\n\n# In[1]:\n\n\nimport pandas as pd\nimport numpy as np\n\n\n# In[2]:\n\n\ndata = pd.read_csv(\"./DATA/Algerian_forest_fires_dataset.csv\")\n\n\n# In[3]:\n\n\ndata\n\n\n# In[4]:\n\n\ndata.shape\n\n\n# In[5]:\n\n\ndata.columns\n\n\n# In[6]:\n\n\ndata.head()\n\n\n# In[7]:\n\n\ndata.tail()\n\n\n# In[8]:\n\n\ndata.describe()\n\n\n# In[9]:\n\n\ndata.info()\n\n\n# In[10]:\n\n\ndata.day.unique()\n\n\n# In[11]:\n\n\ndata[data.isnull().any(axis=1)] # inorder to check the row which is having the missing values\n\n\n# # Here after 123 we have the data set of new region\n\n# In[12]:\n\n\ndata.loc[:122,'Region']=1 #upto 122 ,Region =1 but intialize as 1.0\ndata.loc[122:,'Region']=2 #After 122 , Region = 2 this as 2.0\ndata[['Region']] = data[['Region']].astype(int) #1.0 is coverted to 1 astype integer\n#It is used to convert the data in the \"Region\" column of a pandas DataFrame to integer type.\ndata.head()\n\n\n# In[13]:\n\n\ndata.tail()\n\n\n# In[14]:\n\n\ndata.drop([122,123,168], axis=0, inplace=True)#.reset_index(drop=True)\ndata.day.unique()\n# Drop rows at index labels 122 and 123,168\n\n\n# In[15]:\n\n\nfiltered_rows = data[data['day'] == 'day']\nfiltered_rows\n\n\n# In[16]:\n\n\ndata.drop([124], axis=0, inplace=True)#.reset_index(drop=True)\ndata.day.unique()\n\n\n# In[17]:\n\n\ndata.shape\n\n\n# In[18]:\n\n\ndata.month.unique()\n\n\n# In[19]:\n\n\ndata.year.unique()\n\n\n# In[20]:\n\n\ndata.columns.unique()\n\n\n# In[21]:\n\n\ndata\n\n\n# In[22]:\n\n\ndata.shape\n\n\n# In[23]:\n\n\ndata.columns\n\n\n# In[24]:\n\n\n# Spaces were fixed in the column names\n# The str.strip() function is applied to each column name using the str accessor, which allows you to perform string operations on the elements of a column.\n# After applying the strip() function, the modified column names are assigned back to the data.columns attribute, updating the column names in the DataFrame.\ndata.columns = data.columns.str.strip()\ndata.columns \n\n\n# In[25]:\n\n\ndata.isna().sum()\n\n\n# In[26]:\n\n\ndata.reset_index()\n\n\n# In[27]:\n\n\ndata.reset_index(drop=True, inplace=True)\ndata\n\n\n# In[28]:\n\n\n# Check if the default index is in proper format without gaps\nis_proper_index = data.index.equals(pd.RangeIndex(len(data)))\n\nprint(is_proper_index)\n\n\n# \n\n# In[29]:\n\n\ndata.isna().sum()\n\n\n# In[30]:\n\n\nprint(data.duplicated())\nprint(data[data.duplicated()])\n\n\n# In[31]:\n\n\ndata.shape\n\n\n# ### ANALYSE\n\n# In[32]:\n\n\ndata.describe(include = 'all')\n\n\n# In[33]:\n\n\ndata[\"Classes\"].value_counts()\n\n\n# our dependent feature(Classes) containig only two categories but due to misspace it is showing multiple category so need to change the spaceing in order to make two category\n\n# In[34]:\n\n\ndata.Classes = data.Classes.str.strip()\n\n\n# In[35]:\n\n\ndata[\"Classes\"].value_counts()\n\n\n# In[36]:\n\n\ndata[\"Region\"].value_counts()\n\n\n# In[37]:\n\n\ndata.describe(include='all')\n\n\n# In[38]:\n\n\ndata[['month', 'day', 'year', 'Temperature', 'RH', 'Ws']] = data[['month', 'day', 'year', 'Temperature', 'RH', 'Ws']].astype(int)\n\nobjects = [features for features in data.columns if data[features].dtypes == 'O'] #\"o\" object type\nfor i in objects:\n if i != 'Classes': #exxcept classes\n data[i] = data[i].astype(float)\n\n\n# In[39]:\n\n\nprint(data.dtypes)\n\n\n# In pandas, the object data type is commonly used to represent categorical data. While the object data type can also be used to store other types of data (such as strings), it is often used to represent variables with a limited number of discrete categories or labels.\n\n# In[40]:\n\n\ndata.describe(include=\"all\")\n\n\n# In[41]:\n\n\ndata[:122]\n\n\n# In[42]:\n\n\ndata[122:]\n\n\n# In[43]:\n\n\n# Encoding Not fire as 0 and Fire as 1\ndata['Classes']= np.where(data['Classes']== 'not fire',0,1)\ndata.head(10)\n#If the condition is true, \n# the corresponding element in the 'Classes' column is assigned the value 0.\n# Otherwise, it is assigned the value 1.\n\n\n# In[44]:\n\n\ndata.tail()\n\n\n# In[45]:\n\n\n# Check counts\ndata.Classes.value_counts()\n\n\n# In[46]:\n\n\ncorrelation = data.corr()\ncorrelation\n\n\n# The value of correlation helps in determining the strength and direction of the relationship between two variables. When using correlation for feature selection, the value of correlation can guide you in selecting relevant features\n\n# # Visualize\n\n# In[47]:\n\n\nimport matplotlib.pyplot as plt\nimport seaborn as sns\n\n\n# In[48]:\n\n\nplt.figure(figsize=(20,15)) #size of figure\nsns.heatmap(correlation,annot= True,linewidths=1, linecolor=\"white\", cbar=True, cmap = \"Paired\",xticklabels=\"auto\", yticklabels=\"auto\")\n \n\n\n# In[49]:\n\n\ndata.to_csv('./DATA/ForestFireDataCleaned.csv', index=False)\n\n\n# In[50]:\n\n\nsns.countplot(x='Classes',data=data)\nplt.title('Class Distributions \\n 0: No Fire || 1: Fire', fontsize=14)\nplt.show()\n\n\n# In[51]:\n\n\nsns.countplot(x='Region',hue='Classes',data=data)\nplt.title('Region Distributions \\n 1: Region 1 || 2: Region 2', fontsize=14)\nplt.show()\n\n\n# In[52]:\n\n\n# PLot density plot for all features\n#plt.style.use('seaborn')\ndata.hist(bins=50, figsize=(20,15), ec = 'b')\nplt.show()\n\n\n# In[53]:\n\n\n# countplot over target variable\n\nsns.countplot(x='Temperature',hue='Classes', data=data)\nplt.title('Class Distribution \\n 0: No Fire || 1: Fire', fontsize=14)\n\n\n# In[54]:\n\n\ndata.month.unique()\n\n\n# In[55]:\n\n\n#month wise fire analysis for region 1\ndftemp= data.loc[data['Region']== 1]\nplt.subplots(figsize=(13,6))\nsns.set_style('whitegrid')\nsns.countplot(x='month',hue='Classes',data= dftemp,ec = 'black', palette= 'Set2')#ec='black' sets the edge color of the categorical plot elements to black, and palette='Set2' sets the color palette to the 'Set2' palette.\nplt.title('Fire Analysis Month wise for Region-1', fontsize=18, weight='bold')\nplt.ylabel('Count', weight = 'bold')\nplt.xlabel('Months', weight= 'bold')\nplt.legend(loc='upper right')\nplt.xticks(np.arange(4), ['June','July', 'August', 'September',])\nplt.grid(alpha = 0.5,axis = 'y')\nplt.show()\n\n\n# In[56]:\n\n\n#month wise fire analysis for region 2\ndftemp= data.loc[data['Region']== 2]\nplt.subplots(figsize=(13,6))\nsns.set_style('whitegrid')\nsns.countplot(x='month',hue='Classes',data= dftemp,ec = 'black', palette= 'Set2')#ec='black' sets the edge color of the categorical plot elements to black, and palette='Set2' sets the color palette to the 'Set2' palette.\nplt.title('Fire Analysis Month wise for Region-2', fontsize=18, weight='bold')\nplt.ylabel('Count', weight = 'bold')\nplt.xlabel('Months', weight= 'bold')\nplt.legend(loc='upper right')\nplt.xticks(np.arange(4), ['June','July', 'August', 'September',])\nplt.grid(alpha = 0.5,axis = 'y')\nplt.show()\n\n\n# Yearly plot\n\n# In[57]:\n\n\nplt.subplots(figsize=(13,6))\nsns.set_style('whitegrid')\nsns.countplot(x='year',hue='Classes',data= data,ec = 'black', palette= 'Set2')#ec='black' sets the edge color of the categorical plot elements to black, and palette='Set2' sets the color palette to the 'Set2' palette.\nplt.title('Fire Analysis of Year 2012', fontsize=18, weight='bold')\nplt.ylabel('Count', weight = 'bold')\nplt.xlabel('Year', weight= 'bold')\nplt.legend(loc='upper right')\nplt.grid(alpha = 0.5,axis = 'y')\nplt.show()\n\n\n# In[58]:\n\n\n# ! pip install scikit-learn\n\n\n# In[59]:\n\n\nplt.subplots(figsize=(15,6))\nsns.set_style('whitegrid')\nsns.countplot(x='Rain', hue='Classes', data=data)\n\n# Set the labels and title\nplt.xlabel('Rain')\nplt.ylabel('Count')\nplt.title('Distribution of Classes by Rain')\n\n# Show the plot\nplt.show()\n\n\n# ### With the increase in Rain Fire chance is descreases\n\n# In[60]:\n\n\nplt.subplots(figsize=(15,6))\nsns.set_style('whitegrid')\nsns.countplot(x='Ws', hue='Classes', data=data)\n\n# Set the labels and title\nplt.xlabel('WS')\nplt.ylabel('Count')\nplt.title('Distribution of Classes by Wind Speed:6 to 29')\n\n# Show the plot\nplt.show()\n\n\n# In[61]:\n\n\nplt.subplots(figsize=(15,6))\nsns.set_style('whitegrid')\nsns.countplot(x='RH', hue='Classes', data=data)\n\n# Set the labels and title\nplt.xlabel('RH')\nplt.ylabel('Count')\nplt.title('Distribution of Classes by humidity rate')\n\n# Show the plot\nplt.show()\n\n\n# In[62]:\n\n\nplt.subplots(figsize=(15,6))\nsns.set_style('whitegrid')\nsns.countplot(x='month', hue='Classes', data=data)\n\n# Set the labels and title\nplt.xlabel('Month')\nplt.ylabel('Count')\nplt.title('Distribution of Classes by Month')\n\n# Show the plot\nplt.show()\n\n\n# In[63]:\n\n\nimport pandas as pd\nfrom sklearn.ensemble import RandomForestClassifier\n\n# Assuming 'data' is the DataFrame containing the dataset\n# Splitting the features and target variable\nX = data.drop('Classes', axis=1)\ny = data['Classes']\n\n# Calculate the correlation matrix\ncorrelation_matrix = X.corr()\n\n# Display the correlation matrix\nprint(correlation_matrix)\n\n# Train a Random Forest classifier\nrfmodel = RandomForestClassifier()\nrfmodel.fit(X, y)\n\n# Get feature importances\nfeature_importances = pd.Series(rfmodel.feature_importances_, index=X.columns).sort_values(ascending=False)\n\n# Display feature importances\nprint(feature_importances)\n\n\n# In[64]:\n\n\nplt.subplots(figsize=(18,10))\n\nsns.heatmap(correlation_matrix , annot=True, cmap=plt.cm.CMRmap_r)\nplt.show()\n\n\n# From above anlysis we find some non importance features:\n# - Year (due to missing values)\n# - Ws (wind speed) low correlation\n# - DAY(low corr)\n# - Month(low corr)\n\n# In[65]:\n\n\ndf = data.drop(['day','month','year','Ws'], axis=1)\ndf.head(10)\n\n\n# Spliting Data Set \n\n# In[66]:\n\n\nfrom sklearn.model_selection import train_test_split\n\n\n# In[67]:\n\n\ny = df['Classes']\nX = df.drop('Classes',axis=1)\ny.tail()\n\n\n# In[68]:\n\n\nX.head()\n\n\n# In[69]:\n\n\nX_train, X_test, y_train, y_test = train_test_split( X, y, test_size=0.33, random_state=42)\nX_train.head()\n\n\n# In[70]:\n\n\nX_train.info()\n\n\n# In[71]:\n\n\ndel data,X,y,df\n\n\n# In[72]:\n\n\nfrom sklearn.model_selection import GridSearchCV\n#TRain Function is defined\ndef train(X_train, y_train, model, hyperparameters):\n grid_search = GridSearchCV(estimator=model,param_grid=hyperparameters, cv = 5)\n grid_search.fit(X_train, y_train)\n \n \n #print the best hyperparameters found\n best_params = grid_search.best_params_\n print(\"Best Hyperparameters:\", best_params)\n \n # Train the model with best hyperparametres\n best_model = model.set_params(**best_params)\n best_model.fit(X_train, y_train)\n\n # Print the intercept and coefficients of the best model\n # print('Intercept is :', best_model.best_estimator_.intercept_)\n # print('Coefficient is :', best_model.best_estimator_.coef_)\n\n # Evaluate the best model on the test data\n scores = best_model.score(X_test, y_test)\n print('Score_test_data:', scores)\n \n return best_params, best_model\n\n\n# In[73]:\n\n\nfrom sklearn.metrics import mean_squared_error, r2_score, mean_absolute_error\n\n# EVALUATION\n\ndef evaluate_model(X_test, y_test, best_model):\n # it will evaluate the score by taking testing data with best model\n \n #predict the target values for the best set\n y_pred = best_model.predict(X_test)\n \n # Calculate the MSE\n mse = mean_squared_error(y_test, y_pred)\n\n# Calculate the R-squared\n r2 = r2_score(y_test, y_pred)\n\n# Calculate the adjusted R-squared\n # adjusted_r2 = adjusted_r2_score(y_test, y_pred)\n\n# Calculate the MAE\n mae = mean_absolute_error(y_test, y_pred)\n\n# Print the scores\n print(\"MSE:\", mse)\n print(\"R-squared:\", r2)\n # print(\"Adjusted R-squared:\", adjusted_r2)\n print(\"MAE:\", mae)\n\n return mse,r2,mae\n\n\n# # Linear Regression\n\n# In[74]:\n\n\nfrom sklearn.linear_model import LinearRegression\n# Define the hyperparameters to tune\nhyperparameters = {\n # \"regularization\": [\"l1\", \"l2\"],\n # \"learning_rate\": [0.01, 0.001, 0.0001],\n # \"number_of_epochs\": [10, 50, 100],\n}\nmodel = LinearRegression()\n_,best_model = train(X_train,y_train,model,hyperparameters)\n# print('Intercept is :',best_model.intercept_)\n# print('Coefficient is :',best_model.coef_)\nscores = evaluate_model(X_test,y_test,best_model) \n\n\n# # Ridge\n\n# In[75]:\n\n\nfrom sklearn.linear_model import Ridge\n# Define the hyperparameters to tune\nhyperparameters = {\n \"alpha\": np.logspace(-4, 4, 10),\n}\n\n# Create a Ridge model\nmodel = Ridge()\n_,best_model = train(X_train,y_train,model,hyperparameters)\nscores = evaluate_model(X_test,y_test,best_model) \n\n\n\n# # Lasso\n\n# In[76]:\n\n\nfrom sklearn.linear_model import Lasso\n# Define the hyperparameters to tune\nhyperparameters = {\n \"alpha\": np.logspace(-4, 4, 10),\n}\n\n# Create a Lasso model\nmodel = Lasso()\n_,best_model = train(X_train,y_train,model,hyperparameters)\nscores = evaluate_model(X_test,y_test,best_model) \n\n\n\n# # Decision Tree\n\n# In[77]:\n\n\nfrom sklearn.tree import DecisionTreeRegressor\n# Define the hyperparameters to tune\nhyperparameters = {\n \"max_depth\": [3, 5, 10],\n \"min_samples_split\": [2, 5, 10],\n}\n\n# Create a decision tree regressor\nmodel = DecisionTreeRegressor()\n_,best_model = train(X_train,y_train,model,hyperparameters)\nscores = evaluate_model(X_test,y_test,best_model) \n\n\n\n# In[78]:\n\n\nimport matplotlib.pyplot as plt\n\n# Scores of each model\nlinear_score = 0.6840775270198252\nlasso_score = 0.6695115483312191 \nridge_score = 0.6753837888234437\ndecision_tree_score = 0.9861363636363636\n\n# Models names\nmodels = ['Linear Regression', 'Lasso', 'Ridge', 'Decision Tree']\n\n# Scores for each model\nscores = [linear_score, lasso_score, ridge_score, decision_tree_score]\n\n# Plotting the scores\nplt.bar(models, scores)\nplt.title('Comparison of Model Scores')\nplt.xlabel('Models')\nplt.ylabel('Scores')\nplt.show()\n\n\n# In[ ]:\n\n\n\n\n", "repo_name": "Ashimeon/DATA-SCIENCE", "sub_path": "ForestFireAnalysis (1).py", "file_name": "ForestFireAnalysis (1).py", "file_ext": "py", "file_size_in_byte": 13470, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "pandas.read_csv", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.RangeIndex", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 305, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 346, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 346, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 347, "usage_type": "call"}, {"api_name": "seaborn.countplot", "line_number": 360, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 361, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 361, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 362, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 362, "usage_type": "name"}, {"api_name": "seaborn.countplot", "line_number": 368, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 369, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 369, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 370, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 370, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 379, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 379, "usage_type": "name"}, {"api_name": "seaborn.countplot", "line_number": 387, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 388, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 388, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 402, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 402, "usage_type": "name"}, {"api_name": "seaborn.set_style", "line_number": 403, "usage_type": "call"}, {"api_name": "seaborn.countplot", "line_number": 404, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 405, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 405, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 406, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 406, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 407, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 407, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 408, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 408, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 409, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 409, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 409, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 410, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 410, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 411, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 411, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 419, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 419, "usage_type": "name"}, {"api_name": "seaborn.set_style", "line_number": 420, "usage_type": "call"}, {"api_name": "seaborn.countplot", "line_number": 421, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 422, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 422, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 423, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 423, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 424, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 424, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 425, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 425, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 426, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 426, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 426, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 427, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 427, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 428, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 428, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 436, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 436, "usage_type": "name"}, {"api_name": "seaborn.set_style", "line_number": 437, "usage_type": "call"}, {"api_name": "seaborn.countplot", "line_number": 438, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 439, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 439, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 440, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 440, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 441, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 441, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 442, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 442, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 443, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 443, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 444, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 444, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 456, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 456, "usage_type": "name"}, {"api_name": "seaborn.set_style", "line_number": 457, "usage_type": "call"}, {"api_name": "seaborn.countplot", "line_number": 458, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 461, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 461, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 462, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 462, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 463, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 463, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 466, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 466, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 474, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 474, "usage_type": "name"}, {"api_name": "seaborn.set_style", "line_number": 475, "usage_type": "call"}, {"api_name": "seaborn.countplot", "line_number": 476, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 479, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 479, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 480, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 480, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 481, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 481, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 484, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 484, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 490, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 490, "usage_type": "name"}, {"api_name": "seaborn.set_style", "line_number": 491, "usage_type": "call"}, {"api_name": "seaborn.countplot", "line_number": 492, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 495, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 495, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 496, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 496, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 497, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 497, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 500, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 500, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 506, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 506, "usage_type": "name"}, {"api_name": "seaborn.set_style", "line_number": 507, "usage_type": "call"}, {"api_name": "seaborn.countplot", "line_number": 508, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 511, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 511, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 512, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 512, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 513, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 513, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 516, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 516, "usage_type": "name"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 537, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 541, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 550, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 550, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 552, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 552, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 552, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 553, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 553, "usage_type": "name"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 594, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 616, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 653, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 656, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_error", "line_number": 662, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 685, "usage_type": "call"}, {"api_name": "numpy.logspace", "line_number": 700, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Ridge", "line_number": 704, "usage_type": "call"}, {"api_name": "numpy.logspace", "line_number": 718, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Lasso", "line_number": 722, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeRegressor", "line_number": 741, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 765, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 765, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 766, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 766, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 767, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 767, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 768, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 768, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 769, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 769, "usage_type": "name"}]}
+{"seq_id": "73777497342", "text": "from model import areaXYDao\nfrom service.api.sensibletemperature.formatConvertor import formatConvertor\nfrom service.api.sensibletemperature.sensibleTemperatureApi import sensibleTemperatureApi\nfrom datetime import datetime, timedelta\n\nfrom service.api.shorttermforecast.ForecastTypeEnum import ForecastType\nfrom service.api.shorttermforecast.shortTermWeatherForecastApi import ShortTermWeatherForecastApi\nfrom service.util.definitionRisk import check_risk\n\n\ndef get_today_minus_1():\n convertor = formatConvertor()\n today = convertor.datetime_to_string(datetime.today() - timedelta(hours=1))\n return today\n\n\n\ndef get_SKY_PTY_T1H_in_short_term_forecast(x=37, y=126):\n today = get_today_minus_1()\n ymd = today[0:6]\n time = today[6:] + '00'\n api = ShortTermWeatherForecastApi(x, y, ymd, time) # API 기본 Configuration (x, y, 검색날짜, 검색시간)\n response = api.request_api() # response 값 API 요청\n SKY = api.find_category_from_response(response,\n ForecastType.SKY.name) # response 값에서 카테고리에 따른 값 (카테고리 확인 : ForecastTypeEnum), Default=RN1\n\n PTY = api.find_category_from_response(response,\n ForecastType.PTY.name) # response 값에서 카테고리에 따른 값 (카테고리 확인 : ForecastTypeEnum), Default=RN1\n T1H = api.find_category_from_response(response,\n ForecastType.T1H.name) # response 값에서 카테고리에 따른 값 (카테고리 확인 : ForecastTypeEnum), Default=RN1\n return SKY, PTY, T1H\n\n\ndef get_sensible_temp_present(time_range):\n convertor = formatConvertor()\n today = get_today_minus_1()\n\n api = sensibleTemperatureApi(today) # api configuration\n response = api.request_api() # api 신청\n convertor.setData(response) # 컨버터 등록\n response = convertor.execute() # 컨버터 실행\n\n return convertor.get_predict_list_data(time_range, today)\n\n\ndef get_info_from_address(address): # 구, 동 으로 정보를 찾아냄. 정보가 없으면 구로만 검색한 결과를 뿌려줌\n addList = address.split(' ')\n gu, dong = addList[1], addList[2]\n return areaXYDao.findareaXYBygudong(gu, dong)\n\n\ndef check_weather(data):\n return check_risk(data)\n\n\nif __name__ == '__main__':\n print(get_info_from_address(address=\"제주특별자치도 서귀포시 가가로 14\"))\n get_sensible_temp_present(6)\n", "repo_name": "JangDongHa/miniProject_1week", "sub_path": "service/tempService.py", "file_name": "tempService.py", "file_ext": "py", "file_size_in_byte": 2478, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "service.api.sensibletemperature.formatConvertor.formatConvertor", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 13, "usage_type": "call"}, {"api_name": "service.api.shorttermforecast.shortTermWeatherForecastApi.ShortTermWeatherForecastApi", "line_number": 22, "usage_type": "call"}, {"api_name": "service.api.shorttermforecast.ForecastTypeEnum.ForecastType.SKY", "line_number": 25, "usage_type": "attribute"}, {"api_name": "service.api.shorttermforecast.ForecastTypeEnum.ForecastType", "line_number": 25, "usage_type": "name"}, {"api_name": "service.api.shorttermforecast.ForecastTypeEnum.ForecastType.PTY", "line_number": 28, "usage_type": "attribute"}, {"api_name": "service.api.shorttermforecast.ForecastTypeEnum.ForecastType", "line_number": 28, "usage_type": "name"}, {"api_name": "service.api.shorttermforecast.ForecastTypeEnum.ForecastType.T1H", "line_number": 30, "usage_type": "attribute"}, {"api_name": "service.api.shorttermforecast.ForecastTypeEnum.ForecastType", "line_number": 30, "usage_type": "name"}, {"api_name": "service.api.sensibletemperature.formatConvertor.formatConvertor", "line_number": 35, "usage_type": "call"}, {"api_name": "service.api.sensibletemperature.sensibleTemperatureApi.sensibleTemperatureApi", "line_number": 38, "usage_type": "call"}, {"api_name": "model.areaXYDao.findareaXYBygudong", "line_number": 49, "usage_type": "call"}, {"api_name": "model.areaXYDao", "line_number": 49, "usage_type": "name"}, {"api_name": "service.util.definitionRisk.check_risk", "line_number": 53, "usage_type": "call"}]}
+{"seq_id": "157080379", "text": "import json\nimport datetime\nimport socket\nimport requests\n\nclass TimeShift(object):\n\tdef __init__(self):\n\t\ts=socket.socket(socket.AF_INET,socket.SOCK_DGRAM)\n\t\ts.connect(('8.8.8.8',80))\n\t\tself.address=s.getsockname()[0]\n\n\t\t'''SAVING NEXT PROCESSING TIME FOR EACH PATIENT'''\n\t\tself.scheduling={\n\t\t\"2\":[],\n\t\t\"3\":[],\n\t\t\"4\":[],\n\t\t\"5\":[],\n\t\t}\n\n\t\t#self.catalog=\"http://192.168.1.103:8080\"\n\t\tself.catalog=json.loads(open(\"catalog.json\").read())[\"catalog\"]\n\t\t'''\n\t\tself.my_data={\n\t\t\"time_shift\":\n\t\t{\t\n\t\t\"ip\":socket.gethostbyname(socket.gethostname()),\n\t\t\"port\":8087\n\t\t}\n\t\t}'''\n\t\tself.my_data=json.loads(open(\"timeShiftData.json\").read())\n\t\tself.my_data[\"time_shift\"][\"ip\"]=self.address\n\n\tdef getAddress(self):\n\t\treturn self.address\n\n\tdef getData(self):\n\t\treturn self.my_data\n\n\tdef setData(self,data):\n\t\tself.ip_others=data\n\n\n\tdef configure(self):\n\t\tself.result=requests.post(self.catalog,json.dumps(self.my_data))\n\t\tself.ip_others=self.result.json()\n\n\tdef sendAlert(self):\n\n\t\tfor key in self.scheduling.keys():\n\t\t\tif(key==\"2\"):\n\t\t\t\tfor elem in self.scheduling[key]:\n\t\t\t\t\tif((datetime.datetime.now()-datetime.datetime.strptime(elem[\"last_measurement\"],'%Y-%m-%d %H:%M:%S')).total_seconds()>=60):\n\t\t\t\t\t\tr=requests.get(\"http://\"+self.ip_others[\"queue_server\"][0]+\":\"+self.ip_others[\"queue_server\"][1]+\"/retrieve?pressure_id=\"+elem[\"pressure_id\"]+\"&heart_id=\"+elem[\"heart_id\"]+\"&glucose_id=\"+elem[\"glucose_id\"])\n\t\t\telif(key==\"3\"):\n\t\t\t\tfor elem in self.scheduling[key]:\n\t\t\t\t\tif((datetime.datetime.now()-datetime.datetime.strptime(elem[\"last_measurement\"],'%Y-%m-%d %H:%M:%S')).total_seconds()>=240):\n\t\t\t\t\t\tr=requests.get(\"http://\"+self.ip_others[\"queue_server\"][0]+\":\"+self.ip_others[\"queue_server\"][1]+\"/retrieve?pressure_id=\"+elem[\"pressure_id\"]+\"&heart_id=\"+elem[\"heart_id\"]+\"&glucose_id=\"+elem[\"glucose_id\"])\n\t\t\telif(key==\"4\"):\n\t\t\t\tfor elem in self.scheduling[key]:\n\t\t\t\t\tif((datetime.datetime.now()-datetime.datetime.strptime(elem[\"last_measurement\"],'%Y-%m-%d %H:%M:%S')).total_seconds()>=480):\n\t\t\t\t\t\tr=requests.get(\"http://\"+self.ip_others[\"queue_server\"][0]+\":\"+self.ip_others[\"queue_server\"][1]+\"/retrieve?pressure_id=\"+elem[\"pressure_id\"]+\"&heart_id=\"+elem[\"heart_id\"]+\"&glucose_id=\"+elem[\"glucose_id\"])\n\t\t\telif(key==\"5\"):\n\t\t\t\tfor elem in self.scheduling[key]:\n\t\t\t\t\tif((datetime.datetime.now()-datetime.datetime.strptime(elem[\"last_measurement\"],'%Y-%m-%d %H:%M:%S')).total_seconds()>=960):\n\t\t\t\t\t\tr=requests.get(\"http://\"+self.ip_others[\"queue_server\"][0]+\":\"+self.ip_others[\"queue_server\"][1]+\"/retrieve?pressure_id=\"+elem[\"pressure_id\"]+\"&heart_id=\"+elem[\"heart_id\"]+\"&glucose_id=\"+elem[\"glucose_id\"])\n\n\t\n\tdef addToScheduling(self, data):\n\t\tdata=json.loads(data)\n\t\tobj={\n\t\t\"pressure_id\":data[\"pressure_id\"],\n\t\t\"heart_id\":data[\"heart_id\"],\n\t\t\"glucose_id\":data[\"glucose_id\"],\n\t\t\"last_measurement\":data[\"time_stamp\"]\n\t\t}\n\n\t\tself.scheduling[data[\"code\"]].append(obj)\n\n\tdef removeFromScheduling(self,code,pressure_id,heart_id,glucose_id):\n\t\tfor i in range(len(self.scheduling[str(code)])):\n\t\t\tif(self.scheduling[str(code)][i][\"pressure_id\"]==pressure_id and self.scheduling[str(code)][i][\"heart_id\"]==heart_id and self.scheduling[str(code)][i][\"glucose_id\"]==glucose_id):\n\t\t\t\tdel self.scheduling[str(code)][i]\n\n", "repo_name": "alessandrobaldo/IoTProject", "sub_path": "Code/Timer/TimeShift.py", "file_name": "TimeShift.py", "file_ext": "py", "file_size_in_byte": 3203, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "24", "api": [{"api_name": "socket.socket", "line_number": 8, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 8, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 8, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 21, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 30, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 44, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 44, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 52, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 52, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 56, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 56, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 56, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 57, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 60, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 64, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 64, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 65, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 69, "usage_type": "call"}]}
+{"seq_id": "33196846487", "text": "import pytest\r\nfrom UI_test.UI_Test_3.MY_python_pytest.src.common.webdriver_handler import WebDriverHandler\r\n\r\n\r\n@pytest.fixture(scope=\"function\")\r\ndef webdriver_handler():\r\n webdriver=WebDriverHandler()\r\n webdriver.setup()\r\n yield webdriver\r\n webdriver.quit()\r\n", "repo_name": "vaniffatiy/UI_Test_3", "sub_path": "MY_python_pytest/src/conftest.py", "file_name": "conftest.py", "file_ext": "py", "file_size_in_byte": 274, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "UI_test.UI_Test_3.MY_python_pytest.src.common.webdriver_handler.WebDriverHandler", "line_number": 7, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 5, "usage_type": "call"}]}
+{"seq_id": "17303629959", "text": "from django.shortcuts import render, redirect, reverse\nfrom django.contrib import messages\nfrom .forms import ContactForm\n\n\n# Create your views here.\ndef contact(request):\n \"\"\" A view to show the contact page form \"\"\"\n\n if request.method == 'POST':\n form_data = {\n 'name': request.POST['name'],\n 'email': request.POST['email'],\n 'message': request.POST['message'],\n }\n contact_form = ContactForm(form_data)\n\n if contact_form.is_valid():\n contact_form.save()\n messages.success(request, 'Message sent successfully!')\n return redirect(reverse('contact'))\n else:\n messages.error(request, 'Message failed to send. Check if form is valid.')\n else:\n contact_form = ContactForm()\n\n context = {\n 'contact_form': contact_form,\n }\n\n return render(request, 'contact/contact.html', context)\n", "repo_name": "oksanaokhten/ambrosia", "sub_path": "contact/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 922, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "forms.ContactForm", "line_number": 16, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 20, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 20, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 21, "usage_type": "call"}, {"api_name": "django.shortcuts.reverse", "line_number": 21, "usage_type": "call"}, {"api_name": "django.contrib.messages.error", "line_number": 23, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 23, "usage_type": "name"}, {"api_name": "forms.ContactForm", "line_number": 25, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 31, "usage_type": "call"}]}
+{"seq_id": "38164832642", "text": "import asyncio\nimport re\n\nimport socks_client.tcp_async as socks\n\n\nasync def tcp_client_through_socks(proxy_host, proxy_port, target_host, target_port):\n tcp_socks = socks.socksocket(\n proxy_type=socks.SOCKS5,\n proxy_host=proxy_host,\n proxy_port=proxy_port,\n username=\"my_username\",\n password=\"my_password\",\n rdns=False,\n )\n await tcp_socks.settimeout(5)\n sock = await tcp_socks.connect(dest_host=target_host, dest_port=target_port)\n\n reader, writer = await asyncio.open_connection(\n host=None,\n port=None,\n sock=sock,\n )\n request = (\n b\"GET / HTTP/1.1\\r\\n\" b\"Host: ip.sb\\r\\n\" b\"User-Agent: curl/7.64.0\\r\\n\\r\\n\"\n )\n writer.write(request)\n\n response = await asyncio.wait_for(reader.read(1024), timeout=1)\n\n response_headers = response.decode(\"utf-8\")\n ip_address = re.search(\n r\"\\d{1,3}\\.\\d{1,3}\\.\\d{1,3}\\.\\d{1,3}\", response_headers\n ).group()\n print(ip_address)\n\n\nasync def main():\n proxy_host = \"\"\n proxy_port = 1080\n target_host = \"ip.sb\"\n target_port = 80\n\n await tcp_client_through_socks(proxy_host, proxy_port, target_host, target_port)\n\n\nif __name__ == \"__main__\":\n asyncio.run(main())\n", "repo_name": "plattanus/socks-client", "sub_path": "try_tcp_async.py", "file_name": "try_tcp_async.py", "file_ext": "py", "file_size_in_byte": 1229, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "socks_client.tcp_async.socksocket", "line_number": 8, "usage_type": "call"}, {"api_name": "socks_client.tcp_async", "line_number": 8, "usage_type": "name"}, {"api_name": "socks_client.tcp_async.SOCKS5", "line_number": 9, "usage_type": "attribute"}, {"api_name": "socks_client.tcp_async", "line_number": 9, "usage_type": "name"}, {"api_name": "asyncio.open_connection", "line_number": 19, "usage_type": "call"}, {"api_name": "asyncio.wait_for", "line_number": 29, "usage_type": "call"}, {"api_name": "re.search", "line_number": 32, "usage_type": "call"}, {"api_name": "asyncio.run", "line_number": 48, "usage_type": "call"}]}
+{"seq_id": "25579514405", "text": "from tensorflow.keras.models import load_model\r\nimport matplotlib.pyplot as plt\r\n\r\nfrom sklearn.decomposition import PCA\r\nfrom sklearn.model_selection import train_test_split\r\nfrom skimage import io,transform #skimage模块下的io transform(图像的形变与缩放)模块\r\nimport glob #glob 文件通配符模块\r\nimport os #os 处理文件和目录的模块\r\nfrom sklearn.utils import shuffle\r\nimport scipy.io as sio\r\nfrom os import listdir\r\nfrom os.path import isfile, join\r\nimport pickle\r\nimport scipy.io as scio\r\nimport numpy as np\r\ndef load_dataset():\r\n def read_img(path):\r\n w = 64\r\n h = 64\r\n c = 3\r\n cate = [path + x for x in os.listdir(path) if os.path.isdir(path + x)]\r\n print(os.listdir(path))\r\n print(cate)\r\n imgs = []\r\n labels = []\r\n for idx, folder in enumerate(cate):\r\n for im in glob.glob(folder + '/*.png'):\r\n img = io.imread(im)\r\n img = transform.resize(img, (w, h))\r\n imgs.append(img)\r\n labels.append(idx)\r\n return np.asarray(imgs, np.float32), np.asarray(labels, np.int32)\r\n\r\n data, label = read_img('new_3/')\r\n data, label = shuffle(data, label)\r\n\r\n def normalization(data):\r\n _range = np.max(data) - np.min(data)\r\n return (data - np.min(data)) / _range\r\n\r\n data = normalization(data) # data / 255\r\n X_train, X_test, y_train, y_test = train_test_split(data, label, random_state=42, test_size=0.2)\r\n return X_test, y_test\r\n\r\n\r\nfrom sklearn.manifold import TSNE\r\n\r\n\r\ndef visualize(embedding, value):\r\n z = TSNE(n_components=2).fit_transform(embedding.detach().cpu().numpy())\r\n plt.figure(figsize=(16, 12))\r\n plt.xticks([])\r\n plt.yticks([])\r\n plt.scatter(z[:, 0], z[:, 1], s=30, c=value, cmap=\"Set2\", zorder=2)\r\n plt.show()\r\n\r\n\r\ndef eeg_test():\r\n X_test, y_test = load_dataset()\r\n output_dir = 'models/gai11'\r\n encoder = load_model(join(output_dir, 'encoder.h5'))\r\n x_preds = encoder.predict(X_test)\r\n #y_preds = encoder.predict(y_test)\r\n print('x_preds:', x_preds.shape)\r\n print('y_test:', y_test.shape)\r\n scio.savemat(r'D:\\contrastive-predictive-coding-master\\result_xy11tu.mat',\r\n {'matrix_1':x_preds,'matrix_2':y_test})\r\n # pca = PCA(n_components=4)\r\n # projected_config = pca.fit_transform(x_preds)\r\n # plt.scatter(projected_config[:, 0], projected_config[:, 1], c=)\r\n # plt.colorbar()\r\n# model.eval()\r\n# out = model(data.x)\r\n visualize(x_preds, color = y_test)\r\neeg_test()\r\n", "repo_name": "yanyy202108/CPC-EEG", "sub_path": "CPC-EEG/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 2553, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "os.listdir", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 22, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 27, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 28, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 28, "usage_type": "name"}, {"api_name": "skimage.transform.resize", "line_number": 29, "usage_type": "call"}, {"api_name": "skimage.transform", "line_number": 29, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 32, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sklearn.utils.shuffle", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 39, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 42, "usage_type": "call"}, {"api_name": "sklearn.manifold.TSNE", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 61, "usage_type": "call"}, {"api_name": "scipy.io.savemat", "line_number": 66, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 66, "usage_type": "name"}]}
+{"seq_id": "38237190943", "text": "import webbrowser\r\nimport urllib.parse\r\nimport requests\r\nfrom bs4 import BeautifulSoup\r\n\r\n\r\n\r\n\r\nprint(''' \r\n _____ _ _ _____ \r\n| _ |_|___| |_| _ |_ _ ___ \r\n| __| | _| '_| |_'_| -_|\r\n|__| |_|___|_,_|__|__|_,_|___| v.1 \r\n The Internet Digger \r\n\r\n +----------------+\r\n | By ZKRIM |\r\n | 2K23 |\r\n +----------------+\r\n\r\n''')\r\n\r\ndef search_on_browser(search_term):\r\n # Encodage du terme de recherche pour les URL\r\n encoded_search_term = urllib.parse.quote(search_term)\r\n\r\n # URL de recherche sur Google (vous pouvez modifier l'URL pour utiliser un autre moteur de recherche)\r\n search_url = f\"https://www.google.com/search?q={encoded_search_term}\"\r\n\r\n # Ouvrir le navigateur avec l'URL de recherche\r\n webbrowser.open(search_url)\r\n\r\ndef get_number_of_results(search_term):\r\n encoded_search_term = urllib.parse.quote(search_term)\r\n search_url = f\"https://www.google.com/search?q={encoded_search_term}\"\r\n response = requests.get(search_url)\r\n\r\n if response.status_code == 200:\r\n soup = BeautifulSoup(response.text, \"html.parser\")\r\n result_stats = soup.find(\"div\", {\"id\": \"result-stats\"})\r\n if result_stats:\r\n num_results = result_stats.get_text()\r\n return num_results\r\n return \"Nombre de résultats non disponible.\"\r\n\r\ndef sous_menu_fonction_1():\r\n prenom = input(\"Prénom : \")\r\n nom = input(\"Nom : \")\r\n search_term = f'intext:\"{prenom} {nom}\"'\r\n num_results = get_number_of_results(search_term)\r\n print(f\"Nombre de résultats trouvés sur Google : {num_results}\")\r\n search_on_browser(search_term)\r\n\r\ndef sous_menu_fonction_2():\r\n prenom = input(\"Prénom : \")\r\n nom = input(\"Nom : \")\r\n search_term = f'inurl:\"{prenom} {nom}\"'\r\n num_results = get_number_of_results(search_term)\r\n print(f\"Nombre de résultats trouvés sur Google : {num_results}\")\r\n search_on_browser(search_term)\r\n\r\ndef sous_menu_fonction_3():\r\n prenom = input(\"Prénom : \")\r\n nom = input(\"Nom : \")\r\n search_term = f'{prenom} {nom} site:youtube.com | site:instagram.com | site:twitter.com | site:linkedin.com | site:tiktok.com | site:facebook.com'\r\n num_results = get_number_of_results(search_term)\r\n print(f\"Nombre de résultats trouvés sur Google : {num_results}\")\r\n search_on_browser(search_term)\r\n\r\ndef sous_menu_fonction_4():\r\n prenom = input(\"Prénom : \")\r\n nom = input(\"Nom : \")\r\n search_term = f'intext:\"{prenom} {nom}\" filetype:pdf'\r\n num_results = get_number_of_results(search_term)\r\n print(f\"Nombre de résultats trouvés sur Google : {num_results}\")\r\n search_on_browser(search_term)\r\n\r\ndef afficher_sous_menu_fonction_1():\r\n while True:\r\n print(\"\\n----------> IDENTITY TRACKING MENU <----------\\n\")\r\n print(\"-> 1. Text Searching\")\r\n print(\"-> 2. Title Searching\")\r\n print(\"-> 3. Social Networks Searching\")\r\n print(\"-> 4. Document Searching\")\r\n print(\"-> 5. Main Menu\\n\")\r\n\r\n choix_sous_menu_fonction_1 = input(\"Choose An Option (1-5) : \")\r\n\r\n if choix_sous_menu_fonction_1 == \"1\":\r\n sous_menu_fonction_1()\r\n elif choix_sous_menu_fonction_1 == \"2\":\r\n sous_menu_fonction_2()\r\n elif choix_sous_menu_fonction_1 == \"3\":\r\n sous_menu_fonction_3()\r\n elif choix_sous_menu_fonction_1 == \"4\":\r\n sous_menu_fonction_4()\r\n elif choix_sous_menu_fonction_1 == \"5\":\r\n break\r\n else:\r\n print(\"Invalid Choice. Please Enter A Valid Number (1-5).\")\r\n\r\ndef fonction_1():\r\n while True:\r\n print(\"\\n----------> MENU <----------:\")\r\n print(\"-> 1. Identity Tracking\")\r\n print(\"-> 2. Username Tracking\")\r\n print(\"-> 3. Quit\\n\")\r\n\r\n choix_fonction_1 = input(\"Choose An Option (1-3) : \")\r\n\r\n if choix_fonction_1 == \"1\":\r\n afficher_sous_menu_fonction_1()\r\n elif choix_fonction_1 == \"2\":\r\n fonction_2()\r\n elif choix_fonction_1 == \"3\":\r\n break\r\n else:\r\n print(\"Invalid Choice. Please Enter A Valid Number (1-3).\")\r\n\r\ndef fonction_2():\r\n username = input(\"Username : \")\r\n search_term = f'{username} site:youtube.com | site:instagram.com | site:twitter.com | site:tiktok.com | site:facebook.com'\r\n num_results = get_number_of_results(search_term)\r\n print(f\"Nombre de résultats trouvés sur Google : {num_results}\")\r\n search_on_browser(search_term)\r\n\r\nif __name__ == \"__main__\":\r\n fonction_1()\r\n", "repo_name": "Warning33/PickAxe", "sub_path": "PickAxe.py", "file_name": "PickAxe.py", "file_ext": "py", "file_size_in_byte": 4581, "program_lang": "python", "lang": "fr", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "urllib.parse.parse.quote", "line_number": 25, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 25, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 25, "usage_type": "name"}, {"api_name": "webbrowser.open", "line_number": 31, "usage_type": "call"}, {"api_name": "urllib.parse.parse.quote", "line_number": 34, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 34, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 34, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 36, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 39, "usage_type": "call"}]}
+{"seq_id": "29878707370", "text": "\r\nfrom setuptools import setup, find_packages\r\n\r\nwith open(\"requirements.txt\", \"r\") as requirement_file:\r\n requirements = requirement_file.read().splitlines()\r\n print(requirements)\r\n\r\nsetup(\r\n name='demo',\r\n version='0.0.1',\r\n package_dir={'': 'src'},\r\n packages=find_packages(\"src\"),\r\n install_requires=requirements,\r\n)", "repo_name": "Xophe92/py.mut-issue-pytest-package", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 341, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "setuptools.setup", "line_number": 8, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 12, "usage_type": "call"}]}
+{"seq_id": "27960654064", "text": "from functools import partial\n\nimport grpc\nimport jwt\n\nfrom internal.dependencies import user_id_context_var\nfrom security.jwt import validate_jwt_token\n\n__all__ = (\n 'AuthInterceptor',\n)\n\n\ndef abort(ignored_request, context, text):\n context.abort(grpc.StatusCode.UNAUTHENTICATED, text)\n\n\n_abort_no_token = partial(abort, text='No JWT token present in request')\n_abort_wrong_token = partial(abort, text='JWT token is not valid')\n\n_abortion_no_token = grpc.unary_unary_rpc_method_handler(_abort_no_token)\n_abortion_wrong_token = grpc.unary_unary_rpc_method_handler(_abort_wrong_token)\n\n\nclass AuthInterceptor(grpc.aio.ServerInterceptor):\n \"\"\"\n Аутентифицирует запросы по переданному JWT токену.\n Пример Bearer eyJhbGciOiJIUzI1NiJ9.eyJzdWIiOjM0fQ.UwT8lPNUcHRAEEGesEIBcYKItbrp04maOG03F22ec0Q\n \"\"\"\n async def intercept_service(self, continuation, handler_call_details):\n for md in handler_call_details.invocation_metadata:\n if md.key == 'authorization':\n _, value = md.value.split(' ')\n try:\n user_id = validate_jwt_token(value.strip())\n except jwt.PyJWTError:\n return _abortion_wrong_token\n else:\n user_id_context_var.set(user_id)\n return await continuation(handler_call_details)\n else:\n return _abortion_no_token\n", "repo_name": "AlekseiKhatkevich/strike", "sub_path": "grpc_services/interceptors.py", "file_name": "interceptors.py", "file_ext": "py", "file_size_in_byte": 1451, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "grpc.StatusCode", "line_number": 15, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 18, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 19, "usage_type": "call"}, {"api_name": "grpc.unary_unary_rpc_method_handler", "line_number": 21, "usage_type": "call"}, {"api_name": "grpc.unary_unary_rpc_method_handler", "line_number": 22, "usage_type": "call"}, {"api_name": "grpc.aio", "line_number": 25, "usage_type": "attribute"}, {"api_name": "security.jwt.validate_jwt_token", "line_number": 35, "usage_type": "call"}, {"api_name": "jwt.PyJWTError", "line_number": 36, "usage_type": "attribute"}, {"api_name": "internal.dependencies.user_id_context_var.set", "line_number": 39, "usage_type": "call"}, {"api_name": "internal.dependencies.user_id_context_var", "line_number": 39, "usage_type": "name"}]}
+{"seq_id": "22576199246", "text": "import pygame\nimport config\nimport terrain\n\noverworld_chunks = [\n ['a1', 'a2', 'a3'],\n ['b1', 'b2', 'b3'],\n ['c1', 'c2', 'c3']\n]\n\noverworld = [['0' for x in range(config.MAP_WIDTH)] for y in range(config.MAP_HEIGHT)]\noverworld_layer_1 = [[' ' for x in range(config.MAP_WIDTH)] for y in range(config.MAP_HEIGHT)]\n\n\noverworld_2 = [['0' for x in range(config.MAP_WIDTH)] for y in range(config.MAP_HEIGHT)]\noverworld2_layer_1 = [[' ' for x in range(config.MAP_WIDTH)] for y in range(config.MAP_HEIGHT)]\n\nunderground = [['+' for x in range(config.MAP_WIDTH)] for y in range(config.MAP_HEIGHT)]\n\noverworld2_layer_1[10][10] = 'p'\noverworld_2[4][4] = 'f'\n\noverworld2_layers = [overworld_2, overworld2_layer_1, False]\n\ndef create_boundary(area, id):\n new_area = area\n \n for x in range(len(area[0])):\n new_area[0][x] = id\n new_area[len(area[0]) - 1][x] = id\n for y in range(len(area)):\n new_area[y][0] = id\n new_area[y][len(area) - 1] = id\n\n return new_area\n\noverworld_layer_1 = create_boundary(overworld_layer_1, 'h')\noverworld2_layer_1 = create_boundary(overworld2_layer_1, 'h')\n\n\noverworld_layer_1[3][2] = 'p'\noverworld_layer_1[0][5] = 'H'\noverworld_layer_1[5][5] = 's'\noverworld2_layer_1[99][5] = 'H'\n\noverworld_layers = [overworld, overworld_layer_1, False]\n\ndef generate_underground_layer():\n return terrain.create_terrain(config.MAP_WIDTH, config.MAP_HEIGHT, 10000)\n\nunderground_1 = generate_underground_layer()\nunderground_1 = create_boundary(underground_1, '#')\n\n# for y in range(len(underground_1)):\n# row = ''\n# for x in range(len(underground_1[y])):\n# row += underground_1[y][x]\n\n# print(row)\nunderground_1[50][50] = 'p'\nunderground_layers = [underground, underground_1, True]\n\ndef get_level(level):\n if level == 'overworld':\n return overworld_layers\n \ncurrent_area = overworld_layers", "repo_name": "taloncouture/Sandbox-Game", "sub_path": "map.py", "file_name": "map.py", "file_ext": "py", "file_size_in_byte": 1874, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "config.MAP_WIDTH", "line_number": 11, "usage_type": "attribute"}, {"api_name": "config.MAP_HEIGHT", "line_number": 11, "usage_type": "attribute"}, {"api_name": "config.MAP_WIDTH", "line_number": 12, "usage_type": "attribute"}, {"api_name": "config.MAP_HEIGHT", "line_number": 12, "usage_type": "attribute"}, {"api_name": "config.MAP_WIDTH", "line_number": 15, "usage_type": "attribute"}, {"api_name": "config.MAP_HEIGHT", "line_number": 15, "usage_type": "attribute"}, {"api_name": "config.MAP_WIDTH", "line_number": 16, "usage_type": "attribute"}, {"api_name": "config.MAP_HEIGHT", "line_number": 16, "usage_type": "attribute"}, {"api_name": "config.MAP_WIDTH", "line_number": 18, "usage_type": "attribute"}, {"api_name": "config.MAP_HEIGHT", "line_number": 18, "usage_type": "attribute"}, {"api_name": "terrain.create_terrain", "line_number": 49, "usage_type": "call"}, {"api_name": "config.MAP_WIDTH", "line_number": 49, "usage_type": "attribute"}, {"api_name": "config.MAP_HEIGHT", "line_number": 49, "usage_type": "attribute"}]}
+{"seq_id": "2462881040", "text": "\"\"\"Utils for monoDepth.\"\"\"\nimport re\nimport sys\n\nimport cv2\nimport numpy as np\nimport torch\n\nfrom imaginairy.modules.midas.api import load_midas_transform\n\n\nclass AddMiDaS:\n def __init__(self, model_type=\"dpt_hybrid\"):\n self.transform = load_midas_transform(model_type)\n\n def pt2np(self, x):\n x = ((x + 1.0) * 0.5).detach().cpu().numpy()\n return x\n\n def np2pt(self, x):\n x = torch.from_numpy(x) * 2 - 1.0\n return x\n\n def __call__(self, img):\n # sample['jpg'] is tensor hwc in [-1, 1] at this point\n img = self.pt2np(img)\n img = self.transform({\"image\": img})[\"image\"]\n return img\n\n\ndef read_pfm(path):\n \"\"\"\n Read pfm file.\n\n Args:\n path (str): path to file\n\n Returns:\n tuple: (data, scale)\n \"\"\"\n with open(path, \"rb\") as file:\n header = file.readline().rstrip()\n if header.decode(\"ascii\") == \"PF\":\n color = True\n elif header.decode(\"ascii\") == \"Pf\":\n color = False\n else:\n raise ValueError(\"Not a PFM file: \" + path)\n\n dim_match = re.match(r\"^(\\d+)\\s(\\d+)\\s$\", file.readline().decode(\"ascii\"))\n if dim_match:\n width, height = list(map(int, dim_match.groups()))\n else:\n raise RuntimeError(\"Malformed PFM header.\")\n\n scale = float(file.readline().decode(\"ascii\").rstrip())\n if scale < 0:\n # little-endian\n endian = \"<\"\n scale = -scale\n else:\n # big-endian\n endian = \">\"\n\n data = np.fromfile(file, endian + \"f\")\n shape = (height, width, 3) if color else (height, width)\n\n data = np.reshape(data, shape)\n data = np.flipud(data)\n\n return data, scale\n\n\ndef write_pfm(path, image, scale=1):\n \"\"\"\n Write pfm file.\n\n Args:\n path (str): pathto file\n image (array): data\n scale (int, optional): Scale. Defaults to 1.\n \"\"\"\n\n with open(path, \"wb\") as file:\n if image.dtype.name != \"float32\":\n raise ValueError(\"Image dtype must be float32.\")\n\n image = np.flipud(image)\n\n if len(image.shape) == 3 and image.shape[2] == 3: # color image\n color = True\n elif (\n len(image.shape) == 2 or len(image.shape) == 3 and image.shape[2] == 1\n ): # greyscale\n color = False\n else:\n msg = \"Image must have H x W x 3, H x W x 1 or H x W dimensions.\"\n raise ValueError(msg)\n\n file.write(\"PF\\n\" if color else b\"Pf\\n\")\n file.write(b\"%d %d\\n\" % (image.shape[1], image.shape[0]))\n\n endian = image.dtype.byteorder\n\n if endian == \"<\" or endian == \"=\" and sys.byteorder == \"little\":\n scale = -scale\n\n file.write(b\"%f\\n\" % scale)\n\n image.tofile(file)\n\n\ndef read_image(path):\n \"\"\"\n Read image and output RGB image (0-1).\n\n Args:\n path (str): path to file\n\n Returns:\n array: RGB image (0-1)\n \"\"\"\n img = cv2.imread(path)\n\n if img.ndim == 2:\n img = cv2.cvtColor(img, cv2.COLOR_GRAY2BGR)\n\n img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB) / 255.0\n\n return img\n\n\ndef resize_image(img):\n \"\"\"\n Resize image and make it fit for network.\n\n Args:\n img (array): image\n\n Returns:\n tensor: data ready for network\n \"\"\"\n height_orig = img.shape[0]\n width_orig = img.shape[1]\n\n scale = width_orig / 384 if width_orig > height_orig else height_orig / 384\n\n height = (np.ceil(height_orig / scale / 32) * 32).astype(int)\n width = (np.ceil(width_orig / scale / 32) * 32).astype(int)\n\n img_resized = cv2.resize(img, (width, height), interpolation=cv2.INTER_AREA)\n\n img_resized = (\n torch.from_numpy(np.transpose(img_resized, (2, 0, 1))).contiguous().float()\n )\n img_resized = img_resized.unsqueeze(0)\n\n return img_resized\n\n\ndef resize_depth(depth, width, height):\n \"\"\"\n Resize depth map and bring to CPU (numpy).\n\n Args:\n depth (tensor): depth\n width (int): image width\n height (int): image height\n\n Returns:\n array: processed depth\n \"\"\"\n depth = torch.squeeze(depth[0, :, :, :]).to(\"cpu\")\n\n depth_resized = cv2.resize(\n depth.numpy(), (width, height), interpolation=cv2.INTER_CUBIC\n )\n\n return depth_resized\n\n\ndef write_depth(path, depth, bits=1):\n \"\"\"\n Write depth map to pfm and png file.\n\n Args:\n path (str): filepath without extension\n depth (array): depth\n \"\"\"\n write_pfm(path + \".pfm\", depth.astype(np.float32))\n\n depth_min = depth.min()\n depth_max = depth.max()\n\n max_val = (2 ** (8 * bits)) - 1\n\n if depth_max - depth_min > np.finfo(\"float\").eps:\n out = max_val * (depth - depth_min) / (depth_max - depth_min)\n else:\n out = np.zeros(depth.shape, dtype=depth.type)\n\n if bits == 1:\n cv2.imwrite(path + \".png\", out.astype(\"uint8\"))\n elif bits == 2:\n cv2.imwrite(path + \".png\", out.astype(\"uint16\"))\n", "repo_name": "brycedrennan/imaginAIry", "sub_path": "imaginairy/modules/midas/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 5014, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 7408, "dataset": "github-code", "pt": "24", "api": [{"api_name": "imaginairy.modules.midas.api.load_midas_transform", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 21, "usage_type": "call"}, {"api_name": "re.match", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.fromfile", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.flipud", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.flipud", "line_number": 88, "usage_type": "call"}, {"api_name": "sys.byteorder", "line_number": 105, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 123, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 126, "usage_type": "call"}, {"api_name": "cv2.COLOR_GRAY2BGR", "line_number": 126, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 128, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 128, "usage_type": "attribute"}, {"api_name": "numpy.ceil", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 149, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 151, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 151, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 154, "usage_type": "call"}, {"api_name": "torch.squeeze", "line_number": 173, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 175, "usage_type": "call"}, {"api_name": "cv2.INTER_CUBIC", "line_number": 176, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 190, "usage_type": "attribute"}, {"api_name": "numpy.finfo", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 200, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 203, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 205, "usage_type": "call"}]}
+{"seq_id": "71679106622", "text": "from django.db import models\nimport json\n\nclass Car(models.Model):\n # mã của giấy đăng ký\n registration_id = models.TextField(primary_key=True, null=False, max_length=20, unique=True)\n # biển số đăng ký\n plate_number = models.TextField(null=False, max_length=20, unique=True)\n # nơi đăng ký\n registration_place = models.TextField(null=False)\n # ngày cấp\n registration_date = models.DateField(null=False)\n # mục đích\n purpose = models.CharField(\n max_length=100,\n choices=[\n ('personal', 'Đi lại cá nhân'),\n ('passenger_service', 'Dịch vụ chở khách'),\n ('transportation_service', 'Dịch vụ vận tải'),\n ],\n default='personal',\n )\n # loại xe\n type = models.TextField(null=False)\n # hãng xe\n manufacturer = models.TextField(null=False)\n # mẫu xe\n model = models.TextField(null=False)\n\n # extra \n engine_number = models.TextField(null=False)\n chassis_number = models.TextField(null=False)\n # power = models.FloatField(blank=True, null=True)\n # torque = models.FloatField(blank=True, null=True)\n # transmission = models.CharField(max_length=50, blank=True, null=True)\n # seating_capacity = models.PositiveIntegerField(blank=True, null=True)\n # length = models.FloatField(blank=True, null=True)\n # width = models.FloatField(blank=True, null=True)\n # height = models.FloatField(blank=True, null=True)\n # weight = models.FloatField(blank=True, null=True)\n # fuel_consumption = models.FloatField(blank=True, null=True)\n # suspension = models.CharField(max_length=100, blank=True, null=True)\n # braking_system = models.CharField(max_length=100, blank=True, null=True)\n\n # chủ sở hữu\n owner = models.ForeignKey('owner.Owner', on_delete=models.CASCADE, null=True, blank=True)\n # đã đăng kiểm\n inspection_status = models.CharField(max_length=100, default='Chưa đăng kiểm')\n", "repo_name": "minhduc1122002/registry-total", "sub_path": "backend/car/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1992, "program_lang": "python", "lang": "vi", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.db.models.Model", "line_number": 4, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 4, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 6, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 46, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 46, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 46, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 48, "usage_type": "name"}]}
+{"seq_id": "75226318142", "text": "import smtplib\nimport datetime as dt\nimport random\nimport mailchimp_marketing as MailchimpMarketing\nfrom mailchimp_marketing.api_client import ApiClientError\n\n\n# |------------------------------------GETTING EMAILS FROM MAILCHIMP API------------------------------------|#\n\n# This is where subscribed members' emails will be saved to\nemails = []\n\n# Gets member info from MailChimp\ntry:\n client = MailchimpMarketing.Client()\n client.set_config({\n \"api_key\": \"ENTER_API_KEY_HERE\",\n \"server\": \"ENTER_YOUR_SERVER_HERE\"})\n\n # If members are currently subscribed, adds them to \"emails\" list\n response = client.lists.get_list_members_info(\"ENTER_YOUR_LIST_ID_HERE\")\n for member in response['members']:\n if member['status'] == \"subscribed\":\n emails.append(member['email_address'])\n\n\nexcept ApiClientError as error:\n print(\"Error: {}\".format(error.text))\n\n\n# |------------------------------------LOG IN FOR GMAIL------------------------------------|#\n\nmy_email = \"ENTER_YOUR_GMAIL_HERE\"\npassword = \"ENTER_YOUR_GMAIL_APP_KEY_HERE\"\nunsub_link_url = \"tinyurl.com/unsub321\"\n\n\n# |------------------------------------DATE FORMATTED FOR EMAIL TO BE SENT------------------------------------|#\n\nnow = dt.datetime.now()\nyear = now.year\nmonth = now.month\nday = now.day\nweekday = now.weekday() # 0 = monday, 1 = tuesday, etc\ntoday = f\"{month}/{day}/{year}\"\n\n\n# |------------------------------------WORKOUTS------------------------------------|#\nupper_body_beg = (\n \"Push-Ups: 2 sets of 20\",\n \"Chin-Ups: 2 sets of 5\",\n \"Wall Handstand: 2 sets of 30 seconds\",\n \"Wide-Grip Push-Ups: 2 sets of 20\",\n \"Close-Grip Push-Ups: 2 sets of 20\",\n)\n\nupper_body_int = (\n \"Feet Elevated Push-Ups: 2 sets of 20\",\n \"Pull-Ups: 2 sets of 10\",\n \"Feet Elevated Pike Push-Ups: 2 sets of 15\",\n \"Chin-Ups: 2 sets of 10\",\n \"Wall Handstand: 2 sets of 60 seconds\",\n \"Diamond Push-Ups: 2 sets of 15\",\n\n)\n\nupper_body_adv = (\n \"Feet Elevated Push-Ups: 3 sets of 25\",\n \"Archer Push-Ups: 3 sets of 12 (6 per side)\",\n \"Archer Pull-Ups: 3 sets of 6 (3 per side)\",\n \"One-Arm Push-Ups: 3 sets of 12 (6 per side)\",\n \"Diamond Push-Ups: 3 sets of 20\",\n \"Handstand Push-Ups: 3 sets of 10\",\n)\n\nlower_body_beg = (\n \"Squats: 2 sets of 20\",\n \"Lunges: 2 sets of 10 (each side)\",\n \"Hip Bridge: 2 sets of 20\",\n \"Lying Knee Tuck: 2 sets of 20\",\n \"Wall Sits: 2 sets of 60 seconds\",\n)\n\nlower_body_int = (\n \"Jump Squats: 2 sets of 20\",\n \"Lunges: 2 sets of 20 (each side)\",\n \"Jump Lunges: 2 sets of 12 (each side)\",\n \"Close-Feet Squats: 2 sets of 20\",\n \"Wall Sits: 2 sets of 2 minutes\",\n)\n\nlower_body_adv = (\n \"Pistol Squats: 3 sets of 10 (each side)\",\n \"Jump Squats: 3 sets of 15\",\n \"Jumping Lunges: 3 sets of 15 (each side)\",\n \"Shrimp Squat: 3 sets of 10 (each side)\",\n)\n\n\ncore = (\n \"Plank: 2 sets of 2 minutes\",\n \"Side Plank: 2 minutes (each side)\",\n \"Bicycle Crunches: 2 sets of 30 (each side)\",\n \"V-Ups: 2 sets of 30 reps\",\n \"Crunches: 2 sets of 30 reps\",\n \"Sit-Ups: 2 sets of 30 reps\",\n \"Russian Twists: 2 sets of 30 (each side)\",\n \"Scissor Kicks: 2 sets of 30 (each side)\",\n \"Leg Raises: 2 sets of 15\",\n\n)\n\n\n# |------------------------------------FUNCTION TO GENERATE UNIQUE WORKOUT------------------------------------|#\n\ndef generate_workout():\n upper_body_beg_choice = random.choice(upper_body_beg)\n upper_body_int_choice = random.choice(upper_body_int)\n upper_body_adv_choice = random.choice(upper_body_adv)\n\n lower_body_beg_choice = random.choice(lower_body_beg)\n lower_body_int_choice = random.choice(lower_body_int)\n lower_body_adv_choice = random.choice(lower_body_adv)\n\n core_choice = random.choice(core)\n core_choice_2 = random.choice(core)\n\n return f\"Beginner Workout: \\n\\n{upper_body_beg_choice} \\n{lower_body_beg_choice} \\n{core_choice}\\n\\n\\n\" \\\n f\"-------------------------------------------------------------------------------\\n\\n\\n\" \\\n f\"Intermediate Workout: \\n\\n{upper_body_int_choice} \\n{lower_body_int_choice} \\n{core_choice}\\n\\n\\n\" \\\n f\"-------------------------------------------------------------------------------\\n\\n\\n\" \\\n f\"Advanced Workout: \\n\\n{upper_body_adv_choice} \\n{lower_body_adv_choice} \\n{core_choice} \\n{core_choice_2}\\n\\n\\n\" \\\n f\"-------------------------------------------------------------------------------\\n\\n\\n\" \\\n f\"How to Use: If you're unable to complete the sets and reps required, \" \\\n f\"feel free to split the reps up and complete more sets to finish all the reps. \" \\\n f\"Once you're able to easily complete all sets and reps as written, move on to the next difficulty level.\\n\\n\\n\\n\\n\\n\" \\\n\n\n# |---------------------------------EMAILING THE EMAIL-LIST THE GENERATED WORKOUT---------------------------------|#\n\n\nwith smtplib.SMTP(\"smtp.gmail.com\", 587) as connection:\n connection.starttls()\n connection.login(user=my_email, password=password)\n connection.sendmail(from_addr=my_email,\n to_addrs=emails,\n msg=f\"Subject:Wake Up, Workout! {today} \\n\\n{generate_workout()} \\n\\n To unsubscribe, click here: {unsub_link_url}\"\n )\n\n", "repo_name": "deikaplan/Wake-Up-Workout-Project", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 5230, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "mailchimp_marketing.Client", "line_number": 15, "usage_type": "call"}, {"api_name": "mailchimp_marketing.api_client.ApiClientError", "line_number": 27, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 40, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 117, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 118, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 119, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 121, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 122, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 123, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 125, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 126, "usage_type": "call"}, {"api_name": "smtplib.SMTP", "line_number": 142, "usage_type": "call"}]}
+{"seq_id": "1267170662", "text": "from . import init_app, db\nfrom .models import Variables\nfrom .models import Geography\nfrom .models import GeographyTypes\nfrom .aggregate import Aggregator\nfrom .color import (color, labeled_color_map)\nfrom .legend import (make_numerical_legend, make_labeled_legend)\n\nimport argparse\nfrom pathlib import Path\nimport yaml\nfrom yaml.loader import SafeLoader\nimport pandas\n\ndef main():\n parser = argparse.ArgumentParser()\n parser.add_argument('--update-legends', action='store_true', default=False,\n help=\"don't perform database commits, only update legends\")\n parser.add_argument('config', type=Path, help=\"name of configuration file\")\n args = parser.parse_args()\n\n app = init_app()\n\n # read configuration\n with open(args.config) as f:\n cfg = yaml.load(f, Loader=SafeLoader)\n\n # read data\n if cfg[\"type\"] == \"excel\":\n data = pandas.read_excel(cfg[\"path\"])\n elif cfg[\"type\"] == \"csv\":\n data = pandas.read_csv(cfg[\"path\"])\n else:\n parser.error(f\"unkown data type {cfg['type']}\")\n\n with app.app_context():\n # check datafile\n error = False\n for variable in cfg[\"variables\"]:\n if variable[\"geometry\"] not in data.columns:\n print(f'Error: geometry column {variable[\"geometry\"]}'\n ' not in file')\n error = True\n if variable[\"file_var\"] not in data.columns:\n print(f'Error: geometry column {variable[\"file_var\"]}'\n ' not in file')\n error = True\n v = db.session.query(Variables).filter(\n Variables.id == variable[\"db_var\"])\n if v.one_or_none() is None:\n print(f'Error: db variable {variable[\"db_var\"]} not defined')\n error = True\n if error:\n parser.error(f'errors in variable description file {cfg[\"path\"]}')\n\n agg = Aggregator()\n for variable in cfg[\"variables\"]:\n print(variable['db_var'])\n values = data[[variable[\"geometry\"], variable[\"file_var\"]]]\n values = values.rename(columns={variable[\"geometry\"]: 'gss_id',\n variable[\"file_var\"]: 'value'})\n values[\"year\"] = variable[\"year\"]\n values[\"variable_id\"] = variable[\"db_var\"]\n\n # Calculate color values\n layer_name = '{db_var}_{year}_S01'.format(**variable)\n if variable[\"colormethod\"] == 'labeled':\n # TODO check that 'label_var' is provided in variable\n values_and_labels = data[[variable['file_var'], variable['label_var']]]\n values_and_labels = values_and_labels.rename(columns={\n variable['file_var']: 'value',\n variable['label_var']: 'label',\n })\n values_and_labels = values_and_labels.drop_duplicates()\n values_and_labels.sort_values('value', inplace=True)\n values[\"color\"], cmap, limits = labeled_color_map(\n variable['colormap'],\n values['value'].to_numpy(),\n values_and_labels['value'],\n reverse_colors=variable.get('reverse_color', False),\n )\n make_labeled_legend(layer_name, cmap, values_and_labels['label'].to_numpy(), width=240)\n else:\n values[\"color\"], cmap, limits = color(variable, values[\"value\"].to_numpy())\n make_numerical_legend(layer_name, cmap, limits)\n\n # Aggregate data for composite geometries\n if 'aggregatemethod' in variable:\n meta_column_label = variable.get(\"aggregatemeta\", \"population\")\n meta_column = cfg[\"metadata\"][meta_column_label][variable[\"year\"]]\n population = data[[\n variable[\"geometry\"],\n variable[\"file_var\"],\n meta_column,\n ]]\n population = population.rename(columns={\n variable[\"geometry\"]: 'gss_id',\n variable[\"file_var\"]: 'value',\n meta_column: 'population'\n })\n population.set_index('gss_id', inplace=True)\n\n geography_types = db.session.query(GeographyTypes).where(\n (GeographyTypes.column_name != None),\n (GeographyTypes.gss_code != 'S01')\n )\n for geo_type in geography_types:\n print( 'aggregate {db_var} in {0} using {aggregatemethod}'.format(\n geo_type.gss_code,\n **variable,\n ))\n composite_geographies = db.session.query(Geography).where(\n Geography.gss_code == geo_type.gss_code\n )\n agg_values = agg.aggregate(\n variable['aggregatemethod'],\n composite_geographies,\n population,\n variable[\"year\"],\n variable[\"db_var\"],\n )\n agg_values['color'], cmap, limits = color(\n variable,\n agg_values['value'].to_numpy(),\n )\n layer_name = '{db_var}_{year}_{0}'.format(geo_type.gss_code, **variable)\n make_numerical_legend(layer_name, cmap, limits)\n values = pandas.concat((values, agg_values))\n\n if not args.update_legends:\n values.to_sql(\"data\", con=db.session.get_bind(),\n index=False, if_exists=\"append\", method=\"multi\")\n\n\n\nif __name__ == \"__main__\":\n main()\n", "repo_name": "EnvironmentSocietyHealth/CRESHMap", "sub_path": "CRESHMap/load_variables.py", "file_name": "load_variables.py", "file_ext": "py", "file_size_in_byte": 5827, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 19, "usage_type": "name"}, {"api_name": "yaml.load", "line_number": 26, "usage_type": "call"}, {"api_name": "yaml.loader.SafeLoader", "line_number": 26, "usage_type": "name"}, {"api_name": "pandas.read_excel", "line_number": 30, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 32, "usage_type": "call"}, {"api_name": "models.Variables", "line_number": 48, "usage_type": "argument"}, {"api_name": "models.Variables.id", "line_number": 49, "usage_type": "attribute"}, {"api_name": "models.Variables", "line_number": 49, "usage_type": "name"}, {"api_name": "aggregate.Aggregator", "line_number": 56, "usage_type": "call"}, {"api_name": "color.labeled_color_map", "line_number": 76, "usage_type": "call"}, {"api_name": "legend.make_labeled_legend", "line_number": 82, "usage_type": "call"}, {"api_name": "color.color", "line_number": 84, "usage_type": "call"}, {"api_name": "legend.make_numerical_legend", "line_number": 85, "usage_type": "call"}, {"api_name": "models.GeographyTypes", "line_number": 103, "usage_type": "argument"}, {"api_name": "models.GeographyTypes.column_name", "line_number": 104, "usage_type": "attribute"}, {"api_name": "models.GeographyTypes", "line_number": 104, "usage_type": "name"}, {"api_name": "models.GeographyTypes.gss_code", "line_number": 105, "usage_type": "attribute"}, {"api_name": "models.GeographyTypes", "line_number": 105, "usage_type": "name"}, {"api_name": "models.Geography", "line_number": 112, "usage_type": "argument"}, {"api_name": "models.Geography.gss_code", "line_number": 113, "usage_type": "attribute"}, {"api_name": "models.Geography", "line_number": 113, "usage_type": "name"}, {"api_name": "color.color", "line_number": 122, "usage_type": "call"}, {"api_name": "legend.make_numerical_legend", "line_number": 127, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 128, "usage_type": "call"}]}
+{"seq_id": "13912849809", "text": "import discord\nimport os\nfrom discord.ext import commands\n\nclass ServerInfo(commands.Cog):\n \n def __inti__(self, client):\n self.client = client\n \n @commands.Cog.listener()\n async def on_ready(self):\n print('Server Info Cog is Running')\n \n @commands.command()\n async def serverinfo(self, ctx):\n \"\"\"Shows server info\"\"\"\n\n server = ctx.guild\n\n roles = str(len(server.roles))\n emojis = str(len(server.emojis))\n channels = str(len(server.channels))\n\n embeded = discord.Embed(title=server.name, description='Server Info', color=0xEE8700)\n embeded.set_thumbnail(url=server.icon_url)\n embeded.add_field(name=\"Created on:\", value=server.created_at.strftime('%d %B %Y at %H:%M UTC+3'), inline=False)\n embeded.add_field(name=\"Server ID:\", value=server.id, inline=False)\n embeded.add_field(name=\"Users on server:\", value=server.member_count, inline=True)\n\n embeded.add_field(name=\"Server Region:\", value=server.region, inline=True)\n embeded.add_field(name=\"Verification Level:\", value=server.verification_level, inline=True)\n\n embeded.add_field(name=\"Role Count:\", value=roles, inline=True)\n embeded.add_field(name=\"Emoji Count:\", value=emojis, inline=True)\n embeded.add_field(name=\"Channel Count:\", value=channels, inline=True)\n\n await ctx.channel.send(embed=embeded)\n\ndef setup(client):\n client.add_cog(ServerInfo(client))\n \ndef teardown(client):\n print('Server Info Cog is now not running')", "repo_name": "Archers007/DesertedBot", "sub_path": "Cog1/ServerInfo.py", "file_name": "ServerInfo.py", "file_ext": "py", "file_size_in_byte": 1548, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 5, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 5, "usage_type": "name"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 10, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 10, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 10, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 24, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 14, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 14, "usage_type": "name"}]}
+{"seq_id": "71043301822", "text": "# import the necessary packages\nfrom end2end.transform import four_point_transform\nfrom skimage.filters import threshold_local\nimport numpy as np\nimport argparse\nimport cv2\nimport imutils\n\ndef preprocess(image):\n # load the image and compute the ratio of the old height\n # to the new height, clone it, and resize it\n\n # image = cv2.imread(img_dir)\n # Replace with image data input\n\n ratio = image.shape[0] / 500.0\n orig = image.copy()\n image = imutils.resize(image, height = 500)\n # convert the image to grayscale, blur it, and find edges\n # in the image\n gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n # gray = cv2.GaussianBlur(gray, (5, 5), 0)\n kernel = np.array([[-1,-1,-1], \n [-1, 9,-1],\n [-1,-1,-1]])\n gray = cv2.filter2D(gray, -1, kernel)\n gray = cv2.GaussianBlur(gray, (7, 7), 0)\n edged = cv2.Canny(gray, 75, 300)\n\n # find the contours in the edged image, keeping only the\n # largest ones, and initialize the screen contour\n cnts = cv2.findContours(edged.copy(), cv2.RETR_LIST, cv2.CHAIN_APPROX_SIMPLE)\n cnts = imutils.grab_contours(cnts)\n cnts = sorted(cnts, key = cv2.contourArea, reverse = True)[:5]\n # loop over the contours\n for c in cnts:\n # approximate the contour\n peri = cv2.arcLength(c, True)\n approx = cv2.approxPolyDP(c, 0.02 * peri, True)\n # if our approximated contour has four points, then we\n # can assume that we have found our screen\n if len(approx) == 4:\n screenCnt = approx\n break\n \n try:\n screenCnt\n except:\n raise ValueError(\"Image has no clear 4 sides\")\n finally:\n return orig\n \n # show the contour (outline) of the piece of paper\n # cv2.drawContours(image, [screenCnt], -1, (0, 255, 0), 2)\n\n # apply the four point transform to obtain a top-down\n # view of the original image\n warped = four_point_transform(orig, screenCnt.reshape(4, 2) * ratio)\n # convert the warped image to grayscale, then threshold it\n # to give it that 'black and white' paper effect\n warped = cv2.cvtColor(warped, cv2.COLOR_BGR2GRAY)\n T = threshold_local(warped, 11, offset = 10, method = \"gaussian\")\n warped = (warped > T).astype(\"uint8\") * 255\n\n warped = cv2.cvtColor(warped, cv2.COLOR_GRAY2RGB)\n\n return warped\n\n", "repo_name": "ikhovryak/LeggoDutch", "sub_path": "web_app/end2end/preprocess_image.py", "file_name": "preprocess_image.py", "file_ext": "py", "file_size_in_byte": 2372, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "imutils.resize", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.GaussianBlur", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.Canny", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.RETR_LIST", "line_number": 32, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 32, "usage_type": "attribute"}, {"api_name": "imutils.grab_contours", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.contourArea", "line_number": 34, "usage_type": "attribute"}, {"api_name": "cv2.arcLength", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.approxPolyDP", "line_number": 39, "usage_type": "call"}, {"api_name": "end2end.transform.four_point_transform", "line_number": 58, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 61, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 61, "usage_type": "attribute"}, {"api_name": "skimage.filters.threshold_local", "line_number": 62, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.COLOR_GRAY2RGB", "line_number": 65, "usage_type": "attribute"}]}
+{"seq_id": "42058879323", "text": "from typing import Dict\n\nfrom pyspark.sql import SparkSession, Column, DataFrame\n\n# noinspection PyUnresolvedReferences\nfrom pyspark.sql.functions import col, lit\n\nfrom spark_auto_mapper.automappers.automapper import AutoMapper\nfrom spark_auto_mapper.helpers.automapper_helpers import AutoMapperHelpers as A\nfrom spark_auto_mapper.helpers.expression_comparer import assert_compare_expressions\n\n\ndef test_auto_mapper_boolean(spark_session: SparkSession) -> None:\n # Arrange\n spark_session.createDataFrame(\n [\n (1, \"Qureshi\", \"Imran\", \"0\"),\n (2, \"Vidal\", \"Michael\", \"1\"),\n ],\n [\"member_id\", \"last_name\", \"first_name\", \"my_age\"],\n ).createOrReplaceTempView(\"patients\")\n\n source_df: DataFrame = spark_session.table(\"patients\")\n\n df = source_df.select(\"member_id\")\n df.createOrReplaceTempView(\"members\")\n\n # Act\n mapper = AutoMapper(\n view=\"members\", source_view=\"patients\", keys=[\"member_id\"]\n ).columns(\n age=A.boolean(A.column(\"my_age\")),\n is_active=A.boolean(\"False\"),\n )\n\n assert isinstance(mapper, AutoMapper)\n sql_expressions: Dict[str, Column] = mapper.get_column_specs(source_df=source_df)\n for column_name, sql_expression in sql_expressions.items():\n print(f\"{column_name}: {sql_expression}\")\n\n assert_compare_expressions(\n sql_expressions[\"age\"], col(\"b.my_age\").cast(\"boolean\").alias(\"age\")\n )\n assert_compare_expressions(\n sql_expressions[\"is_active\"], lit(\"False\").cast(\"boolean\").alias(\"is_active\")\n )\n\n result_df: DataFrame = mapper.transform(df=df)\n\n # Assert\n result_df.printSchema()\n result_df.show()\n\n assert result_df.where(\"member_id == 1\").select(\n \"age\",\n \"is_active\",\n ).collect()[0][\n :\n ] == (False, False)\n assert result_df.where(\"member_id == 2\").select(\n \"age\",\n \"is_active\",\n ).collect()[0][\n :\n ] == (True, False)\n\n assert dict(result_df.dtypes)[\"age\"] == \"boolean\"\n assert dict(result_df.dtypes)[\"is_active\"] == \"boolean\"\n", "repo_name": "icanbwell/SparkAutoMapper", "sub_path": "tests/boolean/test_automapper_boolean.py", "file_name": "test_automapper_boolean.py", "file_ext": "py", "file_size_in_byte": 2064, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 5, "dataset": "github-code", "pt": "24", "api": [{"api_name": "pyspark.sql.SparkSession", "line_number": 13, "usage_type": "name"}, {"api_name": "pyspark.sql.DataFrame", "line_number": 23, "usage_type": "name"}, {"api_name": "spark_auto_mapper.automappers.automapper.AutoMapper", "line_number": 29, "usage_type": "call"}, {"api_name": "spark_auto_mapper.helpers.automapper_helpers.AutoMapperHelpers.boolean", "line_number": 32, "usage_type": "call"}, {"api_name": "spark_auto_mapper.helpers.automapper_helpers.AutoMapperHelpers", "line_number": 32, "usage_type": "name"}, {"api_name": "spark_auto_mapper.helpers.automapper_helpers.AutoMapperHelpers.column", "line_number": 32, "usage_type": "call"}, {"api_name": "spark_auto_mapper.helpers.automapper_helpers.AutoMapperHelpers.boolean", "line_number": 33, "usage_type": "call"}, {"api_name": "spark_auto_mapper.helpers.automapper_helpers.AutoMapperHelpers", "line_number": 33, "usage_type": "name"}, {"api_name": "spark_auto_mapper.automappers.automapper.AutoMapper", "line_number": 36, "usage_type": "argument"}, {"api_name": "typing.Dict", "line_number": 37, "usage_type": "name"}, {"api_name": "pyspark.sql.Column", "line_number": 37, "usage_type": "name"}, {"api_name": "spark_auto_mapper.helpers.expression_comparer.assert_compare_expressions", "line_number": 41, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 42, "usage_type": "call"}, {"api_name": "spark_auto_mapper.helpers.expression_comparer.assert_compare_expressions", "line_number": 44, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.lit", "line_number": 45, "usage_type": "call"}, {"api_name": "pyspark.sql.DataFrame", "line_number": 48, "usage_type": "name"}]}
+{"seq_id": "40829735034", "text": "import cv2\nfrom flask import Flask, Response\nfrom flask_sock import Sock\n\nimport RPi.GPIO as GPIO\nimport time\n# ! /usr/bin/python\n# -*- coding: utf-8 -*-\n\n# 17\n# 27\n# 22\n# 10\nRIGHT_STRAIGHT = 27\nRIGHT_REVERSE = 17\nLEFT_REVERSE = 22\nLEFT_STRAIGHT = 10\n\n\nGPIO.setmode(GPIO.BCM)\nGPIO.setup(RIGHT_STRAIGHT, GPIO.OUT)\nGPIO.setup(RIGHT_REVERSE, GPIO.OUT)\nGPIO.setup(LEFT_STRAIGHT, GPIO.OUT)\nGPIO.setup(LEFT_REVERSE, GPIO.OUT)\n\nconnected = set()\napp = Flask('__name__')\nsock = Sock(app)\nvideo = cv2.VideoCapture(0)\n\n\ndef video_stream():\n while True:\n ret, frame = video.read()\n if not ret:\n break\n else:\n ret, buffer = cv2.imencode('.jpeg', frame)\n frame = buffer.tobytes()\n yield (b' --frame\\r\\n' b'Content-type: imgae/jpeg\\r\\n\\r\\n' + frame + b'\\r\\n')\n\n\n@app.route('/video_feed')\ndef video_feed():\n return Response(video_stream(), mimetype='multipart/x-mixed-replace; boundary=frame')\n\n@sock.route(\"/sock\")\ndef server(ws):\n while True:\n try:\n message = ws.receive()\n ws.send(f' recebido: {message}')\n if (message == \"up\"):\n GPIO.output(RIGHT_STRAIGHT, GPIO.HIGH)\n GPIO.output(LEFT_STRAIGHT, GPIO.HIGH)\n time.sleep(0.2)\n GPIO.output(RIGHT_STRAIGHT, GPIO.LOW)\n GPIO.output(LEFT_STRAIGHT, GPIO.LOW)\n ws.send(f' recebido: {message}')\n elif (message == \"down\"):\n GPIO.output(RIGHT_REVERSE, GPIO.HIGH)\n GPIO.output(LEFT_REVERSE, GPIO.HIGH)\n time.sleep(0.2)\n GPIO.output(RIGHT_REVERSE, GPIO.LOW)\n GPIO.output(LEFT_REVERSE, GPIO.LOW)\n ws.send(f' recebido: {message}')\n elif (message == \"left\"):\n GPIO.output(LEFT_STRAIGHT, GPIO.HIGH)\n time.sleep(0.2)\n GPIO.output(LEFT_STRAIGHT, GPIO.LOW)\n ws.send(f' recebido: {message}')\n elif (message == \"right\"):\n GPIO.output(RIGHT_STRAIGHT, GPIO.HIGH)\n time.sleep(0.2)\n GPIO.output(RIGHT_STRAIGHT, GPIO.LOW)\n ws.send(f' recebido: {message}')\n except:\n print('erro na execução')\n connected.remove(ws)\n GPIO.cleanup()\n\n\nif __name__ == \"__main__\":\n app.run(host='0.0.0.0', port=5000, debug=False)\n video.release()\n", "repo_name": "lucassenazuza/roboto", "sub_path": "server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 2431, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "RPi.GPIO.setmode", "line_number": 20, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 20, "usage_type": "name"}, {"api_name": "RPi.GPIO.BCM", "line_number": 20, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 21, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 21, "usage_type": "name"}, {"api_name": "RPi.GPIO.OUT", "line_number": 21, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 22, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 22, "usage_type": "name"}, {"api_name": "RPi.GPIO.OUT", "line_number": 22, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 23, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 23, "usage_type": "name"}, {"api_name": "RPi.GPIO.OUT", "line_number": 23, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 24, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 24, "usage_type": "name"}, {"api_name": "RPi.GPIO.OUT", "line_number": 24, "usage_type": "attribute"}, {"api_name": "flask.Flask", "line_number": 27, "usage_type": "call"}, {"api_name": "flask_sock.Sock", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.imencode", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 45, "usage_type": "call"}, {"api_name": "RPi.GPIO.output", "line_number": 54, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 54, "usage_type": "name"}, {"api_name": "RPi.GPIO.HIGH", "line_number": 54, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.output", "line_number": 55, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 55, "usage_type": "name"}, {"api_name": "RPi.GPIO.HIGH", "line_number": 55, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 56, "usage_type": "call"}, {"api_name": "RPi.GPIO.output", "line_number": 57, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 57, "usage_type": "name"}, {"api_name": "RPi.GPIO.LOW", "line_number": 57, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.output", "line_number": 58, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 58, "usage_type": "name"}, {"api_name": "RPi.GPIO.LOW", "line_number": 58, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.output", "line_number": 61, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 61, "usage_type": "name"}, {"api_name": "RPi.GPIO.HIGH", "line_number": 61, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.output", "line_number": 62, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 62, "usage_type": "name"}, {"api_name": "RPi.GPIO.HIGH", "line_number": 62, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 63, "usage_type": "call"}, {"api_name": "RPi.GPIO.output", "line_number": 64, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 64, "usage_type": "name"}, {"api_name": "RPi.GPIO.LOW", "line_number": 64, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.output", "line_number": 65, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 65, "usage_type": "name"}, {"api_name": "RPi.GPIO.LOW", "line_number": 65, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.output", "line_number": 68, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 68, "usage_type": "name"}, {"api_name": "RPi.GPIO.HIGH", "line_number": 68, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 69, "usage_type": "call"}, {"api_name": "RPi.GPIO.output", "line_number": 70, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 70, "usage_type": "name"}, {"api_name": "RPi.GPIO.LOW", "line_number": 70, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.output", "line_number": 73, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 73, "usage_type": "name"}, {"api_name": "RPi.GPIO.HIGH", "line_number": 73, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 74, "usage_type": "call"}, {"api_name": "RPi.GPIO.output", "line_number": 75, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 75, "usage_type": "name"}, {"api_name": "RPi.GPIO.LOW", "line_number": 75, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.cleanup", "line_number": 80, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 80, "usage_type": "name"}]}
+{"seq_id": "26448470763", "text": "import hashlib\nimport logging\nimport math\nimport os\nimport re\nfrom pathlib import Path\n\nfrom utils.constants import HASH_BUFFER_LEN, TEMP_FOLDER_PATH # MESSAGE_MAX_LEN,\nfrom utils.types import CompressionMethod, DirData, ItemSearchResult, Message, TransferProgress, TransferStatus\n\n\ndef generate_transfer_progress() -> dict[Path, TransferProgress]:\n \"\"\"Generate transfer progress data in absence of dump.\n\n Parses the user's tmp folder to find offsets for incommplete files.\n\n Returns\n -------\n dict[Path, TransferProgress]\n Returns transfer progress dictionary as generated\n \"\"\"\n transfer_progress: dict[Path, TransferProgress] = {}\n for root, _, files in os.walk(str(TEMP_FOLDER_PATH)):\n for file in files:\n path = Path(root).joinpath(file)\n transfer_progress[path] = {\n \"progress\": path.stat().st_size,\n \"status\": TransferStatus.PAUSED,\n }\n return transfer_progress\n\n\ndef path_to_dict(path: Path, share_folder_path: str) -> DirData:\n \"\"\"Converts a given folder path to a dictionary representation of the entire directory structure\n\n Recursively constructs the output dictionary.\n Works relative to the user's share folder.\n\n Parameters\n ----------\n path : Path\n Path to an item to be added to dictionary\n share_folder_path : str\n string path to user's share directory which contains the item at [path]\n\n Returns\n -------\n DirData\n Returns dictionary representation as defined by the DirData custom type\n \"\"\"\n d: DirData = {\n \"path\": str(path).removeprefix(share_folder_path + \"/\"),\n \"name\": path.name,\n \"hash\": None,\n \"compression\": CompressionMethod.NONE.value,\n \"type\": \"\",\n \"size\": None,\n \"children\": [],\n }\n if path.is_dir():\n d[\"type\"] = \"directory\"\n d[\"children\"] = [path_to_dict(item, share_folder_path) for item in path.iterdir()]\n else:\n d[\"type\"] = \"file\"\n d[\"size\"] = Path(path).stat().st_size\n\n return d\n\n\ndef get_files_in_dir(dir: list[DirData] | None, files: list[DirData]):\n \"\"\"Obtain only the file items in a given directory dictionary\n\n Recursively parses dictionary to obtain file items.\n Output is given in the [files] parameter.\n\n Parameters\n ----------\n dir : list[DirData]\n Directory structure starting from immediate children of the desired folder.\n files : list[DirData]\n Empty list which holds the output of this function.\n \"\"\"\n if dir is None:\n return\n for item in dir:\n if item[\"type\"] == \"file\":\n files.append(item)\n else:\n get_files_in_dir(item[\"children\"], files)\n\n\ndef item_search(dir: list[DirData] | None, items: list[ItemSearchResult], search_query: str, owner: str):\n \"\"\"Item search utility.\n\n Recurses a given file structure of a directory to find items that match a search string.\n On each item, the function performs a regex search for exact matches followed by a fuzzy search to capture potential spelling errors.\n Output is given in the [items] parameter.\n\n Parameters\n ----------\n dir : list[DirData]\n Directory structure starting from immediate children of the desired folder.\n items : list[ItemSearchResult]\n Empty list which holds the search results.\n search_query : str\n User provided keyword used for the search process.\n owner : str\n Username of the owner of given [dir]\n\n \"\"\"\n from fuzzysearch import find_near_matches\n\n if dir is None:\n return\n for item in dir:\n if re.search(search_query, item[\"name\"].lower()) is not None or find_near_matches(\n search_query, item[\"name\"].lower(), max_l_dist=1\n ):\n items.append(\n {\n \"owner\": owner,\n \"data\": item,\n }\n )\n if item[\"type\"] == \"directory\":\n item_search(item[\"children\"], items, search_query, owner)\n\n\ndef display_share_dict(share: list[DirData] | None, indents: int = 0):\n \"\"\"Utility to print a dir structure to stdout\"\"\"\n if share is None:\n return\n for item in share:\n if item[\"type\"] == \"file\":\n print(\" \" * indents + item[\"name\"])\n else:\n print(\" \" * indents + item[\"name\"] + \"/\")\n display_share_dict(item[\"children\"], indents + 1)\n\n\ndef update_file_hash(share: list[DirData], file_path: str, new_hash: str):\n \"\"\"Utility to set a new hash value for a specified item in a dir structure.\n\n Recurses a given folder structure and updates the hash attribute when the specified item is found.\n\n Parameters\n ----------\n share : list[DirData]\n Dir structure that comntains item to update\n file_path : str\n Path attribute of item to update\n new_hash: str\n New hash value to be set\n \"\"\"\n for item in share:\n if item[\"type\"] == \"file\" and item[\"path\"] == file_path:\n item[\"hash\"] = new_hash\n return\n elif item[\"children\"]:\n update_file_hash(item[\"children\"], file_path, new_hash)\n return\n\n\ndef find_file(share: list[DirData] | None, path: str) -> DirData | None:\n \"\"\"Utility to find a file item given the file path.\"\"\"\n if share is None:\n return None\n for item in share:\n if item[\"path\"] == path:\n return item\n else:\n s = find_file(item[\"children\"], path)\n if s is not None:\n return s\n return None\n\n\ndef get_file_hash(filepath: str) -> str:\n \"\"\"Calculate hash for a given file on disk.\n\n Reads the given file in chunks and calculates a rolling hash for the same.\n\n Parameters\n ----------\n filepath : str\n Path to a file for which to calculate hash.\n \"\"\"\n hash = hashlib.sha1()\n with open(filepath, \"rb\") as file:\n while True:\n file_bytes = file.read(HASH_BUFFER_LEN)\n hash.update(file_bytes)\n if len(file_bytes) < HASH_BUFFER_LEN:\n break\n return hash.hexdigest()\n\n\ndef get_unique_filename(path: Path) -> Path:\n \"\"\"Utility to generate a unique filename if a desired name already exists on disk.\n\n Adds an incremental numeric suffix to the filename if the original or a previous iteration of the name exists in the user's downloads folder.\n Prevents accidental overwriting that may occur if different files happen to have the same name.\n\n Parameters\n ----------\n path : Path\n Desired path name for the file\n\n Returns\n -------\n Path\n Unique-ified path name for the file\n \"\"\"\n parent, filename, extension = path.parent, path.stem, path.suffix\n counter = 1\n logging.debug(f\"parent: {parent}\")\n logging.debug(f\"making unique file for {path}\")\n while path.exists():\n path = parent / Path(filename + \"_\" + str(counter) + extension)\n counter += 1\n\n logging.debug(f\"unique file name is {path}\")\n return path\n\n\ndef get_pending_downloads(transfer_progress: dict[Path, TransferProgress]) -> str:\n \"\"\"Utility to get a displayable string populated with incomplete downloads\"\"\"\n return \"\\n\".join(\n [\n f\"{str(file).removeprefix(str(TEMP_FOLDER_PATH) + '/')}: {progress['status'].name}\"\n for (file, progress) in transfer_progress.items()\n if progress[\"status\"] in [TransferStatus.DOWNLOADING, TransferStatus.PAUSED, TransferStatus.NEVER_STARTED]\n ]\n )\n\n\ndef get_directory_size(directory: DirData, size: int, count: int) -> tuple[int, int]:\n \"\"\"Calculate directory size and contained files count for a given directory.\n\n Recurses a given directory to calculate the total size of the folder as well as the number of files present in it or its sub folders.\n\n Parameters\n ----------\n directory : DirData\n Directory structure for which to calculate the statistics\n size : int\n Parent level size value, helper param for recursive call\n count : int\n Parent level count value, helper param for recursive call\n\n Returns\n -------\n tuple[int, int]\n Returns a pair of calculated size, count\n \"\"\"\n count = 0\n size = 0\n if directory[\"children\"] is None:\n count += 1\n size += directory[\"size\"]\n else:\n for child in directory[\"children\"]:\n if child[\"type\"] == \"file\":\n count += 1\n size += child[\"size\"]\n else:\n child_size, child_count = get_directory_size(child, size, count)\n size += child_size\n count += child_count\n return size, count\n\n\ndef import_file_to_share(file_path: Path, share_folder_path: Path) -> Path | None:\n \"\"\"Utility to generate symlink to a given file in the user's share folder path.\n\n Parameters\n ----------\n file_path : Path\n Path to a file for which symlink will be generated.\n share_folder_path : Path\n Path to user's share folder where the symlink will be saved.\n \"\"\"\n if file_path.exists():\n imported_file = share_folder_path / file_path.name\n imported_file.symlink_to(file_path, target_is_directory=file_path.is_dir())\n return imported_file\n else:\n logging.error(f\"Attempted to import file {str(file_path)} that does not exist\")\n return None\n\n\ndef construct_message_html(message: Message, is_self: bool) -> str:\n \"\"\"Utility to construct markup for a given message object.\n\n Parameters\n ----------\n message : Message\n A message object to be rendered.\n is_self : bool\n Boolean representing whether the sender of the given message is the current user.\n This is used for rendering a \"You\" in the markup instead of a sender username.\n\n Returns\n -------\n str\n Generated message html as a string\n \"\"\"\n return f\"\"\"
\n{\"You\" if is_self else message[\"sender\"]}: \n{message[\"content\"]}\n
\n\"\"\"\n\n\ndef convert_size(size_bytes: int) -> str:\n \"\"\"Utility to convert a size (bytes) value to a human readable string.\n\n Generates a size string suffixed with a unit like B, KB, MB and so on.\n\n Parameters\n ----------\n size_bytes : int\n Size to be converted as number of bytes\n\n Returns\n -------\n str\n Human readable size string\n \"\"\"\n if size_bytes == 0:\n return \"0B\"\n size_name = [\"B\", \"KB\", \"MB\", \"GB\", \"TB\", \"PB\", \"EB\", \"ZB\", \"YB\"]\n i = int(math.floor(math.log(size_bytes, 1024)))\n p = math.pow(1024, i)\n s = round(size_bytes / p, 2)\n return f\"{s} {size_name[i]}\"\n", "repo_name": "hs2361/Drizzle", "sub_path": "src/utils/helpers.py", "file_name": "helpers.py", "file_ext": "py", "file_size_in_byte": 10844, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 8, "dataset": "github-code", "pt": "27", "api": [{"api_name": "pathlib.Path", "line_number": 22, "usage_type": "name"}, {"api_name": "utils.types.TransferProgress", "line_number": 22, "usage_type": "name"}, {"api_name": "os.walk", "line_number": 23, "usage_type": "call"}, {"api_name": "utils.constants.TEMP_FOLDER_PATH", "line_number": 23, "usage_type": "argument"}, {"api_name": "pathlib.Path", "line_number": 25, "usage_type": "call"}, {"api_name": "utils.types.TransferStatus.PAUSED", "line_number": 28, "usage_type": "attribute"}, {"api_name": "utils.types.TransferStatus", "line_number": 28, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 12, "usage_type": "name"}, {"api_name": "utils.types.TransferProgress", "line_number": 12, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 33, "usage_type": "name"}, {"api_name": "utils.types.DirData", "line_number": 51, "usage_type": "name"}, {"api_name": "utils.types.CompressionMethod.NONE", "line_number": 55, "usage_type": "attribute"}, {"api_name": "utils.types.CompressionMethod", "line_number": 55, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 65, "usage_type": "call"}, {"api_name": "utils.types.DirData", "line_number": 33, "usage_type": "name"}, {"api_name": "utils.types.DirData", "line_number": 70, "usage_type": "name"}, {"api_name": "utils.types.DirData", "line_number": 92, "usage_type": "name"}, {"api_name": "utils.types.ItemSearchResult", "line_number": 92, "usage_type": "name"}, {"api_name": "re.search", "line_number": 116, "usage_type": "call"}, {"api_name": "fuzzysearch.find_near_matches", "line_number": 116, "usage_type": "call"}, {"api_name": "utils.types.DirData", "line_number": 129, "usage_type": "name"}, {"api_name": "utils.types.DirData", "line_number": 141, "usage_type": "name"}, {"api_name": "utils.types.DirData", "line_number": 164, "usage_type": "name"}, {"api_name": "hashlib.sha1", "line_number": 188, "usage_type": "call"}, {"api_name": "utils.constants.HASH_BUFFER_LEN", "line_number": 191, "usage_type": "argument"}, {"api_name": "utils.constants.HASH_BUFFER_LEN", "line_number": 193, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 198, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 216, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 217, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 219, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 222, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 226, "usage_type": "name"}, {"api_name": "utils.types.TransferProgress", "line_number": 226, "usage_type": "name"}, {"api_name": "utils.constants.TEMP_FOLDER_PATH", "line_number": 230, "usage_type": "argument"}, {"api_name": "utils.types.TransferStatus.DOWNLOADING", "line_number": 232, "usage_type": "attribute"}, {"api_name": "utils.types.TransferStatus", "line_number": 232, "usage_type": "name"}, {"api_name": "utils.types.TransferStatus.PAUSED", "line_number": 232, "usage_type": "attribute"}, {"api_name": "utils.types.TransferStatus.NEVER_STARTED", "line_number": 232, "usage_type": "attribute"}, {"api_name": "utils.types.DirData", "line_number": 237, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 273, "usage_type": "name"}, {"api_name": "logging.error", "line_number": 288, "usage_type": "call"}, {"api_name": "utils.types.Message", "line_number": 292, "usage_type": "name"}, {"api_name": "math.floor", "line_number": 333, "usage_type": "call"}, {"api_name": "math.log", "line_number": 333, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 334, "usage_type": "call"}]}
+{"seq_id": "515732697", "text": "\"\"\"empty message\n\nRevision ID: 23bde1fa9e41\nRevises: \nCreate Date: 2019-02-03 19:23:48.182688\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = '23bde1fa9e41'\ndown_revision = None\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n op.create_table('blacklist_token',\n sa.Column('id', sa.Integer(), nullable=False),\n sa.Column('token', sa.String(length=255), nullable=False),\n sa.Column('blacklisted_on', sa.DateTime(), nullable=False),\n sa.PrimaryKeyConstraint('id'),\n sa.UniqueConstraint('token')\n )\n op.create_table('item',\n sa.Column('id', sa.Integer(), nullable=False),\n sa.Column('name', sa.String(length=255), nullable=False),\n sa.Column('brand', sa.String(length=255), nullable=True),\n sa.Column('category', sa.String(length=255), nullable=True),\n sa.Column('productCode', sa.String(length=255), nullable=True),\n sa.PrimaryKeyConstraint('id'),\n sa.UniqueConstraint('name')\n )\n op.create_table('users',\n sa.Column('id', sa.Integer(), nullable=False),\n sa.Column('email', sa.String(length=255), nullable=False),\n sa.Column('password', sa.String(length=255), nullable=False),\n sa.Column('registered_on', sa.DateTime(), nullable=False),\n sa.PrimaryKeyConstraint('id'),\n sa.UniqueConstraint('email')\n )\n op.create_table('user_action',\n sa.Column('id', sa.Integer(), nullable=False),\n sa.Column('username', sa.String(length=255), nullable=False),\n sa.Column('item_name', sa.String(length=255), nullable=True),\n sa.Column('variant_name', sa.String(length=255), nullable=True),\n sa.Column('action', sa.String(length=255), nullable=False),\n sa.Column('timestamp', sa.DateTime(), nullable=False),\n sa.ForeignKeyConstraint(['username'], ['users.email'], ),\n sa.PrimaryKeyConstraint('id')\n )\n op.create_table('variant',\n sa.Column('id', sa.Integer(), nullable=False),\n sa.Column('name', sa.String(length=255), nullable=True),\n sa.Column('item_id', sa.Integer(), nullable=True),\n sa.Column('sellingPrice', sa.String(length=255), nullable=True),\n sa.Column('costPrice', sa.String(length=255), nullable=True),\n sa.Column('quantity', sa.Integer(), nullable=True),\n sa.ForeignKeyConstraint(['item_id'], ['item.id'], ),\n sa.PrimaryKeyConstraint('id')\n )\n # ### end Alembic commands ###\n\n\ndef downgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n op.drop_table('variant')\n op.drop_table('user_action')\n op.drop_table('users')\n op.drop_table('item')\n op.drop_table('blacklist_token')\n # ### end Alembic commands ###\n", "repo_name": "pmishra06/Inventory", "sub_path": "migrations/versions/23bde1fa9e41_.py", "file_name": "23bde1fa9e41_.py", "file_ext": "py", "file_size_in_byte": 2709, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "alembic.op.create_table", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.UniqueConstraint", "line_number": 26, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 28, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 28, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.UniqueConstraint", "line_number": 35, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 37, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 37, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 40, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 40, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.UniqueConstraint", "line_number": 43, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 45, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 45, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 50, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 50, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 51, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 51, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 52, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 53, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 55, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 55, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 56, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 56, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 58, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 58, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 59, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 59, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 60, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 60, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 61, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 61, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 62, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 63, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 70, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 70, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 71, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 71, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 72, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 72, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 73, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 73, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 74, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 74, "usage_type": "name"}]}
+{"seq_id": "41265278314", "text": "import json\nimport copy\nINPUT_DATASET_FILES = ['filtered_spatial_commonsense/posrel.json', 'filtered_spatial_commonsense/height.json', 'filtered_spatial_commonsense/size.json']\nOUTPUT_DATASET_FILES = ['filtered_spatial_commonsense/posrel_coco.json', 'filtered_spatial_commonsense/height_coco.json', 'filtered_spatial_commonsense/size_coco.json']\nLABEL_TO_COCO_LABEL_GENERATION_FOLDER = 'filtered_spatial_commonsense'\nDATASET_LABEL_TO_COCO_LABEL = 'filtered_spatial_commonsense/dataset_label_to_coco_label.json'\ndef help_generate_class_to_coco_class():\n \"\"\"\n This function will generate a json file containing every object class in all INPUT_DATASET_FILES as keys.\n The idea is that the user is to then fill this file's values in order for this program to be able to make the change from the datasets' classes to coco classes.\n So this function will for example leave \"Man\" or \"Girl\" as keys and the user would have to fill \"person\" as the appropiate coco class to them.\n \"\"\"\n print(\"As no DATASET_LABEL_TO_LABEL_CLASS was given, a file helping the process of making it will be made in \"+LABEL_TO_COCO_LABEL_GENERATION_FOLDER+\"/empty_dataset_label_to_coco_label.json\")\n print(\"Every label present in the datasets provided will be saved as a json file's keys. The user is then expecteds to fill these with their respective coco object label for the translation to be made in a later execution.\")\n print(\"If a dataset label doesn't have a coco label counterpart, the entire key should be deleted. When using the file later the ammount of data points deleted because no coco label existed for them will be reported in the statistics.\")\n print()\n all_labels = []\n for dataset_path in INPUT_DATASET_FILES:\n with open(dataset_path, \"r\") as dataset_file:\n for jsonObj in dataset_file:\n data_point = json.loads(jsonObj)\n all_labels.append(data_point['obj_a'])\n all_labels.append(data_point['obj_b'])\n all_labels = set(all_labels)\n label_dict = {}#will contain every label as keys and an empty string as value. It will be saved and the user is supposed to fill the values with the corresponding coco label\n for label in all_labels:\n label_dict[label] = \"\"\n with open(LABEL_TO_COCO_LABEL_GENERATION_FOLDER+\"/empty_dataset_label_to_coco_label.json\",\"w\") as f:\n json.dump(label_dict, f, indent=4, sort_keys=True)\n print(\"all datset labels correctly saved\")\n show_coco_labels = input(\"would you like to get the list of all coco labels?[y/n]\")\n if show_coco_labels == 'y' or show_coco_labels=='Y':\n all_coco_labels = ['person', 'bicycle', 'car', 'motorcycle', 'airplane', 'bus', 'train', 'truck', 'boat', 'traffic light', 'fire hydrant', 'stop sign', 'parking meter', 'bench', 'bird', 'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear', 'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie', 'suitcase', 'frisbee', 'skis', 'snowboard', 'sports ball', 'kite', 'baseball bat', 'baseball glove', 'skateboard', 'surfboard', 'tennis racket', 'bottle', 'wine glass', 'cup', 'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple', 'sandwich', 'orange', 'broccoli', 'carrot', 'hot dog', 'pizza', 'donut', 'cake', 'chair', 'couch', 'potted plant', 'bed', 'dining table', 'toilet', 'tv', 'laptop', 'mouse', 'remote', 'keyboard', 'cell phone', 'microwave', 'oven', 'toaster', 'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors', 'teddy bear', 'hair drier', 'toothbrush']\n all_coco_labels = sorted(all_coco_labels) #order alphabetically\n print(all_coco_labels)\ndef change_labels():\n print(\"initilizing object label translation for:\")\n for dataset_path in INPUT_DATASET_FILES:\n print(\" \"+dataset_path)\n print(\"resuts will be saved in:\")\n for dataset_path in OUTPUT_DATASET_FILES:\n print(\" \"+dataset_path)\n print(\"stats will be saved in \"+LABEL_TO_COCO_LABEL_GENERATION_FOLDER+\"/label_translation_stats.json\")\n #load label translator from any label to coco label\n with open(DATASET_LABEL_TO_COCO_LABEL, 'r') as f:\n label_to_coco_label = json.load(f)\n scs_datasets = []\n #load all datasets\n for dataset_path in INPUT_DATASET_FILES:\n current_dataset = []\n with open(dataset_path, \"r\") as dataset_file:\n for jsonObj in dataset_file:\n data_point = json.loads(jsonObj)\n current_dataset.append(data_point)\n scs_datasets.append(current_dataset)\n #time to do the actual filtering:\n #if any object's label is not in the translation dict, that data point will be removed.\n #when removing a data_point, we save in stats what label caused that removal and the data_point's label removed\n\n stats = {}\n stats['removed_labels'] = []\n for i,dataset in enumerate(scs_datasets):\n #for this dataset initialize stats and future dataset\n filtered_dataset = []\n stats[INPUT_DATASET_FILES[i]] = {}\n stats[INPUT_DATASET_FILES[i]]['n_removed_labels'] = {}\n stats[INPUT_DATASET_FILES[i]]['previous_total'] = len(dataset)\n stats[INPUT_DATASET_FILES[i]]['n_removed_labels']['total'] = 0\n for data_point in dataset:#initialize every possible data point label count to 0\n if data_point['label'] not in stats[INPUT_DATASET_FILES[i]]['n_removed_labels'].keys():\n stats[INPUT_DATASET_FILES[i]]['n_removed_labels'][data_point['label']]=0\n for data_point in dataset:\n data_point = copy.deepcopy(data_point)\n remove = False\n #check whether both object labels are in the translation dict.\n #if they are, trasnlate them individually\n if data_point['obj_a'] not in label_to_coco_label.keys():\n stats['removed_labels'].append(data_point['obj_a'])\n remove=True\n else:\n data_point['obj_a'] = label_to_coco_label[data_point['obj_a']]\n if data_point['obj_b'] not in label_to_coco_label.keys():\n stats['removed_labels'].append(data_point['obj_b'])\n remove=True\n else:\n data_point['obj_b'] = label_to_coco_label[data_point['obj_b']]\n if not remove:#if both object labels are in the translation dict\n filtered_dataset.append(data_point)\n else:# if they're not, it will be removed and stats must be updated\n stats[INPUT_DATASET_FILES[i]]['n_removed_labels'][data_point['label']] += 1\n stats[INPUT_DATASET_FILES[i]]['n_removed_labels']['total'] +=1\n stats[INPUT_DATASET_FILES[i]]['new_total'] = len(filtered_dataset)\n #save new dataset. (stats for all the datsets will be saved in the same place)\n with open(OUTPUT_DATASET_FILES[i],\"w\") as output_file:\n for data_point in filtered_dataset:\n json.dump(data_point, output_file)\n print(file=output_file)\n #save stats\n stats['removed_labels'] = list(set(stats['removed_labels']))\n with open(LABEL_TO_COCO_LABEL_GENERATION_FOLDER+\"/label_translation_stats.json\", \"w\") as f:\n json.dump(stats, f, indent=4)\ndef main():\n if DATASET_LABEL_TO_COCO_LABEL is None:\n help_generate_class_to_coco_class()\n else:\n change_labels()\nif __name__ == '__main__':\n main()", "repo_name": "EXUPLOOOOSION/text-to-layout", "sub_path": "Spatial CommonSense Test/change_labels_to_coco.py", "file_name": "change_labels_to_coco.py", "file_ext": "py", "file_size_in_byte": 7351, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "json.loads", "line_number": 21, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 29, "usage_type": "call"}, {"api_name": "json.load", "line_number": 46, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 53, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 73, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 96, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 101, "usage_type": "call"}]}
+{"seq_id": "6229093463", "text": "from typing import Optional, List, Tuple\nfrom tqdm import tqdm\nfrom pathlib import Path\nfrom concurrent.futures import ThreadPoolExecutor\nimport os\nimport numpy as np\nimport argparse\n\nDEFAULT_FILTER_TAGS = [\"monochrome\", \"greyscale\"]\n\nBATCH_SIZE = 200\n\n\ndef check_contains_tags(caption_file: Path, filter_tags):\n with open(caption_file, \"r\") as f:\n caption = f.read()\n for filter_tag in filter_tags:\n if filter_tag in caption:\n return True\n return False\n\n\ndef move_caption_and_image(\n caption_file: Path,\n image_file: Path,\n caption_output_dir: Path,\n image_output_dir: Path,\n):\n os.rename(caption_file, caption_output_dir / caption_file.name)\n os.rename(image_file, image_output_dir / image_file.name)\n\n\ndef process_captions(\n caption_files: List[Path],\n filter_tags,\n images_dir,\n caption_output_dir,\n image_output_dir,\n pbar,\n):\n for caption_file in tqdm(caption_files):\n if check_contains_tags(caption_file, filter_tags):\n # search image file. image file has same stem but unknown extension\n image_file = next(images_dir.glob(f\"{caption_file.stem}.*\"))\n move_caption_and_image(\n caption_file, image_file, caption_output_dir, image_output_dir\n )\n # print(caption_file, image_file)\n pbar.update(1)\n\n\ndef main(args):\n captions_dir = args.captions_dir\n images_dir = args.images_dir\n if images_dir is None:\n images_dir = captions_dir\n\n caption_output_dir = Path(args.caption_output_dir)\n image_output_dir = Path(args.image_output_dir)\n\n filter_tags = args.filter_tags\n\n caption_files = list(captions_dir.glob(\"*.txt\"))\n\n print(\"Total captions:\", len(caption_files))\n\n if not caption_output_dir.exists():\n caption_output_dir.mkdir()\n\n if not image_output_dir.exists():\n image_output_dir.mkdir()\n\n chunks = np.array_split(caption_files, BATCH_SIZE)\n\n with tqdm(total=len(caption_files)) as pbar:\n with ThreadPoolExecutor(max_workers=BATCH_SIZE) as executor:\n futures = []\n for chunk in chunks:\n futures.append(\n executor.submit(\n process_captions,\n chunk,\n images_dir,\n filter_tags,\n caption_output_dir,\n image_output_dir,\n pbar,\n )\n )\n\n for future in futures:\n future.result()\n\n print(\"Done\")\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser()\n parser.add_argument(\n \"--captions_dir\",\n type=str,\n required=True,\n help=\"Path to directory containing captions\",\n )\n parser.add_argument(\n \"--images_dir\",\n type=str,\n help=\"Path to directory containing images\",\n )\n parser.add_argument(\n \"--caption_output_dir\",\n type=str,\n help=\"Path to directory to save captions\",\n )\n parser.add_argument(\n \"--image_output_dir\",\n type=str,\n help=\"Path to directory to save images\",\n )\n parser.add_argument(\n \"--filter_tags\",\n type=str,\n nargs=\"+\",\n default=DEFAULT_FILTER_TAGS,\n help=\"Tags to filter\",\n )\n args = parser.parse_args()\n main(args)\n", "repo_name": "p1atdev/sd-annotators", "sub_path": "caption_filter.py", "file_name": "caption_filter.py", "file_ext": "py", "file_size_in_byte": 3404, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "24", "api": [{"api_name": "pathlib.Path", "line_number": 14, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 24, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 25, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 26, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 27, "usage_type": "name"}, {"api_name": "os.rename", "line_number": 29, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 30, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 34, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 34, "usage_type": "name"}, {"api_name": "tqdm.tqdm", "line_number": 41, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 58, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.array_split", "line_number": 73, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 75, "usage_type": "call"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 76, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 98, "usage_type": "call"}]}
+{"seq_id": "71695072384", "text": "\nimport os\nfrom os import listdir\nfrom skimage import io\nfrom skimage.color import rgb2gray\nfrom skimage.data import stereo_motorcycle\nfrom skimage.registration import phase_cross_correlation\nfrom skimage.transform import AffineTransform, warp\nimport numpy as np\nimport pandas as pd\nimport argparse\nfrom pathlib import Path\nfrom skimage.util import img_as_ubyte\nfrom skimage.util import img_as_uint\nfrom skimage import exposure\n\n \ndef parseArgs():\n parser = argparse.ArgumentParser(description=\"Image registration parameters\")\n parser.add_argument('-path_to_round1', default= \"/Users/isabelmo/Downloads/registration/testdata/round1\")\n parser.add_argument('-path_to_round2', default= \"/Users/isabelmo/Downloads/registration/testdata/round2\")\n parser.add_argument('-outputs', default= \"/Users/isabelmo/Downloads/registration/testdata/outputs\")\n args = parser.parse_args()\n return args\n\ndef cal_phase_correlate(fixed, moving):\n # calculate phase correlations\n shift, error, phasediff = phase_cross_correlation(fixed, moving)\n shift = [shift[1], shift[0]]\n shift = np.array(shift)\n shift = shift*-1\n print(shift, error, phasediff)\n return shift\n\ndef transform_phase_correlate(moving, shift):\n # calculate correction transform\n transform = AffineTransform(translation=shift)\n # apply it to round 2 image\n return warp(moving, transform)\n\ndef main():\n # load all image names into lists\n args = parseArgs()\n path_to_round1 = args.path_to_round1\n path_to_round2 = args.path_to_round2\n ls_imgs1_names = os.listdir(path_to_round1)\n ls_imgs2_names = os.listdir(path_to_round2)\n path_to_outputs = args.outputs \n Path(path_to_outputs + \"/regs\").mkdir(parents=True, exist_ok=True)\n\n image_r1_names = []\n\n # Create tables for each experiment\n for image_r1 in ls_imgs1_names:\n # For each DAPI image in round 1\n if '_DAPI_ORG' in image_r1:\n # add round 1 path list\n image_r1_names.append(image_r1)\n # store sample id\n spot_str = image_r1[-7:-4]\n\n image_r2_names = []\n # Find corresponding channel images from round 2\n for image_r2 in ls_imgs2_names:\n if f'roi{spot_str}' in image_r2:\n image_r2_names.append(image_r2)\n\n # for dapi from 2nd round calculate transformation\n for image_r2 in image_r2_names:\n if f'_DAPI_ORG_roi{spot_str}' in image_r2:\n # open round 1 & round 2 DAPIs\n fixed = io.imread(f\"{path_to_round1}/{image_r1}\")\n moving = io.imread(f\"{path_to_round2}/{image_r2}\")\n \n # calculate transformation\n shift = cal_phase_correlate(fixed, moving)\n\n # transform dapi\n moving = transform_phase_correlate(moving.astype(np.uint8), shift)\n #moving = moving.astype(np.uint8)\n #fixed = fixed.astype(np.uint8)\n \n\n # create tumbnails\n r1image = np.expand_dims(fixed, axis=-1)\n r2image = np.expand_dims(moving, axis=-1)\n thumbnail = np.concatenate([r1image, r2image, r2image], axis=-1)\n # print(thumbnail.shape)\n # exit()\n io.imsave(f'{path_to_outputs}/DAPIaftertransform_roi{spot_str}.jpg', thumbnail)\n break\n\n transformed_images = []\n # transform all round 2 images\n for image_r2 in image_r2_names:\n # open image\n moving = io.imread(f\"{path_to_round2}/{image_r2}\")\n #moving = moving.astype(np.uint8)\n #print(moving.dtype)\n\n # before\n\n # create tumbnails\n r1image = np.expand_dims(fixed, axis=-1)\n r2image = np.expand_dims(moving, axis=-1)\n r3image = np.zeros_like(r2image)\n thumbnail = np.concatenate([r1image, r2image, r3image], axis=-1)\n # print(thumbnail.shape)\n # exit()\n io.imsave(f'{path_to_outputs}/before_{image_r2}.jpg', thumbnail)\n\n # transform image\n #img_as_ubyte(moving)\n moving = transform_phase_correlate(moving, shift) #astype(np.uint8)\n moving = img_as_uint(moving)\n #image = exposure.rescale_intensity(moving, in_range='uint8')\n transformed_images.append(moving)\n #moving = moving.astype(np.uint8)\n print(moving.dtype)\n\n # after\n\n # create tumbnails\n r1image = np.expand_dims(fixed, axis=-1)\n r2image = np.expand_dims(moving, axis=-1)\n thumbnail = np.concatenate([r1image, r2image, r3image], axis=-1)\n # print(thumbnail.shape)\n # exit()\n io.imsave(f'{path_to_outputs}/after_{image_r2}.jpg', thumbnail)\n \n \n io.imsave(f'{path_to_outputs}/regs/{image_r2}', moving) #astype(np.uint16))\n #io.imsave(f'{path_to_outputs}/regs/{image_r2}', moving.astype(np.uint8))\n \n print(fixed.dtype)\n \n\nif __name__ == \"__main__\":\n main()\n ", "repo_name": "isabellamermaid/image_registration", "sub_path": "imagereg.py", "file_name": "imagereg.py", "file_ext": "py", "file_size_in_byte": 5409, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 19, "usage_type": "call"}, {"api_name": "skimage.registration.phase_cross_correlation", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 30, "usage_type": "call"}, {"api_name": "skimage.transform.AffineTransform", "line_number": 37, "usage_type": "call"}, {"api_name": "skimage.transform.warp", "line_number": 39, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 46, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 47, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 49, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 72, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 72, "usage_type": "name"}, {"api_name": "skimage.io.imread", "line_number": 73, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 73, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 79, "usage_type": "attribute"}, {"api_name": "numpy.expand_dims", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 87, "usage_type": "call"}, {"api_name": "skimage.io.imsave", "line_number": 90, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 90, "usage_type": "name"}, {"api_name": "skimage.io.imread", "line_number": 97, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 97, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 107, "usage_type": "call"}, {"api_name": "skimage.io.imsave", "line_number": 110, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 110, "usage_type": "name"}, {"api_name": "skimage.util.img_as_uint", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 126, "usage_type": "call"}, {"api_name": "skimage.io.imsave", "line_number": 129, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 129, "usage_type": "name"}, {"api_name": "skimage.io.imsave", "line_number": 132, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 132, "usage_type": "name"}]}
+{"seq_id": "43593169248", "text": "import statistics\n\nfrom datetime import date\nfrom datetime import datetime\nfrom datetime import timedelta\n\nfrom lib.jira import Jira\n\nclass ServiceDesk():\n\n def __init__(self, cache, config):\n self.cache = cache\n self.config = config\n\n def buildDatesPast12Months(self):\n\n # Build date ranges filter\n dates = []\n\n # Current month\n last = date.today()\n first = last.replace(day=1)\n\n dates.append({\n \"first\": f\"{first.year}/{first.month}/{first.day}\",\n \"last\": f\"{last.year}/{last.month}/{last.day}\",\n \"label\": f\"{last.year}/{last.month}\",\n })\n\n # First day of this month\n dt = date.today().replace(day=1)\n\n # Loop past last 12 months\n for i in range(12):\n\n # Last of previous month\n last = dt - timedelta(days=1)\n # First of previous month\n dt = first = last.replace(day=1)\n\n dates.append({\n \"first\": f\"{first.year}/{first.month}/{first.day}\",\n \"last\": f\"{last.year}/{last.month}/{last.day}\",\n \"label\": f\"{last.year}/{last.month}\",\n })\n\n dates.reverse()\n\n return dates\n\n def buildDatesYears(self):\n dates = []\n\n for d in range(self.config.jira['start_year'], date.today().year + 1):\n dates.append({\n \"first\": f\"{d}/01/01\",\n \"last\": f\"{d}/12/31\",\n \"label\": f\"{d}\",\n })\n\n return dates\n\n def buildDatesYearlyMonths(self):\n\n # Dates dict\n dates = {}\n\n # Today's date for comparison\n dt = date.today()\n\n # Date for iterate\n di = date(self.config.jira['start_year'], 1, 1)\n\n # Up until current month\n while di < dt:\n # Get first date\n first = di.replace(day=1)\n\n # Update di to next month\n di = (di + timedelta(days=35)).replace(day=1)\n\n # Get last date\n last = di - timedelta(days=1)\n\n # Add new year list\n if first.year not in dates:\n dates[first.year] = []\n\n # Add range to dates\n dates[first.year].append({\n \"first\": f\"{first.year}/{first.month}/{first.day}\",\n \"last\": f\"{last.year}/{last.month}/{last.day}\",\n \"label\": f\"{last.year}/{last.month}\",\n })\n\n return dates\n\n def buildDatesYearlySprints(self):\n\n # Dates dict\n dates = {}\n\n # Today's date for comparison\n dt = date.today()\n\n # Date for iterate\n di = date(2022, 1, 6)\n\n # Up until current month\n while di < dt:\n # Get first date\n first = di\n\n # Get last date\n last = first + timedelta(days=13)\n\n # Update di\n di = last + timedelta(days=1)\n\n # Add new year list\n if first.year not in dates:\n dates[first.year] = []\n i = 1\n\n # Add range to dates\n dates[first.year].append({\n \"first\": f\"{first.year}/{first.month}/{first.day}\",\n \"last\": f\"{last.year}/{last.month}/{last.day}\",\n \"label\": f\"{first.year}-{first.month}-{first.day}\",\n })\n i += 1\n\n return dates\n\n def calculateStats(self, values):\n # Size\n stats = {\n 'size': len(values)\n }\n\n # Maximum, Mean\n if len(values) > 0:\n stats['maximum'] = round(max(values), 2)\n stats['mean'] = round(statistics.mean(values), 2)\n else:\n stats['maximum'] = 0\n stats['mean'] = 0\n\n # Stdev\n if len(values) > 1:\n stats['stdev'] = round(statistics.stdev(values), 2)\n else:\n stats['stdev'] = 0\n\n # Minimum\n while (0.0 in values):\n values.remove(0.0)\n\n if len(values) > 0:\n stats['minimum'] = round(min(values), 2)\n else:\n stats['minimum'] = 0\n\n return stats\n\n def closedIssuesYearly(self, project, extra, cache):\n\n # Build projects filter\n project = self.formatProject(project)\n\n # Build date ranges filter\n dates = self.buildDatesYears()\n\n return self.queryClosedIssues(project, dates, extra, cache)\n\n def closedIssuesYearlyMonthly(self, project, extra, cache):\n\n # Build projects filter\n project = self.formatProject(project)\n\n # Build date ranges filter\n dates = self.buildDatesYearlyMonths()\n\n results = {}\n\n for y, d in dates.items():\n # Query Jira\n results[y] = self.queryClosedIssues(project, d, extra, cache)\n\n return results\n\n def formatProject(self, project):\n\n # Build projects filter\n if type(project) is str:\n return f\"'{project}'\"\n\n elif type(project) is list:\n return \",\".join(list(map(lambda x: f\"'{x}'\", project)))\n\n def queryClosedIssues(self, project, dates, extra, cache=False):\n\n results = {}\n\n fields = [\n \"summary\",\n \"status\",\n ]\n\n jira = Jira(self.config)\n\n for dt in dates:\n\n first = dt['first']\n last = dt['last']\n\n # Prepare jql query\n jql = f\"project in ({project}) AND status in (Closed, Resolved, Done) AND status changed to (Closed, Resolved, Done) DURING ('{first} 00:00','{last} 23:59') {extra}\"\n\n search = jira.apiSearch(jql, fields)\n\n if search:\n result = search.json()\n\n results[dt['label']] = result['total']\n\n if cache:\n self.cache.write(\n run=cache['run'],\n label=cache['label'],\n index=dt['label'],\n value=result['total']\n )\n\n print(f\"results: {results}\")\n\n return results\n\n def queryIssueTimingsCustom(self, project, dates, extra, cache):\n\n # Initiate results dict\n results = {}\n for k in self.config.jira['custom_fields'].keys():\n results[k] = {}\n\n # Initiate fields list\n fields = [\n \"summary\",\n \"status\",\n \"created\",\n ] + list(map(lambda x: x['field'], self.config.jira['custom_fields'].values()))\n\n jira = Jira(self.config)\n\n # Loop dates\n for item in dates.values():\n for dt in item:\n\n first = dt['first']\n last = dt['last']\n\n # Prepare jql query\n jql = f\"project in ({project}) AND status in (Closed, Resolved, Done) AND status changed to (Closed, Resolved, Done) DURING ('{first} 00:00','{last} 23:59') {extra}\"\n\n search = jira.apiSearch(jql, fields)\n\n if search:\n result = search.json()\n\n # Init results lists\n for k in self.config.jira['custom_fields'].keys():\n results[k][dt['label']] = []\n\n for i in result['issues']:\n # Loop custom fields\n # Time to: resolution, first response, close after resolution\n for k in self.config.jira['custom_fields'].keys():\n field = self.config.jira['custom_fields'][k]['field']\n if i ['fields'][field] is not None:\n if len(i['fields'][field]['completedCycles']) > 0:\n field_value = 0\n for j in i['fields'][field]['completedCycles']:\n field_value += j['elapsedTime']['millis']\n\n results[k][dt['label']].append(round(field_value / 1000 / 3600, 2))\n\n for k, v in results.items():\n for dt in v.keys():\n results[k][dt] = self.calculateStats(results[k][dt])\n\n self.cache.write(\n run=cache['run'],\n label=k,\n index=dt,\n value=results[k][dt]\n )\n\n print(f\"results: {k}: {v}\")\n\n return results\n\n def queryIssueTimingsResolution(self, project, dates, extra, cache):\n\n # Initiate results dict\n results = {}\n\n # Initiate fields list\n fields = [\n 'summary',\n 'status',\n 'created',\n 'resolutiondate',\n ]\n\n jira = Jira(self.config)\n\n # Loop dates\n for item in dates.values():\n for dt in item:\n\n first = dt['first']\n last = dt['last']\n\n # Prepare jql query\n jql = f\"project in ({project}) AND status in (Closed, Resolved, Done) AND status changed to (Closed, Resolved, Done) DURING ('{first} 00:00','{last} 23:59') {extra}\"\n\n search = jira.apiSearch(jql, fields)\n\n if search:\n result = search.json()\n\n # Init results list\n results[dt['label']] = []\n\n for i in result['issues']:\n if i['fields']['resolutiondate'] is not None:\n ts = datetime.strptime(i['fields']['resolutiondate'], '%Y-%m-%dT%H:%M:%S.000%z') - \\\n datetime.strptime(i['fields']['created'], '%Y-%m-%dT%H:%M:%S.000%z')\n\n results[dt['label']].append(round(ts.days + ts.seconds / 86400, 2))\n\n for k, v in results.items():\n results[k] = self.calculateStats(v)\n\n self.cache.write(\n run=cache['run'],\n label=cache['label'],\n index=k,\n value=results[k]\n )\n\n print(f\"results: {results}\")\n\n return results\n\n def runCoreTimingsServiceDesk(self, queries, run):\n\n # Init data dict\n data = {\n 'data': [],\n 'labels': [],\n }\n\n # Build date ranges filter\n dates = self.buildDatesYearlySprints()\n\n # Loop queries\n for q in queries:\n\n project = self.formatProject(project=q['project'])\n\n result = self.queryIssueTimingsCustom(\n project=project,\n dates=dates,\n extra=q['extra'],\n cache={\n 'run': run\n }\n )\n\n # Loop result by custom_field key\n for k in result.keys():\n\n # Set labels\n data['labels'] = list(result[k].keys())\n\n # Append data\n data['data'].append({\n 'data': {\n 'maximum': list(map(lambda x: x['maximum'], result[k].values())),\n 'mean': list(map(lambda x: x['mean'], result[k].values())),\n 'minimum': list(map(lambda x: x['minimum'], result[k].values())),\n 'size': list(map(lambda x: x['size'], result[k].values())),\n 'stdev': list(map(lambda x: x['stdev'], result[k].values())),\n },\n 'label': self.config.jira['custom_fields'][k]['description'],\n })\n\n return data\n\n def runCoreTimingsTeams(self, queries, run):\n\n # Init data dict\n data = {\n 'data': [],\n 'labels': []\n }\n\n # Build date ranges filter\n dates = self.buildDatesYearlySprints()\n\n # Loop queries\n for q in queries:\n\n project = self.formatProject(project=q['project'])\n\n result = self.queryIssueTimingsResolution(\n project=project,\n dates=dates,\n extra=q['extra'],\n cache={\n 'label': q['label'],\n 'run': run\n }\n )\n\n # Set labels\n data['labels'] = list(result.keys())\n\n # Append data\n data['data'].append({\n 'data': {\n 'maximum': list(map(lambda x: x['maximum'], result.values())),\n 'mean': list(map(lambda x: x['mean'], result.values())),\n 'minimum': list(map(lambda x: x['minimum'], result.values())),\n 'size': list(map(lambda x: x['size'], result.values())),\n 'stdev': list(map(lambda x: x['stdev'], result.values())),\n },\n 'label': q['label']\n })\n\n return data\n\n def runMonthly(self, queries):\n\n data = {\n 'data': [],\n 'labels': [],\n }\n\n # Build date ranges filter\n dates = self.buildDatesPast12Months()\n\n for q in queries:\n\n # Build projects filter\n project = self.formatProject(project=q['project'])\n\n # Query Jira\n result = self.queryClosedIssues(project, dates, q['extra'])\n\n # Add results to data.json\n data['labels'] = list(result.keys())\n data['data'].append({\n 'data': list(result.values()),\n 'label': f\"{q['label']} ({sum(result.values())})\"\n })\n\n return data\n\n def runYearlyComponents(self, queries, run):\n\n data = {}\n\n results = {}\n\n for q in queries:\n result = self.closedIssuesYearly(\n project=self.config.queries['sd']['project'],\n extra=f\"{self.config.queries['sd']['extra']} {q['extra']}\",\n cache={\n 'label': q['label'],\n 'run': run\n }\n )\n\n # Build results\n for y, i in result.items():\n # Skip zero entries\n if i != 0:\n if y in results:\n results[y][q['label']] = i\n else:\n results[y] = {}\n results[y][q['label']] = i\n\n # Add results to data.json\n for y, i in results.items():\n # Skip if all results sum to 0\n if sum(i.values()) != 0:\n # Initiate object\n if y not in data:\n data[y] = {\n 'data': [],\n 'labels': [],\n }\n\n # Calculate percentages\n p = {}\n for k, j in i.items():\n p[k] = j / sum(i.values()) * 100\n\n data[y]['labels'] = list(map(lambda x: f\"{x} ({p[x] :.1f}%)\", p.keys()))\n data[y]['data'].append({\n 'data': list(i.values()),\n 'label': y\n })\n\n return data\n\n def runYearlyMonthlyComponents(self, queries, run):\n\n data = {}\n\n results = {}\n\n for q in queries:\n result = self.closedIssuesYearlyMonthly(\n project=q['project'],\n extra=q['extra'],\n cache={\n 'label': q['label'],\n 'run': run\n }\n )\n\n # Build results\n for y, i in result.items():\n if y in results:\n results[y][q['label']] = i\n else:\n results[y] = {}\n results[y][q['label']] = i\n\n # Add results to data.json\n for y, i in results.items():\n\n # Check if total year sum > 0\n year_sum = 0\n\n for a in i.values():\n year_sum += sum(a.values())\n\n if year_sum == 0:\n continue\n\n # Initiate object\n if y not in data:\n data[y] = {\n 'data': [],\n 'labels': [],\n }\n\n # Monthly sums dict\n sums = {}\n\n for k, j in i.items():\n # Skip if all results sum to 0\n if sum(j.values()) != 0:\n # Calculate monthly sums\n for m in j.keys():\n if m in sums:\n sums[m] += j[m]\n else:\n sums[m] = j[m]\n\n data[y]['data'].append({\n 'data': list(j.values()),\n 'label': f\"{k} ({sum(j.values())})\"\n })\n\n # Apply labels\n data[y]['labels'] = list(map(lambda x: f\"{datetime.strptime(x, '%Y/%m').strftime('%b')} ({sums[x]})\", sums.keys()))\n\n return data\n\n def runYearlyStatistics(self, queries, run):\n\n stats = {}\n for q in queries:\n # Query Jira\n result = self.closedIssuesYearly(\n project=q['project'],\n extra=q['extra'],\n cache={\n 'label': q['label'],\n 'run': run\n }\n )\n\n # Build stats from result\n for k, v in result.items():\n if k in stats:\n stats[k][q['label']] = v\n else:\n stats[k] = {'Year': k}\n stats[k][q['label']] = v\n\n return stats\n", "repo_name": "accesspc/jira-reports", "sub_path": "app/app/servicedesk.py", "file_name": "servicedesk.py", "file_ext": "py", "file_size_in_byte": 17468, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "datetime.date.today", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 21, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 31, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 37, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 54, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 54, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 69, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 69, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 72, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 80, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 83, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 104, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 104, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 107, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 115, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 118, "usage_type": "call"}, {"api_name": "statistics.mean", "line_number": 144, "usage_type": "call"}, {"api_name": "statistics.stdev", "line_number": 151, "usage_type": "call"}, {"api_name": "lib.jira.Jira", "line_number": 210, "usage_type": "call"}, {"api_name": "lib.jira.Jira", "line_number": 253, "usage_type": "call"}, {"api_name": "lib.jira.Jira", "line_number": 315, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 337, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 337, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 338, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 338, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 583, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 583, "usage_type": "name"}]}
+{"seq_id": "19022932612", "text": "import argparse\nimport platform\nimport sys\n\nimport requests\n\nimport pytwot\n\n\ndef show_version():\n entries = []\n\n entries.append(\"- Python v{0.major}.{0.minor}.{0.micro}-{0.releaselevel}\".format(sys.version_info))\n entries.append(\"- pytwot v{0.major}.{0.minor}.{0.micro}-{0.releaselevel}\".format(pytwot.version_info))\n entries.append(f\"- requests v{requests.__version__}\")\n uname = platform.uname()\n entries.append(\"- system info: {0.system} {0.release} {0.version}\".format(uname))\n print(\"\\n\".join(entries))\n\n\ndef core(args):\n if args.version:\n show_version()\n else:\n print(\n \"Hi Thank you for using pytwot! You can do can do `python3 -m pytwot --version` for version info!\\n\\nDocs: https://py-tweet.readthedocs.io/ \\nGithub: https://github.com/sengolda/pytwot/\"\n )\n\n\ndef parse_args():\n argparser = argparse.ArgumentParser()\n argparser.add_argument(\"-v\", \"--version\", action=\"store_true\", help=\"shows the library versioninfo\")\n argparser.set_defaults(func=core)\n return argparser.parse_args()\n\n\ndef main():\n args = parse_args()\n args.func(args)\n\n\nif __name__ == \"__main__\":\n main()\n", "repo_name": "Sengolda/pytwot", "sub_path": "pytwot/__main__.py", "file_name": "__main__.py", "file_ext": "py", "file_size_in_byte": 1162, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "sys.version_info", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pytwot.version_info", "line_number": 14, "usage_type": "attribute"}, {"api_name": "requests.__version__", "line_number": 15, "usage_type": "attribute"}, {"api_name": "platform.uname", "line_number": 16, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 31, "usage_type": "call"}]}
+{"seq_id": "72945464703", "text": "import linajea\nfrom linajea.tracking import TrackGraph\nimport daisy\nimport logging\nimport argparse\n\nlogger = logging.getLogger(__name__)\nlogging.basicConfig(level=logging.INFO)\n\n\ndef check_for_duplicate_gt_tracks(\n gt_db_name,\n db_host,\n start_frame,\n end_frame,\n cell_radius,\n node_overlap_threshold):\n roi = daisy.Roi((start_frame, 0, 0, 0),\n (end_frame - start_frame, 10000, 10000, 10000))\n gt_db = linajea.CandidateDatabase(gt_db_name, db_host)\n gt_graph = gt_db[roi]\n if not gt_graph.number_of_edges():\n logger.info(\"No edges in database. Skipping track formation.\")\n return\n\n track_graph = TrackGraph(gt_graph)\n tracks = track_graph.get_tracks()\n\n logger.info(\"Found {} tracks\".format(len(tracks)))\n for track_id, track in enumerate(tracks):\n for cell_id in track.nodes():\n track_graph.nodes[cell_id]['track_id'] = track_id\n\n # Count how many times each pair of tracks has cells close to each other\n dup_counts = {}\n for frame, cells in track_graph._cells_by_frame.items():\n for i in range(len(cells)):\n cell1 = cells[i]\n for j in range(i + 1, len(cells)):\n cell2 = cells[j]\n if 'track_id' not in track_graph.nodes[cell1]:\n print(cell1)\n print(track_graph.nodes[cell1])\n track1 = track_graph.nodes[cell1]['track_id']\n track2 = track_graph.nodes[cell2]['track_id']\n if track1 == track2:\n continue\n if close(track_graph.nodes[cell1],\n track_graph.nodes[cell2],\n cell_radius):\n logger.info(\"These cells are close together: %d %s %d %s\"\n % (cell1, track_graph.nodes[cell1],\n cell2, track_graph.nodes[cell2]))\n tup = (min(track1, track2), max(track1, track2))\n dup_counts[tup] = 1 if tup not in dup_counts\\\n else dup_counts[tup] + 1\n print(dup_counts)\n\n for pair in dup_counts:\n if dup_counts[pair] < node_overlap_threshold:\n continue\n print(\"Tracks {} and {} are duplicates\".format(*pair))\n t1 = tracks[pair[0]]\n t2 = tracks[pair[1]]\n print(\"%d has %d cells, %d has %d cells\"\n % (pair[0], t1.number_of_nodes(),\n pair[1], t2.number_of_nodes()))\n to_delete = pair[0] if t1.number_of_nodes() < t2.number_of_nodes()\\\n else pair[1]\n print(\"Would delete track {}\".format(to_delete))\n\n\ndef close(c1, c2, cell_radius):\n p1 = [c1['z'], c1['y'], c1['x']]\n p2 = [c2['z'], c2['y'], c2['x']]\n assert len(p1) == len(p2)\n for dim in range(len(p1)):\n if abs(p1[dim] - p2[dim]) > cell_radius:\n return False\n return True\n\n\ndef check_missing_edges(\n gt_db_name,\n db_host,\n start_frame,\n end_frame,\n move_distance):\n roi = daisy.Roi((start_frame, 0, 0, 0),\n (end_frame - start_frame, 10000, 10000, 10000))\n gt_db = linajea.CandidateDatabase(gt_db_name, db_host)\n gt_graph = gt_db[roi]\n if not gt_graph.number_of_edges():\n logger.info(\"No edges in database. Skipping track formation.\")\n return\n\n track_graph = TrackGraph(gt_graph)\n tracks = track_graph.get_tracks()\n\n logger.info(\"Found {} tracks\".format(len(tracks)))\n for track_id, track in enumerate(tracks):\n for cell_id in track.nodes():\n track_graph.nodes[cell_id]['track_id'] = track_id\n\n missing_edges = {}\n for frame, cells in track_graph._cells_by_frame.items():\n prev_frame = frame - 1\n if prev_frame not in track_graph._cells_by_frame:\n continue\n prev_cells = track_graph._cells_by_frame[prev_frame]\n for i in range(len(cells)):\n cell1 = cells[i]\n if len(track_graph.prev_edges(cell1)) > 0:\n continue\n for j in range(len(prev_cells)):\n cell2 = prev_cells[j]\n if 'track_id' not in track_graph.nodes[cell1]:\n print(cell1)\n print(track_graph.nodes[cell1])\n track1 = track_graph.nodes[cell1]['track_id']\n track2 = track_graph.nodes[cell2]['track_id']\n if track1 == track2:\n continue\n if close(track_graph.nodes[cell1],\n track_graph.nodes[cell2],\n move_distance):\n logger.info(\"These cells are close together and \"\n \"potentially missing an edge between them: \"\n \"%d %s %d %s\"\n % (cell1, track_graph.nodes[cell1],\n cell2, track_graph.nodes[cell2]))\n missing_edges[(cell1, cell2)] = (track_graph.nodes[cell1],\n track_graph.nodes[cell2])\n\n print(\"Found %d potential missing edges:\" % len(missing_edges))\n approved_ids = []\n for ids, locs in missing_edges.items():\n print(\"%s %s %s %s\" % (ids[0], locs[0], ids[1], locs[1]))\n merge = input().strip()\n while merge not in ['y', 'n']:\n merge = input(\"Please enter y or n\")\n if merge == 'y':\n approved_ids.append(ids)\n print(\"Approved %d missing edges: %s\" % (len(approved_ids), approved_ids))\n\n\ndef check_degree(\n gt_db_name,\n db_host,\n start_frame,\n end_frame):\n roi = daisy.Roi((start_frame, 0, 0, 0),\n (end_frame - start_frame, 10000, 10000, 10000))\n gt_db = linajea.CandidateDatabase(gt_db_name, db_host)\n logger.info(\"Reading GT cells and edges in %s\" % roi)\n gt_subgraph = gt_db[roi]\n node_degrees = {node: degree for node, degree in gt_subgraph.in_degree()}\n max_node_degree = max(node_degrees.values())\n logger.info(\"Max node degree for subgraph: %s\" % max_node_degree)\n if max_node_degree > 2:\n logger.info(\"Expected max in_degree <=2, got %d. \\n %s\"\n % (max_node_degree, str(node_degrees)))\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument(\"gt_db_name\")\n parser.add_argument(\"cell_radius\")\n parser.add_argument(\"node_overlap_threshold\")\n parser.add_argument(\"move_threshold\")\n parser.add_argument(\"-f\", \"--frames\", type=int, nargs=2, default=[0, 10e5])\n args = parser.parse_args()\n gt_db_name = args.gt_db_name\n start_frame, end_frame = args.frames\n cell_radius = args.cell_radius\n node_overlap_threshold = args.node_overlap_threshold\n move_threshold = args.move_threshold\n\n db_host = \"localhost\"\n\n check_for_duplicate_gt_tracks(\n gt_db_name, db_host,\n start_frame, end_frame,\n int(cell_radius), int(node_overlap_threshold))\n check_missing_edges(\n gt_db_name, db_host,\n start_frame, end_frame,\n int(move_threshold))\n check_degree(gt_db_name, db_host, start_frame, end_frame)\n", "repo_name": "funkelab/linajea_experiments", "sub_path": "data/scripts/check_duplicate_gt_and_missing_edges.py", "file_name": "check_duplicate_gt_and_missing_edges.py", "file_ext": "py", "file_size_in_byte": 7230, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3, "dataset": "github-code", "pt": "24", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 8, "usage_type": "attribute"}, {"api_name": "daisy.Roi", "line_number": 18, "usage_type": "call"}, {"api_name": "linajea.CandidateDatabase", "line_number": 20, "usage_type": "call"}, {"api_name": "linajea.tracking.TrackGraph", "line_number": 26, "usage_type": "call"}, {"api_name": "daisy.Roi", "line_number": 89, "usage_type": "call"}, {"api_name": "linajea.CandidateDatabase", "line_number": 91, "usage_type": "call"}, {"api_name": "linajea.tracking.TrackGraph", "line_number": 97, "usage_type": "call"}, {"api_name": "daisy.Roi", "line_number": 152, "usage_type": "call"}, {"api_name": "linajea.CandidateDatabase", "line_number": 154, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 166, "usage_type": "call"}]}
+{"seq_id": "40201359012", "text": "from matplotlib import pyplot as plt\nimport numpy as np \n\nf = open('idea1.txt', 'r')\na = f.read()\nf.close()\n\n\na = a.split()\n\ndic = {}\n\nfor j,i in enumerate(a):\n\tif j%2==0 and j%4!=0:\n\t\tx = int(a[j])\n\t\ty = int(a[j+1])\n\t\ttry:\n\t\t\tr = abs(x-y)/x\n\t\texcept:\n\t\t\tr = 0\n\t\tif r in dic:\n\t\t\tdic[r] += 1\n\t\telse:\n\t\t\tdic[r] = 1\n\narr = sorted(dic)\ndic1 = {(i):dic[i] for i in arr[1:]}\n\nplt.xlabel('Relative Error')\nplt.ylabel('Frequency')\nplt.plot(list(dic1.keys()), list(dic1.values()))\nplt.show()\n\n# plt.bar(dic1.keys(), dic1.values(), width=0.1,color='b')\n# plt.show()", "repo_name": "vinusankarsiitgn/My-approx", "sub_path": "distrib.py", "file_name": "distrib.py", "file_ext": "py", "file_size_in_byte": 557, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "matplotlib.pyplot.xlabel", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}]}
+{"seq_id": "22264288368", "text": "import torch\nfrom torch import nn\nfrom torchsummary import summary\n\nclass CNN(nn.Module):\n\n def __init__(self):\n super().__init__()\n\n # 4 conv layer -> flatten -> linear -> softmax\n self.conv1 = nn.Sequential(\n nn.Conv2d(\n in_channels=1,\n out_channels=16,\n kernel_size=3,\n stride=1,\n padding=2\n ),\n nn.ReLU(),\n nn.MaxPool2d(kernel_size=2)\n )\n self.conv2 = nn.Sequential(\n nn.Conv2d(\n in_channels=16,\n out_channels=32,\n kernel_size=3,\n stride=1,\n padding=2\n ),\n nn.ReLU(),\n nn.MaxPool2d(kernel_size=2)\n )\n self.conv3 = nn.Sequential(\n nn.Conv2d(\n in_channels=32,\n out_channels=64,\n kernel_size=3,\n stride=1,\n padding=2\n ),\n nn.ReLU(),\n nn.MaxPool2d(kernel_size=2)\n )\n self.conv4 = nn.Sequential(\n nn.Conv2d(\n in_channels=64,\n out_channels=128,\n kernel_size=3,\n stride=1,\n padding=2\n ),\n nn.ReLU(),\n nn.MaxPool2d(kernel_size=2)\n )\n self.flatten = nn.Flatten()\n self.linear = nn.Linear(in_features=128 * 5 * 4, out_features=8)\n\n def forward(self, input_data):\n x = self.conv1(input_data)\n x = self.conv2(x)\n x = self.conv3(x)\n x = self.conv4(x)\n x = self.flatten(x)\n logits = self.linear(x)\n\n return logits\n\n\nif __name__ == \"__main__\":\n cnn = CNN()\n summary(cnn.cuda(), (1, 64, 44))\n", "repo_name": "kshipra-jadav/pytorch_audio_classification", "sub_path": "final/cnn.py", "file_name": "cnn.py", "file_ext": "py", "file_size_in_byte": 1801, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "torch.nn.Module", "line_number": 5, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 5, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 11, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.nn.Flatten", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 55, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 56, "usage_type": "name"}, {"api_name": "torchsummary.summary", "line_number": 71, "usage_type": "call"}]}
+{"seq_id": "22913294085", "text": "#!/usr/bin/python3\n\"\"\"The ``5-filter_cities`` module takes in the name of a states as an\n argument and lists all cities of that state, using the database\n 'hbtn_0e_4_usa'\n\"\"\"\nimport sys\nimport MySQLdb\n\nif __name__ == '__main__':\n if len(sys.argv) < 5:\n sys.exit(1)\n db = MySQLdb.connect(host=\"localhost\", port=3306, user=sys.argv[1],\n passwd=sys.argv[2], db=sys.argv[3])\n cur = db.cursor()\n cur.execute(\"\"\"SELECT cities.name FROM cities INNER JOIN states\n ON states.id = cities.state_id WHERE states.name LIKE %s\n ORDER BY cities.id ASC\"\"\", (sys.argv[4],))\n cities = cur.fetchall()\n j = 0\n for city in cities:\n for i in city:\n if j == 0:\n print(\"{}\".format(i), end=\"\")\n j += 1\n continue\n print(\", {}\".format(i), end=\"\")\n print()\n\n cur.close()\n db.close()\n", "repo_name": "budiong054/alx-higher_level_programming", "sub_path": "0x0F-python-object_relational_mapping/5-filter_cities.py", "file_name": "5-filter_cities.py", "file_ext": "py", "file_size_in_byte": 926, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "sys.argv", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 11, "usage_type": "call"}, {"api_name": "MySQLdb.connect", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 13, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 17, "usage_type": "attribute"}]}
+{"seq_id": "8279209707", "text": "import numpy as np\nimport torch\n\nclass AApLayer(torch.nn.Module):\n def __init__(self, n_in, n_out, aging_generator, args):\n super().__init__()\n self.args = args\n self.device = args.DEVICE\n \n theta = torch.rand([n_in + 2, n_out])/100. + args.gmin\n theta[-1, :] = theta[-1, :] + args.gmax\n theta[-2, :] = args.ACT_eta3/(1.-args.ACT_eta3)*(torch.sum(theta[:-2,:], axis=0)+theta[-1,:])\n self.theta_ = torch.nn.Parameter(theta, requires_grad=True)\n \n # initialize time sampling\n # t = 0 equals nominal training\n self.K = args.K_train\n if args.MODE == 'nominal':\n self.t = torch.tensor([0.])\n else:\n self.t = torch.linspace(0, 1, self.K)\n # initialize aging model\n self.M = args.M_train\n self.aging_generator = aging_generator\n \n # initialization for variation\n self.N = args.N_train\n self.epsilon = args.e_train\n \n @property\n def agingmodels(self):\n return self.aging_generator.get_models(self.M*self.theta_.numel()*self.N) # M aging models for each sampled theta (N variations)\n\n @property\n def theta_ideal(self):\n self.theta_.data.clamp_(-self.args.gmax, self.args.gmax)\n theta_temp = self.theta_.clone()\n theta_temp[theta_temp.abs() < self.args.gmin] = 0.\n return theta_temp.detach() + self.theta_ - self.theta_.detach()\n \n @property\n def theta(self):\n mean = self.theta_ideal.repeat(self.N, 1, 1)\n variation = (torch.rand(mean.shape)*2. - 1.) * self.epsilon + 1.\n return mean.to(self.device) * variation.to(self.device)\n \n @property\n def theta_aged(self):\n # generate aging decay coefficient [M, K, N, n_in, n_out]\n aging_decay = torch.tensor([m(self.t) for m in self.agingmodels]) # [M*N*n_in*n_out, K]\n aging_decay = aging_decay.reshape(self.M,self.theta.shape[0],self.theta.shape[1],self.theta.shape[2],self.K).permute(0,4,1,2,3)\n # broad casting: [M, K, N, n_in, n_out] * [N, n_in, n_out] -> multiply for the last 2 dimension\n return self.theta * aging_decay.to(self.device)\n \n @property\n def W(self):\n return self.theta_aged.abs() / torch.sum(self.theta_aged.abs(), axis=3, keepdim=True)\n\n def INV(self, x):\n return -(self.args.NEG_eta1 + self.args.NEG_eta2 * torch.tanh((x - self.args.NEG_eta3) * self.args.NEG_eta4))\n \n def MAC(self, a):\n # 0 and positive thetas are corresponding to no negative weight circuit\n positive = self.theta.clone().to(self.device)\n positive[positive >= 0] = 1.\n positive[positive < 0] = 0.\n negative = 1. - positive\n # a in [M, K, N, E, n_in]\n a_extend = torch.cat([a,\n torch.ones( [a.shape[0], a.shape[1], a.shape[2], a.shape[3], 1]).to(self.device),\n torch.zeros([a.shape[0], a.shape[1], a.shape[2], a.shape[3], 1]).to(self.device)], dim=4)\n a_neg = self.INV(a_extend)\n a_neg[:,:,:,:,-1] = 0.\n z = torch.matmul(a_extend, self.W * positive) + torch.matmul(a_neg, self.W * negative)\n return z\n \n def ACT(self, z):\n return self.args.ACT_eta1 + self.args.ACT_eta2 * torch.tanh((z - self.args.ACT_eta3) * self.args.ACT_eta4)\n \n def forward(self, a_previous):\n z_new = self.MAC(a_previous)\n a_new = self.ACT(z_new)\n return a_new\n \n def SetParameter(self, name, value):\n # set time sampling and update K\n if name == 't':\n self.t = value\n self.K = self.t.shape[0]\n # set number of aging-model sampling M\n elif name == 'M':\n self.M = value\n # set number of samples\n elif name == 'N':\n self.N = value\n # set variations\n elif name == 'epsilon':\n self.epsilon = value\n # set device\n elif name == 'device':\n self.device = value\n \n\nclass AApNN(torch.nn.Module):\n def __init__(self, topology, aging_generator, args):\n super().__init__()\n self.args = args\n self.M = args.M_train\n self.K = args.K_train\n if args.MODE == 'nominal':\n self.t = torch.tensor([0.])\n else:\n self.t = torch.linspace(0, 1, self.K)\n self.N = args.N_train\n self.epsilon = args.e_train\n self.model = torch.nn.Sequential()\n self.device = args.DEVICE\n for i in range(len(topology)-1):\n self.model.add_module(f'{i}-th pLayer', AApLayer(topology[i], topology[i+1], aging_generator, args))\n \n def forward(self, X):\n X_extend = X.repeat(self.M, self.K, self.N, 1, 1)\n return self.model(X_extend)\n \n def SetParameter(self, name, value):\n # set time sampling and update K\n if name == 't':\n self.t = value\n self.K = self.t.shape[0]\n for m in self.model:\n m.SetParameter('t', self.t)\n # set number of time sampling K and generate random time sampling\n elif name == 'K':\n self.K = value\n self.t = torch.rand(self.K)\n for m in self.model:\n m.SetParameter('t', self.t)\n # set number of aging-model sampling M\n elif name == 'M':\n self.M = value\n for m in self.model:\n m.SetParameter('M', self.M)\n # set number of samples\n elif name == 'N':\n self.N = value\n for m in self.model:\n m.SetParameter('N', self.N)\n # set variations\n elif name == 'epsilon':\n self.epsilon = value\n for m in self.model:\n m.SetParameter('epsilon', self.epsilon)\n # set device\n elif name == 'device':\n self.device = value\n for m in self.model:\n m.SetParameter('device', self.device)\n \n\nclass Lossfunction(torch.nn.Module):\n def __init__(self, args):\n super().__init__()\n self.args = args\n\n def standard(self, prediction, label): \n label = label.reshape(-1, 1)\n fy = prediction.gather(1, label).reshape(-1, 1)\n fny = prediction.clone()\n fny = fny.scatter_(1, label, -10 ** 10)\n fnym = torch.max(fny, axis=1).values.reshape(-1, 1)\n l = torch.max(self.args.m + self.args.T - fy, torch.tensor(0)) + torch.max(self.args.m + fnym, torch.tensor(0))\n L = torch.mean(l)\n return L\n \n def MonteCarlo(self, prediction, label):\n M = prediction.shape[0]\n K = prediction.shape[1]\n N = prediction.shape[2]\n loss = torch.tensor(0.).to(self.args.DEVICE)\n for m in range(M):\n for k in range(K):\n for n in range(N):\n loss += self.standard(prediction[m,k,n,:,:], label)\n return loss / M / K / N\n \n def GaussianQuadrature(self, prediction, label):\n return torch.tensor(0.)\n \n def forward(self, prediction, label):\n if self.args.integration == 'MC':\n return self.MonteCarlo(prediction, label)\n elif self.args.integration == 'GQ':\n return self.GaussianQuadrature(prediction, label)", "repo_name": "Neuromophic/Aging-aware-training", "sub_path": "pNN_aging.py", "file_name": "pNN_aging.py", "file_ext": "py", "file_size_in_byte": 7254, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "torch.nn", "line_number": 4, "usage_type": "attribute"}, {"api_name": "torch.rand", "line_number": 10, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.nn.Parameter", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 13, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.linspace", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.tanh", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.tanh", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 104, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.linspace", "line_number": 113, "usage_type": "call"}, {"api_name": "torch.nn.Sequential", "line_number": 116, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 116, "usage_type": "attribute"}, {"api_name": "torch.rand", "line_number": 135, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 160, "usage_type": "attribute"}, {"api_name": "torch.max", "line_number": 170, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 172, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 187, "usage_type": "call"}]}
+{"seq_id": "25616515380", "text": "import pickle as pkl\nfrom pathlib import Path\nimport numpy as np\n\nfrom aes670hw2 import enhance as enh\nfrom aes670hw2 import geo_plot as gp\n\ndef wrap_pixels(X, cycle_num, cycle_size, num_px):\n \"\"\"\n Based on the cycle size, the number of concatenated \"cycle split\" cycles,\n and the number of appended pixels per cycle in a 1D dataset, split the\n dataset back into the original continuous time series per pixel\n \"\"\"\n X = X[cycle_num*cycle_size:cycle_num*cycle_size+cycle_size]\n return np.split(X, num_px, axis=0)\n\n\nif __name__==\"__main__\":\n model_dir = Path(\"models/set003\")\n\n set_label = \"silty-loam_set003\"\n\n \"\"\"\n Parameters for 'wrapping' a dataset generated with the\n cycle-split technique datasets\n \"\"\"\n # Size of the\n cycle_size = 8064 # set003 training data\n #cycle_size = 2016 # set003 validation/testing data\n num_px = 12 # for unraveling pixels in each dataset\n #CNUM = 2 # cycle number\n #PNUM = 10\n for CNUM in range(4):\n for PNUM in range(12):\n TITLE = f\"Set 3 model on training data, C{CNUM}/P{PNUM}\"\n\n fig_dir = Path(f\"figures/output_curves/\")\n fig_path = fig_dir.joinpath(\n Path(f\"{set_label}_train_C{CNUM}_P{PNUM}.png\"))\n\n #checkpoint_file = Path(\"data/model_check/set001\")\n checkpoint_file = model_dir.joinpath(\"checkpoint\")\n\n t_out, v_out, s_out = pkl.load(model_dir.joinpath(\n f\"output/{set_label}_out.pkl\").open(\"rb\"))\n\n \"\"\" Load the pickles of model input data \"\"\"\n t_pkl = model_dir.joinpath(f\"input/{set_label}_training.pkl\")\n v_pkl = model_dir.joinpath(f\"input/{set_label}_validation.pkl\")\n s_pkl = model_dir.joinpath(f\"input/{set_label}_testing.pkl\")\n\n t_feat,t_truth,t_times = pkl.load(t_pkl.open(\"rb\"))\n v_feat,v_truth,v_times = pkl.load(v_pkl.open(\"rb\"))\n s_feat,s_truth,s_times = pkl.load(s_pkl.open(\"rb\"))\n\n \"\"\" Configure the dataset to model and plot \"\"\"\n TRUTH = t_truth\n OUT = t_out\n TIMES = t_times\n\n print(\"TRUTH:\",TRUTH.shape)\n print(\"MODEL:\",OUT.shape)\n\n # Extract the truth basis sample and split it into curves for each\n # independent soil depth level\n TRUTH = wrap_pixels(TRUTH, CNUM, cycle_size, num_px)[PNUM]\n TRUTH = [TRUTH[:,i] for i in range(TRUTH.shape[1])]\n OUT = wrap_pixels(OUT, CNUM, cycle_size, num_px)[PNUM]\n OUT = [OUT[:,i] for i in range(OUT.shape[1])]\n ipos = CNUM*cycle_size\n TIMES = [ t.strftime(\"%j\") for t in TIMES[CNUM] ]\n\n gp.plot_lines(\n TIMES,\n TRUTH+OUT,\n show=False,\n plot_spec={\n \"yrange\":(0,400),\n \"title\":TITLE,\n \"xlabel\":\"Day of the year\",\n \"ylabel\":\"Liquid soil moisture ($\\\\frac{kg}{m^2}$)\",\n \"line_width\":1.2,\n \"colors\":[\"blue\" for i in range(len(TRUTH))] + \\\n [\"red\" for i in range(len(OUT))],\n #\"legend_ncols\":2,\n \"dpi\":200,\n },\n image_path=fig_path\n )\n\n # tmp\n", "repo_name": "Mitchell-D/testbed", "sub_path": "old/evaluate_model.py", "file_name": "evaluate_model.py", "file_ext": "py", "file_size_in_byte": 3406, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "numpy.split", "line_number": 15, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 19, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 37, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 39, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 44, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 52, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 53, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 54, "usage_type": "call"}, {"api_name": "aes670hw2.geo_plot.plot_lines", "line_number": 73, "usage_type": "call"}, {"api_name": "aes670hw2.geo_plot", "line_number": 73, "usage_type": "name"}]}
+{"seq_id": "28577807333", "text": "# Functionalities related to searching in a file\n\nimport modules.file.read as fr\nfrom __init__ import IDENTIFIER_COL, PATH\n\ndef get_item_details(file, identifier):\n \"\"\"Extract the item details from database\n \n Args:\n (str): the file to search for the item in\n (str|int): the identifier of the item being queried\n \n Returns:\n (list|None): list of the item details if exist and None otherwise\"\"\"\n\n # Extracting users data from database as list of lists\n items_data = fr.get_lines(PATH[file], type='list')\n\n # Iterating over the users data\n for item in items_data:\n\n # Checking the items identifiers against the queried identifier\n if item[IDENTIFIER_COL[file]] == identifier:\n # if found return the data\n return item\n\n # If the identifier is not found return None\n return None", "repo_name": "AbdElrahman-A-Eid/CrowdFundApp", "sub_path": "modules/file/search.py", "file_name": "search.py", "file_ext": "py", "file_size_in_byte": 868, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "27", "api": [{"api_name": "modules.file.read.get_lines", "line_number": 17, "usage_type": "call"}, {"api_name": "modules.file.read", "line_number": 17, "usage_type": "name"}, {"api_name": "__init__.PATH", "line_number": 17, "usage_type": "name"}, {"api_name": "__init__.IDENTIFIER_COL", "line_number": 23, "usage_type": "name"}]}
+{"seq_id": "31152739892", "text": "from django.urls import path \nfrom . import views \n\napp_name = 'user'\n\nurlpatterns = [\n path('dashboard/', views.dashboard_view, name='dashboard'),\n path('profile/', views.profile_view, name='profile'),\n path('diagnosis/', views.diagonzie_symptoms, name='diagnosis'),\n path('diagnosis-details/', views.diagnosis_details, name='diagnosis-details')\n]", "repo_name": "Prosperibe12/medproject", "sub_path": "userapp/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 360, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}]}
+{"seq_id": "25094526232", "text": "\"\"\"\nScript used for training the normalnet using tfutils\n\"\"\"\n\nfrom __future__ import division, print_function, absolute_import\nimport os, sys\nimport numpy as np\n\nimport tensorflow as tf\n\nimport normal_encoder_asymmetric_with_bypass\n\nfrom tfutils import base, data, model, optimizer\n\nimport json\nimport copy\nimport argparse\n\n#os.environ[\"CUDA_VISIBLE_DEVICES\"]=\"2\"\n\nhost = os.uname()[1]\n\nDATA_PATH = {}\nif host == 'freud': # freud\n DATA_PATH['train'] = '/media/data/one_world_dataset/randomperm.hdf5'\n DATA_PATH['val'] = '/media/data/one_world_dataset/randomperm_test1.hdf5'\n\nelif host.startswith('node') or host == 'openmind7': # OpenMind\n DATA_PATH['train'] = '/om/user/chengxuz/Data/one_world_dataset/randomperm.hdf5'\n DATA_PATH['val'] = '/om/user/chengxuz/Data/one_world_dataset/randomperm_test1.hdf5'\nelse:\n print(\"Not supported yet!\")\n exit()\n\n\n\ndef online_agg(agg_res, res, step):\n if agg_res is None:\n agg_res = {k:[] for k in res}\n for k,v in res.items():\n agg_res[k].append(np.mean(v))\n return agg_res\n\n\ndef exponential_decay(global_step,\n learning_rate=.01,\n decay_factor=.95,\n decay_steps=1,\n ):\n # Decay the learning rate exponentially based on the number of steps.\n if decay_factor is None:\n lr = learning_rate # just a constant.\n else:\n # Calculate the learning rate schedule.\n lr = tf.train.exponential_decay(\n learning_rate, # Base learning rate.\n global_step, # Current index into the dataset.\n decay_steps, # Decay step\n decay_factor, # Decay rate.\n staircase=True)\n return lr\n\n\nclass Threedworld(data.HDF5DataProvider):\n\n #N_TRAIN = 2048000 - 102400\n #N_VAL = 102400 \n N_TRAIN = 2048000\n N_VAL = 128000\n\n def __init__(self,\n data_path,\n group='train',\n batch_size=1,\n crop_size=None,\n *args,\n **kwargs):\n \"\"\"\n A specific reader for Threedworld generated dataset stored as a HDF5 file\n\n Args:\n - data_path\n path to raw hdf5 data\n Kwargs:\n - group (str, default: 'train')\n Which subset of the dataset you want: train, val.\n The latter contains 50k images from the train set,\n so that you can directly compare performance on the validation set\n to the performance on the train set to track overfitting.\n - batch_size (int, default: 1)\n Number of images to return when `next` is called. By default set\n to 1 since it is expected to be used with queues where reading one\n image at a time is ok.\n - crop_size (int or None, default: None)\n For center crop (crop_size x crop_size). If None, no cropping will occur.\n - *args, **kwargs\n Extra arguments for HDF5DataProvider\n \"\"\"\n self.group = group\n self.images = 'images'\n self.labels = 'normals'\n '''\n if self.group=='train':\n subslice = range(self.N_TRAIN)\n else:\n subslice = range(self.N_TRAIN, self.N_TRAIN + self.N_VAL)\n '''\n super(Threedworld, self).__init__(\n data_path[group],\n [self.images, self.labels],\n batch_size=batch_size,\n postprocess={self.images: self.postproc, self.labels: self.postproc},\n pad=True,\n *args, **kwargs)\n if crop_size is None:\n self.crop_size = 256\n else:\n self.crop_size = crop_size\n\n self.off = None\n self.now_num = 0\n\n def postproc(self, ims, f):\n norm = ims.astype(np.float32) / 255\n if self.group=='train':\n #print('In train')\n if self.now_num==0:\n off = np.random.randint(0, 256 - self.crop_size, size=2)\n self.off = off\n else:\n off = self.off\n else:\n off = int((256 - self.crop_size)/2)\n off = [off, off]\n images_batch = norm[:,\n off[0]: off[0] + self.crop_size,\n off[1]: off[1] + self.crop_size]\n if self.now_num==0:\n self.now_num = 1\n else:\n self.now_num = 0\n\n return images_batch\n\n def next(self):\n batch = super(Threedworld, self).next()\n feed_dict = {'images': np.squeeze(batch[self.images]),\n 'labels': np.squeeze(batch[self.labels])}\n return feed_dict\n\n#BATCH_SIZE = 256\n#BATCH_SIZE = 192\nBATCH_SIZE = 128\nNUM_BATCHES_PER_EPOCH = Threedworld.N_TRAIN // BATCH_SIZE\nIMAGE_SIZE_CROP = 224\nNUM_CHANNELS = 3\nNORM_NUM = (IMAGE_SIZE_CROP**2) * NUM_CHANNELS * BATCH_SIZE\n\ndef loss_ave_l2(output, labels):\n loss = tf.nn.l2_loss(output - labels) / NORM_NUM\n return loss\n\ndef rep_loss(inputs, outputs, target):\n loss = tf.nn.l2_loss(outputs - inputs[target]) / NORM_NUM\n return {'loss': loss}\n\ndef postprocess_config(cfg):\n cfg = copy.deepcopy(cfg)\n for k in ['encode', 'decode', 'hidden']:\n if k in cfg:\n ks = cfg[k].keys()\n for _k in ks:\n cfg[k][int(_k)] = cfg[k].pop(_k)\n return cfg\n\n\ndef preprocess_config(cfg):\n cfg = copy.deepcopy(cfg)\n for k in ['encode', 'decode', 'hidden']:\n if k in cfg:\n ks = cfg[k].keys()\n for _k in ks:\n #assert isinstance(_k, int), _k\n cfg[k][str(_k)] = cfg[k].pop(_k)\n return cfg\n\ndef main(args):\n #cfg_initial = postprocess_config(json.load(open(cfgfile)))\n if args.gpu>-1:\n os.environ['CUDA_VISIBLE_DEVICES'] = str(args.gpu)\n cfg_initial = preprocess_config(json.load(open(args.pathconfig)))\n exp_id = args.expId\n cache_dir = os.path.join(args.cacheDirPrefix, '.tfutils', 'localhost:'+ str(args.nport), 'normalnet-test', 'normalnet', exp_id)\n params = {\n 'save_params': {\n 'host': 'localhost',\n #'port': 31001,\n 'port': args.nport,\n 'dbname': 'normalnet-test',\n 'collname': 'normalnet',\n #'exp_id': 'trainval0',\n 'exp_id': exp_id,\n #'exp_id': 'trainval2', # using screen?\n\n 'do_save': True,\n #'do_save': False,\n 'save_initial_filters': True,\n 'save_metrics_freq': 2000, # keeps loss from every SAVE_LOSS_FREQ steps.\n 'save_valid_freq': 10000,\n 'save_filters_freq': 30000,\n 'cache_filters_freq': 10000,\n 'cache_dir': cache_dir, # defaults to '~/.tfutils'\n },\n\n 'load_params': {\n 'host': 'localhost',\n # 'port': 31001,\n # 'dbname': 'alexnet-test',\n # 'collname': 'alexnet',\n # 'exp_id': 'trainval0',\n 'port': args.nport,\n 'dbname': 'normalnet-test',\n 'collname': 'normalnet',\n #'exp_id': 'trainval0',\n 'exp_id': exp_id,\n #'exp_id': 'trainval2', # using screen?\n 'do_restore': True,\n 'load_query': None\n },\n\n 'model_params': {\n 'func': normal_encoder_asymmetric_with_bypass.normalnet_tfutils,\n 'seed': args.seed,\n 'cfg_initial': cfg_initial\n },\n\n 'train_params': {\n 'data_params': {\n 'func': Threedworld,\n 'data_path': DATA_PATH,\n 'group': 'train',\n 'crop_size': IMAGE_SIZE_CROP,\n 'batch_size': 1\n },\n 'queue_params': {\n 'queue_type': 'fifo',\n 'batch_size': BATCH_SIZE,\n 'n_threads': 4,\n 'seed': 0,\n },\n 'thres_loss': 1000,\n 'num_steps': 90 * NUM_BATCHES_PER_EPOCH # number of steps to train\n },\n\n 'loss_params': {\n 'targets': 'labels',\n 'agg_func': tf.reduce_mean,\n 'loss_per_case_func': loss_ave_l2,\n 'loss_per_case_func_params': {}\n },\n\n 'learning_rate_params': {\n 'func': tf.train.exponential_decay,\n 'learning_rate': .01,\n 'decay_rate': .95,\n 'decay_steps': NUM_BATCHES_PER_EPOCH, # exponential decay each epoch\n 'staircase': True\n },\n\n 'optimizer_params': {\n 'func': optimizer.ClipOptimizer,\n 'optimizer_class': tf.train.MomentumOptimizer,\n 'clip': True,\n 'momentum': .9\n },\n 'validation_params': {\n 'topn': {\n 'data_params': {\n 'func': Threedworld,\n 'data_path': DATA_PATH, # path to image database\n 'group': 'val',\n 'crop_size': IMAGE_SIZE_CROP, # size after cropping an image\n },\n 'targets': {\n 'func': rep_loss,\n 'target': 'labels',\n },\n 'queue_params': {\n 'queue_type': 'fifo',\n 'batch_size': BATCH_SIZE,\n 'n_threads': 4,\n 'seed': 0,\n },\n 'num_steps': Threedworld.N_VAL // BATCH_SIZE + 1,\n 'agg_func': lambda x: {k:np.mean(v) for k,v in x.items()},\n 'online_agg_func': online_agg\n },\n },\n\n 'log_device_placement': False, # if variable placement has to be logged\n }\n #base.get_params()\n base.train_from_params(**params)\n\nif __name__ == '__main__':\n #base.get_params()\n #base.train_from_params(**params)\n parser = argparse.ArgumentParser(description='The script to train the normalnet')\n parser.add_argument('--nport', default = 22334, type = int, action = 'store', help = 'Port number of mongodb')\n parser.add_argument('--pathconfig', default = \"normals_config_winner0.cfg\", type = str, action = 'store', help = 'Path to config file')\n parser.add_argument('--expId', default = \"trainval2\", type = str, action = 'store', help = 'Name of experiment id')\n parser.add_argument('--seed', default = 0, type = int, action = 'store', help = 'Random seed for model')\n parser.add_argument('--gpu', default = -1, type = int, action = 'store', help = 'Index of gpu, currently only one gpu is allowed')\n parser.add_argument('--cacheDirPrefix', default = \"/home/chengxuz\", type = str, action = 'store', help = 'Prefix of cache directory')\n\n args = parser.parse_args()\n\n main(args)\n", "repo_name": "neuroailab/barrel", "sub_path": "normals_relat/normal_pred/train_normalnet_hdf5.py", "file_name": "train_normalnet_hdf5.py", "file_ext": "py", "file_size_in_byte": 10725, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 4, "dataset": "github-code", "pt": "24", "api": [{"api_name": "os.uname", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.train.exponential_decay", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 55, "usage_type": "attribute"}, {"api_name": "tfutils.data.HDF5DataProvider", "line_number": 64, "usage_type": "attribute"}, {"api_name": "tfutils.data", "line_number": 64, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 124, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 128, "usage_type": "attribute"}, {"api_name": "numpy.squeeze", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.nn.l2_loss", "line_number": 160, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 160, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.l2_loss", "line_number": 164, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 164, "usage_type": "attribute"}, {"api_name": "copy.deepcopy", "line_number": 168, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 178, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 190, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 191, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 193, "usage_type": "call"}, {"api_name": "os.path", "line_number": 193, "usage_type": "attribute"}, {"api_name": "normal_encoder_asymmetric_with_bypass.normalnet_tfutils", "line_number": 232, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 257, "usage_type": "attribute"}, {"api_name": "tensorflow.train", "line_number": 263, "usage_type": "attribute"}, {"api_name": "tfutils.optimizer.ClipOptimizer", "line_number": 271, "usage_type": "attribute"}, {"api_name": "tfutils.optimizer", "line_number": 271, "usage_type": "name"}, {"api_name": "tensorflow.train", "line_number": 272, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 295, "usage_type": "call"}, {"api_name": "tfutils.base.train_from_params", "line_number": 303, "usage_type": "call"}, {"api_name": "tfutils.base", "line_number": 303, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 308, "usage_type": "call"}]}
+{"seq_id": "74392918141", "text": "from src.config import Config\nfrom lightning.pytorch.callbacks.early_stopping import EarlyStopping\nfrom lightning.pytorch.callbacks.model_checkpoint import ModelCheckpoint\nfrom src.dataloaders import KFoldDataModuleContainer\nfrom src.models import S3Rec\nimport torch\nimport lightning as L\nimport copy\nimport numpy as np\nimport wandb\n\n\nclass HoldoutTrainer:\n def __init__(self, config: Config, model: L.LightningModule, data_module: L.LightningDataModule, metric: str, mode=\"max\") -> None:\n self.config = config\n self.model = model\n self.data_module = data_module\n self.is_cv = self.config.trainer.cv\n\n self.device = torch.device(\"cuda\" if config.cuda_condition else \"cpu\")\n\n self.early_stop = EarlyStopping(monitor=metric, patience=10, verbose=True, mode=mode)\n\n if config.trainer.is_pretrain:\n checkpoint_file = f\"{config.timestamp}_{config.model.model_name}_pretrain_{{{metric}:.2f}}\"\n else:\n checkpoint_file = f\"{config.timestamp}_{config.model.model_name}_{{{metric}:.2f}}\"\n\n self.checkpoint = ModelCheckpoint(monitor=metric, mode=mode, dirpath=config.path.output_dir, filename=checkpoint_file)\n\n if config.cuda_condition:\n self.trainer = L.Trainer(max_epochs=self.config.trainer.epoch, callbacks=[self.early_stop, self.checkpoint], accelerator=\"cuda\")\n else:\n self.trainer = L.Trainer(max_epochs=self.config.trainer.epoch, callbacks=[self.early_stop, self.checkpoint], accelerator=\"cpu\")\n\n def train(self):\n print(\"start training\")\n self.trainer.fit(self.model, datamodule=self.data_module)\n\n def predict(self):\n # inference\n preds = self.trainer.predict(self.model, datamodule=self.data_module)\n\n if self.is_cv:\n return torch.concatenate(preds)\n else:\n rating_pred = np.array(torch.concatenate(preds))\n ind = np.argpartition(rating_pred, -10)[:, -10:]\n arr_ind = rating_pred[np.arange(len(rating_pred))[:, None], ind]\n arr_ind_argsort = np.argsort(arr_ind)[np.arange(len(rating_pred)), ::-1]\n pred_list = ind[np.arange(len(rating_pred))[:, None], arr_ind_argsort]\n\n return pred_list\n\n def test(self):\n self.trainer.test(self.model, datamodule=self.data_module)\n\n\nclass PretrainTrainer(HoldoutTrainer):\n def __init__(self, config: Config, model: S3Rec, data_module: L.LightningDataModule, metric: str, mode: str, pretrain_path: str) -> None:\n super().__init__(config, model, data_module, metric, mode)\n self.pretrain_path = pretrain_path\n\n def train(self):\n super().train()\n\n self.save_best_pretrained_module()\n\n def save_best_pretrained_module(self):\n # load and\n self.model.load_from_checkpoint(self.checkpoint.best_model_path, config=self.config, name2attr_size=self.model.name2attr_size)\n self.model.save_pretrained_module(self.pretrain_path)\n\n\nclass KFoldTrainer:\n def __init__(\n self, config: Config, model: L.LightningModule, kfold_data_module_container: KFoldDataModuleContainer, metric: str, mode: str\n ) -> None:\n self.config = config\n self.n_fold = config.trainer.k\n\n self.kfold_data_module_container = kfold_data_module_container\n\n self.fold_trainers: list[HoldoutTrainer] = self.__fold_trainers(self.n_fold, config, model, kfold_data_module_container, metric, mode)\n self.sub_result_csv_list = []\n self.val_result_csv_list = []\n\n def __fold_trainers(\n self, n_fold: int, config: Config, model: L.LightningModule, kfold_data_module_container: KFoldDataModuleContainer, metric: str, mode: str\n ) -> list[HoldoutTrainer]:\n fold_trainers = []\n\n for fold in range(n_fold):\n fold_model = copy.deepcopy(model)\n kfold_data_module = kfold_data_module_container.kfold_data_module(fold)\n\n trainer = HoldoutTrainer(config, fold_model, kfold_data_module, metric, mode)\n fold_trainers.append(trainer)\n\n return fold_trainers\n\n def train(self):\n cv_score = 0.0\n\n for fold, fold_trainer in enumerate(self.fold_trainers):\n print(f\"------------- Train Fold {fold} -------------\")\n\n fold_trainer.train()\n fold_model = fold_trainer.model\n\n print(\n \"check tr_result, val_result: \",\n len(fold_model.tr_result),\n len(fold_model.val_result),\n )\n tr_avg_loss = torch.stack([x[\"rec_avg_loss\"] for x in fold_model.tr_result]).mean()\n tr_cur_loss = torch.stack([x[\"rec_cur_loss\"] for x in fold_model.tr_result]).mean()\n\n val_recall = fold_model.val_result.mean()\n\n print(f\">>> tr_avg_loss: {tr_avg_loss},tr_cur_loss: {tr_cur_loss}, val_recall@10: {val_recall}\")\n cv_score += val_recall\n\n cv_score /= self.n_fold\n print(f\"-----------------cv_recall@10_score: {cv_score}-----------------\")\n wandb.log({\"cv_score\": cv_score})\n\n return cv_score\n\n def predict(self):\n fold = 0\n while self.fold_trainers:\n fold_trainer = self.fold_trainers.pop(0)\n\n print(f\"------------- Predict Fold {fold} -------------\")\n if fold == 0:\n output = fold_trainer.predict()\n else:\n output = output + fold_trainer.predict()\n\n fold += 1\n\n rating_pred = np.array(output / self.n_fold)\n\n ind = np.argpartition(rating_pred, -10)[:, -10:]\n arr_ind = rating_pred[np.arange(len(rating_pred))[:, None], ind]\n arr_ind_argsort = np.argsort(arr_ind)[np.arange(len(rating_pred)), ::-1]\n oof_pred_list = ind[np.arange(len(rating_pred))[:, None], arr_ind_argsort]\n\n return oof_pred_list\n\n def test(self):\n pass\n", "repo_name": "boostcampaitech5/level2_movierecommendation-recsys-03", "sub_path": "sequential/src/trainers.py", "file_name": "trainers.py", "file_ext": "py", "file_size_in_byte": 5865, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 4, "dataset": "github-code", "pt": "24", "api": [{"api_name": "src.config.Config", "line_number": 14, "usage_type": "name"}, {"api_name": "lightning.LightningModule", "line_number": 14, "usage_type": "attribute"}, {"api_name": "lightning.LightningDataModule", "line_number": 14, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 20, "usage_type": "call"}, {"api_name": "lightning.pytorch.callbacks.early_stopping.EarlyStopping", "line_number": 22, "usage_type": "call"}, {"api_name": "lightning.pytorch.callbacks.model_checkpoint.ModelCheckpoint", "line_number": 29, "usage_type": "call"}, {"api_name": "lightning.Trainer", "line_number": 32, "usage_type": "call"}, {"api_name": "lightning.Trainer", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.concatenate", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.concatenate", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.argpartition", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 51, "usage_type": "call"}, {"api_name": "src.config.Config", "line_number": 60, "usage_type": "name"}, {"api_name": "src.models.S3Rec", "line_number": 60, "usage_type": "name"}, {"api_name": "lightning.LightningDataModule", "line_number": 60, "usage_type": "attribute"}, {"api_name": "src.config.Config", "line_number": 77, "usage_type": "name"}, {"api_name": "lightning.LightningModule", "line_number": 77, "usage_type": "attribute"}, {"api_name": "src.dataloaders.KFoldDataModuleContainer", "line_number": 77, "usage_type": "name"}, {"api_name": "src.config.Config", "line_number": 89, "usage_type": "name"}, {"api_name": "lightning.LightningModule", "line_number": 89, "usage_type": "attribute"}, {"api_name": "src.dataloaders.KFoldDataModuleContainer", "line_number": 89, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 116, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 117, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.argpartition", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 148, "usage_type": "call"}]}
+{"seq_id": "27928567950", "text": "import scapy.all as scapy\nimport argparse\n\ndef get_arguments():\n parser = argparse.ArgumentParser()\n parser.add_argument(\"-i\", \"--ip\", dest=\"target\", help=\"Target IP / IP Range\")\n options = parser.parse_args()\n if not options.target:\n parser.error(\"[!] Please add an interface to proceed (like : 192.168.1.1/24), --help for more informations.\")\n return options\n\ndef scan(ip):\n arp_request = scapy.ARP(pdst = ip)\n broadcast = scapy.Ether(dst = \"ff:ff:ff:ff:ff:ff\")\n packet = broadcast/arp_request\n ask_list = scapy.srp(packet, timeout = 1, verbose = False)[0]\n \n packet_list = []\n for i in ask_list:\n packet_dict = {\"ip\" : i[1].psrc, \"mac\" : i[1].hwsrc}\n packet_list.append(packet_dict)\n return(packet_list)\n\ndef print_res(res):\n print(\"\"\" __ _ ___ _____ _ _ __ ___ _ __ __ ___ __ __ _ __ _ ___ ___ \n| \\| | __|_ _| | | |/__\\| _ \\ |/ / /' _/ / _// \\| \\| | \\| | __| _ \\ \n| | ' | _| | | | 'V' | \\/ | v / < `._`.| \\_| /\\ | | ' | | ' | _|| v / \n|_|\\__|___| |_| !_/ \\_!\\__/|_|_\\_|\\_\\ |___/ \\__/_||_|_|\\__|_|\\__|___|_|_\\ \"\"\")\n print(\"=========================================\")\n print(\"IP\\t\\t\\tMAC Address\\n=========================================\")\n for n in res:\n print(n[\"ip\"] + \"\\t\\t\" + n[\"mac\"])\n \noptions = get_arguments()\nscan_result = scan(options.target)\nprint_res(scan_result)", "repo_name": "LasCC/Network-Scanner", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1390, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 15, "dataset": "github-code", "pt": "27", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 5, "usage_type": "call"}, {"api_name": "scapy.all.ARP", "line_number": 13, "usage_type": "call"}, {"api_name": "scapy.all", "line_number": 13, "usage_type": "name"}, {"api_name": "scapy.all.Ether", "line_number": 14, "usage_type": "call"}, {"api_name": "scapy.all", "line_number": 14, "usage_type": "name"}, {"api_name": "scapy.all.srp", "line_number": 16, "usage_type": "call"}, {"api_name": "scapy.all", "line_number": 16, "usage_type": "name"}]}
+{"seq_id": "23382301169", "text": "from .legion_tools import *\nimport scipy.sparse.linalg as lin\n\ndef liouvillian_sim(job_index, output_directory='./results'):\n\n with open('stack.csv', 'r') as f:\n header = f.readline()\n stack_name = header.split('\\n')[0]\n stack_frame = pd.read_csv(f)\n\n stack_directory = output_directory\n\n kappa_phi = 0.0\n\n sys_params = stack_frame.iloc[job_index]\n frame_params = sys_params\n packaged_params = Parameters(frame_params.fc, frame_params.Ej, frame_params.g, frame_params.Ec, frame_params.eps,\n frame_params.fd, frame_params.kappa, frame_params.gamma, frame_params.t_levels,\n frame_params.c_levels, frame_params.gamma_phi, kappa_phi, frame_params.n_t,\n frame_params.n_c)\n #directory = stack_directory + '/' + sys_params.group_folder + '/' + str(job_index)\n\n print(stack_directory)\n\n directory = stack_directory + '/' + sys_params.group_folder + '/' + str(sys_params.job_index)\n\n if not os.path.exists(directory):\n os.makedirs(directory)\n cwd = os.getcwd()\n os.chdir(directory)\n print(directory)\n sys_params.to_csv('settings.csv', header=False)\n\n H = hamiltonian(packaged_params)\n c_ops = collapse_operators(packaged_params)\n\n L = liouvillian(H, c_ops)\n data = L.data\n csc = data.tocsc()\n\n values, states = lin.eigs(csc, k=2, sigma=0)\n values = pd.DataFrame(values)\n values.columns = ['eigenvalues']\n states = pd.DataFrame(states)\n values.to_csv('eigenvalues.csv',index=False)\n states.to_csv('states.csv',index=False)\n\n os.chdir(cwd)\n", "repo_name": "paulsbrookes/bistability_tools", "sub_path": "cqed_lib/cqed_tools/simulation/liouvillian_sim.py", "file_name": "liouvillian_sim.py", "file_ext": "py", "file_size_in_byte": 1631, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "scipy.sparse.linalg.eigs", "line_number": 41, "usage_type": "call"}, {"api_name": "scipy.sparse.linalg", "line_number": 41, "usage_type": "name"}]}
+{"seq_id": "37902003689", "text": "import os\nfrom setuptools import setup\n\n# Utility function to read the README file.\n# Used for the long_description.\ndef read(fname):\n return open(os.path.join(os.path.dirname(__file__), fname)).read()\n\nsetup(\n name = \"mpsolver\",\n version = \"1.0.0\",\n author = \"Darius Braziunas\",\n author_email = \"darius.braziunas@gmail.com\",\n description = \"Mathematical programming (MP) interface to CPLEX and GLPK solvers\",\n long_description=read('README.txt'),\n #keywords = \"example documentation tutorial\",\n url = \"http://cs.toronto.edu/~darius\",\n packages=['mpsolver'],\n platforms = ['any'],\n install_requires =['numpy']\n)\n\n", "repo_name": "dariux/mpsolver", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 649, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "os.path.join", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 9, "usage_type": "call"}]}
+{"seq_id": "13971204172", "text": "from google.transit import gtfs_realtime_pb2\nimport requests\nimport math\nfrom datetime import datetime\nfrom pytz import timezone\nimport configparser\nimport operator\nfrom operator import itemgetter\n\nconfig = configparser.ConfigParser()\nconfig.read(\"nyc.ini\")\napikey = config[\"api\"][\"key\"]\n\ndef printRoute(id,name,time):\n if time == '0':\n print('\\x1b[0;30;41m ' + id + ' \\x1b[0m', '\\x1b[1;31;40m' + name + '\\x1b[0m', '\\x1b[1;31;40m' + time + '\\x1b[0m', '\\x1b[1;31;40m' + \"min\" + '\\x1b[0m', sep=\"\")\n else:\n print('\\x1b[0;30;41m ' + id + ' \\x1b[0m', '\\x1b[1;32;40m' + name + '\\x1b[0m', '\\x1b[1;32;40m' + time + '\\x1b[0m', '\\x1b[1;32;40m' + \"min\" + '\\x1b[0m', sep=\"\")\n\ndict = {}\ndictCount = 0\ndestinations = {'1': \" Van Cort Park 242 St \",'2':\" Wakefield-241 St \",'3':\" Harlem-148 St \"}\n\nfeed = gtfs_realtime_pb2.FeedMessage()\nresponse = requests.get(apikey)\nresponseString = response.content\nfeed.ParseFromString(responseString)\nif not responseString:\n print(\"MTA is Busy\")\nfor entity in feed.entity:\n if entity.HasField('trip_update'):\n tu = entity.trip_update\n tr = tu.trip\n rt = tr.route_id\n sc = len(tu.stop_time_update)\n count = 0\n if rt == '1' or rt == '2' or rt == '3':\n while count < sc:\n stu = tu.stop_time_update[count]\n sid = stu.stop_id\n arr = stu.arrival.time\n dep = stu.departure.time\n at = datetime.fromtimestamp(arr)\n dt = datetime.fromtimestamp(dep)\n now = datetime.now()\n tdelta = at - now\n seconds = tdelta.total_seconds()\n minutes = seconds / 60\n final = math.floor(minutes)\n if final < 0:\n final=1000\n count = count + 1\n if sid == '127N':\n if rt == '1':\n dict[dictCount] = final, \"1\"\n dictCount = dictCount + 1\n elif rt == '2':\n dict[dictCount] = final, \"2\"\n dictCount = dictCount + 1\n elif rt == '3':\n dict[dictCount] = final, \"3\"\n dictCount = dictCount + 1\n else:\n pass\n else:\n pass\ndictSorted = sorted(dict.items(), key=operator.itemgetter(1))\ntubListItem1 = dictSorted[0]\ntubListItem2 = dictSorted[1]\ntub1 = tubListItem1[1]\ntub2 = tubListItem2[1]\ntime1 = str(tub1[0])\ntime2 = str(tub2[0])\nnew_york = timezone('America/New_York')\nny_time = datetime.now(new_york)\n\n#Output\nprint (\"Times Sq-42 St -\", ny_time.strftime('%H:%M:%S'))\nprintRoute(tub1[1],destinations[tub1[1]],time1)\nprintRoute(tub2[1],destinations[tub2[1]],time2)\n", "repo_name": "simonbmadsen/settings", "sub_path": "nyc.py", "file_name": "nyc.py", "file_ext": "py", "file_size_in_byte": 2847, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "configparser.ConfigParser", "line_number": 10, "usage_type": "call"}, {"api_name": "google.transit.gtfs_realtime_pb2.FeedMessage", "line_number": 24, "usage_type": "call"}, {"api_name": "google.transit.gtfs_realtime_pb2", "line_number": 24, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 43, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 44, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 44, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 45, "usage_type": "name"}, {"api_name": "math.floor", "line_number": 49, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 67, "usage_type": "call"}, {"api_name": "pytz.timezone", "line_number": 74, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 75, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 75, "usage_type": "name"}]}
+{"seq_id": "5058239061", "text": "from django.shortcuts import render, redirect\nfrom django.http import HttpResponse, JsonResponse, HttpResponseRedirect\nfrom .models import Book\nfrom .forms import BookForm\nfrom django.views.decorators.csrf import csrf_exempt\n\ndef index(request):\n all_books = Book.objects.all()\n return render(request, 'bookstore_list.html', context={\"books\": all_books})\n\nbookstore_list = [\n {\n 'index': 0,\n 'id': 1,\n 'name': 'Harry potter',\n 'price': 100,\n 'description': \"Learning Learnin gJSfffjk dfjdklg jkdgjdkgjdkgjd kgjdkgjdglk jdgkljfjfjejekgjekgjekgjekgjgekjgekLearningJSfffjkdfjdklgjkdgjdkgjdkgjdkgjdkgjdglkjdgkljfjfjejekgjekgjekgjekgjgekjgekLearningJSfffjkdfjdklgjkdgjdkgjdkgjdkgjdkgjdglkjdgkljfjfjejekgjekgjekgjekgjgekjgek\",\n },\n {\n 'index': 1,\n 'id': 2,\n 'name': 'harry potter',\n 'price': 400,\n 'description': \"Learning LearningJS fffjkdfjd klgjkdgjdkgjdkgjdkgjd kgjdglkjdgk ljfj fjejekgjekgjekgjekgjgekjgek\",\n },\n {\n 'index': 2,\n 'id': 3,\n 'name': 'Harry potter 3',\n 'price': 200,\n 'description': \"LearningJS fffjk dfjdklgjkdg jdkgjdkgjd kgjdkgjdglkjdgkl jfjfjejekg jekgjekgjekgjgekjgek\",\n },\n]\n\n\ndef bookstore_details(request, *args, **kwrgs):\n book_id = kwrgs.get('book_id')\n book = Book.objects.get(pk=book_id)\n return render(request, 'bookstore_details.html', context={\"book\": book})\n\ndef bookstore_delete(request, **kwargs):\n book_id = kwargs.get('book_id')\n Book.objects.get(pk=book_id).delete()\n return redirect(\"bookstore:bookstore-list\")\n\ndef bookstore_update(request, **kwargs):\n book_id = kwargs.get('book_id') \n book = Book.objects.get(pk=book_id)\n form = BookForm(instance=book)\n if request.method == \"PUT\":\n form = BookForm(data=request.POST, instance=book)\n if form.is_valid():\n form.save()\n return redirect(\"bookstore:bookstore-details\", pk=book.id)\n \n return render(request, 'bookstore_update.html', context={\n 'form': form, \n 'book': book\n })\n\ndef create_new_book(request):\n form = BookForm()\n if request.method == \"POST\":\n form = BookForm(data=request.POST)\n if form.is_valid():\n form.save()\n return redirect(\"bookstore:bookstore-list\")\n return render(request, 'bookstore_create.html', context={\n 'form': form\n })", "repo_name": "radwanabil/Django", "sub_path": "bookstoreLab1/bookstore/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2407, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "models.Book.objects.all", "line_number": 8, "usage_type": "call"}, {"api_name": "models.Book.objects", "line_number": 8, "usage_type": "attribute"}, {"api_name": "models.Book", "line_number": 8, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 9, "usage_type": "call"}, {"api_name": "models.Book.objects.get", "line_number": 38, "usage_type": "call"}, {"api_name": "models.Book.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "models.Book", "line_number": 38, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 39, "usage_type": "call"}, {"api_name": "models.Book.objects.get", "line_number": 43, "usage_type": "call"}, {"api_name": "models.Book.objects", "line_number": 43, "usage_type": "attribute"}, {"api_name": "models.Book", "line_number": 43, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 44, "usage_type": "call"}, {"api_name": "models.Book.objects.get", "line_number": 48, "usage_type": "call"}, {"api_name": "models.Book.objects", "line_number": 48, "usage_type": "attribute"}, {"api_name": "models.Book", "line_number": 48, "usage_type": "name"}, {"api_name": "forms.BookForm", "line_number": 49, "usage_type": "call"}, {"api_name": "forms.BookForm", "line_number": 51, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 54, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 56, "usage_type": "call"}, {"api_name": "forms.BookForm", "line_number": 62, "usage_type": "call"}, {"api_name": "forms.BookForm", "line_number": 64, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 67, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 68, "usage_type": "call"}]}
+{"seq_id": "30245029166", "text": "import pandas as pd\nimport matplotlib.pyplot as plt\n\n# 用来正常显示中文标签\nplt.rcParams[\"font.sans-serif\"] = [\"SimHei\"]\n# 用来正常显示负号\nplt.rcParams[\"axes.unicode_minus\"] = False\n# 图片输出目录\nfs_images_dir = \"images/3-4/\"\n\n\n# 评分记录数据查看\ndef get_ratings_mess(file_path):\n print(\"文件路径:{}\".format(file_path))\n # drop 删除函数,这里删除第0行\n events = pd.read_table(file_path, header=0, sep=\"|\").drop(0)\n print(\"原始events的key为:\\n{}\".format(events.keys()))\n # print(events.columns.values)\n print(\"数据的前5条为:\\n{}\".format(events.head(5)))\n # 去空格\n events.columns = events.rename(columns=lambda x: x.strip()).keys()\n events = events.replace(' ', '')\n print(\"去除标题空格后,events的key为:\\n{}\".format(events.keys()))\n print(\"去掉空格后,数据的前5条为:\\n{}\".format(events.head(5)))\n # 因为原始数据的原因,按照|分割后,字段前后多了空格\n rate_ser = events[\"rating\"].groupby(events[\"rating\"]).count()\n print(\"events的值有:\\n{}\".format(rate_ser))\n\n plt.axes(aspect=1)\n plt.pie(x=rate_ser.values, labels=rate_ser.keys(), autopct=\"%3.1f %%\")\n plt.legend(bbox_to_anchor=(1.0, 1.0))\n plt.title(\"评分记录信息\")\n plt.savefig(fs_images_dir + \"ratings.png\")\n plt.show()\n\n\nif __name__ == '__main__':\n fs_file_path = \"data/foursquare-2013/ratings.dat\"\n get_ratings_mess(fs_file_path)\n", "repo_name": "legend1412/Recommend", "sub_path": "Chapter03/3-4.py", "file_name": "3-4.py", "file_ext": "py", "file_size_in_byte": 1464, "program_lang": "python", "lang": "zh", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "matplotlib.pyplot.rcParams", "line_number": 5, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 5, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 7, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 7, "usage_type": "name"}, {"api_name": "pandas.read_table", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pie", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}]}
+{"seq_id": "35468360686", "text": "import time\nimport tornado.web\nimport json\n\nfrom lib import history\n\n\nclass MainUIRequestHandler(tornado.web.RequestHandler):\n name = None\n data_queues = None\n worker_handle = None\n components = None\n config = None\n historic_task_list = None\n\n def initialize(self, data_queues, worker_handle, settings):\n self.name = 'main'\n self.data_queues = data_queues\n self.worker_handle = worker_handle\n self.components = []\n self.config = settings\n self.historic_task_list = []\n\n def get(self, path):\n if self.get_query_arguments('ajax'):\n # Print out the json based on the call\n self.handle_ajax_call(self.get_query_arguments('ajax')[0])\n else:\n self.get_historical_tasks()\n self.set_header(\"Content-Type\", \"text/html\")\n self.render(\"main/main.html\", historic_task_list=self.historic_task_list, time_now=time.time())\n\n def handle_ajax_call(self, query):\n self.set_header(\"Content-Type\", \"application/json\")\n if query == 'workersInfo':\n if self.get_query_arguments('workerId'):\n worker_info = self.get_workers_info(self.get_query_arguments('workerId')[0])\n self.set_header(\"Content-Type\", \"text/html\")\n self.render(\"main/main-worker-pie-chart.html\", worker_info=worker_info)\n else:\n self.write(json.dumps(self.get_workers_info()))\n if query == 'pendingTasks':\n if self.get_query_arguments('format') and 'html' in self.get_query_arguments('format'):\n self.set_header(\"Content-Type\", \"text/html\")\n self.render(\"main/main-pending-tasks.html\", time_now=time.time())\n else:\n self.write(json.dumps(self.get_pending_tasks()))\n if query == 'historicalTasks':\n if self.get_query_arguments('format') and 'html' in self.get_query_arguments('format'):\n self.get_historical_tasks()\n self.set_header(\"Content-Type\", \"text/html\")\n self.render(\"main/main-completed-tasks-list.html\", historic_task_list=self.historic_task_list,\n time_now=time.time())\n else:\n self.set_header(\"Content-Type\", \"application/json\")\n self.write(json.dumps(self.get_historical_tasks()))\n\n def get_workers_info(self, worker_id=None):\n if worker_id is not None:\n workers_info = self.worker_handle.get_worker_status(worker_id)\n else:\n workers_info = self.worker_handle.get_all_worker_status()\n return workers_info\n\n def get_workers_count(self):\n return len(self.worker_handle.get_all_worker_status())\n\n def get_pending_tasks(self):\n return self.worker_handle.job_queue.list_all_incoming_items()\n\n def get_historical_tasks(self):\n history_logging = history.History(self.config)\n self.historic_task_list = list(history_logging.get_historic_task_list(20))\n return self.historic_task_list\n", "repo_name": "BrianPugh/unmanic", "sub_path": "webserver/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3061, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "github-code", "pt": "25", "api": [{"api_name": "tornado.web.web", "line_number": 8, "usage_type": "attribute"}, {"api_name": "tornado.web", "line_number": 8, "usage_type": "name"}, {"api_name": "time.time", "line_number": 31, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 41, "usage_type": "call"}, {"api_name": "time.time", "line_number": 45, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 47, "usage_type": "call"}, {"api_name": "time.time", "line_number": 53, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 56, "usage_type": "call"}, {"api_name": "lib.history.History", "line_number": 72, "usage_type": "call"}, {"api_name": "lib.history", "line_number": 72, "usage_type": "name"}]}
+{"seq_id": "12828860704", "text": "import logging\n#from logging.handlers import RotatingFileHandler\nfrom recommender import config\n\ndef getLogger(obj):\n if not isinstance(obj, str):\n if hasattr(obj, '__name__'):\n obj = obj.__name__\n elif hasattr(obj, '__class__'):\n obj = obj.__class__.__name__\n else:\n obj = type(obj).__name__\n return configureLogger(logging.getLogger(obj))\n\ndef configureLogger(logger, level=config.LOG_LEVEL):\n if not len(logger.handlers):\n formatter = logging.Formatter(config.LOG_FORMAT)\n logger.setLevel(level)\n\n # Setup console logging\n ch = logging.StreamHandler()\n ch.setLevel(level)\n ch.setFormatter(formatter)\n logger.addHandler(ch)\n\n # # Setup file logging as well\n # fh = RotatingFileHandler(config.LOG_FILE, maxBytes=config.LOG_FILE_MAX_BYTES,\n # backupCount=config.LOG_FILE_BACKUP_COUNT)\n # fh.setLevel(level)\n # fh.setFormatter(formatter)\n # logger.addHandler(fh)\n\n return logger", "repo_name": "ScJa/projectr", "sub_path": "recommender/recommender/util/log.py", "file_name": "log.py", "file_ext": "py", "file_size_in_byte": 1059, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "27", "api": [{"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "recommender.config.LOG_LEVEL", "line_number": 15, "usage_type": "attribute"}, {"api_name": "recommender.config", "line_number": 15, "usage_type": "name"}, {"api_name": "logging.Formatter", "line_number": 17, "usage_type": "call"}, {"api_name": "recommender.config.LOG_FORMAT", "line_number": 17, "usage_type": "attribute"}, {"api_name": "recommender.config", "line_number": 17, "usage_type": "name"}, {"api_name": "logging.StreamHandler", "line_number": 21, "usage_type": "call"}]}
+{"seq_id": "30480180520", "text": "from django.test import TestCase\nfrom django.urls import reverse\nfrom django.contrib.auth.models import User\n\nfrom rest_framework.test import APIClient\nfrom rest_framework import status\n\nfrom articles.models import Author\n\n\nclass AuthorViewTestCase(TestCase):\n \"\"\"Test suite for the api endpoints related to authors.\"\"\"\n\n def setUp(self):\n \"\"\"Define the test client and other test variables.\"\"\"\n self.client = APIClient()\n\n self.user = User.objects.create(username='nerd', is_staff=True)\n self.author = Author.objects.create(name='Dennis Ritchie')\n\n self.epoint = reverse('author', kwargs={'pk': 1})\n\n\n def test_delete_author_unlogged(self):\n \"\"\"Test if an unlogged user can delete an author.\"\"\"\n request = self.client.delete(self.epoint)\n self.assertEqual(request.status_code, status.HTTP_403_FORBIDDEN)\n\n\n def test_add_author_unlogged(self):\n \"\"\"Test if an unlogged user can add an author.\"\"\"\n data = {\n 'name': 'Linus Torvalds'\n }\n\n request = self.client.post(reverse('author-create'), data)\n\n self.assertEqual(request.status_code, status.HTTP_403_FORBIDDEN)\n\n\n def test_update_author_unlogged(self):\n \"\"\"Test if an unlogged user can delete an author.\"\"\"\n data = {'name': 'Ken Thompson'}\n\n request = self.client.patch(self.epoint, data)\n\n self.assertEqual(request.status_code, status.HTTP_403_FORBIDDEN)\n\n\n def test_retrieve_author_unlogged(self):\n \"\"\"Test if an unlogged user can retrieve an author.\"\"\"\n request = self.client.get(self.epoint)\n\n self.assertEqual(request.status_code, status.HTTP_200_OK)\n\n\n def test_update_author_logged(self):\n \"\"\"Test if an logged user can delete an author.\"\"\"\n self.client.force_authenticate(user=self.user)\n\n data = {'name': 'Ken Thompson'}\n\n request = self.client.patch(self.epoint, data)\n\n self.assertEqual(request.status_code, status.HTTP_200_OK)\n\n\n def test_delete_author_logged(self):\n \"\"\"Test if an logged user can delete an author.\"\"\"\n self.client.force_authenticate(user=self.user)\n\n request = self.client.delete(self.epoint)\n self.assertEqual(request.status_code, status.HTTP_204_NO_CONTENT)\n\n\n def test_add_author_logged(self):\n \"\"\"Test if an logged user can add an author.\"\"\"\n self.client.force_authenticate(user=self.user)\n\n data = {\n 'name': 'Donald Knuth',\n }\n\n request = self.client.post(reverse('author-create'), data)\n\n self.assertEqual(request.status_code, status.HTTP_201_CREATED)\n", "repo_name": "murilocamargos/ckl-challenge", "sub_path": "articles/tests/views/author.py", "file_name": "author.py", "file_ext": "py", "file_size_in_byte": 2634, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.test.TestCase", "line_number": 11, "usage_type": "name"}, {"api_name": "rest_framework.test.APIClient", "line_number": 16, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.create", "line_number": 18, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 18, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 18, "usage_type": "name"}, {"api_name": "articles.models.Author.objects.create", "line_number": 19, "usage_type": "call"}, {"api_name": "articles.models.Author.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "articles.models.Author", "line_number": 19, "usage_type": "name"}, {"api_name": "django.urls.reverse", "line_number": 21, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_403_FORBIDDEN", "line_number": 27, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 27, "usage_type": "name"}, {"api_name": "django.urls.reverse", "line_number": 36, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_403_FORBIDDEN", "line_number": 38, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 38, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_403_FORBIDDEN", "line_number": 47, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 47, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 54, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 54, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 65, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 65, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_204_NO_CONTENT", "line_number": 73, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 73, "usage_type": "name"}, {"api_name": "django.urls.reverse", "line_number": 84, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 86, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 86, "usage_type": "name"}]}
+{"seq_id": "24291139185", "text": "import sys\nimport os\nimport glob\nimport shutil\nimport tarfile\nimport boto3\nimport threading\nfrom caldp import process\nfrom caldp import log\nfrom caldp import exit_codes\nfrom caldp import sysexit\n\n\ndef s3_split_uri(uri):\n \"\"\"\n >>> s3_split_uri('s3://the-bucket/prefix/parts/come/next')\n ('the-bucket', 'prefix/parts/come/next')\n\n >>> s3_split_uri('s3://the-bucket')\n ('the-bucket', '')\n \"\"\"\n parts = uri[5:].split(\"/\")\n bucket, prefix = parts[0], \"/\".join(parts[1:])\n # print(\"s3_split_uri:\", uri, \"->\", bucket, prefix)\n return bucket, prefix\n\n\ndef get_input_path(input_uri, dataset, make=False):\n \"\"\"Fetches the path to input files\"\"\"\n cwd = os.getcwd()\n if input_uri.startswith(\"file\"):\n input_path = input_uri.split(\":\")[-1]\n else:\n input_path = os.path.join(cwd, \"inputs\", dataset)\n if make is True:\n os.makedirs(input_path, exist_ok=True)\n return input_path\n\n\ndef get_output_dir(output_uri):\n \"\"\"Returns full path to output folder\"\"\"\n if output_uri.startswith(\"file\"):\n output_dir = output_uri.split(\":\")[-1]\n elif output_uri.startswith(\"s3\"):\n output_dir = os.path.abspath(\"outputs\")\n return output_dir\n\n\ndef get_input_dir(input_uri):\n if input_uri.startswith(\"file\"):\n input_dir = input_uri.split(\":\")[-1]\n else:\n input_dir = os.path.join(os.getcwd(), \"inputs\")\n return input_dir\n\n\ndef find_output_files(dataset):\n if process.IPPPSSOOT_RE.match(dataset):\n search_fits = f\"{dataset}/*.fits\"\n search_tra = f\"{dataset}/*.tra\"\n output_files = list(glob.glob(search_fits))\n output_files.extend(list(glob.glob(search_tra)))\n return output_files\n elif process.SVM_RE.match(dataset) and dataset.split(\"_\")[0] in list(process.SVM_INSTR.keys()):\n ipppss = process.get_svm_obs_set(dataset)\n search_fits = f\"{dataset}/hst_*{ipppss}*.fits\"\n search_txt = f\"{dataset}/hst_*{ipppss}*.txt\"\n search_ecsv = f\"{dataset}/hst_*{ipppss}*.ecsv\"\n search_manifest = f\"{dataset}/{dataset}_manifest.txt\"\n search_log = f\"{dataset}/astrodrizzle.log\"\n output_files = list(glob.glob(search_fits))\n output_files.extend(list(glob.glob(search_txt)))\n output_files.extend(list(glob.glob(search_ecsv)))\n output_files.extend(list(glob.glob(search_manifest)))\n output_files.extend(list(glob.glob(search_log)))\n return output_files\n elif process.MVM_RE.match(dataset):\n search_fits = f\"{dataset}/hst_{dataset}*.fits\"\n search_txt = f\"{dataset}/hst_{dataset}*.txt\"\n search_manifest = f\"{dataset}/{dataset}_manifest.txt\"\n output_files = list(glob.glob(search_fits))\n output_files.extend(list(glob.glob(search_txt)))\n output_files.extend(list(glob.glob(search_manifest)))\n return output_files\n else:\n raise ValueError(\"Invalid dataset name {dataset}, dataset must be an ipppssoot, SVM, or MVM dataset\")\n\n\ndef find_previews(dataset, output_files):\n search_prev = f\"{dataset}/previews/*\"\n output_files.extend(list(glob.glob(search_prev)))\n return output_files\n\n\ndef find_input_files(dataset):\n \"\"\"If job fails (no outputs), tar the input files instead for debugging purposes.\"\"\"\n search_inputs = f\"{dataset}/*\"\n file_list = list(glob.glob(search_inputs))\n return file_list\n\n\ndef make_tar(file_list, dataset):\n tar = dataset + \".tar.gz\"\n log.info(\"Creating tarfile: \", tar)\n if os.path.exists(tar):\n os.remove(tar) # clean up from prev attempts\n with tarfile.open(tar, \"x:gz\") as t:\n for f in file_list:\n print(os.path.basename(f))\n t.add(f)\n log.info(\"Tar successful: \", tar)\n tar_dest = os.path.join(dataset, tar)\n shutil.copy(tar, dataset) # move tarfile to outputs/{ipst}\n os.remove(tar)\n return tar_dest\n\n\ndef upload_tar(tar, output_path):\n with sysexit.exit_on_exception(exit_codes.S3_UPLOAD_ERROR, \"S3 tar upload of\", tar, \"to\", output_path, \"FAILED.\"):\n client = boto3.client(\"s3\")\n parts = output_path[5:].split(\"/\")\n bucket, prefix = parts[0], \"/\".join(parts[1:])\n objectname = prefix + \"/\" + os.path.basename(tar)\n log.info(f\"Uploading: s3://{bucket}/{objectname}\")\n if output_path.startswith(\"s3\"):\n with open(tar, \"rb\") as f:\n client.upload_fileobj(f, bucket, objectname, Callback=ProgressPercentage(tar))\n\n\nclass ProgressPercentage(object):\n def __init__(self, filename):\n self._filename = filename\n self._size = float(os.path.getsize(filename))\n self._seen_so_far = 0\n self._lock = threading.Lock()\n\n def __call__(self, bytes_amount):\n # To simplify, assume this is hooked up to a single filename\n with self._lock:\n self._seen_so_far += bytes_amount\n percentage = (self._seen_so_far / self._size) * 100\n sys.stdout.write(\"\\r%s %s / %s (%.2f%%)\" % (self._filename, self._seen_so_far, self._size, percentage))\n sys.stdout.flush()\n\n\ndef clean_up(file_list, dataset, dirs=None):\n print(\"\\nCleaning up...\")\n for f in file_list:\n try:\n os.remove(f)\n except FileNotFoundError:\n print(f\"file {f} not found\")\n if dirs is not None:\n for d in dirs:\n subdir = os.path.abspath(f\"{dataset}/{d}\")\n try:\n shutil.rmtree(subdir)\n except OSError:\n print(f\"dir {subdir} not found\")\n print(\"Done.\")\n\n\ndef tar_outputs(dataset, input_uri, output_uri):\n working_dir = os.getcwd()\n output_path = process.get_output_path(output_uri, dataset)\n output_dir = get_output_dir(output_uri)\n os.chdir(output_dir) # create tarfile with ipst/*fits (ipst is parent dir)\n output_files = find_output_files(dataset)\n if len(output_files) == 0:\n log.info(\"No output files found. Tarring inputs for debugging.\")\n os.chdir(working_dir)\n input_dir = get_input_dir(input_uri)\n os.chdir(input_dir)\n file_list = find_input_files(dataset)\n else:\n file_list = find_previews(dataset, output_files)\n tar = make_tar(file_list, dataset)\n upload_tar(tar, output_path)\n clean_up(file_list, dataset, dirs=[\"previews\", \"env\"])\n os.chdir(working_dir)\n return tar, file_list\n", "repo_name": "spacetelescope/caldp", "sub_path": "caldp/file_ops.py", "file_name": "file_ops.py", "file_ext": "py", "file_size_in_byte": 6351, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3, "dataset": "github-code", "pt": "27", "api": [{"api_name": "os.getcwd", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 53, "usage_type": "call"}, {"api_name": "caldp.process.IPPPSSOOT_RE.match", "line_number": 58, "usage_type": "call"}, {"api_name": "caldp.process.IPPPSSOOT_RE", "line_number": 58, "usage_type": "attribute"}, {"api_name": "caldp.process", "line_number": 58, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 61, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 62, "usage_type": "call"}, {"api_name": "caldp.process.SVM_RE.match", "line_number": 64, "usage_type": "call"}, {"api_name": "caldp.process.SVM_RE", "line_number": 64, "usage_type": "attribute"}, {"api_name": "caldp.process", "line_number": 64, "usage_type": "name"}, {"api_name": "caldp.process.SVM_INSTR.keys", "line_number": 64, "usage_type": "call"}, {"api_name": "caldp.process.SVM_INSTR", "line_number": 64, "usage_type": "attribute"}, {"api_name": "caldp.process.get_svm_obs_set", "line_number": 65, "usage_type": "call"}, {"api_name": "caldp.process", "line_number": 65, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 71, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 72, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 73, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 74, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 75, "usage_type": "call"}, {"api_name": "caldp.process.MVM_RE.match", "line_number": 77, "usage_type": "call"}, {"api_name": "caldp.process.MVM_RE", "line_number": 77, "usage_type": "attribute"}, {"api_name": "caldp.process", "line_number": 77, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 81, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 82, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 83, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 91, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 98, "usage_type": "call"}, {"api_name": "caldp.log.info", "line_number": 104, "usage_type": "call"}, {"api_name": "caldp.log", "line_number": 104, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 106, "usage_type": "call"}, {"api_name": "tarfile.open", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "caldp.log.info", "line_number": 111, "usage_type": "call"}, {"api_name": "caldp.log", "line_number": 111, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 113, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 114, "usage_type": "call"}, {"api_name": "caldp.sysexit.exit_on_exception", "line_number": 119, "usage_type": "call"}, {"api_name": "caldp.sysexit", "line_number": 119, "usage_type": "name"}, {"api_name": "caldp.exit_codes.S3_UPLOAD_ERROR", "line_number": 119, "usage_type": "attribute"}, {"api_name": "caldp.exit_codes", "line_number": 119, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "caldp.log.info", "line_number": 124, "usage_type": "call"}, {"api_name": "caldp.log", "line_number": 124, "usage_type": "name"}, {"api_name": "os.path.getsize", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path", "line_number": 133, "usage_type": "attribute"}, {"api_name": "threading.Lock", "line_number": 135, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 142, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 142, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 143, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 143, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 155, "usage_type": "call"}, {"api_name": "os.path", "line_number": 155, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 157, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 164, "usage_type": "call"}, {"api_name": "caldp.process.get_output_path", "line_number": 165, "usage_type": "call"}, {"api_name": "caldp.process", "line_number": 165, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 167, "usage_type": "call"}, {"api_name": "caldp.log.info", "line_number": 170, "usage_type": "call"}, {"api_name": "caldp.log", "line_number": 170, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 171, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 173, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 180, "usage_type": "call"}]}
+{"seq_id": "42484257292", "text": "import cv2\nimport numpy as np\nimport math\nimport config\n\nclass DrawAngle:\n def __init__(self, margin_=30, size_w_=300):\n # 初始化一个空画布 300×300 三通道 背景色为白色\n self.margin = margin_ # 上下左右边距\n self.size_w = size_w_\n self.radius = int((self.size_w - 2 * self.margin) / 2) # 圆的半径\n (self.center_x, self.center_y) = (int(self.radius + self.margin), int(self.radius + self.margin)) # 圆心\n self.center = (self.center_x, self.center_y)\n\n\n def draw_y(self, angle_y=80,b_show=False):\n \"\"\"\n 绘制横滚图\n :return:\n \"\"\"\n img = np.ones((self.size_w, self.size_w, 3), dtype=\"uint8\")\n # img *= 255\n # 蓝色底\n # rgb = [153, 194, 255]373d5f\n rgb = [55, 61, 95]\n img[:, :, 0] = np.squeeze(np.ones((self.size_w, self.size_w, 1), dtype=\"uint8\") * rgb[2])\n img[:, :, 1] = np.squeeze(np.ones((self.size_w, self.size_w, 1), dtype=\"uint8\") * rgb[1])\n img[:, :, 2] = np.squeeze(np.ones((self.size_w, self.size_w, 1), dtype=\"uint8\") * rgb[0])\n # 绘制一个绿色的圆\n cv2.circle(img, center=self.center, radius=self.radius, color=(255, 0, 0), thickness=1)\n # 绘制两条垂直辅助线\n cv2.line(img, (int(self.center_x - self.radius), int(self.center_y)),\n (int(self.center_x + self.radius), int(self.center_y)), (128, 128, 128),\n thickness=1)\n cv2.line(img, (int(self.center_x), int(self.center_y - self.radius)),\n (int(self.center_x), int(self.center_y + self.radius)), (128, 128, 128),\n thickness=1)\n\n # 绘制角度线\n x_0 = self.center_x - (self.radius - self.margin) * math.cos(angle_y * np.pi / 180.0)\n y_0 = self.center_y - (self.radius - self.margin) * math.sin(angle_y * np.pi / 180.0)\n x_1 = self.center_x + (self.radius - self.margin) * math.cos(angle_y * np.pi / 180.0)\n y_1 = self.center_y + (self.radius - self.margin) * math.sin(angle_y * np.pi / 180.0)\n cv2.line(img, (int(x_0), int(y_0)), (int(x_1), int(y_1)), (0, 255, 0), thickness=3)\n\n # 绘制角度线\n cv2.imwrite(config.save_angle_y_path, img)\n if b_show:\n cv2.imshow(\"circle\", img)\n\n cv2.waitKey(0)\n cv2.destroyAllWindows()\n\n def draw_z(self,angle_z,b_show=False):\n img = np.ones((self.size_w, self.size_w, 3), dtype=\"uint8\")\n # img *= 255\n # 蓝色底\n # rgb = [153, 194, 255]\n rgb = [55, 61, 95]\n img[:, :, 0] = np.squeeze(np.ones((self.size_w, self.size_w, 1), dtype=\"uint8\") * rgb[2])\n img[:, :, 1] = np.squeeze(np.ones((self.size_w, self.size_w, 1), dtype=\"uint8\") * rgb[1])\n img[:, :, 2] = np.squeeze(np.ones((self.size_w, self.size_w, 1), dtype=\"uint8\") * rgb[0])\n # 绘制一个绿色的圆\n cv2.circle(img, center=self.center, radius=self.radius, color=(0, 255, 0), thickness=1)\n pt1 = []\n\n # 3. 画出60条秒和分钟的刻线\n for i in range(60):\n # 最外部圆,计算A点\n x1 = self.center_x + (self.radius - 5) * math.cos(i * 6 * np.pi / 180.0)\n y1 = self.center_y + (self.radius - 5) * math.sin(i * 6 * np.pi / 180.0)\n pt1.append((int(x1), int(y1)))\n\n # 同心小圆,计算B点\n x2 = self.center_x + (self.radius - self.margin) * math.cos(i * 6 * np.pi / 180.0)\n y2 = self.center_y + (self.radius - self.margin) * math.sin(i * 6 * np.pi / 180.0)\n\n cv2.line(img, pt1[i], (int(x2), int(y2)), (0, 0, 0), thickness=1)\n\n # 4. 画出12条小时的刻线\n for i in range(12):\n # 12条小时刻线应该更长一点\n x = self.center_x + (self.radius - int(self.margin * 1.5)) * math.cos(i * 30 * np.pi / 180.0)\n y = self.center_y + (self.radius - int(self.margin * 1.5)) * math.sin(i * 30 * np.pi / 180.0)\n # 这里用到了前面的pt1\n cv2.line(img, pt1[i * 5], (int(x), int(y)), (0, 0, 0), thickness=2)\n\n # 5 绘制 角度 绘制文字\n for i in range(12):\n # 12条小时刻线应该更长一点\n if 0 <= i <= 3 or 9 <= i < 12:\n delta_margin = int(self.margin / 3)\n else:\n delta_margin = int(self.margin * 4 / 5)\n x = self.center_x + (self.radius + delta_margin) * math.cos(i * 30 * np.pi / 180.0)\n y = self.center_y + (self.radius + delta_margin) * math.sin(i * 30 * np.pi / 180.0)\n font = cv2.FONT_HERSHEY_SIMPLEX\n show_angle = str((i * 30 + 90) % 360)\n # show_angle = str(i * 30)\n cv2.putText(img, show_angle, (int(x), int(y)), font, 0.3, (0, 0, 0), 1)\n show_n = False\n if i == 0:\n show_n = True\n str_n = 'E'\n elif i == 3:\n show_n = True\n str_n = 'S'\n elif i == 6:\n show_n = True\n str_n = 'W'\n elif i == 9:\n show_n = True\n str_n = 'N'\n else:\n show_n = False\n str_n = ''\n if show_n:\n x_n = self.center_x + (self.radius - int(self.margin * 2)) * math.cos(i * 30 * np.pi / 180.0)\n y_n = self.center_y + (self.radius - int(self.margin * 2)) * math.sin(i * 30 * np.pi / 180.0)\n cv2.putText(img, str_n, (int(x_n), int(y_n)), font, 0.7, (0, 0, 0), 1)\n\n # 6 绘制角度线\n angle_z = angle_z - 90\n if angle_z < 0:\n angle_z = angle_z + 360\n x = self.center_x + (self.radius - self.margin * 2) * math.cos(angle_z * np.pi / 180.0)\n y = self.center_y + (self.radius - self.margin * 2) * math.sin(angle_z * np.pi / 180.0)\n # 这里用到了前面的pt1\n cv2.line(img, self.center, (int(x), int(y)), (0, 0, 255), thickness=3)\n # # 绘制箭头\n # if i==0:\n # x_0 = center_x + (radius + delta_margin) * math.cos(i * 30 * np.pi / 180.0)\n # y_0 = center_y + (radius + delta_margin) * math.sin(i * 30 * np.pi / 180.0)\n # points =\n # cv2.polylines(img=img, pts=[points], isClosed=True, color=(0, 0, 255), thickness=3)\n\n # 绘制角度线\n cv2.imwrite(config.save_angle_z_path, img)\n if b_show:\n cv2.imshow(\"circle\", img)\n cv2.waitKey(0)\n cv2.destroyAllWindows()\n\n\nif __name__ == '__main__':\n draw_angle_obj = DrawAngle(20,200)\n draw_angle_obj.draw_y(20,b_show=True)\n draw_angle_obj.draw_z(30,b_show=True)\n", "repo_name": "ndkjing/uuv", "sub_path": "common/draw_angle.py", "file_name": "draw_angle.py", "file_ext": "py", "file_size_in_byte": 6695, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "numpy.ones", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.line", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.line", "line_number": 35, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 40, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 41, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 42, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 43, "usage_type": "attribute"}, {"api_name": "cv2.line", "line_number": 44, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 47, "usage_type": "call"}, {"api_name": "config.save_angle_y_path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 62, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 64, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 70, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 71, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 75, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 76, "usage_type": "attribute"}, {"api_name": "cv2.line", "line_number": 78, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 83, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 84, "usage_type": "attribute"}, {"api_name": "cv2.line", "line_number": 86, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 95, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 96, "usage_type": "attribute"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 97, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 100, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 118, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 119, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 120, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 126, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 127, "usage_type": "attribute"}, {"api_name": "cv2.line", "line_number": 129, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 138, "usage_type": "call"}, {"api_name": "config.save_angle_z_path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 140, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 141, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 142, "usage_type": "call"}]}
+{"seq_id": "32612635360", "text": "import pandas as pd\nimport pathlib\nfrom enum import Enum\n\nclass FeatureType(Enum):\n CATEGORICAL = 1\n NUMERICAL = 2\n TIMESERIES = 3\n\n\n\ndef get_numeric_features(data, uniqueness_threshold=0.05):\n \"\"\"\n Given a pandas DataFrame, retuns the list of numeric columns from it.\n\n :param data: (pandas DataFrame), the target dataset\n :param uniqueness_threshold: (float), the uniqueness of occurance to treat numeric labeled feature to exclude from numeric\n :return: (list) list of numeric columns in the DataFrame\n \"\"\"\n\n num_cols = data._get_numeric_data().columns\n\n # finding out numeric labeled categorical features \n # Reference: https://stackoverflow.com/questions/35826912/what-is-a-good-heuristic-to-detect-if-a-column-in-a-pandas-dataframe-is-categori\n likely_categorical = []\n for var in num_cols:\n unique_values_ratio = 1.*data[var].nunique()/data[var].count()\n\n if unique_values_ratio < uniqueness_threshold:\n likely_categorical.append(var)\n\n\n return list(set(num_cols) - set(likely_categorical))\n\n\ndef get_categorical_features(data, uniqueness_threshold=0.05):\n \"\"\"\n Given a pandas DataFrame, retuns the list of categorical columns from it.\n\n :param data: (pandas DataFrame), the target dataset\n :param uniqueness_threshold: (float), the uniqueness of occurance to treat numeric labeled feature to exclude from numeric\n :return: (list) list of categorical columns in the DataFrame\n \"\"\"\n\n all_columns_set = set(data.columns)\n numeric_columns_set = set(get_numeric_features(data=data, uniqueness_threshold=uniqueness_threshold))\n\n return list(all_columns_set - numeric_columns_set)\n\n\ndef get_all_features(data):\n \"\"\"\n Given a Pandas DataFrame, returns all its column names.\n\n :param data: (Pandas DataFrame), the target DataFrame\n :return: (list) list of all the column names in the DataFrame\n \"\"\"\n\n return list(data.columns)\n\n\ndef get_feature_type(data, target_feature)->str:\n \"\"\" Returns the type of the supplied feature. Such as, Categorical, Neumerical, TimeSeries.\n\n Args:\n data (Pandas DataFrame): The target DataFrame \n target_feature (string, column name): the target feature to detect and return type of.\n\n Returns:\n str: \"\" \n \"\"\"\n\n if target_feature not in get_all_features(data):\n return None\n\n if target_feature in get_numeric_features(data):\n return FeatureType.NUMERICAL\n elif target_feature in get_categorical_features(data):\n return FeatureType.CATEGORICAL\n \n ## TODO: Add Time series data type\n\n return None\n\n\n\nif __name__ == '__main__':\n WEBAPP_DIR = str(pathlib.Path(__file__).parent.parent.parent.absolute()) + \"/webapp/\"\n\n df = pd.read_csv(WEBAPP_DIR+\"/datasets/titanic.csv\")\n\n print(get_numeric_features(df))\n print(get_categorical_features(df))\n\n print(get_feature_type(df, \"sex\") == get_feature_type(df, \"survived\"))\n\n print(FeatureType.CATEGORICAL == FeatureType.NUMERICAL)\n\n # n = get_all_features(data=df)\n\n # dataType = get_feature_type(data=df, target_feature=\"sex\")\n\n # print(dataType)\n", "repo_name": "pseudoPixels/vizAI", "sub_path": "graphaite/core/utils/dataFrameUtils.py", "file_name": "dataFrameUtils.py", "file_ext": "py", "file_size_in_byte": 3137, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "enum.Enum", "line_number": 5, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 88, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 90, "usage_type": "call"}]}
+{"seq_id": "40243842182", "text": "import streamlit as st\nimport pandas as pd\nimport time\nfrom DAO import *\nfrom model import *\nimport pdfkit as pdf\nimport matplotlib.pyplot as plt\n\n# Função para gerar o relatório a partir dos critérios selecionados\ndef geraRelatorio():\n if not relatorio_fields:\n return -1\n \n # Verifica se foi adicionado algum filtro \n if st.session_state.filtros == \"\":\n st.session_state.filtros = None\n \n session = DAO.getSession()\n session.expire_on_commit = False\n\n # Verifica sobre qual tabela do banco será emitido um relatório\n if relatorio_type == 'Videos':\n query = DAORelatorioVideos.select(session, st.session_state.filtros, st.session_state.ordenacao, relatorio_fields)\n if relatorio_type == 'Streams':\n query = DAORelatorioStreams.select(session, st.session_state.filtros, st.session_state.ordenacao, relatorio_fields)\n if relatorio_type == 'Canais':\n query = DAORelatorioCanais.select(session, st.session_state.filtros, st.session_state.ordenacao, relatorio_fields)\n if relatorio_type == 'Usuários':\n query = DAORelatorioUsuarios.select(session, st.session_state.filtros, st.session_state.ordenacao, relatorio_fields)\n if relatorio_type == 'Categorias':\n query = DAORelatorioCategories.select(session, st.session_state.filtros, st.session_state.ordenacao, relatorio_fields)\n\n # Conexão\n connection = session.connection()\n df = pd.read_sql_query(query.statement, con = connection)\n session.commit()\n session.close()\n\n # Fazer com que o dataframe receba os dados da sql query recebida na linha 34\n st.session_state.dataframe = df\n\n # Convertendo o dataframe do relatorio para excel e html\n df.to_excel(\"DB/relatorio.xlsx\", index=False)\n df.to_html('DB/relatorio.html', index=False)\n\n# Função para limpar o campo do input do valor do filtro\ndef clear_form():\n st.session_state[\"bar\"] = \"\"\n\n# Define os campos que podem ser incluídos nos filtros e no relatório\ndef defineCampos():\n st.session_state.filtros = \"\"\n st.session_state.query = False\n st.session_state.qtdFiltros = 0\n\ndef defineOrdenacao():\n st.session_state.ordenacao = True\n\n# Converter o relatório para pdf\ndef relatorioPDF():\n path_to_wkhtmltopdf = r'C:\\Program Files\\wkhtmltopdf\\bin\\wkhtmltopdf.exe'\n config = pdf.configuration(wkhtmltopdf=path_to_wkhtmltopdf)\n pdf.from_file('DB/relatorio.html', 'relatorio.pdf', configuration=config)\n\n# Aqui está sendo mostrado o título da página (na aba do navegador)\nst.set_page_config(page_title=\"Relatório Twitch API\")\n\n# Estilização da página\napp_style = \"\"\"\n \n \"\"\"\nst.markdown(app_style, unsafe_allow_html=True) \n\n# Barra lateral\nst.sidebar.header(\"Twitch API\")\nst.sidebar.image('./img/twitch.png')\nst.sidebar.write(\"\\n\")\nst.sidebar.header(\"Desenvolvido por:\")\nst.sidebar.markdown('''''', unsafe_allow_html=True)\nst.sidebar.write(\"Guilherme Ribeiro\")\nst.sidebar.write(\"Tales Oliveira\")\nst.sidebar.markdown('''''', unsafe_allow_html=True)\n\n# Título da página\nst.title('Relatórios Twitch API\\n')\n\n# ===================================================================\n# Se pagina=0, mostra a página inicial. Se pagina=1, mostra o relatório\nif 'pagina' not in st.session_state:\n st.session_state.pagina = 0\n\n# Para exibir o dataframe\nif 'dataframe' not in st.session_state:\n st.session_state.dataframe = False\n\n# Dados do dataframe do relatório\nif 'df' not in st.session_state:\n st.session_state.df = False\n\n# Filtros\nif 'filtros' not in st.session_state:\n st.session_state.filtros = \"\"\n\n# Conta a quantidade de filtros\nif 'qtdFiltros' not in st.session_state:\n st.session_state.qtdFiltros = 0\n\n# Ordenação\nif 'ordenacao' not in st.session_state:\n st.session_state.ordenacao = None\n\n# Query SQL a ser executada\nif 'query' not in st.session_state:\n st.session_state.query = False\n\n# Relatóio\nif 'relatorio' not in st.session_state:\n st.session_state.relatorio = False\n\n# Dados em PDF\nif 'dataPdf' not in st.session_state:\n st.session_state.dataPdf = None\n\n# Dados em XLSX\nif 'dataXlsx' not in st.session_state:\n st.session_state.dataXlsx = None\n\n# ==============================================================================================\n# Página inicial, onde são selecionados os campos, filtros e ordenação para gerar o relatório\nif st.session_state.pagina == 0:\n relatorio_type = st.selectbox(\n 'Selecione a tabela para gerar o relatório:',\n ('Videos', 'Streams', 'Canais', 'Usuários', 'Categorias'), on_change=defineCampos)\n\n session = DAO.getSession()\n session.expire_on_commit = False\n\n if st.session_state.query is False:\n if relatorio_type == 'Videos':\n query = DAORelatorioVideos.select(session, None, None, None) \n elif relatorio_type == 'Streams':\n query = DAORelatorioStreams.select(session, None, None, None)\n elif relatorio_type == 'Canais':\n query = DAORelatorioCanais.select(session, None, None, None)\n elif relatorio_type == 'Usuários':\n query = DAORelatorioUsuarios.select(session, None, None, None)\n elif relatorio_type == 'Categorias':\n query = DAORelatorioCategories.select(session, None, None, None)\n\n st.session_state.df = pd.read_sql_query(query.statement, con=session.bind)\n session.commit()\n session.close()\n st.session_state.query = True\n\n # Selecionar os campos das tabelas\n relatorio_fields = st.multiselect(f'Selecione os campos do relatório de {relatorio_type}:', options = st.session_state.df.columns, placeholder = 'Selecionar campo')\n \n # Formulário dos filtros\n st.write('\\n')\n st.write(\"Filtrar por campos:\")\n with st.form(\"myform\"):\n f1, f2, f3 = st.columns([1, 1, 1])\n with f1:\n field = st.selectbox(\"Campo:\", options = st.session_state.df.columns)\n with f2:\n comparison = st.selectbox(\"Comparação:\", options = ('igual', 'maior', 'menor', 'maior ou igual', 'menor ou igual', 'diferente de', 'contendo a string'))\n with f3:\n comparison_value = st.text_input(\"Valor\")\n\n f1, f2, f3 = st.columns([1, 1, 1])\n \n with f2:\n st.write('\\n')\n submit = st.form_submit_button(label=\"Adicionar filtro\", on_click=clear_form)\n\n # Tipo de comparação a ser feita\n if submit and comparison_value:\n map_operation = {\n 'igual': f'= ',\n 'maior': f'> ',\n 'menor': f'< ',\n 'maior ou igual': f'>= ',\n 'menor ou igual': f'<= ',\n 'diferente de': f'!= ',\n 'contendo a string': 'LIKE \\'%'\n }\n\n operation = map_operation[comparison]\n \n if st.session_state.df[f'{field}'].dtypes == 'object' and not comparison == 'contendo a string':\n value = f\"'{comparison_value}'\"\n elif comparison == 'contendo a string':\n value = f\"{comparison_value}\"\n else:\n value = f'{comparison_value}'\n\n # Adiciona os filtros\n if st.session_state.qtdFiltros == 0:\n st.session_state.filtros += f\"{field} {operation}{value}\"\n else:\n st.session_state.filtros += f\" AND {field} {operation}{value}\"\n\n # Se a opção \"contendo a string\" for utilizada\n if comparison == 'contendo a string':\n st.session_state.filtros += '%\\''\n\n # Adiciona filtro\n st.session_state.qtdFiltros = 1\n container = st.empty()\n container.success('Filtro adicionado com sucesso!') \n time.sleep(3) \n container.empty() \n\n # Caso o valor a ser comparado com o campo não seja preenchido\n if submit and not comparison_value: \n container = st.empty()\n container.error('Preencha o valor da comparação!') \n time.sleep(3) \n container.empty() \n\n st.write('\\n\\n\\n\\n\\n\\n')\n st.write('Filtros adicionados:')\n with st.expander(\" \"):\n st.write(st.session_state.filtros)\n\n\n # Formulário de ordenação do relatório\n st.write('\\n')\n st.write(\"Ordenar relatório:\")\n with st.form(\"myform2\"):\n o1, o2 = st.columns([1.5, 1.5])\n with o1:\n camposOrdenacao = st.selectbox('Ordenar por:', options = st.session_state.df.columns)\n with o2:\n tipoOrdenacao = st.selectbox(f'Campo {camposOrdenacao} ordenado de modo:', options=('Crescente', 'Decrescente'), index=0)\n #tipoOrdenacao = st.radio(f'Campo {camposOrdenacao} ordenado de modo:', options = ('Crescente', 'Decrescente'), horizontal = True)\n \n st.write('\\n\\n\\n\\n\\n')\n o1, o2 = st.columns([1, 1])\n \n st.write('\\n')\n ordenacao_relatorio = st.form_submit_button(label='Ordenar relatório', on_click=defineOrdenacao)\n\n if ordenacao_relatorio == 'Crescente':\n ordenacao = 'ASC'\n else:\n ordenacao = 'DESC'\n \n if ordenacao_relatorio:\n st.session_state.ordenacao = f'{camposOrdenacao} {ordenacao}'\n container = st.empty()\n container.success(f'Relatório será ordenado pelo campo {camposOrdenacao} de forma {tipoOrdenacao}!') \n time.sleep(3) \n container.empty() \n\n st.write('\\n\\n\\n')\n f1, f2, f3 = st.columns([1, 1, 1])\n\n with f2:\n st.write('\\n\\n\\n\\n\\n')\n relatorio = st.button('Gerar relatório')\n\n st.write('\\n\\n\\n\\n\\n')\n\n if relatorio:\n status = geraRelatorio()\n if status == -1:\n st.error(\"Selecione os campos do relatório!\")\n else:\n st.session_state.pagina = 1\n st.rerun()\n\n# ==============================================================================================\n# Página do relatório\nelse:\n if st.session_state.relatorio == False:\n # Gerar pdf do relatório\n relatorioPDF()\n with open(\"relatorio.pdf\", \"rb\") as pdf_file:\n st.session_state.pdfData = pdf_file.read()\n\n # Gerar xlsx do relatório\n with open(\"DB/relatorio.xlsx\", \"rb\") as xlsx_file:\n st.session_state.xlsxData = xlsx_file.read()\n\n # Exibe dataframe\n st.session_state.relatorio = st.dataframe(st.session_state.dataframe, width=1000, height=500)\n\n f1, f2, f3 = st.columns([1, 1, 1])\n st.write('\\n')\n\n with f1:\n relatorio_pdf = st.download_button('Exportar relátorio para PDF', data = st.session_state.pdfData,\n file_name=\"relatorio.pdf\")\n\n with f2:\n relatorio_xlsx= st.download_button('Exportar relátorio para XLSX', data = st.session_state.xlsxData,\n file_name=\"relatorio.xlsx\")\n\n with f3:\n new_relatorios = st.button(\"Criar mais relatórios\")\n \n # Caso seja selecionada a opção de gerar mais relatórios\n if new_relatorios:\n st.session_state.pagina = 0\n st.session_state.qtdFiltros = 0\n st.session_state.filtros = \"\"\n st.session_state.ordenacao = None\n st.session_state.relatorio = False\n relatorio_fields = []\n st.rerun()\n\n with f1:\n relatorio_type = st.selectbox(\n 'Selecione o relatório:',\n ('Videos', 'Streams', 'Canais', 'Usuários', 'Categorias'), on_change=defineCampos)\n\n with f2:\n st.session_state.relatorio_fields = st.multiselect(f'Selecione os campos do relatório de {relatorio_type}:', options=st.session_state.df.columns)\n\n# ==============================================================================================\n# Lógica para o gráfico\n if st.session_state.relatorio and 'relatorio_fields' in st.session_state and len(st.session_state.relatorio_fields) >= 2:\n x_column = st.session_state.relatorio_fields[0]\n y_column = st.session_state.relatorio_fields[1]\n grouped_data = st.session_state.df.groupby(x_column)[y_column].count().reset_index()\n\n st.write('\\n\\n')\n st.write(f\"Gráfico de contagem de {y_column} agrupado por {x_column}:\")\n\n # Use st.pyplot() para exibir a figura Matplotlib\n fig, ax = plt.subplots()\n grouped_data.plot(kind='bar', x=x_column, y=y_column, ax=ax)\n st.pyplot(fig)\n", "repo_name": "Gksnoda/API-twitch", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 15456, "program_lang": "python", "lang": "pt", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "streamlit.session_state", "line_number": 15, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 16, "usage_type": "attribute"}, {"api_name": "DAO.getSession", "line_number": 18, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 23, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 25, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 27, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 29, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 31, "usage_type": "attribute"}, {"api_name": "pandas.read_sql_query", "line_number": 35, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 40, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 48, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 52, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 53, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 54, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 57, "usage_type": "attribute"}, {"api_name": "pdfkit.configuration", "line_number": 62, "usage_type": "call"}, {"api_name": "pdfkit.from_file", "line_number": 63, "usage_type": "call"}, {"api_name": "streamlit.set_page_config", "line_number": 66, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 184, "usage_type": "call"}, {"api_name": "streamlit.sidebar.header", "line_number": 187, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 187, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.image", "line_number": 188, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 188, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.write", "line_number": 189, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 189, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.header", "line_number": 190, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 190, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.markdown", "line_number": 191, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 191, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.write", "line_number": 192, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 192, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.write", "line_number": 193, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 193, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.markdown", "line_number": 194, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 194, "usage_type": "attribute"}, {"api_name": "streamlit.title", "line_number": 197, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 201, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 202, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 205, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 206, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 209, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 210, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 213, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 214, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 217, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 218, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 221, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 222, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 225, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 226, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 229, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 230, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 233, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 234, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 237, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 238, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 242, "usage_type": "attribute"}, {"api_name": "streamlit.selectbox", "line_number": 243, "usage_type": "call"}, {"api_name": "DAO.getSession", "line_number": 247, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 250, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 262, "usage_type": "attribute"}, {"api_name": "pandas.read_sql_query", "line_number": 262, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 265, "usage_type": "attribute"}, {"api_name": "streamlit.multiselect", "line_number": 268, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 268, "usage_type": "attribute"}, {"api_name": "streamlit.write", "line_number": 271, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 272, "usage_type": "call"}, {"api_name": "streamlit.form", "line_number": 273, "usage_type": "call"}, {"api_name": "streamlit.columns", "line_number": 274, "usage_type": "call"}, {"api_name": "streamlit.selectbox", "line_number": 276, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 276, "usage_type": "attribute"}, {"api_name": "streamlit.selectbox", "line_number": 278, "usage_type": "call"}, {"api_name": "streamlit.text_input", "line_number": 280, "usage_type": "call"}, {"api_name": "streamlit.columns", "line_number": 282, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 285, "usage_type": "call"}, {"api_name": "streamlit.form_submit_button", "line_number": 286, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 302, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 310, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 311, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 313, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 317, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 320, "usage_type": "attribute"}, {"api_name": "streamlit.empty", "line_number": 321, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 323, "usage_type": "call"}, {"api_name": "streamlit.empty", "line_number": 328, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 330, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 333, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 334, "usage_type": "call"}, {"api_name": "streamlit.expander", "line_number": 335, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 336, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 336, "usage_type": "attribute"}, {"api_name": "streamlit.write", "line_number": 340, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 341, "usage_type": "call"}, {"api_name": "streamlit.form", "line_number": 342, "usage_type": "call"}, {"api_name": "streamlit.columns", "line_number": 343, "usage_type": "call"}, {"api_name": "streamlit.selectbox", "line_number": 345, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 345, "usage_type": "attribute"}, {"api_name": "streamlit.selectbox", "line_number": 347, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 350, "usage_type": "call"}, {"api_name": "streamlit.columns", "line_number": 351, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 353, "usage_type": "call"}, {"api_name": "streamlit.form_submit_button", "line_number": 354, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 362, "usage_type": "attribute"}, {"api_name": "streamlit.empty", "line_number": 363, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 365, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 368, "usage_type": "call"}, {"api_name": "streamlit.columns", "line_number": 369, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 372, "usage_type": "call"}, {"api_name": "streamlit.button", "line_number": 373, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 375, "usage_type": "call"}, {"api_name": "streamlit.error", "line_number": 380, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 382, "usage_type": "attribute"}, {"api_name": "streamlit.rerun", "line_number": 383, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 388, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 392, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 396, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 399, "usage_type": "attribute"}, {"api_name": "streamlit.dataframe", "line_number": 399, "usage_type": "call"}, {"api_name": "streamlit.columns", "line_number": 401, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 402, "usage_type": "call"}, {"api_name": "streamlit.download_button", "line_number": 405, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 405, "usage_type": "attribute"}, {"api_name": "streamlit.download_button", "line_number": 409, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 409, "usage_type": "attribute"}, {"api_name": "streamlit.button", "line_number": 413, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 417, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 418, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 419, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 420, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 421, "usage_type": "attribute"}, {"api_name": "streamlit.rerun", "line_number": 423, "usage_type": "call"}, {"api_name": "streamlit.selectbox", "line_number": 426, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 431, "usage_type": "attribute"}, {"api_name": "streamlit.multiselect", "line_number": 431, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 435, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 436, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 437, "usage_type": "attribute"}, {"api_name": "streamlit.session_state.df.groupby", "line_number": 438, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 438, "usage_type": "attribute"}, {"api_name": "streamlit.write", "line_number": 440, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 441, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 444, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 444, "usage_type": "name"}, {"api_name": "streamlit.pyplot", "line_number": 446, "usage_type": "call"}]}
+{"seq_id": "71118427913", "text": "from datetime import timedelta\nfrom typing import List, Tuple\n\n\ndef humanize_timedelta(timedelta: timedelta) -> str:\n result: List[str] = []\n\n days = timedelta.days\n mm, ss = divmod(timedelta.seconds, 60)\n hh, mm = divmod(mm, 60)\n\n def plural(n: int) -> Tuple[int, str]:\n return n, \"s\" if abs(n) != 1 else \"\"\n\n if days > 0:\n result.append(\"%d day%s\" % plural(days))\n if hh > 0 or result:\n result.append(\"%d hour%s\" % plural(hh))\n if mm > 0 or result:\n result.append(\"%d min%s\" % plural(mm))\n if len(result) <= 1:\n result.append(\"%d sec%s\" % plural(ss))\n\n return \", \".join(result)\n\n\ndef humanize_bytes(bytes: int) -> str:\n units = [\"B\", \"kB\", \"MB\", \"GB\"]\n\n factor = 1\n unit = \"\"\n for unit in units:\n next_factor = factor << 10\n if bytes < next_factor:\n break\n factor = next_factor\n\n return \"%.2f %s\" % (float(bytes) / factor, unit)\n", "repo_name": "dmitmel/dotfiles", "sub_path": "script-resources/welcome/humanize.py", "file_name": "humanize.py", "file_ext": "py", "file_size_in_byte": 875, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 66, "dataset": "github-code", "pt": "27", "api": [{"api_name": "datetime.timedelta", "line_number": 5, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 6, "usage_type": "name"}, {"api_name": "datetime.timedelta.days", "line_number": 8, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 8, "usage_type": "name"}, {"api_name": "datetime.timedelta.seconds", "line_number": 9, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 9, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 12, "usage_type": "name"}]}
+{"seq_id": "28771183549", "text": "import os\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom shared import database\nfrom shared import printlog\nfrom datetime import datetime\nimport matplotlib.dates as mdates\n\n\nclass Consumption:\n \"\"\" plotter implementation \"\"\"\n\n def __init__(self, from_date, to_date,\n filename='hourly.png',\n printer=printlog.PrintLog()):\n \"\"\" run plotter once \"\"\"\n\n # prefix filename with date\n filename = from_date[0:10] + '-' + filename\n\n # open database\n db = database.Database(printer)\n\n # create x axis as numpy datetime objects\n xlist = db.get_hours(from_date, to_date)\n xl = []\n for s in xlist:\n xl.append(datetime.strptime(s, '%Y-%m-%d %H:%M:%S'))\n x = np.array(xl)\n\n # create & decorate canvas\n fig, ax = plt.subplots(figsize=(11.6, 8.2)) # object oriented IF\n ax.set_xlabel('Datetime')\n ax.set_ylabel('kg (pellets)')\n ax.set_title(filename) # title is filename\n ax.grid(linestyle='dotted', linewidth='0.2', color='grey') # grid\n\n # create hourly chart\n ylist = db.get_hourly_consumption(from_date, to_date)\n y = np.array(ylist)\n charttype = 'bar'\n if charttype == 'line':\n hrs = mdates.HourLocator() # every hour\n ax.xaxis.set_minor_locator(hrs) # minor ticks\n ax.plot(x, y, label='kgh (pellets)', color='brown', linestyle='solid')\n ax.legend()\n #\n elif charttype == 'bar':\n # render xaxis labels\n plt.xticks(rotation=90)\n labels = []\n for s in xlist:\n labels.append(s[:13])\n # render bar chart\n ax.bar(np.array(labels, dtype=object), y)\n # add statistics to bar chart\n kgs = 0\n for y in ylist:\n kgs += y\n hrs = len(ylist)\n avg = round(kgs / hrs * 24, 1)\n stats = ' (' + str(kgs) + ' kg in ' + str(hrs) + ' hr, avg= ' + str(avg) + ' kg/day)'\n ax.set_title(filename + stats)\n ax.axhline(kgs / hrs, color='red', linewidth=2)\n #\n elif charttype == 'step':\n # see matplotlib.pyplot.step\n raise Exception('Step chart not implemented yet')\n #\n # create output file\n printer.print('Created file: ' + filename)\n fig.savefig(printer.get_foldername() + printer.get_slash() + filename)\n fig.show()\n\n\nif __name__ == '__main__':\n print('So sorry, the ' + os.path.basename(__file__) + ' module does not run as a standalone.')\n", "repo_name": "majo48/connect-web-logger", "sub_path": "plotter/consumption.py", "file_name": "consumption.py", "file_ext": "py", "file_size_in_byte": 2631, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 6, "dataset": "github-code", "pt": "27", "api": [{"api_name": "shared.printlog.PrintLog", "line_number": 15, "usage_type": "call"}, {"api_name": "shared.printlog", "line_number": 15, "usage_type": "name"}, {"api_name": "shared.database.Database", "line_number": 22, "usage_type": "call"}, {"api_name": "shared.database", "line_number": 22, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.dates.HourLocator", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}]}
+{"seq_id": "41963925015", "text": "'''\r\n1. 이 성에 있는 방의 개수\r\n2. 가장 넓은 방의 넓이\r\n3. 하나의 벽을 제거하여 얻을 수 있는 가장 넓은 방 크기\r\n\r\nN M\r\nM 개 줄 정수로 벽에 대한 정보\r\n\r\n서쪽 벽 1 L\r\n북쪽 벽 2 U\r\n동쪽 벽 4 R\r\n남쪽 벽 8 D\r\n1 \r\n2\r\n4 \r\n8\r\n1 2 => 3\r\n1 4 => 5\r\n1 8 => 9\r\n2 4 => 6\r\n2 8 => 10\r\n4 8 => 12\r\n1 2 4 => 7\r\n1 2 8 => 11\r\n1 4 8 => 13\r\n2 4 8 => 14\r\n1 2 4 8 => 15\r\n# 테두리에 위치한 노드는 벽이 있다\r\n'''\r\nfrom collections import deque\r\nM,N = map(int,input().split())\r\ngrid = [list(map(int,input().split())) for _ in range(N)]\r\nvisited = [[-1 for _ in range(M)] for _ in range(N)]\r\nque = deque([])\r\ndic = { # 갈 수 있는 곳 상하좌우 북2 /남8 /서1/ 동4 # 15 #1 2 4 8 서 북 동 남 / 좌 상 우 하\r\n '0' : [1,1,1,1],\r\n '1' : [0,1,1,1],\r\n '2' : [1,0,1,1],\r\n '4' : [1,1,0,1],\r\n '8' : [1,1,1,0],\r\n '3' : [0,0,1,1],\r\n '5' : [0,1,0,1],\r\n '9' : [0,1,1,0],\r\n '6' : [1,0,0,1],\r\n '10': [1,0,1,0],\r\n '12': [1,1,0,0],\r\n '7' : [0,0,0,1],\r\n '11' : [0,0,1,0],\r\n '13' : [0,1,0,0],\r\n '14' : [1,0,0,0],\r\n '15' : [0,0,0,0]\r\n}\r\n\r\ndef in_range(x,y):\r\n return 0<= x < N and 0<=y 30:\n pyautogui.click(screenWidth - center[0]*widthScalar, center[1]*heightScalar)\n #cv2.circle(frame, center, radius, (255, 0, 255), 3)\n\n #cv2.imshow(\"detected circles\", frame)\n if cv2.waitKey(1) & 0xFF == ord('q'):\n break\n # When everything done, release the capture\n cap.release()\n cv2.destroyAllWindows()\n return 0\n\n\nif __name__ == \"__main__\":\n main(sys.argv[1:])\n", "repo_name": "SunnyHarjani/fingerStylus", "sub_path": "fingerStylus.py", "file_name": "fingerStylus.py", "file_ext": "py", "file_size_in_byte": 2040, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "cv2.VideoCapture", "line_number": 9, "usage_type": "call"}, {"api_name": "pyautogui.size", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.HoughCircles", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.HOUGH_GRADIENT", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.uint16", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 35, "usage_type": "call"}, {"api_name": "pyautogui.moveTo", "line_number": 42, "usage_type": "call"}, {"api_name": "pyautogui.click", "line_number": 48, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 56, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 61, "usage_type": "attribute"}]}
+{"seq_id": "30302242595", "text": "from functools import wraps\nfrom inspect import Parameter, Signature\nfrom typing import Any, Callable, List, Optional\n\nfrom fastapi import Query\nfrom makefun import with_signature\nfrom pydantic import NonNegativeInt, conint\n\nfrom fastapi_rest_framework.errors import BizError, ErrorCode\nfrom fastapi_rest_framework.schemas import FilterQuery, OrderingQuery, PaginationQuery\n\n\ndef parse_pagination_query(\n max_limit: Optional[NonNegativeInt] = 500, default_limit: NonNegativeInt = 10\n) -> Callable[..., PaginationQuery]:\n LimitType = conint(ge=0, le=max_limit)\n\n @wraps\n def wrapped(\n offset: NonNegativeInt = Query(0),\n limit: LimitType = Query(default_limit), # type: ignore\n ) -> PaginationQuery:\n return PaginationQuery(offset=offset, limit=NonNegativeInt(limit))\n\n return wrapped\n\n\ndef parse_ordering_query(\n fields: Optional[List[str]] = None,\n) -> Callable[..., OrderingQuery]:\n if fields is not None:\n fields_set = set(fields)\n\n @wraps\n def wrapped(\n ordering: str = Query(\n \"\",\n description=\"Comma seperated list of ordering the results.\\n\"\n \"You may also specify reverse orderings by prefixing the field name with '-'.\",\n example=\"-updated_at\",\n ),\n ) -> OrderingQuery:\n orderings = list(filter(None, ordering.split(\",\")))\n if fields is not None:\n for x in orderings:\n name = x.startswith(\"-\") and x[1:] or x\n if name not in fields_set:\n raise BizError(\n ErrorCode.IllegalFieldError,\n f\"{x} is not available in ordering fields\",\n )\n return OrderingQuery(orderings=orderings)\n\n return wrapped\n\n\ndef parse_filter_query(\n fields: Optional[List[str]] = None,\n) -> Callable[..., FilterQuery]:\n if fields is None:\n fields = []\n parameters = [\n Parameter(\n name=field,\n kind=Parameter.KEYWORD_ONLY,\n annotation=str,\n default=Query(\"\", description=f\"F{field} filter\"),\n )\n for field in fields\n ]\n signature = Signature(parameters)\n\n @with_signature(signature)\n def wrapped(**kwargs: Any) -> FilterQuery:\n return FilterQuery(fields=kwargs)\n\n return wrapped\n", "repo_name": "joint-online-judge/fastapi-rest-framework", "sub_path": "fastapi_rest_framework/parsers.py", "file_name": "parsers.py", "file_ext": "py", "file_size_in_byte": 2329, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "typing.Optional", "line_number": 14, "usage_type": "name"}, {"api_name": "pydantic.NonNegativeInt", "line_number": 14, "usage_type": "name"}, {"api_name": "pydantic.conint", "line_number": 16, "usage_type": "call"}, {"api_name": "pydantic.NonNegativeInt", "line_number": 20, "usage_type": "name"}, {"api_name": "fastapi.Query", "line_number": 20, "usage_type": "call"}, {"api_name": "fastapi.Query", "line_number": 21, "usage_type": "call"}, {"api_name": "fastapi_rest_framework.schemas.PaginationQuery", "line_number": 23, "usage_type": "call"}, {"api_name": "pydantic.NonNegativeInt", "line_number": 23, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 18, "usage_type": "name"}, {"api_name": "fastapi_rest_framework.schemas.PaginationQuery", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 15, "usage_type": "name"}, {"api_name": "fastapi_rest_framework.schemas.PaginationQuery", "line_number": 15, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 29, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 29, "usage_type": "name"}, {"api_name": "fastapi.Query", "line_number": 36, "usage_type": "call"}, {"api_name": "fastapi_rest_framework.errors.BizError", "line_number": 48, "usage_type": "call"}, {"api_name": "fastapi_rest_framework.errors.ErrorCode.IllegalFieldError", "line_number": 49, "usage_type": "attribute"}, {"api_name": "fastapi_rest_framework.errors.ErrorCode", "line_number": 49, "usage_type": "name"}, {"api_name": "fastapi_rest_framework.schemas.OrderingQuery", "line_number": 52, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 34, "usage_type": "name"}, {"api_name": "fastapi_rest_framework.schemas.OrderingQuery", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 30, "usage_type": "name"}, {"api_name": "fastapi_rest_framework.schemas.OrderingQuery", "line_number": 30, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 58, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 58, "usage_type": "name"}, {"api_name": "inspect.Parameter", "line_number": 63, "usage_type": "call"}, {"api_name": "inspect.Parameter.KEYWORD_ONLY", "line_number": 65, "usage_type": "attribute"}, {"api_name": "inspect.Parameter", "line_number": 65, "usage_type": "name"}, {"api_name": "fastapi.Query", "line_number": 67, "usage_type": "call"}, {"api_name": "inspect.Signature", "line_number": 71, "usage_type": "call"}, {"api_name": "typing.Any", "line_number": 74, "usage_type": "name"}, {"api_name": "fastapi_rest_framework.schemas.FilterQuery", "line_number": 75, "usage_type": "call"}, {"api_name": "makefun.with_signature", "line_number": 73, "usage_type": "call"}, {"api_name": "fastapi_rest_framework.schemas.FilterQuery", "line_number": 74, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 59, "usage_type": "name"}, {"api_name": "fastapi_rest_framework.schemas.FilterQuery", "line_number": 59, "usage_type": "name"}]}
+{"seq_id": "20964657909", "text": "import io\nimport logging\nimport os\nimport time\nfrom datetime import datetime\nfrom typing import List, Tuple\nfrom urllib.parse import quote\n\nimport PyRSS2Gen\nfrom starlette_context import context\n\nfrom podcast.dao.podcast import Podcast as PodcastDao, PodcastSource\n\nfrom podcast.dao.episode import Episode as EpisodeDao\nfrom podcast.internal.resource.resource import Resource as InternalResource\nfrom podcast.internal.episode.episode import Episode as EpisodeInternal\nfrom podcast.pkg.cipher.jwt import Jwt\nfrom podcast.pkg.errors.biz_error import PodcastShareNotSet, PodcastShareTokenInvalid\nfrom podcast.pkg.parser.podcast import Podcast as PodcastInfo\nfrom podcast.pkg.parser.document import Document\nfrom podcast.pkg.parser.rss import RSS\nfrom podcast.pkg.parser.video import Video\n\nfrom podcast.pkg.type import dict_exclude_keys, str_is_empty\nfrom podcast.pkg.segment.text_segments import TextSegment\nfrom podcast.dao.base import GenStatus\n\n\nclass Podcast:\n def __init__(self, **kwargs):\n self.gid = kwargs.get(\"gid\", \"\")\n self.source = kwargs.get(\"source\", \"\")\n self.title = kwargs.get(\"title\", \"\")\n self.author = kwargs.get(\"author\", \"\")\n self.description = kwargs.get(\"description\", \"\")\n self.cover_resource_id = kwargs.get(\"cover_resource_id\", None)\n self.book_resource_id = kwargs.get(\"book_resource_id\", None)\n self.recent_play_chapter_gid = kwargs.get(\"recent_play_chapter_gid\", None)\n self.category_id = kwargs.get(\"category_id\", None)\n self.language = kwargs.get(\"language\", None)\n self.share_time = kwargs.get(\"share_time\", None)\n self.share_enable = kwargs.get(\"share_enable\", None)\n self.cover_url = kwargs.get(\"cover_url\", None)\n self.url = kwargs.get(\"url\", None)\n self.frequency = kwargs.get(\"frequency\", None)\n self.first_execute_time = kwargs.get(\"first_execute_time\", None)\n self.timer_id = kwargs.get(\"timer_id\", None)\n\n self.blog_rss_url = kwargs.get(\"blog_rss_url\", None)\n\n def __iter__(self):\n keys_map = {\n \"gid\": \"id\",\n \"source\": \"source\",\n \"title\": \"title\",\n \"author\": \"author\",\n \"cover_url\": \"coverUrl\",\n \"url\": \"url\",\n \"description\": \"description\",\n \"recent_play_chapter_gid\": \"recentPlayChapterId\",\n \"language\": \"language\",\n \"share_time\": \"share_time\",\n \"share_enable\": \"share_enable\",\n \"frequency\": \"frequency\",\n \"first_execute_time\": \"firstExecuteTime\",\n \"timer_id\": \"timerId\",\n }\n for key, new_key in keys_map.items():\n yield new_key, self.__dict__.get(key)\n\n def get_podcast(self) -> PodcastDao:\n podcast_dao: PodcastDao = PodcastDao(gid=self.gid)\n return podcast_dao.get_by_gid()\n\n def delete_podcast(self):\n podcast_dao: PodcastDao = PodcastDao(gid=self.gid)\n return podcast_dao.delete()\n\n def get_podcast_detail(self, offset=0, limit=100, order=\"desc\") -> Tuple[PodcastDao, List[EpisodeDao], int]:\n podcast: PodcastDao = self.get_podcast()\n if podcast is None:\n return None, []\n\n episodes = EpisodeDao(podcast_gid=podcast.gid).get_episodes_by_podcast_gid( offset, limit, order)\n\n total = EpisodeDao(podcast_gid=podcast.gid).get_episodes_total_by_podcast_gid()\n return podcast, episodes, total\n\n def get_podcasts_by_gids(self, gids: List[int])->List[PodcastDao]:\n return PodcastDao.get_by_gids(gids)\n\n def _get_podcast_without_voice(self):\n podcast: PodcastDao = self.get_podcast()\n if podcast is None:\n return None, []\n return podcast, EpisodeDao.get_episodes_without_voice(podcast.gid)\n\n @classmethod\n def get_podcasts(cls, offset: int, limit: int):\n return PodcastDao.get_podcasts(offset, limit)\n\n @classmethod\n def get_podcasts_total(cls):\n return PodcastDao.get_podcasts_total()\n\n def _parse_local_book(self) -> PodcastInfo:\n resources = InternalResource.get_resources_by_gid_array([self.book_resource_id])\n book_resource_dict = resources.get(self.book_resource_id)\n\n document = Document(self.language)\n document.loads(book_resource_dict.get(\"name\"), book_resource_dict.get(\"content\"))\n return document.parse()\n\n def _parse_rss(self) -> PodcastInfo:\n rss = RSS(self.url, self.language)\n return rss.parse()\n\n def _parse_video(self) -> PodcastInfo:\n video = Video(self.language, self.url)\n return video.parse()\n\n def parse(self):\n podcast_dao = self.get_podcast()\n\n self.language = podcast_dao.language\n self.source = podcast_dao.source\n self.title = podcast_dao.title\n self.description = podcast_dao.description\n self.author = podcast_dao.author\n\n podcast_info: PodcastInfo = None\n if self.source == PodcastSource.local:\n self.book_resource_id = podcast_dao.book_resource_id\n podcast_info = self._parse_local_book()\n elif self.source == PodcastSource.rss:\n self.url = podcast_dao.url\n podcast_info = self._parse_rss()\n elif self.source == PodcastSource.video:\n self.url = podcast_dao.url\n podcast_info = self._parse_video()\n self.save_podcast(podcast_info)\n return podcast_info\n\n @classmethod\n def _gen_language_segments(cls, podcast: PodcastDao, episodes: List[EpisodeDao]):\n episode_language_segments = {}\n for episode in episodes:\n episode_language_segments[episode.gid] = TextSegment(podcast.language, episode.content).gen_lang_segments()\n return episode_language_segments\n\n def prepare_gen_episode(self):\n podcast_dao, episodes_dao = self._get_podcast_without_voice()\n return self.gid, podcast_dao.source, EpisodeInternal.gen_language_segments(podcast_dao.source,\n podcast_dao.language, episodes_dao)\n\n def add_podcast(self, name):\n if (self.title is None or self.title == \"\") and self.source == PodcastSource.local:\n self.title = os.path.splitext(os.path.basename(name))[0]\n podcast = PodcastDao(**dict_exclude_keys(self.__dict__, \"gid\"))\n podcast.save()\n return podcast\n\n def update_podcast_info(self, **kwargs):\n podcast = PodcastDao(gid=self.gid, **kwargs)\n podcast.update(kwargs)\n\n def _update_podcast(self, podcast_info: PodcastInfo):\n update_fields = {\n \"title\": podcast_info.title if str_is_empty(self.title) else self.title,\n \"author\": podcast_info.author if str_is_empty(self.author) else self.author,\n \"description\": podcast_info.description if str_is_empty(self.description) is None else self.description,\n }\n\n if podcast_info.cover is not None:\n resource_dao = InternalResource(file=podcast_info.cover, type=\"cover\").save()\n update_fields.update({\n \"cover_resource_id\": resource_dao.gid\n })\n\n PodcastDao(gid=self.gid).update(update_fields)\n\n def _update_episodes(self, podcast_info: PodcastInfo):\n keys = [episode.key for episode in podcast_info.episodes]\n existed_episodes = EpisodeDao.get_episode_by_keys(self.gid, keys)\n existed_keys = set([episode.key for episode in existed_episodes])\n\n new_episodes = []\n for episode in podcast_info.episodes:\n if episode.key not in existed_keys:\n new_episodes.append(episode)\n\n episodes_dao = []\n for episode in reversed(new_episodes):\n kwargs = {\n \"podcast_gid\": self.gid,\n } | episode.__dict__\n\n if episode.cover is not None:\n resource_dao = InternalResource(file=episode.cover, type=\"cover\").save()\n kwargs.update({\n \"cover_resource_id\": resource_dao.gid\n })\n episodes_dao.append(EpisodeDao(**(dict_exclude_keys(kwargs, \"gid\"))))\n EpisodeDao.save(episodes_dao)\n\n def save_podcast(self, podcast_info: PodcastInfo):\n self._update_podcast(podcast_info)\n self._update_episodes(podcast_info)\n\n \n @classmethod\n def _gen_share_token(cls, payload: dict) -> str:\n return quote(Jwt(payload=payload).encode_share_token())\n\n @classmethod\n def _gen_resource_token(cls, podcast_dao: PodcastDao, resource_id: str) -> str:\n return quote(Jwt(payload={\n \"podcast_id\": podcast_dao.gid,\n \"share_time\": podcast_dao.share_time,\n \"user_id\": podcast_dao.created_by,\n \"resource_id\": resource_id,\n }).encode_share_token())\n\n def share_podcast(self, share_enable: bool) -> str:\n podcast_dao: PodcastDao = self.get_podcast()\n\n share_time = podcast_dao.share_time if podcast_dao.share_enable else int(time.time())\n if bool(podcast_dao.share_enable) != share_enable:\n podcast_dao.update({\n \"share_enable\": share_enable,\n \"share_time\": share_time,\n })\n\n share_token = self._gen_share_token({\n \"podcast_id\": podcast_dao.gid,\n \"share_time\": share_time,\n \"user_id\": podcast_dao.created_by,\n }) if share_enable else \"\"\n\n return share_token\n\n def gen_rss(self, share_token: str):\n podcast, episodes, _ = self.get_podcast_detail()\n\n if podcast.share_enable == 0:\n raise PodcastShareNotSet()\n\n base_url = context.data.get(\"base_url\")\n cover_url = f\"{base_url}/api/web/podcast_resource/${self._gen_resource_token(podcast, podcast.cover_resource_id)}\"\n rss = PyRSS2Gen.RSS2(title=podcast.title,\n description=podcast.title,\n link=base_url if podcast.url is None else podcast.url,\n image=PyRSS2Gen.Image(\n url=cover_url,\n title=podcast.title,\n link=cover_url,\n ),\n lastBuildDate=datetime.now())\n\n for episode in episodes:\n voice_url = f\"{base_url}/api/web/podcast_resource/${self._gen_resource_token(podcast, episode.voice_resource_id)}\"\n\n cover_resource_id = episode.cover_resource_id\n if str_is_empty(cover_resource_id):\n cover_resource_id = podcast.cover_resource_id\n episode_cover_url = f\"{base_url}/api/web/podcast_resource/${self._gen_resource_token(podcast, cover_resource_id)}\"\n rss.items.append(PyRSS2Gen.RSSItem(\n title=episode.title,\n link=episode.link,\n description=episode.title,\n enclosure=PyRSS2Gen.Enclosure(voice_url,\n episode.episode_size,\n \"episode/mpeg\"),\n guid=PyRSS2Gen.Guid(episode_cover_url, isPermaLink=1),\n pubDate=datetime.fromtimestamp(episode.pub_time if episode.pub_time else episode.created_at)\n ))\n\n try:\n content_io = io.BytesIO()\n rss.write_xml(content_io, \"utf-8\")\n return content_io\n except Exception as ex:\n logging.exception(ex)\n\n @staticmethod\n def check_rss_token(token: str) -> dict:\n data = Jwt(token=token).decode_share_token()\n context.data[\"user_id\"] = data.get(\"user_id\")\n podcast_id = data.get(\"podcast_id\")\n share_time = data.get(\"share_time\")\n\n podcast_dao = Podcast(gid=podcast_id).get_podcast()\n if not podcast_dao.share_enable:\n raise PodcastShareNotSet()\n\n if podcast_dao.share_time != share_time:\n raise PodcastShareTokenInvalid()\n \n return data\n\n def update_gen_status(self, status: int):\n podcast_dao: PodcastDao = self.get_podcast()\n podcast_dao.update({\n \"gen_status\": status,\n })", "repo_name": "PodcastIO/canary", "sub_path": "server/src/podcast/internal/podcast/podcast.py", "file_name": "podcast.py", "file_ext": "py", "file_size_in_byte": 12195, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "24", "api": [{"api_name": "podcast.dao.podcast.Podcast", "line_number": 72, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 71, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 76, "usage_type": "name"}, {"api_name": "podcast.dao.podcast", "line_number": 80, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 80, "usage_type": "name"}, {"api_name": "podcast.dao.podcast", "line_number": 81, "usage_type": "name"}, {"api_name": "podcast.dao.episode.Episode", "line_number": 84, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.gid", "line_number": 84, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast", "line_number": 84, "usage_type": "name"}, {"api_name": "podcast.dao.episode.Episode", "line_number": 86, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.gid", "line_number": 86, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast", "line_number": 86, "usage_type": "name"}, {"api_name": "podcast.dao.podcast", "line_number": 87, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 79, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 79, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 79, "usage_type": "name"}, {"api_name": "podcast.dao.episode.Episode", "line_number": 79, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 89, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.Podcast.get_by_gids", "line_number": 90, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 90, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 89, "usage_type": "name"}, {"api_name": "podcast.dao.podcast", "line_number": 93, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 93, "usage_type": "name"}, {"api_name": "podcast.dao.podcast", "line_number": 94, "usage_type": "name"}, {"api_name": "podcast.dao.podcast", "line_number": 96, "usage_type": "name"}, {"api_name": "podcast.dao.episode.Episode.get_episodes_without_voice", "line_number": 96, "usage_type": "call"}, {"api_name": "podcast.dao.episode.Episode", "line_number": 96, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.gid", "line_number": 96, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast.Podcast.get_podcasts", "line_number": 100, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 100, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.Podcast.get_podcasts_total", "line_number": 104, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 104, "usage_type": "name"}, {"api_name": "podcast.internal.resource.resource.Resource.get_resources_by_gid_array", "line_number": 107, "usage_type": "call"}, {"api_name": "podcast.internal.resource.resource.Resource", "line_number": 107, "usage_type": "name"}, {"api_name": "podcast.pkg.parser.document.Document", "line_number": 110, "usage_type": "call"}, {"api_name": "podcast.pkg.parser.podcast.Podcast", "line_number": 106, "usage_type": "name"}, {"api_name": "podcast.pkg.parser.rss.RSS", "line_number": 115, "usage_type": "call"}, {"api_name": "podcast.pkg.parser.podcast.Podcast", "line_number": 114, "usage_type": "name"}, {"api_name": "podcast.pkg.parser.video.Video", "line_number": 119, "usage_type": "call"}, {"api_name": "podcast.pkg.parser.podcast.Podcast", "line_number": 118, "usage_type": "name"}, {"api_name": "podcast.pkg.parser.podcast.Podcast", "line_number": 131, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.PodcastSource.local", "line_number": 132, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast.PodcastSource", "line_number": 132, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.PodcastSource.rss", "line_number": 135, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast.PodcastSource", "line_number": 135, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.PodcastSource.video", "line_number": 138, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast.PodcastSource", "line_number": 138, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 145, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 145, "usage_type": "name"}, {"api_name": "podcast.dao.episode.Episode", "line_number": 145, "usage_type": "name"}, {"api_name": "podcast.pkg.segment.text_segments.TextSegment", "line_number": 148, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.language", "line_number": 148, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast", "line_number": 148, "usage_type": "name"}, {"api_name": "podcast.internal.episode.episode.Episode.gen_language_segments", "line_number": 153, "usage_type": "call"}, {"api_name": "podcast.internal.episode.episode.Episode", "line_number": 153, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.PodcastSource.local", "line_number": 157, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast.PodcastSource", "line_number": 157, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path", "line_number": 158, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 158, "usage_type": "call"}, {"api_name": "podcast.dao.podcast", "line_number": 159, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 159, "usage_type": "call"}, {"api_name": "podcast.pkg.type.dict_exclude_keys", "line_number": 159, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.save", "line_number": 160, "usage_type": "call"}, {"api_name": "podcast.dao.podcast", "line_number": 160, "usage_type": "name"}, {"api_name": "podcast.dao.podcast", "line_number": 161, "usage_type": "name"}, {"api_name": "podcast.dao.podcast", "line_number": 164, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 164, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.update", "line_number": 165, "usage_type": "call"}, {"api_name": "podcast.dao.podcast", "line_number": 165, "usage_type": "name"}, {"api_name": "podcast.pkg.parser.podcast.Podcast", "line_number": 167, "usage_type": "name"}, {"api_name": "podcast.pkg.type.str_is_empty", "line_number": 169, "usage_type": "call"}, {"api_name": "podcast.pkg.type.str_is_empty", "line_number": 170, "usage_type": "call"}, {"api_name": "podcast.pkg.type.str_is_empty", "line_number": 171, "usage_type": "call"}, {"api_name": "podcast.internal.resource.resource.Resource", "line_number": 175, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 180, "usage_type": "call"}, {"api_name": "podcast.pkg.parser.podcast.Podcast", "line_number": 182, "usage_type": "name"}, {"api_name": "podcast.dao.episode.Episode.get_episode_by_keys", "line_number": 184, "usage_type": "call"}, {"api_name": "podcast.dao.episode.Episode", "line_number": 184, "usage_type": "name"}, {"api_name": "podcast.internal.resource.resource.Resource", "line_number": 199, "usage_type": "call"}, {"api_name": "podcast.dao.episode.Episode", "line_number": 203, "usage_type": "call"}, {"api_name": "podcast.pkg.type.dict_exclude_keys", "line_number": 203, "usage_type": "call"}, {"api_name": "podcast.dao.episode.Episode.save", "line_number": 204, "usage_type": "call"}, {"api_name": "podcast.dao.episode.Episode", "line_number": 204, "usage_type": "name"}, {"api_name": "podcast.pkg.parser.podcast.Podcast", "line_number": 206, "usage_type": "name"}, {"api_name": "urllib.parse.quote", "line_number": 213, "usage_type": "call"}, {"api_name": "podcast.pkg.cipher.jwt.Jwt", "line_number": 213, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 216, "usage_type": "name"}, {"api_name": "urllib.parse.quote", "line_number": 217, "usage_type": "call"}, {"api_name": "podcast.pkg.cipher.jwt.Jwt", "line_number": 217, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 225, "usage_type": "name"}, {"api_name": "time.time", "line_number": 227, "usage_type": "call"}, {"api_name": "podcast.dao.podcast", "line_number": 243, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.share_enable", "line_number": 245, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast", "line_number": 245, "usage_type": "name"}, {"api_name": "podcast.pkg.errors.biz_error.PodcastShareNotSet", "line_number": 246, "usage_type": "call"}, {"api_name": "starlette_context.context.data.get", "line_number": 248, "usage_type": "call"}, {"api_name": "starlette_context.context.data", "line_number": 248, "usage_type": "attribute"}, {"api_name": "starlette_context.context", "line_number": 248, "usage_type": "name"}, {"api_name": "podcast.dao.podcast", "line_number": 249, "usage_type": "argument"}, {"api_name": "podcast.dao.podcast.cover_resource_id", "line_number": 249, "usage_type": "attribute"}, {"api_name": "PyRSS2Gen.RSS2", "line_number": 250, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.title", "line_number": 250, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast", "line_number": 250, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.title", "line_number": 251, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast", "line_number": 251, "usage_type": "name"}, {"api_name": "podcast.dao.podcast.url", "line_number": 252, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast", "line_number": 252, "usage_type": "name"}, {"api_name": "PyRSS2Gen.Image", "line_number": 253, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.title", "line_number": 255, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast", "line_number": 255, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 258, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 258, "usage_type": "name"}, {"api_name": "podcast.dao.podcast", "line_number": 261, "usage_type": "argument"}, {"api_name": "podcast.pkg.type.str_is_empty", "line_number": 264, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.cover_resource_id", "line_number": 265, "usage_type": "attribute"}, {"api_name": "podcast.dao.podcast", "line_number": 265, "usage_type": "name"}, {"api_name": "podcast.dao.podcast", "line_number": 266, "usage_type": "argument"}, {"api_name": "PyRSS2Gen.RSSItem", "line_number": 267, "usage_type": "call"}, {"api_name": "PyRSS2Gen.Enclosure", "line_number": 271, "usage_type": "call"}, {"api_name": "PyRSS2Gen.Guid", "line_number": 274, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 275, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 275, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 279, "usage_type": "call"}, {"api_name": "logging.exception", "line_number": 283, "usage_type": "call"}, {"api_name": "podcast.pkg.cipher.jwt.Jwt", "line_number": 287, "usage_type": "call"}, {"api_name": "starlette_context.context.data", "line_number": 288, "usage_type": "attribute"}, {"api_name": "starlette_context.context", "line_number": 288, "usage_type": "name"}, {"api_name": "podcast.pkg.errors.biz_error.PodcastShareNotSet", "line_number": 294, "usage_type": "call"}, {"api_name": "podcast.pkg.errors.biz_error.PodcastShareTokenInvalid", "line_number": 297, "usage_type": "call"}, {"api_name": "podcast.dao.podcast.Podcast", "line_number": 302, "usage_type": "name"}]}
+{"seq_id": "38970076267", "text": "\"\"\"views.py\"\"\"\nimport numpy as np\nfrom django.shortcuts import render\nfrom joblib import load\n\nclf = load(\"Notebooks/dt_model.joblib\")\n\n\ndef index(req):\n \"\"\"View function for home page of site.\"\"\"\n return render(req, \"index.html\")\n\n\ndef result(pred):\n \"\"\"View function for result page of site.\"\"\"\n data = predict(pred)\n return render(pred, \"result.html\", {\"data\": data})\n\n\ndef predict(req):\n \"\"\"View function for prediction page of site.\"\"\"\n age = req.POST[\"age\"]\n sex = req.POST[\"sex\"]\n cp_ = req.POST[\"cp\"]\n trestbps = req.POST[\"trestbps\"]\n chol = req.POST[\"chol\"]\n fbs = req.POST[\"fbs\"]\n thalach = req.POST[\"thalach\"]\n exang = req.POST[\"exang\"]\n thal = req.POST[\"thal\"]\n x_test = [[age, sex, cp_, trestbps, chol, fbs, thalach, exang, thal]]\n x_test = np.array(x_test).reshape(1, 9)\n x_test = np.array(x_test, dtype=float)\n y_pred = clf.predict(x_test).tolist()\n ans = y_pred[0]\n return ans\n", "repo_name": "debjitpal5040/save-heart", "sub_path": "save_heart/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 956, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "joblib.load", "line_number": 6, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 11, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}]}
+{"seq_id": "8941565816", "text": "from torch import nn\nfrom torch.nn import functional as F\n\nfrom maskrcnn_benchmark.modeling.poolers import Pooler\n\nfrom maskrcnn_benchmark.layers import Conv2d\n\n\nclass KeypointRCNNFeatureExtractor(nn.Module):\n def __init__(self, cfg):\n super(KeypointRCNNFeatureExtractor, self).__init__()\n\n resolution = cfg.MODEL.ROI_KEYPOINT_HEAD.POOLER_RESOLUTION\n scales = cfg.MODEL.ROI_KEYPOINT_HEAD.POOLER_SCALES\n sampling_ratio = cfg.MODEL.ROI_KEYPOINT_HEAD.POOLER_SAMPLING_RATIO\n pooler = Pooler(\n output_size=(resolution, resolution),\n scales=scales,\n sampling_ratio=sampling_ratio,\n )\n self.pooler = pooler\n\n input_features = cfg.MODEL.BACKBONE.OUT_CHANNELS\n layers = cfg.MODEL.ROI_KEYPOINT_HEAD.CONV_LAYERS\n next_feature = input_features\n self.blocks = []\n for layer_idx, layer_features in enumerate(layers, 1):\n layer_name = \"conv_fcn{}\".format(layer_idx)\n module = Conv2d(next_feature, layer_features, 3, stride=1, padding=1)\n nn.init.kaiming_normal_(module.weight, mode=\"fan_out\", nonlinearity=\"relu\")\n nn.init.constant_(module.bias, 0)\n self.add_module(layer_name, module)\n next_feature = layer_features\n self.blocks.append(layer_name)\n\n def forward(self, x, proposals):\n x = self.pooler(x, proposals)\n for layer_name in self.blocks:\n x = F.relu(getattr(self, layer_name)(x))\n return x\n\n\n_ROI_KEYPOINT_FEATURE_EXTRACTORS = {\n \"KeypointRCNNFeatureExtractor\": KeypointRCNNFeatureExtractor\n}\n\n\ndef make_roi_keypoint_feature_extractor(cfg):\n func = _ROI_KEYPOINT_FEATURE_EXTRACTORS[\n cfg.MODEL.ROI_KEYPOINT_HEAD.FEATURE_EXTRACTOR\n ]\n return func(cfg)\n", "repo_name": "mlcommons/training", "sub_path": "object_detection/pytorch/maskrcnn_benchmark/modeling/roi_heads/keypoint_head/roi_keypoint_feature_extractors.py", "file_name": "roi_keypoint_feature_extractors.py", "file_ext": "py", "file_size_in_byte": 1796, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1501, "dataset": "github-code", "pt": "24", "api": [{"api_name": "torch.nn.Module", "line_number": 9, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 9, "usage_type": "name"}, {"api_name": "maskrcnn_benchmark.modeling.poolers.Pooler", "line_number": 16, "usage_type": "call"}, {"api_name": "maskrcnn_benchmark.layers.Conv2d", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn.init.kaiming_normal_", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 30, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 31, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 39, "usage_type": "name"}]}
+{"seq_id": "34485849037", "text": "from django.shortcuts import render, redirect\nfrom django.http import HttpResponse\nfrom django.contrib.auth.forms import UserCreationForm, AuthenticationForm\nfrom django.contrib.auth import authenticate, login as loginuser\n\n# Create your views here.\n\n\"\"\"\ndef home(request):\n \n # print(\"Hello World!...This is Home\")\n \n html = '''\n \n
Home Page
\n \n
\n
Item 1
\n
Item 2
\n
Item 3
\n
Item 4
\n
\n \n
This is Para
\n \n \n \n Go to facebook\n '''\n \n \n # return HttpResponse(\"Response from View File\")\n return HttpResponse(html)\n\n\"\"\"\n\n\ndef home(request):\n \n return render(request,'index.html')\n\n\n\n\ndef login(request):\n \n if request.method == \"GET\":\n form = AuthenticationForm()\n context = { \"form\" : form }\n \n return render(request, 'login.html', context=context)\n \n else:\n form = AuthenticationForm(data=request.POST)\n print(form.is_valid())\n if form.is_valid():\n username = form.cleaned_data.get('username')\n password = form.cleaned_data.get('passowrd')\n user = authenticate(username = username, password = password)\n print(\"Authenticated::\", user) \n if user is not None:\n loginuser(request, user)\n return redirect('home')\n \n else:\n context = { \"form\" : form }\n return render(request, 'login.html', context=context)\n\n\n\n\ndef signUp(request):\n \n if request.method == 'GET':\n \n form = UserCreationForm()\n context = {\n \"form\" : form\n }\n return render(request, 'signup.html', context=context)\n \n else:\n \n print(request.POST)\n form = UserCreationForm(request.POST)\n context = {\n \"form\" : form\n }\n \n \n if form.is_valid():\n user = form.save()\n print(user)\n # return HttpResponse(\"Form is Valid\")\n if user is not None:\n return redirect('login')\n else:\n # return HttpResponse(\"Form is Valid\")\n return render(request, 'signup.html', context=context)\n \n ", "repo_name": "Snoveedh/YouDo_Medo", "sub_path": "app/views_backup.py", "file_name": "views_backup.py", "file_ext": "py", "file_size_in_byte": 2422, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "django.shortcuts.render", "line_number": 40, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.AuthenticationForm", "line_number": 48, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 51, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.AuthenticationForm", "line_number": 54, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 59, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 62, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 63, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 67, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.UserCreationForm", "line_number": 76, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 80, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.UserCreationForm", "line_number": 85, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 96, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 99, "usage_type": "call"}]}
+{"seq_id": "4430129234", "text": "\"\"\"Files.\"\"\"\nimport glob\nimport csv\nfrom datetime import datetime\nfrom datetime import date\n\n\ndef no_data(header, types_dict, i):\n \"\"\"If data is '-', then return this function.\"\"\"\n if header[i] not in types_dict:\n types_dict[header[i]] = ['-']\n else:\n types_dict[header[i]].append('-')\n\n\ndef final_data_types(types_dict):\n \"\"\"Get final data types into dict.\"\"\"\n for key in types_dict:\n val = types_dict[key]\n if 'str' in val:\n types_dict[key] = 'str'\n continue\n if 'int' in val and 'date' in val:\n types_dict[key] = 'str'\n continue\n if 'int' in val and 'str' not in val and 'date' not in val:\n types_dict[key] = 'int'\n continue\n if 'date' in val and 'str' not in val and 'int' not in val:\n types_dict[key] = 'date'\n continue\n if '-' in val and 'str' not in val and 'int' not in val and 'date' not in val:\n types_dict[key] = '-'\n continue\n if '-' in val:\n continue\n return types_dict\n\n\ndef add_str(header, types_dict, i):\n \"\"\"Add 'str' to list.\"\"\"\n if header[i] not in types_dict:\n types_dict[header[i]] = ['str']\n else:\n types_dict[header[i]].append('str')\n\n\ndef add_int(header, types_dict, i):\n \"\"\"Add 'int' to list.\"\"\"\n if header[i] not in types_dict:\n types_dict[header[i]] = ['int']\n else:\n types_dict[header[i]].append('int')\n\n\ndef is_int(value):\n \"\"\"Check for digit.\"\"\"\n return value.isdigit()\n\n\ndef is_date(value):\n \"\"\"Check is correct date.\"\"\"\n format = \"%d.%m.%Y\"\n try:\n datetime.strptime(value, format).date()\n return True\n except ValueError:\n return False\n\n\ndef final(csv_list, types_dict, header):\n \"\"\"Final list of dicts with correct types.\"\"\"\n final_list = []\n for row in csv_list[1:]:\n final_dict = {}\n for i, value in enumerate(row):\n if value == '-':\n final_dict[header[i]] = None\n continue\n if final_data_types(types_dict)[header[i]] == 'str':\n final_dict[header[i]] = str(value)\n continue\n if final_data_types(types_dict)[header[i]] == 'int':\n final_dict[header[i]] = int(value)\n continue\n if final_data_types(types_dict)[header[i]] == 'date':\n final_dict[header[i]] = datetime.strptime(value, \"%d.%m.%Y\").date()\n final_list.append(final_dict)\n return final_list\n\n\ndef read_file_contents(filename: str) -> str:\n \"\"\"Read file contents into string.\"\"\"\n with open(filename) as f: # Opens file with name of \"test.txt\"\n data = f.read() # Reads all the lines from the file and saves it as a string.\n return data\n\n\ndef read_file_contents_to_list(filename: str) -> list:\n \"\"\"Read file contents into list of lines.\"\"\"\n list_of_lines = read_file_contents(filename).split('\\n')\n return list_of_lines\n\n\ndef read_csv_file(filename: str) -> list:\n \"\"\"Read CSV file into list of rows.\"\"\"\n csv_list = []\n with open(filename) as csv_file:\n csv_reader = csv.reader(csv_file, delimiter=',')\n for row in csv_reader:\n csv_list.append(row)\n return csv_list\n\n\ndef write_contents_to_file(filename: str, contents: str) -> None:\n \"\"\"Write contents to file.\"\"\"\n with open(filename, \"w\") as f:\n f.write(contents)\n\n\ndef write_lines_to_file(filename: str, lines: list) -> None:\n \"\"\"Write lines to file.\"\"\"\n with open(filename, \"w\") as f:\n lines = '\\n'.join(lines)\n f.write(lines)\n\n\ndef write_csv_file(filename: str, data: list) -> None:\n \"\"\"\n Write data into CSV file.\n\n Data is a list of lists.\n List represents lines.\n Each element (which is list itself) represents columns in a line.\n\n [[\"name\", \"age\"], [\"john\", \"11\"], [\"mary\", \"15\"]]\n Will result in csv file:\n\n name,age\n john,11\n mary,15\n\n Use csv module here.\n\n :param filename: Name of the file.\n :param data: List of lists to write to the file.\n :return: None\n \"\"\"\n with open(filename, 'w', newline='') as csv_file:\n csv_writer = csv.writer(csv_file)\n for row in data:\n csv_writer.writerow(row)\n\n\ndef merge_dates_and_towns_into_csv(dates_file: str, towns_file: str, csv_output: str) -> None:\n \"\"\"\n Merge information from two files into one CSV file.\n\n dates_file contains names and dates. Separated by colon.\n john:01.01.2001\n mary:06.03.2016\n\n You don't have to validate the date.\n Every line contains name, colon and date.\n\n towns_file contains names and towns. Separated by colon.\n john:london\n mary:new york\n\n Every line contains name, colon and town name.\n\n Those two files should be merged by names.\n The result should be a csv file:\n\n name,town,date\n john,london,01.01.2001\n mary,new york,06.03.2016\n\n Applies for the third part:\n If information about a person is missing, it should be \"-\" in the output file.\n\n The order of the lines should follow the order in dates input file.\n Names which are missing in dates input file, will follow the order\n in towns input file.\n The order of the fields is: name,town,date\n\n name,town,date\n john,-,01.01.2001\n mary,new york,-\n\n Hint: try to reuse csv reading and writing functions.\n When reading csv, delimiter can be specified.\n\n :param dates_file: Input file with names and dates.\n :param towns_file: Input file with names and towns.\n :param csv_output: Output CSV-file with names, towns and dates.\n :return: None\n \"\"\"\n dates = []\n towns = []\n result = [[\"name\", \"town\", \"date\"]]\n with open(dates_file) as csv_file:\n csv_reader = csv.reader(csv_file, delimiter=':')\n for row in csv_reader:\n dates.append(row)\n with open(towns_file) as csv_file:\n csv_reader = csv.reader(csv_file, delimiter=':')\n for row in csv_reader:\n towns.append(row)\n\n for date_person in dates:\n name, date = date_person\n result.append([name, '-', date])\n\n for town_person in towns:\n name2, town = town_person\n for elem in result[1:]:\n if name2 == elem[0]:\n elem[1] = town\n break\n else:\n result.append([name2, town, '-'])\n\n with open(csv_output, 'w', newline='') as csv_file:\n csv_writer = csv.writer(csv_file)\n for row in result:\n csv_writer.writerow(row)\n\n\ndef read_csv_file_into_list_of_dicts(filename: str) -> list:\n \"\"\"\n Read csv file into list of dictionaries.\n\n Header line will be used for dict keys.\n\n Each line after header line will result in a dict inside the result list.\n Every line contains the same number of fields.\n\n Example:\n\n name,age,sex\n John,12,M\n Mary,13,F\n\n Header line will be used as keys for each content line.\n The result:\n [\n {\"name\": \"John\", \"age\": \"12\", \"sex\": \"M\"},\n {\"name\": \"Mary\", \"age\": \"13\", \"sex\": \"F\"},\n ]\n\n If there are only header or no rows in the CSV-file,\n the result is an empty list.\n\n The order of the elements in the list should be the same\n as the lines in the file (the first line becomes the first element etc.)\n\n :param filename: CSV-file to read.\n :return: List of dictionaries where keys are taken from the header.\n \"\"\"\n list_of_dicts = []\n lists = read_csv_file(filename)\n if lists:\n header_list = lists[0]\n lists.remove(header_list)\n for content_list in lists:\n dic = {}\n for i in range(len(header_list)):\n key = header_list[i]\n value = content_list[i]\n dic[key] = value\n list_of_dicts.append(dic)\n return list_of_dicts\n else:\n return []\n\n\ndef write_list_of_dicts_to_csv_file(filename: str, data: list) -> None:\n \"\"\"\n Write list of dicts into csv file.\n\n Data contains a list of dictionaries.\n Dictionary key represents the field.\n\n Example data:\n [\n {\"name\": \"john\", \"age\": \"23\"}\n {\"name\": \"mary\", \"age\": \"44\"}\n ]\n Will become:\n name,age\n john,23\n mary,44\n\n The order of fields/headers is not important.\n The order of lines is important (the same as in the list).\n\n Example:\n [\n {\"name\": \"john\", \"age\": \"12\"},\n {\"name\": \"mary\", \"town\": \"London\"}\n ]\n Will become:\n name,age,town\n john,12,\n mary,,London\n\n Fields which are not present in one line will be empty.\n\n The order of the lines in the file should be the same\n as the order of elements in the list.\n\n :param filename: File to write to.\n :param data: List of dictionaries to write to the file.\n :return: None\n \"\"\"\n list_of_lists = []\n header = [] # header contains e.g name, hobby and etc\n for dic in data: # get dictionary from list\n for key in dic: # get dict key\n if key in header: # if the key is already in header list, then go ahead\n continue\n else:\n header.append(key)\n list_of_lists.append(header)\n\n for dict in data:\n content = []\n for i in header:\n if i in dict.keys(): # check if element i is in dict\n content.append(dict[i])\n else:\n content.append('')\n list_of_lists.append(content)\n if len(data) == 0:\n list_of_lists = ''\n return write_csv_file(filename, list_of_lists)\n else:\n return write_csv_file(filename, list_of_lists)\n\n\ndef read_csv_file_into_list_of_dicts_using_datatypes(filename: str) -> list:\n \"\"\"\n Read data from file and cast values into different datatypes.\n\n If a field contains only numbers, turn this into int.\n If a field contains only dates (in format dd.mm.yyyy), turn this into date.\n Otherwise the datatype is string (default by csv reader).\n\n Example:\n name,age\n john,11\n mary,14\n\n Becomes ('age' is int):\n [\n {'name': 'john', 'age': 11},\n {'name': 'mary', 'age': 14}\n ]\n\n But if all the fields cannot be cast to int, the field is left to string.\n Example:\n name,age\n john,11\n mary,14\n ago,unknown\n\n Becomes ('age' cannot be cast to int because of \"ago\"):\n [\n {'name': 'john', 'age': '11'},\n {'name': 'mary', 'age': '14'},\n {'name': 'ago', 'age': 'unknown'}\n ]\n\n Example:\n name,date\n john,01.01.2020\n mary,07.09.2021\n\n Becomes:\n [\n {'name': 'john', 'date': datetime.date(2020, 1, 1)},\n {'name': 'mary', 'date': datetime.date(2021, 9, 7)},\n ]\n\n Example:\n name,date\n john,01.01.2020\n mary,late 2021\n\n Becomes:\n [\n {'name': 'john', 'date': \"01.01.2020\"},\n {'name': 'mary', 'date': \"late 2021\"},\n ]\n\n Value \"-\" indicates missing value and should be None in the result\n Example:\n name,date\n john,-\n mary,07.09.2021\n\n Becomes:\n [\n {'name': 'john', 'date': None},\n {'name': 'mary', 'date': datetime.date(2021, 9, 7)},\n ]\n\n None value also doesn't affect the data type\n (the column will have the type based on the existing values).\n\n The order of the elements in the list should be the same\n as the lines in the file.\n\n For date, strptime can be used:\n https://docs.python.org/3/library/datetime.html#examples-of-usage-date\n \"\"\"\n csv_list = read_csv_file(filename)\n types_dict = {}\n if len(csv_list) == 0:\n return []\n else:\n header = csv_list[0]\n for row in csv_list[1:]:\n for i, value in enumerate(row):\n if value == '-':\n no_data(header, types_dict, i)\n continue\n if is_date(value):\n if header[i] not in types_dict:\n types_dict[header[i]] = ['date']\n continue\n else:\n types_dict[header[i]].append('date')\n continue\n\n if not is_date(value) and not is_int(value):\n if header[i] not in types_dict:\n types_dict[header[i]] = ['str']\n continue\n else:\n types_dict[header[i]].append('str')\n continue\n\n if is_int(value):\n add_int(header, types_dict, i)\n continue\n else:\n add_str(header, types_dict, i)\n\n return final(csv_list, types_dict, header)\n\n\ndef read_people_data(directory: str) -> dict:\n \"\"\"\n Read people data from files.\n\n Files are inside directory. Read all *.csv files.\n Each file has an int field \"id\" which should be used to merge information.\n The result should be one dict where the key is id (int) and value is\n a dict of all the different values across the the files.\n Missing keys should be in every dictionary.\n Missing value is represented as None.\n\n File: a.csv\n id,name\n 1,john\n 2,mary\n 3,john\n\n File: births.csv\n id,birth\n 1,01.01.2001\n 2,05.06.1990\n\n File: deaths.csv\n id,death\n 2,01.02.2020\n 1,-\n\n Becomes:\n {\n 1: {\"id\": 1, \"name\": \"john\", \"birth\": datetime.date(2001, 1, 1), \"death\": None},\n 2: {\"id\": 2, \"name\": \"mary\", \"birth\": datetime.date(1990, 6, 5),\n \"death\": datetime.date(2020, 2, 1)},\n 3: {\"id\": 3, \"name\": \"john\", \"birth\": None, \"death\": None},\n }\n\n\n :param directory: Directory where the csv files are.\n :return: Dictionary with id as keys and data dictionaries as values.\n \"\"\"\n path = (directory + \"/\" + \"*.csv\")\n dict = {}\n files = [f for f in glob.glob(path)]\n for f in files:\n d = read_csv_file_into_list_of_dicts_using_datatypes(f)\n for line in d:\n if line['id'] not in dict:\n dict[line['id']] = line\n else:\n dict[line['id']].update(line)\n return dict\n\n\ndef generate_people_report(person_data_directory: str, report_filename: str) -> None:\n \"\"\"\n Generate report about people data.\n\n Data should be read using read_people_data().\n\n The input files contain fields \"birth\" and \"death\" which are dates. Those can be in different files. There are no duplicate headers in the files (except for the \"id\").\n\n The report is a CSV file where all the fields are written to\n (along with the headers).\n In addition, there should be two fields:\n - \"status\" this is either \"dead\" or \"alive\" depending on whether\n there is a death date\n - \"age\" - current age or the age when dying.\n The age is calculated as full years.\n Birth 01.01.1940, death 01.01.2020 - age: 80\n Birth 02.01.1940, death 01.01.2020 - age: 79\n\n If there is no birth date, then the age is -1.\n\n When calculating age, dates can be compared.\n\n The lines in the files should be ordered:\n - first by the age ascending (younger before older);\n if the age cannot be calculated, then those lines will come last\n - if the age is the same, then those lines should be ordered\n by birthdate descending (newer birth before older birth)\n - if both the age and birth date are the same,\n then by name ascending (a before b). If name is not available, use \"\" (people with missing name should be before people with name)\n - if the names are the same or name field is missing,\n order by id ascending.\n\n Dates in the report should in the format: dd.mm.yyyy\n (2-digit day, 2-digit month, 4-digit year).\n\n :param person_data_directory: Directory of input data.\n :param report_filename: Output file.\n :return: None\n \"\"\"\n operate_with_dicts = []\n people_data = read_people_data(person_data_directory)\n\n for i in people_data.values():\n ret = {}\n for el in i:\n ret[el] = i[el]\n if i[el] is None:\n ret[el] = '-'\n if type(i[el]) is date:\n ret[el] = ret[el].strftime('%d.%m.%Y')\n\n if 'birth' in i and i['birth'] is not None and 'death' in i and i['death'] is not None:\n age = i['death'].year - i['birth'].year - (\n (i['death'].month, i['death'].day) < (i['birth'].month, i['birth'].day))\n ret['age'] = age\n if 'birth' in i and i['birth'] is not None and 'death' in i and i['death'] is None:\n today = date.today()\n age = today.year - i['birth'].year - ((today.month, today.day) < (i['birth'].month, i['birth'].day))\n ret['age'] = age\n i['birth'] = i['birth'].strftime(\"%d.%m.%Y\")\n\n if ret['birth'] == '-':\n ret['age'] = -1\n\n if ret['death'] != '-':\n ret['status'] = 'dead'\n else:\n ret['status'] = 'alive'\n\n operate_with_dicts.append(ret)\n\n sorted_list = sorted(operate_with_dicts, key=lambda i: (i['age'] if i['age'] > -1 else 10000,\n -date_to_int(i['birth']) if i['birth'] != '-' else i[\n 'birth'],\n i['name'] if 'name' in i else '',\n i['id']))\n\n return write_list_of_dicts_to_csv_file(report_filename, sorted_list)\n\n\ndef date_to_int(date) -> int:\n \"\"\"Convert date to int type number.\"\"\"\n d = date.split('.')\n d.reverse()\n joined_string = ''.join(d)\n return int(joined_string)\n\n\nif __name__ == '__main__':\n # print(read_csv_file_into_list_of_dicts('csv_town.txt'))\n # print(read_people_data('data'))\n print(generate_people_report('data', 'example_report.csv'))\n", "repo_name": "Danwerk/Python-course", "sub_path": "EX/ex07_files/file_handling.py", "file_name": "file_handling.py", "file_ext": "py", "file_size_in_byte": 17815, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 65, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 65, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 87, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 87, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 109, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 150, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 203, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 207, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 212, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 213, "usage_type": "name"}, {"api_name": "csv.writer", "line_number": 225, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 495, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 554, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 562, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 562, "usage_type": "name"}, {"api_name": "datetime.date.split", "line_number": 588, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 588, "usage_type": "name"}]}
+{"seq_id": "17683608701", "text": "import hashlib\nimport json\nimport io\nimport random\n\nimport ddddocr\nfrom PIL import Image\nimport execjs\nimport requests\nimport time\n\nfrom loguru import logger\nsession = requests.session()\nencrypt = open('encrypt.js', encoding='utf-8').read()\nencrypt = execjs.compile(encrypt)\nencrypt_param = open('encrypt_param.js', encoding='utf-8').read()\nencrypt_param = execjs.compile(encrypt_param)\nocr = ddddocr.DdddOcr(det=False, ocr=False, show_ad=False)\n\n\ndef _img_restore(img: bytes) -> bytes:\n \"\"\"\n 底图还原\n \"\"\"\n image = Image.open(io.BytesIO(img))\n standard_img = Image.new(\"RGBA\", (260, 160))\n position = [39, 38, 48, 49, 41, 40, 46, 47, 35, 34, 50, 51, 33, 32, 28, 29, 27, 26, 36, 37, 31, 30, 44, 45, 43,\n 42, 12, 13, 23, 22, 14, 15, 21, 20, 8, 9, 25, 24, 6, 7, 3, 2, 0, 1, 11, 10, 4, 5, 19, 18, 16, 17]\n s, u = 80, 10\n for c in range(52):\n a = position[c] % 26 * 12 + 1\n b = s if position[c] > 25 else 0\n im = image.crop(box=(a, b, a + 10, b + 80))\n standard_img.paste(im, box=(c % 26 * 10, 80 if c > 25 else 0))\n temp_io = io.BytesIO()\n standard_img.save(temp_io, format='png')\n return temp_io.getvalue()\n\n\ndef register_slide():\n params = {\n 't': str(time.time()).replace(\".\", \"\")[:13],\n }\n print(params)\n url = 'https://www.geetest.com/demo/gt/register-slide-official'\n response = session.get(\n url,\n params=params,\n )\n if response.status_code == 200:\n if response.json().get('success') == 1:\n gt = response.json().get('gt')\n challenge = response.json().get('challenge')\n logger.success(f\"注册成功: gt {gt} challenge {challenge}\")\n return gt, challenge\n else:\n logger.error(f\"注册失败\")\n else:\n logger.error(f\"{url} 接口异常非200状态码\")\n\n\ndef get_image(gt, challenge):\n params = {\n 'is_next': 'true',\n 'type': 'slide3',\n 'gt': gt,\n 'challenge': challenge,\n 'lang': 'zh-cn',\n 'https': 'true',\n 'protocol': 'https://',\n 'offline': 'false',\n 'product': 'embed',\n 'api_server': 'api.geevisit.com',\n 'isPC': 'true',\n 'autoReset': 'true',\n 'width': '100%',\n 'callback': 'geetest_1693994409960',\n }\n\n # url = 'https://api.geetest.com/get.php'\n url = 'https://api.geevisit.com/get.php'\n response = session.get(url, params=params)\n # print(response.content.decode(\"unicode_escape\"))\n if response.status_code == 200:\n if 'bg' in response.text and 'fullbg' in response.text:\n data = response.text.replace('geetest_1693994409960(', \"\")[:-1]\n json_Data = json.loads(data)\n bg = 'https://static.geetest.com/' + json_Data.get('bg').replace(\".jpg\", '.webp')\n slice = 'https://static.geetest.com/' + json_Data.get('slice').replace(\".jpg\", '.webp')\n challenge = json_Data.get(\"challenge\")\n try:\n bg_reponse = session.get(bg)\n slice_reponse = session.get(slice)\n hy_image = _img_restore(bg_reponse.content)\n with open(\"底图还原.png\", \"wb\") as f:\n f.write(hy_image)\n with open(\"小滑块.png\", \"wb\") as f:\n f.write(slice_reponse.content)\n logger.success('成功获取混淆底图')\n logger.success(f\"返回新的challenge: {challenge}\")\n return challenge, json_Data.get('s')\n except:\n logger.error('保存混淆底图失败')\n else:\n logger.error(f\"{url} 接口异常非200状态码\")\n\n\ndef ajax(gt, challenge):\n response = session.get(\n f'https://api.geevisit.com/ajax.php?gt={gt}&challenge={challenge}&lang=zh-cn&pt=0&client_type=web',\n )\n # print(response.text)\n\ndef _track(distance: int):\n def __ease_out_expo(step):\n return 1 if step == 1 else 1 - pow(2, -10 * step)\n if not isinstance(distance, int) or distance < 0:\n raise ValueError(f\"distance类型必须是大于等于0的整数: distance: {distance}, type: {type(distance)}\")\n # 初始化轨迹列表\n slide_track = [[random.randint(-50, -10), random.randint(-50, -10), 0], [0, 0, 0],]\n # 共记录count次滑块位置信息\n count = 30 + int(distance / 2)\n # 初始化滑动时间\n t = random.randint(50, 100)\n # 记录上一次滑动的距离\n _x, _y = 0, 0\n for i in range(count):\n # 已滑动的横向距离\n x = round(__ease_out_expo(i / count) * distance)\n # 滑动过程消耗的时间\n t += random.randint(10, 20)\n if x == _x:\n continue\n slide_track.append([x, _y, t])\n _x = x\n slide_track.append(slide_track[-1])\n passTime = slide_track[-1][-1]\n return slide_track, passTime\n\ndef guiji(offset):\n # break_flag = x + offset\n guiji_array = []\n array_one = [-31, -31, 0]\n guiji_array.append(array_one)\n x, y, time_x = 0, 0, 0\n while x < offset:\n i = [x, y, time_x]\n x += random.randint(0, 2)\n y += random.randint(-1, 1)\n time_x += random.randint(0, 15)\n guiji_array.append(i)\n guiji_array.append(i) #\n logger.info(f\"原轨迹长度: {len(guiji_array)}\")\n logger.info(f\"原轨迹: {guiji_array}\")\n return guiji_array\n\n\ndef slide_verify(gt, challenge, guiji, t, passtime, s):\n encrypt_params = encrypt_param.call(\"get_aes_params\", guiji, t, gt, challenge, passtime, s)\n encrypt_params['rp'] = hashlib.md5(encrypt_params.get('rp').encode()).hexdigest()\n w = encrypt.call('get_w', json.dumps(encrypt_params))\n\n params = {\n 'gt': gt,\n 'challenge': challenge,\n 'lang': 'zh-cn',\n '%24_BCX': '0',\n 'client_type': 'web',\n \"w\": w,\n 'callback': 'geetest_1694058231021'\n }\n response = session.get(\n 'https://api.geevisit.com/ajax.php', params=params\n )\n print(response.text)\n\ndef loc_image(bg_path, block_path):\n try:\n # bg_img = Image.open(bg_path)\n # block_img = Image.open(block_path)\n target_content = open(bg_path, \"rb\").read()\n background_content = open(block_path, \"rb\").read()\n res = ocr.slide_match(target_content, background_content, simple_target=True)\n logger.success(\"识别成功: {}\".format(res))\n return res.get('target')[0]\n except:\n logger.error(\"ddddd识别异常\")\n\n\nif __name__ == '__main__':\n for i in range(5):\n gt, challenge = register_slide()\n ajax(gt, challenge)\n challenge, s = get_image(gt, challenge)\n loc = loc_image('底图还原.png', '小滑块.png')\n guiji_array, passtime = _track(loc)\n slide_verify(gt, challenge, guiji_array, loc, passtime, s)\n\n\n\n\n\n\n", "repo_name": "renxiaoyao798/gpss_learn_reverse", "sub_path": "极验三代/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 6819, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "github-code", "pt": "27", "api": [{"api_name": "requests.session", "line_number": 13, "usage_type": "call"}, {"api_name": "execjs.compile", "line_number": 15, "usage_type": "call"}, {"api_name": "execjs.compile", "line_number": 17, "usage_type": "call"}, {"api_name": "ddddocr.DdddOcr", "line_number": 18, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 25, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 25, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 25, "usage_type": "call"}, {"api_name": "PIL.Image.new", "line_number": 26, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 26, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 35, "usage_type": "call"}, {"api_name": "time.time", "line_number": 42, "usage_type": "call"}, {"api_name": "loguru.logger.success", "line_number": 54, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 54, "usage_type": "name"}, {"api_name": "loguru.logger.error", "line_number": 57, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 57, "usage_type": "name"}, {"api_name": "loguru.logger.error", "line_number": 59, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 59, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 87, "usage_type": "call"}, {"api_name": "loguru.logger.success", "line_number": 99, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 99, "usage_type": "name"}, {"api_name": "loguru.logger.success", "line_number": 100, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 100, "usage_type": "name"}, {"api_name": "loguru.logger.error", "line_number": 103, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 103, "usage_type": "name"}, {"api_name": "loguru.logger.error", "line_number": 105, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 105, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 120, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 124, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 131, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 148, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 149, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 150, "usage_type": "call"}, {"api_name": "loguru.logger.info", "line_number": 153, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 153, "usage_type": "name"}, {"api_name": "loguru.logger.info", "line_number": 154, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 154, "usage_type": "name"}, {"api_name": "hashlib.md5", "line_number": 160, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 161, "usage_type": "call"}, {"api_name": "loguru.logger.success", "line_number": 184, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 184, "usage_type": "name"}, {"api_name": "loguru.logger.error", "line_number": 187, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 187, "usage_type": "name"}]}
+{"seq_id": "24766561534", "text": "from django.http import Http404\nfrom rest_framework import status, permissions, generics, filters\nfrom rest_framework.views import APIView\nfrom rest_framework.response import Response\nfrom .models import Project, Pledge, Category\nfrom .serializers import ProjectSerializer, PledgeSerializer, ProjectDetailSerializer, CategoryProjectSerializer, CategorySerializer, ProjectTotalSerializer\nfrom .permissions import IsOwnerOrReadOnly, isSuperUser\n\nclass CategoryList(APIView):\n permission_classes = [isSuperUser,permissions.IsAuthenticatedOrReadOnly]\n\n def get(self, request):\n categories = Category.objects.all()\n serializer = CategorySerializer(categories, many=True)\n return Response(serializer.data)\n \n def post(self, request):\n serializer = CategorySerializer(data=request.data)\n \n if serializer.is_valid():\n print(request.user.is_superuser)\n\n if (request.user.is_superuser):\n\n serializer.save()\n return Response(\n serializer.data,\n status=status.HTTP_201_CREATED\n )\n return Response(\n serializer.errors,\n status=status.HTTP_401_UNAUTHORIZED\n )\n\nclass ProjectList(APIView):\n permission_classes = [permissions.IsAuthenticatedOrReadOnly]\n\n def get(self, request):\n projects = Project.objects.all()\n serializer = ProjectSerializer(projects, many=True)\n return Response(serializer.data)\n\n def post(self, request):\n serializer = ProjectSerializer(data=request.data)\n\n if serializer.is_valid():\n serializer.save(owner=request.user)\n return Response(\n serializer.data,\n status=status.HTTP_201_CREATED\n )\n return Response(\n serializer.errors,\n status=status.HTTP_400_BAD_REQUEST\n )\n\nclass FilterView(generics.ListAPIView):\n serializer_class = ProjectSerializer\n queryset = Project.objects.all()\n filter_backends = [filters.OrderingFilter]\n ordering_fields = ['category', 'deadline', 'owner']\n # def get_queryset(self):\n # queryset = Project.objects.all()\n # ...\n # category = self.request.query_params.get('category', None)\n # date_created = self.request.query_params.get('date_created', None)\n # ...\n # if category is not None:\n # queryset = queryset.filter(category=category)\n # if date_created is not None:\n # queryset = queryset.filter(date_created=date_created)\n # return queryset\n\nclass CategoryProject(generics.RetrieveAPIView):\n permission_classes = [isSuperUser]\n queryset = Category.objects.all()\n serializer_class = CategoryProjectSerializer\n lookup_field = 'category'\n\nclass ProjectDetail(APIView):\n permission_classes = [permissions.IsAuthenticatedOrReadOnly, IsOwnerOrReadOnly]\n queryset = Project.objects.all()\n serializer_class = ProjectDetailSerializer\n\n def get_object(self, pk):\n try:\n return Project.objects.get(pk=pk)\n except Project.DoesNotExist:\n raise Http404\n \n def get(self, request, pk):\n project = self.get_object(pk)\n serializer = ProjectDetailSerializer(project)\n return Response(serializer.data)\n\n def put(self, request, pk):\n project = self.get_object(pk)\n self.check_object_permissions(request, project)\n data = request.data\n serializer = ProjectDetailSerializer(\n instance=project,\n data=data,\n partial=True\n )\n if serializer.is_valid():\n serializer.save()\n return Response(\n serializer.data,\n status=status.HTTP_201_CREATED\n )\n return Response(\n serializer.errors,\n status=status.HTTP_400_BAD_REQUEST\n )\n\n def delete(self, request, pk):\n project = self.get_object(pk)\n self.check_object_permissions(request, project)\n\n try:\n project.delete()\n return Response(status = status.HTTP_204_NO_CONTENT)\n except Project.DoesNotExist:\n raise Http404\n\nclass PledgeList(APIView):\n permission_classes = [permissions.IsAuthenticatedOrReadOnly]\n\n def get(self, request):\n pledges = Pledge.objects.all()\n serializer = PledgeSerializer(pledges, many=True)\n return Response(serializer.data)\n \n def post(self, request):\n serializer = PledgeSerializer(data=request.data)\n\n if serializer.is_valid():\n serializer.save(supporter=request.user)\n return Response(\n serializer.data,\n status=status.HTTP_201_CREATED\n )\n return Response(\n serializer.errors,\n status=status.HTTP_400_BAD_REQUEST\n )\n\nclass PledgeDetail(APIView):\n permission_classes = [permissions.IsAuthenticatedOrReadOnly,IsOwnerOrReadOnly]\n\n def get_object(self, pk):\n try:\n return Pledge.objects.get(pk=pk)\n except Pledge.DoesNotExist:\n raise Http404\n \n def get(self, request, pk):\n pledge = self.get_object(pk)\n serializer = PledgeSerializer(pledge)\n return Response(serializer.data)\n\n def put(self, request, pk):\n pledge = self.get_object(pk)\n self.check_object_permissions(request, pledge)\n data = request.data\n serializer = PledgeSerializer(\n instance=pledge,\n data=data,\n partial=True\n )\n if serializer.is_valid():\n serializer.save()\n return Response(\n serializer.data, \n status = status.HTTP_201_CREATED\n )\n return Response (\n serializer.errors,\n status=status.HTTP_400_BAD_REQUEST\n )\n \n def delete(self, request, pk):\n pledge = self.get_object(pk)\n self.check_object_permissions(request, pledge)\n\n try:\n pledge.delete()\n return Response(status = status.HTTP_204_NO_CONTENT)\n except Pledge.DoesNotExist:\n raise Http404\n\nclass ProjectTotals(APIView):\n \n # permission_classes = [permissions.IsAuthenticatedOrReadOnly]\n\n def get(self, request):\n projects = Project.objects.all()\n serializer = ProjectTotalSerializer(projects, many=True)\n return Response(serializer.data)", "repo_name": "SamaraLove/drf", "sub_path": "crowdfunding/projects/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 6500, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "rest_framework.views.APIView", "line_number": 9, "usage_type": "name"}, {"api_name": "permissions.isSuperUser", "line_number": 10, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticatedOrReadOnly", "line_number": 10, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 10, "usage_type": "name"}, {"api_name": "models.Category.objects.all", "line_number": 13, "usage_type": "call"}, {"api_name": "models.Category.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "models.Category", "line_number": 13, "usage_type": "name"}, {"api_name": "serializers.CategorySerializer", "line_number": 14, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 15, "usage_type": "call"}, {"api_name": "serializers.CategorySerializer", "line_number": 18, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 26, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 28, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 28, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 30, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_401_UNAUTHORIZED", "line_number": 32, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 32, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 35, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticatedOrReadOnly", "line_number": 36, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 36, "usage_type": "name"}, {"api_name": "models.Project.objects.all", "line_number": 39, "usage_type": "call"}, {"api_name": "models.Project.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "models.Project", "line_number": 39, "usage_type": "name"}, {"api_name": "serializers.ProjectSerializer", "line_number": 40, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 41, "usage_type": "call"}, {"api_name": "serializers.ProjectSerializer", "line_number": 44, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 48, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 50, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 50, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 52, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 54, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 54, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 57, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 57, "usage_type": "name"}, {"api_name": "serializers.ProjectSerializer", "line_number": 58, "usage_type": "name"}, {"api_name": "models.Project.objects.all", "line_number": 59, "usage_type": "call"}, {"api_name": "models.Project.objects", "line_number": 59, "usage_type": "attribute"}, {"api_name": "models.Project", "line_number": 59, "usage_type": "name"}, {"api_name": "rest_framework.filters.OrderingFilter", "line_number": 60, "usage_type": "attribute"}, {"api_name": "rest_framework.filters", "line_number": 60, "usage_type": "name"}, {"api_name": "rest_framework.generics.RetrieveAPIView", "line_number": 74, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 74, "usage_type": "name"}, {"api_name": "permissions.isSuperUser", "line_number": 75, "usage_type": "name"}, {"api_name": "models.Category.objects.all", "line_number": 76, "usage_type": "call"}, {"api_name": "models.Category.objects", "line_number": 76, "usage_type": "attribute"}, {"api_name": "models.Category", "line_number": 76, "usage_type": "name"}, {"api_name": "serializers.CategoryProjectSerializer", "line_number": 77, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 80, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticatedOrReadOnly", "line_number": 81, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 81, "usage_type": "name"}, {"api_name": "permissions.IsOwnerOrReadOnly", "line_number": 81, "usage_type": "name"}, {"api_name": "models.Project.objects.all", "line_number": 82, "usage_type": "call"}, {"api_name": "models.Project.objects", "line_number": 82, "usage_type": "attribute"}, {"api_name": "models.Project", "line_number": 82, "usage_type": "name"}, {"api_name": "serializers.ProjectDetailSerializer", "line_number": 83, "usage_type": "name"}, {"api_name": "models.Project.objects.get", "line_number": 87, "usage_type": "call"}, {"api_name": "models.Project.objects", "line_number": 87, "usage_type": "attribute"}, {"api_name": "models.Project", "line_number": 87, "usage_type": "name"}, {"api_name": "models.Project.DoesNotExist", "line_number": 88, "usage_type": "attribute"}, {"api_name": "models.Project", "line_number": 88, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 89, "usage_type": "name"}, {"api_name": "serializers.ProjectDetailSerializer", "line_number": 93, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 94, "usage_type": "call"}, {"api_name": "serializers.ProjectDetailSerializer", "line_number": 100, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 107, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 109, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 109, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 111, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 113, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 113, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 122, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_204_NO_CONTENT", "line_number": 122, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 122, "usage_type": "name"}, {"api_name": "models.Project.DoesNotExist", "line_number": 123, "usage_type": "attribute"}, {"api_name": "models.Project", "line_number": 123, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 124, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 126, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticatedOrReadOnly", "line_number": 127, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 127, "usage_type": "name"}, {"api_name": "models.Pledge.objects.all", "line_number": 130, "usage_type": "call"}, {"api_name": "models.Pledge.objects", "line_number": 130, "usage_type": "attribute"}, {"api_name": "models.Pledge", "line_number": 130, "usage_type": "name"}, {"api_name": "serializers.PledgeSerializer", "line_number": 131, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 132, "usage_type": "call"}, {"api_name": "serializers.PledgeSerializer", "line_number": 135, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 139, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 141, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 141, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 143, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 145, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 145, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 148, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticatedOrReadOnly", "line_number": 149, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 149, "usage_type": "name"}, {"api_name": "permissions.IsOwnerOrReadOnly", "line_number": 149, "usage_type": "name"}, {"api_name": "models.Pledge.objects.get", "line_number": 153, "usage_type": "call"}, {"api_name": "models.Pledge.objects", "line_number": 153, "usage_type": "attribute"}, {"api_name": "models.Pledge", "line_number": 153, "usage_type": "name"}, {"api_name": "models.Pledge.DoesNotExist", "line_number": 154, "usage_type": "attribute"}, {"api_name": "models.Pledge", "line_number": 154, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 155, "usage_type": "name"}, {"api_name": "serializers.PledgeSerializer", "line_number": 159, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 160, "usage_type": "call"}, {"api_name": "serializers.PledgeSerializer", "line_number": 166, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 173, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 175, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 175, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 177, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 179, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 179, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 188, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_204_NO_CONTENT", "line_number": 188, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 188, "usage_type": "name"}, {"api_name": "models.Pledge.DoesNotExist", "line_number": 189, "usage_type": "attribute"}, {"api_name": "models.Pledge", "line_number": 189, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 190, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 192, "usage_type": "name"}, {"api_name": "models.Project.objects.all", "line_number": 197, "usage_type": "call"}, {"api_name": "models.Project.objects", "line_number": 197, "usage_type": "attribute"}, {"api_name": "models.Project", "line_number": 197, "usage_type": "name"}, {"api_name": "serializers.ProjectTotalSerializer", "line_number": 198, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 199, "usage_type": "call"}]}
+{"seq_id": "16047760362", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[26]:\n\n\n# Dependencies and Setup\nfrom bs4 import BeautifulSoup as bs\nfrom splinter import Browser\nimport pandas as pd\nfrom webdriver_manager.chrome import ChromeDriverManager\nimport requests\nimport os\nimport time\n\n\n# In[2]:\n\n\nexecutable_path = {'executable_path': ChromeDriverManager().install()}\nbrowser = Browser('chrome', **executable_path, headless=False)\n\n\n# In[3]:\n\n\n# Visit the NASA Mars News Site\nurl = \"https://mars.nasa.gov/news/\"\nbrowser.visit(url)\n\n\n# In[4]:\n\n\n# Parse Results HTML with BeautifulSoup\n# Find Everything Inside:\n#
\n#
\n\nhtml = browser.html\nnews_soup = bs(html, \"html.parser\")\nslide_element = news_soup.select_one(\"ul.item_list li.slide\")\n\n\n# In[5]:\n\n\nslide_element.find(\"div\", class_=\"content_title\")\n\n\n# In[6]:\n\n\n# Scrape the Latest News Title\n\nnews_title = slide_element.find(\"div\", class_=\"content_title\").get_text()\nprint(news_title)\n\n\n# In[7]:\n\n\n# Scrape the Latest Paragraph Text\nnews_paragraph = slide_element.find(\"div\", class_=\"article_teaser_body\").get_text()\nprint(news_paragraph)\n\n\n# In[8]:\n\n\n# FEATURED IMAGE\n# NASA JPL Site\nexecutable_path = {'executable_path': ChromeDriverManager().install()}\nbrowser = Browser('chrome', **executable_path, headless=False)\n\n\n# In[9]:\n\n\nurl = 'https://www.jpl.nasa.gov/spaceimages/'\nbrowser.visit(url)\n\n\n# In[10]:\n\n\n# Create a Beautiful Soup object\nhtml = browser.html\nsoup = bs(html, 'html.parser')\n\n\n# In[11]:\n\n\n#Get the images\nimages = soup.findAll('img')\nexample = images[0]\nexample\n\n\n# In[12]:\n\n\n# Use Base URL to Create Absolute URL\nimg_url = f\"https://www.jpl.nasa.gov{images}\"\nprint(img_url)\n\n\n# In[13]:\n\n\n#MARS WEATHER\n\nurl = \"https://twitter.com/marswxreport?lang=en\"\nbrowser.visit(url)\n\n\n# In[15]:\n\n\n# HTML Object \nhtml_weather = browser.html\n\n# Parse HTML with Beautiful Soup\nsoup = bs(html_weather, 'html.parser')\n\n# Find elements that contain tweets\nlatest_tweets = soup.find_all('div', class_='js-tweet-text-container')\n\n# Retrieve all elements that contain news title in the specified range\n# Look for entries that display weather related words to exclude non weather related tweets \nfor tweet in latest_tweets: \n weather_tweet = tweet.find('p').text\n if 'Sol' and 'pressure' in weather_tweet:\n print(weather_tweet)\n break\n else: \n pass\n\n\n# In[19]:\n\n\n#MARS FACTS\n\n# Visit Mars facts url \nfacts_url = 'http://space-facts.com/mars/'\n\n# Use Panda's `read_html` to parse the url\nmars_facts = pd.read_html(facts_url)\n\n# Find the mars facts DataFrame in the list of DataFrames as assign it to `mars_df`\nmars_df = mars_facts[0]\n\n# Assign the columns `['Description', 'Value']`\nmars_df.columns = ['Description','Value']\n\n# Set the index to the `Description` column without row indexing\nmars_df.set_index('Description', inplace=True)\n\n# Save html code to folder Assets\nmars_df.to_html()\n\ndata = mars_df.to_dict(orient='records') \n\n# Display mars_df\nmars_df\n\n\n# In[30]:\n\n\n#MARS HEMISPHERES\n\n# Go to hemisphere website through splinter module \nurl = 'https://astrogeology.usgs.gov/search/results?q=hemisphere+enhanced&k1=target&v1=Mars'\nbrowser.visit(url)\n\n\n# In[31]:\n\n\nhemisphere_image_urls = []\n\n# Get a List of All the Hemispheres\nlinks = browser.find_by_css(\"a.product-item h3\")\nfor item in range(len(links)):\n hemisphere = {}\n \n # Find Element on Each Loop to Avoid a Stale Element Exception\n browser.find_by_css(\"a.product-item h3\")[item].click()\n \n # Find Sample Image Anchor Tag & Extract \n sample_element = browser.find_link_by_text(\"Sample\").first\n hemisphere[\"img_url\"] = sample_element[\"href\"]\n \n # Get Hemisphere Title\n hemisphere[\"title\"] = browser.find_by_css(\"h2.title\").text\n \n # Append Hemisphere Object to List\n hemisphere_image_urls.append(hemisphere)\n \n # Navigate Backwards\n browser.back()\n\n\n# In[32]:\n\n\nhemisphere_image_urls\n\n\n# In[ ]:\n\n\n\n\n", "repo_name": "gmichallet/web-scraping-challenge", "sub_path": "Scrape_mars.py", "file_name": "Scrape_mars.py", "file_ext": "py", "file_size_in_byte": 3919, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "webdriver_manager.chrome.ChromeDriverManager", "line_number": 20, "usage_type": "call"}, {"api_name": "splinter.Browser", "line_number": 21, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 41, "usage_type": "call"}, {"api_name": "webdriver_manager.chrome.ChromeDriverManager", "line_number": 73, "usage_type": "call"}, {"api_name": "splinter.Browser", "line_number": 74, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 89, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 125, "usage_type": "call"}, {"api_name": "pandas.read_html", "line_number": 150, "usage_type": "call"}]}
+{"seq_id": "42239669035", "text": "from django.urls import path\nfrom .views import PostFeedbackList,PostList, PostCreate, PostDetailView,PostUpdate,PostFeedback,PostDelete\nfrom . import views\n\nurlpatterns = [\n path('', PostList.as_view(), name='post_list'),\n path('post/new/', PostCreate.as_view(), name='post_new'),\n path('post//', PostDetailView.as_view(), name='post_detail'),\n path('post//edit/', PostUpdate.as_view(), name='post_edit'),\n path('post/feedback/', PostFeedback.as_view(), name='feedback'),\n path('post/feedback/list/', PostFeedbackList.as_view(), name='feedbackList'),\n path('post//delete/', PostDelete.as_view(), name='post_delete'),\n\n]\n", "repo_name": "InesM95/djangogirlForms", "sub_path": "blog/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 666, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "views.PostList.as_view", "line_number": 6, "usage_type": "call"}, {"api_name": "views.PostList", "line_number": 6, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "views.PostCreate.as_view", "line_number": 7, "usage_type": "call"}, {"api_name": "views.PostCreate", "line_number": 7, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "views.PostDetailView.as_view", "line_number": 8, "usage_type": "call"}, {"api_name": "views.PostDetailView", "line_number": 8, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "views.PostUpdate.as_view", "line_number": 9, "usage_type": "call"}, {"api_name": "views.PostUpdate", "line_number": 9, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "views.PostFeedback.as_view", "line_number": 10, "usage_type": "call"}, {"api_name": "views.PostFeedback", "line_number": 10, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "views.PostFeedbackList.as_view", "line_number": 11, "usage_type": "call"}, {"api_name": "views.PostFeedbackList", "line_number": 11, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "views.PostDelete.as_view", "line_number": 12, "usage_type": "call"}, {"api_name": "views.PostDelete", "line_number": 12, "usage_type": "name"}]}
+{"seq_id": "12031757049", "text": "import pyaudio\nimport numpy as np\nfrom presto.channels.channel import Channel\nimport struct\n\npa = pyaudio.PyAudio()\n\nDEFAULT_INFO = pa.get_default_host_api_info()\n\n\nclass MicrophoneAudio(Channel):\n def __init__(\n self,\n frames_per_buffer=1024,\n input_device_index=DEFAULT_INFO[\"defaultInputDevice\"],\n rate=48000,\n data_queue=None,\n ):\n super().__init__()\n self._rate = rate\n self._frames_per_buffer = frames_per_buffer\n self._data_queue = data_queue\n self._input_device_index = input_device_index\n self._stream = None\n\n def run(self):\n # read input stream\n self._stream = pa.open(\n rate=self._rate,\n channels=1,\n format=pyaudio.paInt16,\n input=True,\n input_device_index=self._input_device_index,\n frames_per_buffer=self._frames_per_buffer,\n )\n\n while True:\n data = self._stream.read(self._rate)\n samps = np.fromstring(data, dtype=np.int16)\n if self._data_queue is not None:\n self._data_queue.put(samps)\n\n def shutdown(self):\n if self._stream is not None:\n self._stream.stop_stream()\n self._stream.close()\n\n\nif __name__ == \"__main__\":\n listener = MicrophoneAudio()\n listener.start()\n", "repo_name": "jacquelinegarrahan/presto", "sub_path": "presto/channels/microphone.py", "file_name": "microphone.py", "file_ext": "py", "file_size_in_byte": 1348, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "pyaudio.PyAudio", "line_number": 6, "usage_type": "call"}, {"api_name": "presto.channels.channel.Channel", "line_number": 11, "usage_type": "name"}, {"api_name": "pyaudio.paInt16", "line_number": 31, "usage_type": "attribute"}, {"api_name": "numpy.fromstring", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 39, "usage_type": "attribute"}]}
+{"seq_id": "13587714913", "text": "import dash_core_components as dcc\nimport dash_html_components as html\nfrom dash.dependencies import Input, Output\n\nfrom app import app\nfrom apps import BPD, CLR , NT\n\n\nindex_page = html.Div([\n dcc.Link('Annotate BPD', href='/apps/BPD'),\n html.Br(),\n dcc.Link('Annotate CLR', href='/apps/CLR'),\n html.Br(),\n dcc.Link('Annotate NT', href='/apps/NT'),\n])\n\napp.layout = html.Div([\n dcc.Location(id='url', refresh=False),\n html.Div(id='page-content')\n])\n\n\n@app.callback(Output('page-content', 'children'),\n Input('url', 'pathname'))\ndef display_page(pathname):\n if pathname == '/apps/BPD':\n return BPD.layout\n elif pathname == '/apps/CLR':\n return CLR.layout\n elif pathname == '/apps/NT':\n return NT.layout\n else:\n return index_page\n\nif __name__ == '__main__':\n app.run_server(debug=True)", "repo_name": "dmalagarriga/annotate_ecography_images", "sub_path": "index.py", "file_name": "index.py", "file_ext": "py", "file_size_in_byte": 861, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "25", "api": [{"api_name": "dash_html_components.Div", "line_number": 9, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 10, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 11, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 12, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 13, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 14, "usage_type": "call"}, {"api_name": "app.app.layout", "line_number": 17, "usage_type": "attribute"}, {"api_name": "app.app", "line_number": 17, "usage_type": "name"}, {"api_name": "dash_html_components.Div", "line_number": 17, "usage_type": "call"}, {"api_name": "dash_core_components.Location", "line_number": 18, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 19, "usage_type": "call"}, {"api_name": "apps.BPD.layout", "line_number": 27, "usage_type": "attribute"}, {"api_name": "apps.BPD", "line_number": 27, "usage_type": "name"}, {"api_name": "apps.CLR.layout", "line_number": 29, "usage_type": "attribute"}, {"api_name": "apps.CLR", "line_number": 29, "usage_type": "name"}, {"api_name": "apps.NT.layout", "line_number": 31, "usage_type": "attribute"}, {"api_name": "apps.NT", "line_number": 31, "usage_type": "name"}, {"api_name": "app.app.callback", "line_number": 23, "usage_type": "call"}, {"api_name": "app.app", "line_number": 23, "usage_type": "name"}, {"api_name": "dash.dependencies.Output", "line_number": 23, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 24, "usage_type": "call"}, {"api_name": "app.app.run_server", "line_number": 36, "usage_type": "call"}, {"api_name": "app.app", "line_number": 36, "usage_type": "name"}]}
+{"seq_id": "35117307709", "text": "#!/usr/bin/python3.5\n# -*- coding: utf-8 -*-\n\nfrom multiprocessing import Event,Queue,Pool,Process,Value\nimport io,os,sys,time,random,threading,queue,pgbar\n\ndef eq_put_y(a,b,c):\n\tif a == 0:\n\t\tyield 0\n\tif c == 1:\n\t\tfor i in range(c):\n\t\t\tif a >= b:\n\t\t\t\ta-=b\n\t\t\t\tyield a\n\t\t\telse:\n\t\t\t\tyield 0\n\telif c > 1:\n\t\tfor i in range(c):\n\t\t\tif a >= b*c:\n\t\t\t\ta-=b\n\t\t\t\tyield a\n\t\t\telif a*c >= b/c and a < b*c:\n\t\t\t\tif a > c:\n\t\t\t\t\ta=a-(int(a/c)+a%c)\n\t\t\t\t\tyield a\n\t\t\t\telif a < c:\n\t\t\t\t\tyield 0\n\t\t\telif a*c < b/c:\n\t\t\t\t\tyield 0\n\ndef wq_put_y(a,b):\n\tfor i in range(a,b):\n\t\tyield i+1\n\ndef efunc():\n\tglobal task,wqs,procs,taskend\n\tprint('[efunc]event tid',threading.current_thread().name,'is starting...')\n\twhile True:\n\t\tif task != 0:\n\t\t\twhile not eq.full():\n\t\t\t\tif task == 0:\n\t\t\t\t\tbreak\n\t\t\t\teg=eq_put_y(task,wqs,procs)\n\t\t\t\teql=[]\n\t\t\t\t#print('[eq_put]task =',task)\n\t\t\t\teql.append(task)\n\t\t\t\ttask=next(eg)\n\t\t\t\teql.append(task)\n\t\t\t\t#print('[eq_put]eql :',eql)\n\t\t\t\teq.put(eql)\n\t\t\t\tet.set()\n\t\telif task == 0:\n\t\t\tbreak\n\t\tee.clear()\n\t\tee.wait()\n\t\tet.set()\n\t#print('[efunc]task =',task,'et set',et.is_set(),'| ee set',ee.is_set())\n\tn=0\n\twhile True:\t\n\t\tif not eq.full():\n\t\t\tn+=1\n\t\t\tif n <= procs:\n\t\t\t\teq.put('done')\n\t\t\t\tprint('[efunc]n =',n,'eq empty',eq.empty(),'ee set:',ee.is_set())\n\t\t\telif n > procs:\n\t\t\t\tee.clear()\n\t\t\t\tet.set()\n\t\t\t\ttaskend=True\n\t\t\t\tprint('[efunc]',threading.current_thread().name,'<1> et set:',et.is_set(),'| ee set:',ee.is_set())\n\t\t\t\treturn\n\t\tee.clear()\n\t\tee.wait()\n \ndef eq_get():\n\tglobal wqs,wg,procs,weqget,task\n\twqe=[]\n\twqa=None\n\twqb=None\n\t#print(threading.current_thread().name,'weqget =',weqget,'ee set',ee.is_set())\n\tif weqget:\n\t\tee.set()\n\t\twhile eq.empty():\n\t\t\tprint('[eq_get]',threading.current_thread().name,'wcq empty :',wcq.empty(),'weqget is',weqget,'| we set',we.is_set(),'| eq qsize:',eq.qsize())\n\t\t\tif not weqget or not ee.is_set():\n\t\t\t\tbreak\n\telse:\n\t\twfunc()\n\t\treturn\n\n\tif not eq.empty() and weqget:\n\t\twhile True:\n\t\t\ttry:\n\t\t\t\twqe=eq.get_nowait()\n\t\t\texcept:\n\t\t\t\tprint('[eq_get]',threading.current_thread().name,'wqe get failed|eq empty :',eq.empty(),'| wcq empty :',wcq.empty(),'|weqget:',weqget)\n\t\t\t\tif not weqget:\n\t\t\t\t\twfunc()\n\t\t\t\t\treturn\n\t\t\t\telse:\n\t\t\t\t\tee.set()\n\t\t\t\t\tcontinue\n\t\t\tbreak\n\t\tprint('<%.4f s>' % (time.time()-st),'| [eq_get]',threading.current_thread().name,'wqe=',wqe,'| eq empty :',eq.empty(),'| we set',we.is_set(),'| ee set:',ee.is_set(),'| wcq empty :',wcq.empty(),)\n\t\tif wqe != 'done' and wqe != []:\n\t\t\twqa=wqe.pop()\n\t\t\twqb=wqe.pop()\n\t\t\twg=wq_put_y(wqa,wqb)\n\t\t\tee.set()\n\t\t\twq_put()\n\t\telif wqe == 'done':\n\t\t\tweqget=False\n\t\t\tee.set()\n\t#print('[eq_get]',threading.current_thread().name,'return to wfunc','we set :',we.is_set(),'ee set :',ee.is_set(),'| eq empty :',eq.empty(),'wcq empty :',wcq.empty(),'weqget :',weqget)\n\twe.set()\n\twfunc()\n\ndef wq_put():\n\tglobal wg,weqget\n\tx=None\n\tif not we.is_set() and not wcq.empty():\n\t\twe.set()\n\tif weqget:\n\t\ttry:\n\t\t\tx=next(wg)\n\t\texcept:\n\t\t\tif not wcq.empty():\n\t\t\t\twcq.get_nowait()\n\t\t\t\twe.clear()\n\t\t\tprint('[wq_put]',threading.current_thread().name,'return to wfunc','wcq empty :',wcq.empty(),'| eq empty :',eq.empty(),'weqget =',weqget,'we set',we.is_set())\n\t\t\twfunc()\n\t\t\treturn\n\telse:\n\t\tprint('[wq_put]',threading.current_thread().name,'return to wfunc','wcq empty :',wcq.empty(),'| eq empty :',eq.empty(),'weqget =',weqget,'we set',we.is_set())\n\t\twfunc()\n\t\treturn\n\t#print('[wq_put]',threading.current_thread().name,'x =',x,'we set',we.is_set())\n\tif not wq.full() and x != None:\n\t\ttry:\n\t\t\twq.put_nowait(x)\n\t\texcept:\n\t\t\tpass\n\t#print('[wq_put]check 2 :',threading.current_thread().name)\n\twfunc()\n\ndef wfunc():\n\tglobal ptime,pcount,resbf,threadover,weqget,errlist,th_fin_c\n\tx=None\n\tn=0\n\twhile not wq.empty():\n\t\ttry:\n\t\t\tx=wq.get_nowait()\n\t\texcept:\n\t\t\tbreak\n\t\ttext=''\n\t\t#print('[wfunc]',threading.current_thread().name,'x =',x)\n\t\tr=random.randint(2,6)\n\t\tfor i in range(r):\n\t\t\ttext+='a'\n\t\t\ttime.sleep(1)\n\t\tstd=str(x)+'\\t'+text+'\\n'\n\t\ttry:\n\t\t\tresbf.put_nowait(std)\n\t\texcept:\n\t\t\terrlist.append(std)\n\t\t#print('[wfunc]',threading.current_thread().name,'std :',std)\n\t\twq.task_done()\n\t\tptime+=r\n\t\tpcount+=1\n\t\tn+=1\n\n\tif wcq.empty():\n\t\ttry:\n\t\t\twcq.put_nowait(threading.current_thread().name)\n\t\texcept:\n\t\t\twfunc()\n\t\t\treturn\n\t\tprint('[wfunc]return to eq_get()',threading.current_thread().name,'wcq empty',wcq.empty(),'we set',we.is_set(),'weqget =',weqget)\n\t\twe.clear()\n\t\teq_get()\n\t\treturn\n\telif not weqget and not wcq.empty():\n\t\t#print('[wfunc]check 3 alive thresds :',threading.current_thread().name,threading.current_thread().is_alive())\n\t\tif ee.is_set():\n\t\t\tee.clear()\n\t\treturn\n\twe.wait()\n\t#print('[wfunc]check 1 :',threading.current_thread().name)\n\twq_put()\n\ndef resbf_flush(ps):\n\tglobal resbf,reslog,errline,task,taskend\n\tprint('[resbf_flush]res_save_tid',threading.current_thread().name,'is starting...')\n\twhile True:\n\t\tresbfqs=resbf.qsize()\n\t\t#print('[resbf_flus]task!=0,resbf qsize =',resbfqs)\n\t\tif resbfqs >= ps:\n\t\t\tthqs=int((resbfqs-resbfqs%ps)/ps)\n\t\telse:\n\t\t\tthqs=resbfqs\n\t\t#print('[resbf_flus]',threading.current_thread().name,'task!=0,thqs =',thqs)\n\t\tfor i in range(thqs):\n\t\t\ttry:\n\t\t\t\tv=resbf.get_nowait()\n\t\t\texcept:\n\t\t\t\tcontinue\n\t\t\t#print('[resbf_flus]v =',v)\n\t\t\ttry:\n\t\t\t\treslog.write(v)\n\t\t\texcept:\n\t\t\t\terrline+=1\n\t\t\t\tprint('reslog write erroe at line:',errline,v)\n\t\treslog.flush()\n\t\t#print('[resbf_flush]et set:',et.is_set(),'taskend =',taskend)\n\t\tif taskend == False:\n\t\t\tet.clear()\n\t\t\tet.wait()\n\t\telif taskend == True:\n\t\t\tet.wait()\n\t\t\t#print('[resbf_flush]',threading.current_thread().name,'<2> et set:',et.is_set(),'| end set:',end.is_set())\n\t\t\tend.wait()\n\t\t\t#print('[resbf_flush]',threading.current_thread().name,'<3> et set:',et.is_set(),'| end set:',end.is_set())\n\t\t\t#print('[resbf_flush]',threading.current_thread().name,'last resbf size ;',resbf.qsize())\n\t\t\twhile not resbf.empty():\n\t\t\t\ttry:\n\t\t\t\t\tv=resbf.get_nowait()\n\t\t\t\texcept:\n\t\t\t\t\tcontinue\n\t\t\t\t#print('[resbf_flush]',threading.current_thread().name,'check point.....','et set:',et.is_set(),'| resbf empty :',resbf.empty(),'v=',v)\n\t\t\t\ttry:\n\t\t\t\t\treslog.write(v)\n\t\t\t\texcept:\n\t\t\t\t\terrline+=1\n\t\t\t\t\tprint('reslog write erroe at line:',errline,v)\n\t\t\treslog.flush()\n\t\t\tbreak\n\ndef wfunc_bar():\n\tglobal bartask,st\n\twhile True:\n\t\tcount=bar.value\n\t\tif count < bartask:\n\t\t\tpgbar.bar(bartask,count,50,st)\n\t\telif count >= bartask:\n\t\t\tpgbar.bar(bartask,count,50,st)\n\t\t\tee.set()\n\t\t\tbreak\n\t\t#print('count =',count)\n\t\ttime.sleep(0.5)\n\ndef c_e_th():\n\tglobal wths,procs\n\tthp=[]\n\tpevent=threading.Thread(target=efunc,name='pevent_tid='+str(os.getpid())+'/0')\n\tpevent.start()\n\tx=int(wths/1000)\n\tif wths/1000 <= 1:\n\t\tths=procs\n\telse:\n\t\tths=procs*x\n\tfor i in range(ths):\n\t\trbf=threading.Thread(target=resbf_flush,args=(ths,),name='res_save_tid='+str(os.getpid())+'/'+str(i+1))\n\t\tthp.append(rbf)\n\tfor a in thp:\n\t\ta.start()\n\tpevent.join()\n\tprint('\\n[c_e_th]there is no more task,efunc done,use time:%.2f' % (time.time()-st)+'s')\n\tprint('='*60)\n\tprint('[c_e_th]waiting for resbf thread over...')\n\tfor b in thp:\n\t\tb.join()\n\tprint('[c_e_th]errline =',errline)\n\tprint('[c_e_th]resbf is over......')\n\n\t#wbar=threading.Thread(target=wfunc_bar,name='wbar_tid='+str(os.getpid()))\n\t#wbar.start()\n\t#wbar.join()\n\ndef c_w_th(ths):\n\tthp=[]\n\tfor i in range(ths):\n\t\tt=threading.Thread(target=wfunc,name='tid'+str(os.getpid())+r'/'+str(i))\n\t\tthp.append(t)\n\tfor a in thp:\n\t\ta.start()\n\tfor b in thp:\n\t\tb.join()\n\tprint('\\n[c_w_th]',os.getpid(),'wq unfinished tasks :',wq.unfinished_tasks)\n\tprint('[c_w_th]ee set',ee.is_set(),'| resbf qsize:',resbf.qsize())\n\t\ndef pefunc():\n\tprint(os.getpid(),'pefunc is starting......')\n\tc_e_th()\n\tprint('[pefunc]pefunc done......')\n\ndef pwfunc():\n\tglobal allcount,alltime\n\tprint('[pwfunc]pid =',os.getpid(),'is running...')\n\tc_w_th(wths)\n\tallcount.value+=pcount\n\talltime.value+=ptime\n\tprint('\\n[pwfunc]pid='+str(os.getpid())+' real time: '+str(ptime)+'s\\tcounts:'+str(pcount))\n\tprint('[pwfunc]'+str(os.getpid())+' wfunc is done use time:%.2f' % (time.time()-st)+'s')\n\tprint('[pwfunc]wq empty :',wq.empty(),'|wq size :',wq.qsize(),'| errlist count :',len(errlist))\n\treturn os.getpid()\n\t\ndef delcache():\n\tcachedir='__pycache__'\n\ttry:\n\t\tos.chdir(cachedir)\n\texcept:\n\t\treturn\n\tflist=os.listdir()\n\twhile True:\n\t\ttry:\n\t\t\tos.remove(flist.pop())\n\t\texcept:\n\t\t\tos.rmdir('../'+cachedir)\n\t\t\tos.chdir('../')\n\t\t\treturn\ndef cb_test(test):\n\tglobal p_fin_c,procs\n\tp_fin_c.append(test)\n\tprint('[cb_test]',p_fin_c)\n\tif len(p_fin_c) == procs:\n\t\tpw.terminate()\n\nif __name__=='__main__':\n\tst=time.time()\n#public var set\n\tprocs=int(input('set procs:'))-1\n\tif procs <= 1:\n\t\tprocs=1\n\twths=int(input('set thread count:'))\n\twqs=wths\n\t#procs=os.cpu_count()\n\teq=Queue(procs)\n\ttask=int(input('set task count:'))\n\tbartask=task\n\talltime=Value('i',0)\n\tallcount=Value('i',0)\n\tbar=Value('i',1)\n\n#log file set\n\tdelcache()\n\tfname='./result.log'\n\ttry:\n\t\tos.remove(fname)\n\texcept:\n\t\tpass\n\tos.path.exists(fname)\n\treslog=open(fname,'a')\n\tresbf=Queue()\n\terrline=0\n\n#set var to event procs\n\tee=Event()\n\tend=Event()\n\tet=threading.Event()\n\ttaskend=False\n\tpe=Process(target=pefunc)\n\tpe.start()\n\tdel et,taskend,bartask\n\n#set var to work procs\n\twq=queue.Queue(wqs)\n\twcq=queue.Queue(1)\n\twe=threading.Event()\n\tweqget=True\n\twg=None\n\tptime=0\n\tpcount=0\n\terrlist=[]\n\tp_fin_c=[]\n\tpw=Pool(procs)\n\tfor i in range(procs):\n\t\tpw.apply_async(pwfunc,callback=cb_test)\n\tpw.close()\n\tpw.join()\n\tprint('pw is over......')\n\tend.set()\n\tprint('mainend end set :',end.is_set())\n\tpe.join()\n\tprint('unfinished resbf size ;',resbf.qsize())\n\treslog.close()\n\n\tprint('\\nreal time: '+str(alltime.value)+'s\\tcounts: '+str(allcount.value))\n\tprint('use time: %.2f' % (time.time()-st)+'s')", "repo_name": "seantbs/python3.5", "sub_path": "py-test/event_qw_no_fin.py", "file_name": "event_qw_no_fin.py", "file_ext": "py", "file_size_in_byte": 9515, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "threading.current_thread", "line_number": 37, "usage_type": "call"}, {"api_name": "threading.current_thread", "line_number": 69, "usage_type": "call"}, {"api_name": "threading.current_thread", "line_number": 83, "usage_type": "call"}, {"api_name": "threading.current_thread", "line_number": 95, "usage_type": "call"}, {"api_name": "time.time", "line_number": 103, "usage_type": "call"}, {"api_name": "threading.current_thread", "line_number": 103, "usage_type": "call"}, {"api_name": "threading.current_thread", "line_number": 129, "usage_type": "call"}, {"api_name": "threading.current_thread", "line_number": 133, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 156, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 159, "usage_type": "call"}, {"api_name": "threading.current_thread", "line_number": 173, "usage_type": "call"}, {"api_name": "threading.current_thread", "line_number": 177, "usage_type": "call"}, {"api_name": "threading.current_thread", "line_number": 192, "usage_type": "call"}, {"api_name": "pgbar.bar", "line_number": 242, "usage_type": "call"}, {"api_name": "pgbar.bar", "line_number": 244, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 248, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 253, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 253, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 261, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 261, "usage_type": "call"}, {"api_name": "time.time", "line_number": 266, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 281, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 281, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 287, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 291, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 297, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 301, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 302, "usage_type": "call"}, {"api_name": "time.time", "line_number": 302, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 304, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 309, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 312, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 315, "usage_type": "call"}, {"api_name": "os.rmdir", "line_number": 317, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 318, "usage_type": "call"}, {"api_name": "time.time", "line_number": 328, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 336, "usage_type": "call"}, {"api_name": "multiprocessing.Value", "line_number": 339, "usage_type": "call"}, {"api_name": "multiprocessing.Value", "line_number": 340, "usage_type": "call"}, {"api_name": "multiprocessing.Value", "line_number": 341, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 347, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 350, "usage_type": "call"}, {"api_name": "os.path", "line_number": 350, "usage_type": "attribute"}, {"api_name": "multiprocessing.Queue", "line_number": 352, "usage_type": "call"}, {"api_name": "multiprocessing.Event", "line_number": 356, "usage_type": "call"}, {"api_name": "multiprocessing.Event", "line_number": 357, "usage_type": "call"}, {"api_name": "threading.Event", "line_number": 358, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 360, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 365, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 366, "usage_type": "call"}, {"api_name": "threading.Event", "line_number": 367, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 374, "usage_type": "call"}, {"api_name": "time.time", "line_number": 387, "usage_type": "call"}]}
+{"seq_id": "39395553368", "text": "from __future__ import unicode_literals\nimport os\n\nfrom flask import Flask\nfrom flask import request\nfrom flask import jsonify\n\nfrom pbkdf2 import crypt\n\nfrom services.downloader import youtube_download\n\napp = Flask(__name__)\n\nHASHED_KEY = os.environ.get('YDL_HASHED_KEY', None)\nUSE_AUTH = os.environ.get('YDL_USE_AUTH', False)\nDL_PATH = os.environ.get('YDL_DL_PATH', '.')\n\n\n@app.route(\"/\", methods=[\"GET\"])\ndef handle_index():\n return handle_help()\n\n\n@app.route(\"/ydl\", methods=[\"GET\"])\ndef handle_download():\n\n youtubeId = request.args.get('id')\n onlyAudio = request.args.get('onlyAudio')\n key = request.args.get('key')\n\n if not youtubeId:\n return handle_help(400)\n\n # Authentication\n if USE_AUTH and not HASHED_KEY == crypt(key, HASHED_KEY):\n return response(500, 'incorrect key')\n\n ydl_opts = {\n 'format': 'bestaudio' if onlyAudio else 'best',\n 'outtmpl': DL_PATH+'/%(title)s.%(ext)s',\n 'nocheckcertificate': True\n }\n\n youtube_download(youtubeId, ydl_opts)\n return response(200, 'download started...')\n\n\n@app.route(\"/help\", methods=[\"GET\"])\ndef handle_help(code=200):\n return response(code, 'how to use this service: /ydl?id=&key=onlyAudio=false')\n\n\ndef response(code, message):\n response = jsonify({'message': message})\n response.status_code = code\n return response\n\nif __name__ == \"__main__\":\n app.run()", "repo_name": "mklan/remote-youtube-dl", "sub_path": "src/flask_app.py", "file_name": "flask_app.py", "file_ext": "py", "file_size_in_byte": 1422, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "flask.Flask", "line_number": 12, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 14, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 15, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 16, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask.request.args.get", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 29, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 29, "usage_type": "name"}, {"api_name": "pbkdf2.crypt", "line_number": 35, "usage_type": "call"}, {"api_name": "services.downloader.youtube_download", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 54, "usage_type": "call"}]}
+{"seq_id": "981578836", "text": "import numpy as np\nfrom PIL import Image, ImageFilter\n\nim = Image.open('./LLL.png').convert('L')\nimg = np.asarray(im)\nprint(img.dtype) # データ型\n\npil_img = Image.fromarray(img)\npil_img.save('./out.png')\n\nprint(img.size)", "repo_name": "Taiki-azrs/RaspiGPGPU_guide", "sub_path": "chap05/image_io_cpu.py", "file_name": "image_io_cpu.py", "file_ext": "py", "file_size_in_byte": 225, "program_lang": "python", "lang": "ar", "doc_type": "code", "stars": 22, "dataset": "github-code", "pt": "25", "api": [{"api_name": "PIL.Image.open", "line_number": 4, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 4, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 5, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 8, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 8, "usage_type": "name"}]}
+{"seq_id": "21600105875", "text": "import time\nimport click\nimport contextlib\nfrom typing import Callable, Iterable\n\ndef wait(iterations: int, interval: float, error: Exception=TimeoutError(), predicate: Callable[[], bool]=lambda: False) -> Iterable[int]:\n for iteration in range(iterations):\n time.sleep(interval)\n if predicate():\n break\n yield\n else:\n raise error\n\ndef block_wait(iterations: int, interval: float, error: Exception=TimeoutError(), predicate: Callable[[], bool]=lambda: False) -> None:\n for x in wait(iterations, interval, error, predicate):\n pass\n\ndef negate(f: Callable[[], bool]) -> bool:\n return lambda: not(f())\n\ndef first(iterable):\n return next(iter(iterable))\n\nclass Timer:\n def __init__(self):\n self.start = None\n self.end = None\n def __enter__(self):\n self.start = time.time()\n return self\n def __exit__(self, *exc):\n self.end = time.time()\n @property\n def running(self):\n return self.start and not self.end\n @property\n def elapsed(self):\n if not self.start:\n raise ValueError('not started')\n end = self.end or time.time()\n return end-self.start\n\nclass EchoTimer(Timer):\n def __init__(self, message):\n super().__init__()\n self.message = message\n def __enter__(self):\n super().__enter__()\n click.echo(self.message+' ', nl=False)\n def __exit__(self, *exc):\n super().__exit__(*exc)\n if not any(exc):\n click.echo('(%.2fs)' % self.elapsed)\n\ndef null(obj, component):\n \"A Query `missing` function that always returns None\"\n return None\n\nclass Literal:\n \"Makes restructure_dict return a specific value (without dict lookup)\"\n def __init__(self, value):\n self.value = value\n\nclass Query:\n def __init__(self, path, missing=None, cast=None, reraise=KeyError):\n self.path = path\n self.missing = missing\n self.cast = cast\n self.reraise = reraise\n def __call__(self, obj):\n if isinstance(self.path, Literal):\n return self.path.value\n for component in self.path.split('/'):\n if component.startswith('[') and component.endswith(']'):\n component = int(component[1:-1])\n try:\n obj = obj[component]\n except KeyError as error:\n if self.missing:\n return self.missing(obj, component)\n if self.reraise is KeyError:\n raise\n else:\n raise self.reraise from error\n raise\n return obj if self.cast is None else self.cast(obj)\n\ndef restructure_dict(src, queries):\n dst = {}\n for dst_path, query in queries.items():\n dst_obj = dst\n components = dst_path.split('/')\n prefix, key = components[:-1], components[-1]\n for component in prefix:\n dst_obj = dst_obj.setdefault(component, {})\n dst_obj[key] = query(src)\n return dst\n\ndef shell(ns={}):\n try:\n import IPython\n IPython.start_ipython(user_ns=ns, display_banner=False, argv=[])\n except ImportError:\n import code\n code.interact(local=ns)\n", "repo_name": "yaniv-aknin/fafalytics", "sub_path": "fafalytics/pyutils.py", "file_name": "pyutils.py", "file_ext": "py", "file_size_in_byte": 3222, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3, "dataset": "github-code", "pt": "25", "api": [{"api_name": "typing.Callable", "line_number": 6, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 8, "usage_type": "call"}, {"api_name": "typing.Iterable", "line_number": 6, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 15, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 19, "usage_type": "name"}, {"api_name": "time.time", "line_number": 30, "usage_type": "call"}, {"api_name": "time.time", "line_number": 33, "usage_type": "call"}, {"api_name": "time.time", "line_number": 41, "usage_type": "call"}, {"api_name": "click.echo", "line_number": 50, "usage_type": "call"}, {"api_name": "click.echo", "line_number": 54, "usage_type": "call"}, {"api_name": "IPython.start_ipython", "line_number": 103, "usage_type": "call"}, {"api_name": "code.interact", "line_number": 106, "usage_type": "call"}]}
+{"seq_id": "7345021316", "text": "from flask import Flask, request, render_template, redirect, flash, session\nfrom flask_debugtoolbar import DebugToolbarExtension\nfrom surveys import satisfaction_survey as survey\n\napp = Flask(__name__)\napp.config['SECRET_KEY'] = '0123456'\napp.config['DEBUG_TB_INTERCEPT_REDIRECTS'] = False\napp.debug = True\n\ndebug = DebugToolbarExtension(app)\n\n@app.route('/')\ndef root():\n \"\"\"Create start page\"\"\"\n\n title = survey.title\n instructions = survey.instructions\n\n return render_template('start.html', title=title, instructions=instructions)\n\n\n@app.route('/start', methods=['POST'])\ndef start():\n \"\"\"Start the survey and empty responses in session storage.\"\"\"\n \n session['responses'] = []\n\n return redirect('/questions/0')\n\n\n@app.route('/questions/')\ndef questions(q_id):\n \"\"\"Create questions pages\"\"\"\n \n responses = session['responses']\n\n if len(responses) == 0 and q_id != 0:\n flash('Click start to begin survey!')\n return redirect('/')\n\n if len(responses) == len(survey.questions):\n return redirect('/thanks')\n\n if q_id != len(responses):\n flash('Please answer questions in order!')\n return redirect(f'/questions/{len(responses)}')\n\n question = survey.questions[q_id]\n choices = question.choices\n\n return render_template('questions.html', question=question.question, choices=choices, id=q_id)\n\n\n@app.route('/answer', methods=['POST'])\ndef answers():\n \"\"\"Create answer route\"\"\"\n\n choice = request.form['answer']\n \n responses = session['responses']\n responses.append(choice)\n session['responses'] = responses\n\n if len(survey.questions) == len(responses):\n return redirect('/thanks')\n else:\n return redirect(f'/questions/{len(responses)}')\n\n\n@app.route('/thanks')\ndef thanks():\n \"\"\"Thank user for participating.\"\"\"\n\n return render_template('thanks.html')\n", "repo_name": "jr0dd/usf-bootcamp", "sub_path": "flask-survey/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1888, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "25", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask_debugtoolbar.DebugToolbarExtension", "line_number": 10, "usage_type": "call"}, {"api_name": "surveys.satisfaction_survey.title", "line_number": 16, "usage_type": "attribute"}, {"api_name": "surveys.satisfaction_survey", "line_number": 16, "usage_type": "name"}, {"api_name": "surveys.satisfaction_survey.instructions", "line_number": 17, "usage_type": "attribute"}, {"api_name": "surveys.satisfaction_survey", "line_number": 17, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 26, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 39, "usage_type": "call"}, {"api_name": "surveys.satisfaction_survey.questions", "line_number": 41, "usage_type": "attribute"}, {"api_name": "surveys.satisfaction_survey", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 46, "usage_type": "call"}, {"api_name": "surveys.satisfaction_survey.questions", "line_number": 48, "usage_type": "attribute"}, {"api_name": "surveys.satisfaction_survey", "line_number": 48, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 58, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 60, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 62, "usage_type": "name"}, {"api_name": "surveys.satisfaction_survey.questions", "line_number": 64, "usage_type": "attribute"}, {"api_name": "surveys.satisfaction_survey", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 74, "usage_type": "call"}]}
+{"seq_id": "23259139237", "text": "# Imports\n\nfrom flask import Flask, g\nfrom flask_cors import CORS\nfrom flask_login import LoginManager\nfrom api.user import user\nfrom api.parks import park\n\nimport models\n\n\n# ------------------------------------------------------------------------------------------------------\n\n# Setup and Initialization\n\nDEBUG = True\nPORT = 9000\nlogin_manager = LoginManager()\napp = Flask(__name__, static_url_path=\"\", static_folder=\"static\")\n\n\n# ------------------------------------------------------------------------------------------------------\n\n# Session\n\napp.secret_key = 'RANDOM ASS STRING'\napp.register_blueprint(user)\napp.register_blueprint(park)\n\n\n# CORS(api, origins = ['http://localhost:3000'], supports_credentials = True)\n# CORS(user, origins = ['http://localhost:3000'], supports_credentials = True)\n\n\nlogin_manager.init_app(app)\n\n\n# ------------------------------------------------------------------------------------------------------\n\n# Decorators and Functions\n\n@login_manager.user_loader\ndef load_user(userid):\n try:\n return models.User.get(models.User.id == userid)\n except models.DoesNotExist:\n return None\n\n@app.before_request\ndef before_request():\n \"\"\"Connect to the database before each request\"\"\"\n g.db = models.DATABASE\n g.db.connect()\n\n@app.after_request\ndef after_request(response):\n \"\"\"Close the database connection after each request\"\"\"\n g.db.close()\n return response\n\n@app.route('/')\ndef index():\n return 'hi'\n\n\n# ------------------------------------------------------------------------------------------------------\n\n# Run the App!\n\nif __name__ == '__main__':\n models.initialize()\n app.run(debug=DEBUG, port=PORT)\n", "repo_name": "OladiranJ/Open-Runs-Backend", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1744, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "flask_login.LoginManager", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 19, "usage_type": "call"}, {"api_name": "api.user.user", "line_number": 27, "usage_type": "argument"}, {"api_name": "api.parks.park", "line_number": 28, "usage_type": "argument"}, {"api_name": "models.User.get", "line_number": 45, "usage_type": "call"}, {"api_name": "models.User", "line_number": 45, "usage_type": "attribute"}, {"api_name": "models.DoesNotExist", "line_number": 46, "usage_type": "attribute"}, {"api_name": "flask.g.db", "line_number": 52, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 52, "usage_type": "name"}, {"api_name": "models.DATABASE", "line_number": 52, "usage_type": "attribute"}, {"api_name": "flask.g.db.connect", "line_number": 53, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 53, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 53, "usage_type": "name"}, {"api_name": "flask.g.db.close", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 58, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 58, "usage_type": "name"}, {"api_name": "models.initialize", "line_number": 71, "usage_type": "call"}]}
+{"seq_id": "34861257228", "text": "from os.path import join as op_join\nimport copy\nimport numpy as np\nfrom numpy import newaxis as nax\n\nimport wx\nimport wx.aui\nimport wx.lib.plot.plotcanvas as wlpc\nimport wx.lib.stattext\nimport wx.lib.agw.buttonpanel as bp\nimport wx.lib.agw.foldpanelbar as fpb\nfrom wx.lib.anchors import LayoutAnchors\nfrom wx.lib.stattext import GenStaticText\nfrom wx.lib.buttons import GenToggleButton as wxTogBut\nfrom wx.lib.plot.polyobjects import PolyLine, PlotGraphics\n# from wx.lib.plot.polyobjects import PolyMarker\n\nimport mva.chemometrics as chemtrics\n# noinspection PyProtectedMember\nfrom mva.chemometrics import _index\nfrom plot import PolyEllipse\nfrom commons import error_box\nfrom commons import PolyMarker\n\n[wxID_PCA, wxID_PCAPLCPCALOADSV, wxID_PCAPLCPCASCORE, wxID_PCAPLCPCEIGS,\n wxID_PCAPLCPCVAR, \n ] = [wx.NewId() for _init_ctrls in range(5)]\n\n[ID_RUNPCA, ID_EXPORTPCA, ID_PCATYPE, \n ID_SPNPCS, ID_NUMPCS1, ID_NUMPCS2,\n ] = [wx.NewId() for _init_btnpanel_ctrls in range(6)]\n\n# FR1 (wxID_FRAME1)\n[FR1_, FR1_BTNAPPLY, FR1_CBGRID,\n FR1_SPNFONTSIZEAXES, FR1_SPNXMAX, FR1_SPNXMIN,\n FR1_SPNYMAX, FR1_SPNYMIN, FR1_STFONT,\n FR1_STTITLE, FR1_STXFROM, FR1_STXLABEL,\n FR1_STXTO, FR1_STYFROM, FR1_STYLABEL, FR1_STYTO,\n FR1_TXTTITLE, FR1_TXTXLABEL, FR1_TXTXMAX,\n FR1_TXTXMIN, FR1_TXTYLABEL, FR1_TXTYMAX,\n FR1_TXTYMIN,\n ] = [wx.NewId() for _init_plot_prop_ctrls in range(23)]\n\n[MNUPLOTCOPY, MNUPLOTPRINT, MNUPLOTSAVE, MNUPLOTPROPS, MNUPLOTCOORDS,\n ] = [wx.NewId() for _init_plot_menu_Items in range(5)]\n\n\ndef set_btn_state(s1, s2, tb):\n # toolbar button enabled condition\n if s1 == s2:\n state = False\n else:\n state = True\n \n buttons = [tb.tbLoadLabels, tb.tbLoadLabStd1, \n tb.tbLoadLabStd2, tb.tbLoadSymStd2]\n \n for button in buttons:\n button.Enable(state)\n\ndef create_sym_col_select(canvas, output):\n \"\"\" populate symbol select pop-up\n\n \"\"\"\n print(' in create_sym_col_select pca loc 76')\n # first destroy current\n canvas.tbMain.SymPopUpWin.Destroy()\n # create empty ctrl\n canvas.tbMain.SymPopUpWin = SymColSelectTool(canvas.tbMain)\n spuw = canvas.tbMain.SymPopUpWin\n # create ctrls\n count = 0\n # apply button\n spuw.btnApply = wx.Button(spuw, wx.NewId(), 'Apply')\n spuw.Bind(wx.EVT_BUTTON, spuw.on_btn_apply, spuw.btnApply)\n # close button\n spuw.btnClose = wx.Button(spuw, wx.NewId(), 'Close')\n spuw.Bind(wx.EVT_BUTTON, spuw.on_btn_close, spuw.btnClose)\n # spacer\n spuw.stSpacer = wx.StaticText(spuw, -1, '')\n # dynamic ctrls\n spuw.colctrls = []\n spuw.symctrls = []\n\n sc = str(count)\n for each in output:\n exec('canvas.tbMain.SymPopUpWin.st' + sc + ' = wx.StaticText(canvas.tbMain.SymPopUpWin, -1, ' +\n 'each[0])')\n exec('canvas.tbMain.SymPopUpWin.btn' + sc + ' = wx.BitmapButton(canvas.tbMain.SymPopUpWin, ' +\n 'bitmap=wx.Bitmap(op_join(\"bmp\", \"' + each[1] + '.bmp\"), wx.BITMAP_TYPE_BMP), id_=-1)')\n exec('canvas.tbMain.SymPopUpWin.btn' + sc + '.symname = \"' + each[1] + '\"')\n exec('canvas.tbMain.SymPopUpWin.btn' + sc + '.Bind(wx.EVT_BUTTON, canvas.tbMain.SymPopUpWin.on_btn_symbol' + ')')\n exec('canvas.tbMain.SymPopUpWin.cp' + sc + ' = wx.ColourPickerCtrl(canvas.tbMain.SymPopUpWin, ' +\n '-1, col=' + str(each[2]) + ', style=wx.CLRP_DEFAULT_STYLE)')\n \n # output ctrl names to use later\n spuw.colctrls.append('cp' + sc)\n spuw.symctrls.append('btn' + sc)\n count += 1 \n \n # create sizer\n spuw.grsSelect = wx.GridSizer(cols=3, hgap=2, rows=count+1, vgap=2)\n # add standard ctrls\n spuw.grsSelect.Add(spuw.btnClose, 0, border=0, flag=wx.EXPAND)\n spuw.grsSelect.Add(spuw.btnApply, 0, border=0, flag=wx.EXPAND)\n spuw.grsSelect.Add(spuw.stSpacer, 0, border=0, flag=wx.EXPAND)\n # add dynamic ctrls to sizer\n for nwin in range(count):\n exec('canvas.tbMain.SymPopUpWin.grsSelect.Add(canvas.tbMain.SymPopUpWin.st' +\n str(nwin) + ', 0, border=0, flag=wx.EXPAND)')\n exec('canvas.tbMain.SymPopUpWin.grsSelect.Add(canvas.tbMain.SymPopUpWin.btn' +\n str(nwin) + ', 0, border=0, flag=wx.EXPAND)')\n exec('canvas.tbMain.SymPopUpWin.grsSelect.Add(canvas.tbMain.SymPopUpWin.cp' +\n str(nwin) + ', 0, border=0, flag=wx.EXPAND)')\n \n # set sizer and resize\n # noinspection PyUnresolvedReferences\n canvas.tbMain.SymPopUpWin.SetSizer(canvas.tbMain.SymPopUpWin.grsSelect)\n resize = (canvas.tbMain.SymPopUpWin.GetSize()[0], count * 35)\n canvas.tbMain.SymPopUpWin.SetSize(resize)\n\n# noinspection PyTypeChecker\ndef box_plot(canvas, x, labels, **_attr):\n \"\"\"Box and whisker plot; x is a column vector, labels a list of strings\n\n \"\"\"\n objects, count = [], 1\n uG = np.unique(np.array(labels))\n for each in uG:\n # get values\n group = x[np.array(labels) == each]\n # calculate group median\n m = np.median(group)\n # lower (first) quartile\n lq = np.median(group[group < m])\n # upper (third) quartile\n uq = np.median(group[group > m])\n # interquartile range\n iqr = uq-lq\n # lower whisker\n lw = m - (1.5 * iqr)\n # upper whisker\n uw = m + (1.5 * iqr)\n # lower outlier\n lo = group[group < lw]\n # upper outlier\n uo = group[group > uw]\n # plot b&w\n solid = wx.PENSTYLE_SOLID\n objects.append(PolyLine([[count - .25, m], [count + .25, m]],\n width=1, colour='blue', style=solid))\n objects.append(PolyLine([[count - .25, lq], [count + .25, lq]],\n width=1, colour='black', style=solid))\n objects.append(PolyLine([[count - .25, uq], [count + .25, uq]],\n width=1, colour='black', style=solid))\n objects.append(PolyLine([[count - .25, lq], [count - .25, uq]],\n width=1, colour='black', style=solid))\n objects.append(PolyLine([[count + .25, lq], [count + .25, uq]],\n width=1, colour='black', style=solid))\n objects.append(PolyLine([[count, lq], [count, lw]],\n width=1, colour='black', style=solid))\n objects.append(PolyLine([[count, uq], [count, uw]],\n width=1, colour='black', style=solid))\n objects.append(PolyLine([[count - .1, lw], [count + .1, lw]],\n width=1, colour='black', style=solid))\n objects.append(PolyLine([[count - .1, uw], [count + .1, uw]],\n width=1, colour='black', style=solid))\n if len(lo) > 0:\n objects.append(PolyMarker(np.concatenate(\n (np.ones((len(lo), 1)) * count, lo[:, nax]), 1),\n colour='red', fillcolour='red', marker='circle', size=1))\n if len(uo) > 0:\n objects.append(PolyMarker(np.concatenate(\n (np.ones((len(uo), 1)) * count, uo[:, nax]), 1),\n colour='red', fillcolour='red', marker='circle', size=1))\n count += 1\n \n canvas.xSpec = 'auto'\n # canvas.draw(PlotGraphics(objects, _attr['title'], _attr['xLabel'],\n # _attr['yLabel'], xTickLabels=uG))\n canvas.draw(PlotGraphics(objects, _attr['title'], _attr['xLabel'],\n _attr['yLabel']))\n\n\n# noinspection PyTypeChecker\ndef plot_error_bar(canvas, **_attr):\n \"\"\"Errorbar plot\n Defaults: \n 'x'= None - xaxis values, column vector\n 'y'= None - average, column vector \n 'validation'= None - list of 0's & 1's & 2's \n 'title'= '', - plot title\n 'xLabel'= '', - x-axis label\n 'yLabel'= '', - y-axis label\n 'lsfit'=False, - show linear fit\n 'usesym'=[]\n 'usecol'=[]\n \"\"\"\n \n # defaults\n colours = ['black', 'red', 'blue']\n usesym = ['square', 'circle', 'triangle']\n ledgtext = ['Train', 'Validation', 'Test']\n\n # user defined\n if _attr['usesym']:\n # noinspection PyUnusedLocal\n symbols = _attr['usesym']\n if _attr['usecol']:\n colours = _attr['usecol']\n \n objects = []\n if _attr['lsfit']:\n # show linear fit\n objects.append(PolyLine(np.array([[_attr['x'].min(), _attr['x'].min()],\n [_attr['x'].max(), _attr['x'].max()]]),\n legend='Linear fit', colour='cyan',\n width=1, style=wx.PENSTYLE_SOLID))\n\n for val in range(max(_attr['validation']) + 1):\n # get average and stdev of predictions for each calibration point\n average, stdev = [], []\n xsub = np.take(_attr['x'], _index(_attr['validation'], val), 0)\n uXsub = np.unique(xsub)\n ysub = np.take(_attr['y'], _index(_attr['validation'], val), 0)\n for item in range(len(uXsub)):\n average.append(np.mean(np.take(ysub, _index(xsub, uXsub[item]))))\n stdev.append(np.std(np.take(ysub, _index(xsub, uXsub[item]))))\n \n # markers\n objects.append(PolyMarker(np.concatenate((uXsub[:, nax],\n np.array(average)[:, nax]), 1),\n legend=ledgtext[val], colour=colours[val],\n marker=usesym[val], size=1.5,\n fillstyle=wx.BRUSHSTYLE_SOLID))\n \n # errorbars & horizontal bars\n for line, uxval in enumerate(uXsub):\n avgln = average[line]\n stdln = stdev[line]\n # errorbars\n objects.append(\n PolyLine(np.array([[uxval, avgln - stdln],\n [uxval, avgln + stdln]]),\n colour=colours[val], width=1, style=wx.PENSTYLE_SOLID))\n # horizontal bars +ve\n amxs = .01 * abs(max(uXsub))\n objects.append(\n PolyLine(np.array([[uxval - amxs, avgln + stdln],\n [uxval + amxs, avgln + stdln]]),\n colour=colours[val], width=1, style=wx.PENSTYLE_SOLID))\n # horizontal bars -ve\n objects.append(\n PolyLine(np.array([[uxval - amxs, avgln-stdln],\n [uxval + amxs, avgln-stdln]]),\n colour=colours[val], width=1, style=wx.PENSTYLE_SOLID))\n \n # axis limits\n atx = _attr['x']\n aty = _attr['y']\n\n xAx = (atx.min() - (.05 * atx.max()), atx.max() + (.05 * atx.max()))\n yAx = (aty.min() - (.05 * aty.max()), aty.max() + (.05 * aty.max()))\n \n canvas.draw(PlotGraphics(objects, _attr['title'], _attr['xLabel'],\n _attr['yLabel']), xAx, yAx)\n \n# noinspection PyUnresolvedReferences\ndef plot_pls_model(canvas, model='full', tbar=None, **_attr):\n \"\"\"Plot PLS predictions or scores; model = 'full' for PLSR,\n model = 'ga' for GA-PLS feature selection\n \n **_attr - key word _attributes \n Defaults: \n 'predictions'= None - pls predictions\n 'cL' = None - constituents\n 'scores' = None - pls spectral scores\n 'plScL' = None - false class for pls-da\n 'validation' = None, - split data\n 'factors' = 1, - no. latent variables\n 'type' = 0 - plsr or pls-da\n 'symbols' = False - plot with symbols\n 'usetxt' = True - plot with text labels\n 'RMSEPT' = 0 - RMSE for independent test samples\n 'col1' = 0 - col for xaxis\n 'col2' = 1 - col for yaxis\n \"\"\"\n typex = _attr['type']\n cL = None\n pRed = None\n canvPref = None\n\n if model in ['full']:\n canvPref = 'plcPredPls'\n prnt = canvas.parent.parent\n elif model == 'ga':\n canvPref = 'plcGaModelPlot'\n prnt = canvas.parent.parent.prnt.splitPrnt\n\n nBook = canvas.parent\n\n # Parece que Notebook no tiene 'SetTabSize' en phoenyx\n # if _attr['predictions'].shape[1] > 1:\n # canvas.parent.SetTabSize((80, 15))\n # else:\n # canvas.parent.SetTabSize((0, 1))\n # canvas.parent.SetPageText(0, '')\n \n if typex == 0:\n numPlots = _attr['predictions'].shape[1]\n else:\n numPlots = _attr['predictions'].shape[1] + 1\n \n # delete pages\n nBook.SetSelection(0)\n for page in range(nBook.GetPageCount() - 1, -1, -1):\n nBook.DeletePage(page)\n \n for const in range(numPlots):\n if typex == 0:\n cL = _attr['cL'][:, const][:, nax]\n pRed = _attr['predictions'][:, const][:, nax]\n elif (typex == 1) & (const > 0) is True:\n cL = _attr['plScL'][:, const-1][:, nax]\n pRed = _attr['predictions'][:, const-1][:, nax]\n \n # create new canvas\n sc1 = str(const + 1)\n exec(\"prnt.\" + canvPref + sc1 + \"= MyPlotCanvas(id_=-1, \" +\n \"name='\" + canvPref + sc1 + \"', parent=nBook, \" +\n \"pos=(0, 0), size=(302, 246), \" +\n \"style=0, toolbar=tbar)\")\n exec(\"prnt.\" + canvPref + sc1 + \".font_size_axis = 8\")\n exec(\"prnt.\" + canvPref + sc1 + \".font_size_title = 10\")\n exec(\"prnt.\" + canvPref + sc1 + \".enable_zoom = True\")\n exec(\"prnt.\" + canvPref + sc1 + \".SetToolTip('')\")\n exec(\"prnt.\" + canvPref + sc1 + \".enable_legend = True\")\n exec(\"prnt.\" + canvPref + sc1 + \".font_size_legend = 8\")\n exec(\"prnt.\" + canvPref + sc1 + \".SetAutoLayout(True)\")\n exec(\"prnt.\" + canvPref + sc1 +\n \".SetConstraints(LayoutAnchors(prnt.\" + canvPref + sc1 +\n \", True, True, True, True))\")\n\n # create new nb page\n if _attr['predictions'].shape[1] > 1:\n exec(\"nBook.AddPage(imageId=-1, page=prnt.\" + canvPref +\n sc1 + \", select=False, text='PLS Predictions \" +\n sc1 + \"')\")\n else:\n exec(\"nBook.AddPage(imageId=-1, page=prnt.\" + canvPref +\n sc1 + \", select=False, text='')\")\n \n # use it for plotting\n cmd = \"ncanv = prnt.\" + canvPref + sc1\n exec(cmd, locals(), globals())\n \n if (typex == 1) and (const == 0):\n # plot pls-da scores\n plot_scores(ncanv, _attr['scores'], cl=_attr['cL'][:, 0],\n labels=_attr['label'], validation=_attr['validation'],\n col1=_attr['col1'], col2=_attr['col2'],\n title='PLS Scores',\n xLabel='t[%i]' % _attr['col1']+1,\n yLabel='t[%i]' % _attr['col2']+1,\n xval=True, pconf=False, symb=_attr['symbols'],\n text=_attr['usetxt'], usecol=_attr['usecol'],\n usesym=_attr['usesym'])\n \n else: \n if _attr['symbols']:\n # pls predictions as errorbar plot\n title = 'PLS Predictions: %i factors, RMS(Indep. Test) %.2f' % \\\n (_attr['factors']+1, _attr['RMSEPT'])\n plot_error_bar(ncanv, x=cL, y=pRed,\n validation=_attr['validation'],\n title=title, xLabel='Actual', yLabel='Predicted',\n lsfit=True, usesym=_attr['usesym'],\n usecol=_attr['usecol'])\n else:\n # pls predictions as scatter plot\n TrnPnts = np.zeros((1, 2), 'd')\n ValPnts = np.zeros((1, 2), 'd'),\n TstPnts = np.zeros((1, 2), 'd')\n\n for i in range(len(cL)):\n if int(np.reshape(_attr['validation'][i], ())) == 0:\n y = float(np.reshape(cL[i], ()))\n py = float(np.reshape(pRed[i], ()))\n TrnPnts = np.concatenate((TrnPnts, np.reshape((y, py), (1, 2))), 0)\n elif int(np.reshape(_attr['validation'][i], ())) == 1:\n y = float(np.reshape(cL[i], ()))\n py = float(np.reshape(pRed[i], ()))\n ValPnts = np.concatenate((ValPnts, np.reshape((y, py), (1, 2))), 0)\n elif int(np.reshape(_attr['validation'][i], ())) == 2:\n y = float(np.reshape(cL[i], ()))\n py = float(np.reshape(pRed[i], ()))\n TstPnts = np.concatenate((TstPnts, np.reshape((y, py), (1, 2))), 0)\n \n TrnPnts = TrnPnts[1:len(TrnPnts) + 1]\n ValPnts = ValPnts[1:len(ValPnts) + 1]\n TstPnts = TstPnts[1:len(TstPnts) + 1]\n \n TrnPntObj = PolyMarker(TrnPnts, legend='Train',\n colour='black',\n marker='square', size=1.5,\n fillstyle=wx.BRUSHSTYLE_TRANSPARENT)\n \n ValPntObj = PolyMarker(ValPnts, legend='Cross Val.',\n colour='red',\n marker='circle', size=1.5,\n fillstyle=wx.BRUSHSTYLE_TRANSPARENT)\n \n TstPntObj = PolyMarker(TstPnts, legend='Indep. Test',\n colour='blue',\n marker='triangle', size=1.5,\n fillstyle=wx.BRUSHSTYLE_TRANSPARENT)\n\n # noinspection PyTypeChecker\n LinearObj = PolyLine(np.array([[cL.min(), cL.min()],\n [cL.max(), cL.max()]]),\n legend='Linear fit', colour='cyan',\n width=1, style=wx.PENSTYLE_SOLID)\n \n PlsModel = PlotGraphics([TrnPntObj, ValPntObj, TstPntObj, LinearObj],\n ' '.join(('PLS Predictions:',\n str(_attr['factors'] + 1),\n 'factors, RMS(Indep. Test)',\n '%.2f' % _attr['RMSEPT'])),\n 'Actual', 'Predicted')\n \n xAx = (cL.min() - (0.05 * cL.max()), cL.max() + (0.05 * cL.max()))\n \n ys = np.concatenate((TrnPnts, ValPnts), 0)\n # noinspection PyArgumentList\n yAx = (ys.min() - (0.05 * ys.max()), ys.max() + (0.05 * ys.max()))\n \n ncanv.draw(PlsModel, xAx, yAx)\n\n nBook.SetSelection(0)\n exec(\"canvas = prnt.\" + canvPref + str(1))\n \n return canvas\n \ndef plot_line(plotCanvas, plotArr, **_attr):\n \"\"\"Line plot\n **_attr - key word _attributes\n Defaults:\n 'xaxis' = None, - Vector of x-axis values\n 'rownum' = 0, - Row of plotArr to plot\n 'tit'= '', - A small domestic bird\n 'xLabel'= '', - The desired x-axis label\n 'yLabel'= '', - The desired y-axis label\n 'type'= 'single', - 'single' or 'multi'\n 'ledge'= [], - Figure legend labels\n 'wdth'= 1, - Line width\n \"\"\"\n\n colourList = ['blue', 'red', 'green', 'light_grey', 'cyan', 'black']\n NewplotLine = None\n \n if _attr['type'] == 'single':\n pA = plotArr[_attr['rownum'], 0:len(_attr['xaxis'])][:, nax]\n Line = PolyLine(np.concatenate((_attr['xaxis'], pA), 1),\n colour='black', width=_attr['wdth'],\n style=wx.PENSTYLE_SOLID)\n NewplotLine = PlotGraphics([Line], _attr['tit'],\n _attr['xLabel'], _attr['yLabel'])\n elif _attr['type'] == 'multi':\n ColourCount = 0\n Line = []\n for i in range(plotArr.shape[0]):\n pA = plotArr[i]\n pA = pA[:, nax]\n if _attr['ledge'] is not None:\n Line.append(PolyLine(np.concatenate((_attr['xaxis'], pA), 1),\n legend=_attr['ledge'][i],\n colour=colourList[ColourCount],\n width=_attr['wdth'],\n style=wx.PENSTYLE_SOLID))\n else:\n Line.append(PolyLine(np.concatenate((_attr['xaxis'], pA), 1),\n colour=colourList[ColourCount],\n width=_attr['wdth'],\n style=wx.PENSTYLE_SOLID))\n ColourCount += 1\n if ColourCount == len(colourList):\n ColourCount = 0\n NewplotLine = PlotGraphics(Line, _attr['tit'],\n _attr['xLabel'], _attr['yLabel'])\n \n plotCanvas.draw(NewplotLine) # , xAxis=(_attr['xaxis'].min(), _attr['xaxis'].max()))\n \ndef plot_stem(plotCanvas, plotArr, **_attr):\n \"\"\"Stem plot\n **_attr - key word _attributes\n Defaults:\n 'tit'= '', - Figure title\n 'xLabel'= '', - The desired x-axis label\n 'yLabel'= '', - The desired y-axis label\n 'wdth'= 1, - Line width\n \"\"\"\n \n # plotArr is an n x 2 array\n plotStem = []\n for i in range(plotArr.shape[0]):\n newCoords = np.array([[plotArr[i, 0], 0], [plotArr[i, 0], plotArr[i, 1]]])\n plotStem.append(PolyLine(newCoords, colour='black',\n width=_attr['wdth'], style=wx.PENSTYLE_SOLID))\n # noinspection PyTypeChecker\n plotStem.append(PolyLine(\n np.array([[plotArr[0, 0] - (.1 * plotArr[0, 0]), 0],\n [plotArr[len(plotArr) - 1, 0] + (.1 * plotArr[0, 0]), 0]]),\n colour='black',\n width=1, style=wx.PENSTYLE_SOLID))\n \n plotStem = PlotGraphics(plotStem, _attr['tit'],\n _attr['xLabel'], _attr['yLabel'])\n \n plotCanvas.draw(plotStem)\n\ndef plot_symbols(plotCanvas, coords, **_attr):\n \"\"\"Symbol plot\n **_attr - key word _attributes\n Defaults:\n 'mask' = [], - List of zeros, ones and/or twos to\n define train, cross-validation and\n test samples\n 'cLass' = [], - List of integers from 1:n, where \n n=no. of groups\n 'col1' = 0, - Column to plot along abscissa\n 'col2' = 1, - Column to plot along ordinate\n 'tit'= '', - A small domestic bird\n 'xL'= '', - The desired x-axis label\n 'yL'= '', - The desired y-axis label\n 'text'= [], - List of labels to use in legend\n 'usemask'= True,- Flag to define whether to use 'mask'\n 'usecol'=[], - Use a list of colours\n 'usesym'= [], - List of symbols for plotting\n \"\"\"\n \n desCl = np.unique(_attr['text'])\n eCount = 0\n if not _attr['usecol']:\n colours = ['blue', 'red', 'green', 'cyan', 'black']\n else:\n colours = _attr['usecol']\n \n if not _attr['usesym']:\n symbols = ['circle', 'square', 'plus',\n 'triangle', 'cross', 'triangle_down']\n else:\n symbols = _attr['usesym']\n \n # plot scores using symbols\n valSym = ['circle', 'square']\n plotSym, countSym, countColour, output = [], 0, 0, []\n for each in desCl:\n if countSym > len(symbols)-1:\n countSym = 0\n if countColour > len(colours)-1:\n countColour = 0\n \n # slice coords\n alist = coords[np.array(_attr['text']) == each, :]\n\n if _attr['col1'] != _attr['col2']:\n alist = np.take(alist, (_attr['col1'], _attr['col2']), 1)\n else:\n sCount = copy.deepcopy(eCount)+1 \n eCount = eCount+len(alist)\n alist = np.concatenate((np.arange(sCount, eCount + 1)[:, nax],\n alist[:, _attr['col1']][:, nax]), 1)\n \n # col = wx.Colour(round(np.rand(1).tolist()[0]*255),\n # round(np.rand(1).tolist()[0]*255),\n # round(np.rand(1).tolist()[0]*255))\n \n output.append([each, symbols[countSym], colours[countColour]])\n \n if _attr['usemask'] is False:\n plotSym.append(PolyMarker(alist, marker=symbols[countSym],\n fillcolour=colours[countColour],\n colour=colours[countColour],\n size=2, legend=each))\n \n else:\n listM = _attr['mask'][np.array(_attr['text']) == each]\n for m in range(3):\n if m == 0:\n # include legend entry\n plotSym.append(PolyMarker(alist[listM == m],\n marker=symbols[countSym],\n fillcolour=colours[countColour],\n colour=colours[countColour],\n size=2.5, legend=each))\n else:\n # no legend\n plotSym.append(PolyMarker(alist[listM == m],\n marker=symbols[countSym],\n fillcolour=colours[countColour],\n colour=colours[countColour],\n size=2.5))\n if m > 0:\n if symbols[countSym] not in ['cross', 'plus']:\n # overlay white circle/square to indicate\n # validation/test sample\n plotSym.append(PolyMarker(\n alist[listM == m],\n marker=valSym[m - 1], colour=wx.Colour('white'),\n fillcolour=wx.Colour('white'), size=1))\n else:\n # overlay white square to indicate validation sample\n plotSym.insert(len(plotSym) - 1,\n PolyMarker(alist[listM == m],\n marker=valSym[m - 1],\n colour=wx.Colour('black'),\n fillcolour=wx.Colour('white'),\n size=2.5))\n \n countSym += 1\n countColour += 1\n \n draw_plotSym = PlotGraphics(plotSym, _attr['tit'],\n xLabel=_attr['xL'], yLabel=_attr['yL'])\n \n if plotCanvas is not None:\n plotCanvas.draw(draw_plotSym)\n \n return plotSym, output\n\ndef plot_text(plotCanvas, coords, **_attr):\n \"\"\"Text label plot\n **_attr - key word _attributes\n Defaults:\n 'mask' = [], - List of zeros, ones and/or twos to\n define train, cross-validation and\n test samples\n 'cLass' = [], - List of integers from 1:n, where \n n=no. of groups\n 'col1' = 0, - Column to plot along abscissa\n 'col2' = 1, - Column to plot along ordinate\n 'tit'= '', - A small domestic bird\n 'xL'= '', - The desired x-axis label\n 'yL'= '', - The desired y-axis label\n 'text'= [], - List of labels to use in plotting\n 'usemask'= True,- Flag to define whether to use 'mask' \n \"\"\"\n \n # make sure label string\n nt = [str(i) for i in _attr['text']]\n _attr['text'] = nt\n \n plotText = []\n colours = ['black', 'blue', 'red']\n if _attr['usemask']:\n colRange = 3\n else:\n colRange = 1\n\n # plot 2d\n if (coords.shape[1] > 1) & (_attr['col1'] != _attr['col2']):\n # set text colour - black=train, blue=val, red=test\n for getColour in range(colRange):\n if colRange == 3:\n idx = _index(_attr['mask'], getColour)\n else:\n idx = range(len(coords))\n plotText.append(PolyMarker(\n np.take(np.take(coords, [_attr['col1'], _attr['col2']], 1),\n idx, 0),\n marker='text', legend=np.take(_attr['text'], idx, 0),\n colour=colours[getColour]))\n # plot 1d\n else:\n points = np.take(coords, [_attr['col1']], 1)\n nCl = np.unique(_attr['text'])\n eCount = 0\n for each in nCl:\n aslice = points[np.array(_attr['text']) == each]\n lbls = np.array(_attr['text'])[np.array(_attr['text']) == each]\n \n sCount = copy.deepcopy(eCount) + 1\n eCount = eCount + len(aslice)\n \n pointSub = np.concatenate((np.arange(sCount, eCount + 1)[:, nax],\n aslice), 1)\n \n if _attr['usemask'] is False:\n plotText.append(PolyMarker(pointSub, marker='text',\n legend=lbls.tolist()))\n else:\n msk = np.array(_attr['mask'])\n txt = np.array(_attr['text'])\n for each2 in range(3):\n msk_lst = msk[txt == each2].tolist()\n msk_idx = _index(msk_lst, each2)\n plotText.append(\n PolyMarker(np.take(pointSub, msk_idx, 0),\n marker='text',\n legend=np.take(lbls, msk_idx.tolist(),\n colour=colours[each2])))\n \n if (coords.shape[1] > 1) & (_attr['col1'] != _attr['col2']):\n draw_plot_text = PlotGraphics(plotText, _attr['tit'],\n xLabel=_attr['xL'], yLabel=_attr['yL'])\n else:\n draw_plot_text = PlotGraphics(plotText, _attr['tit'],\n xLabel='', yLabel=_attr['yL'])\n print('plotText ', plotText)\n if plotCanvas is not None:\n plotCanvas.draw(draw_plot_text)\n \n return plotText\n\ndef plot_loads(canvas, loads, **_attr):\n \"\"\"Model loadings plot\n **_attr - key word _attributes\n Defaults:\n 'xaxis' = [], - Vector of x-axis values\n 'col1' = 0, - Column to plot along abscissa\n 'col2' = 1, - Column to plot along ordinate\n 'title'= '', - Figure title\n 'xLabel'= '', - The desired x-axis label\n 'yLabel'= '', - The desired y-axis label\n 'type'= 0, - List of labels to use in plotting\n 'usecol'= [], - List of colours for symbol plot\n 'usesym'= [], - List of symbols for plotting\n \"\"\"\n i = 0\n\n # for model loadings plots\n plot = []\n \n if (_attr['col1'] != _attr['col2']) & (loads is not None) is True:\n # standard deviation\n select = np.concatenate((loads[:, _attr['col1']][:, nax],\n loads[:, _attr['col2']][:, nax]), 1)\n meanCoords = np.reshape(np.mean(select, 0), (1, 2))\n std = np.mean(np.std(select))\n\n if _attr['type'] == 0:\n # plot labels\n textPlot = plot_text(None, select, mask=None, cLass=None,\n text=_attr['xaxis'], usemask=False, col1=0,\n col2=1, tit='', xL='', yL='')\n for each in textPlot:\n plot.append(each)\n \n else:\n test = np.sqrt((loads[:, _attr['col1']]-meanCoords[0, 0]) ** 2 +\n (loads[:, _attr['col2']]-meanCoords[0, 1]) ** 2)\n index = np.arange(len(_attr['xaxis']))\n\n if _attr['type'] == 1: \n # >1*std error & labels\n outIdx = index[test > std]\n getOutliers = np.take(select, outIdx, 0)\n\n # plot labels\n labels = []\n for each in outIdx:\n labels.append(_attr['xaxis'][each])\n textPlot = plot_text(None, getOutliers, mask=None, cLass=None,\n text=labels, usemask=False, col1=0, col2=1,\n tit='', xL='', yL='')\n for each in textPlot:\n plot.append(each)\n \n elif _attr['type'] == 2:\n # >2*std error & labels\n outIdx = index[test > std * 2]\n \n getOutliers = np.take(select, outIdx, 0)\n \n # plot labels\n labels = []\n for each in outIdx:\n labels.append(_attr['xaxis'][each])\n textPlot = plot_text(None, getOutliers, mask=None, cLass=None,\n text=labels, usemask=False, col1=0, col2=1,\n tit='', xL='', yL='')\n for each in textPlot:\n plot.append(each)\n \n elif _attr['type'] == 3:\n # >2*std error & symbols\n outIdx = index[test > std * 2]\n \n # loadings > 2*std\n getOutliers = np.take(select, outIdx, 0)\n \n # identify regions\n # noinspection PyUnusedLocal\n newxvar = np.take(_attr['xaxis'], outIdx)\n regions = [outIdx[0]]\n for i in range(len(outIdx) - 1 ):\n if outIdx[i + 1] - 1 != outIdx[i]:\n regions.append(outIdx[i])\n regions.append(outIdx[i + 1])\n if np.mod(len(regions), 2) == 1:\n regions.append(outIdx[i + 1])\n \n # plot regions by symbol/colour\n cl, labels, i = [], [], 0\n while i < len(regions):\n cl.extend((np.ones(regions[i + 1] - regions[i] + 1, ) * i).tolist())\n for j in range(regions[i + 1]-regions[i]+1):\n labels.append(str(_attr['xaxis'][regions[i]]) + ' - ' +\n str(_attr['xaxis'][regions[i + 1]]))\n i += 2\n \n symPlot, output = plot_symbols(None, getOutliers, mask=None,\n cLass=np.array(cl),\n text=labels, usemask=False,\n col1=0, col2=1, tit='',\n xL='', yL='',\n usecol=_attr['usecol'],\n usesym=_attr['usesym'])\n \n # create window in background for changing symbols/colours\n create_sym_col_select(canvas, output)\n \n for each in symPlot:\n plot.append(each)\n \n # ellipse boundary\n plot.append(PolyMarker([[meanCoords[0, 0] - (std * 2),\n meanCoords[0, 1] - (std * 2)],\n [meanCoords[0, 0] + (std * 2),\n meanCoords[0, 1] + (std * 2)]],\n colour='white', size=1, marker='circle'))\n # centroid\n plot.append(PolyMarker(meanCoords, colour='blue',\n size=2, marker='plus'))\n # plot 1 std\n plot.append(PolyEllipse(meanCoords, colour='green', width=1,\n dim=(std * 2, std * 2),\n style=wx.PENSTYLE_SOLID))\n # plot 2 stds\n plot.append(PolyEllipse(meanCoords, colour='green', width=1,\n dim=(std * 4, std * 4),\n style=wx.PENSTYLE_SOLID))\n \n # draw it\n canvas.draw(PlotGraphics(plot, _attr['title'], _attr['xLabel'],\n _attr['yLabel']))\n \n \ndef plot_scores(canvas, scores, **_attr):\n \"\"\"Model scores plot\n **_attr - key word _attributes\n Defaults:\n 'cl' = [] - List of integers\n 'labels' = [] - List of sample labels\n 'validation' = [] - List of zeros, ones and/or twos\n 'col1' = 0, - Column to plot along abscissa\n 'col2' = 1, - Column to plot along ordinate\n 'title'= '', - Figure title\n 'xLabel'= '', - The desired x-axis label\n 'yLabel'= '', - The desired y-axis label\n 'xval'= False, - Cross-validation used flag\n 'text'= True, - Text label plotting used flag\n 'pconf'= True, - 95% confidence limits plotted flag\n 'symb'= False, - Symbol plotting used flag\n 'usecol'= [], - List of colours to use in plotting\n 'usesym'= [], - List of symbols for plotting\n \"\"\"\n \n # make sure we can plot txt\n \n if (canvas.GetName() not in ['plcDFAscores']) & \\\n (len(canvas.GetName().split('plcGaModelPlot')) == 1):\n canvas.tbMain.tbConf.SetValue(False)\n if (canvas.tbMain.tbPoints.GetValue() is not True) & \\\n (canvas.tbMain.tbSymbols.GetValue() is not True):\n canvas.tbMain.tbPoints.SetValue(True)\n _attr['text'] = True\n \n # get mean centres\n # nb for a dfa/cva plot scaled to unit variance\n # 95% confidence radius is 2.15\n shapex = scores.shape\n nCl = np.unique(_attr['cl'])\n\n plot = []\n if (shapex[1] > 1) & (_attr['col1'] != _attr['col2']):\n canvas.xSpec = 'auto'\n \n scores = np.concatenate((scores[:, _attr['col1']][:, nax],\n scores[:, _attr['col2']][:, nax]), 1)\n \n mScores = np.zeros((1, 2))\n for i in range(len(nCl)):\n mScores = np.concatenate(\n (mScores, np.mean(np.take(scores, _index(_attr['cl'], nCl[i]),\n 0), 0)[nax, :]), 0)\n mScores = mScores[1: len(mScores)]\n \n if _attr['symb'] is True:\n # plot symbols\n sym_plot, output = plot_symbols(None, scores,\n mask=_attr['validation'],\n cLass=_attr['cl'],\n text=_attr['labels'],\n usemask=_attr['xval'], col1=0,\n col2=1, tit='', xL='', yL='',\n usecol=_attr['usecol'],\n usesym=_attr['usesym'])\n \n # create window in background for changing symbols/colours\n create_sym_col_select(canvas, output)\n \n for each in sym_plot:\n plot.append(each)\n \n if _attr['text']:\n # plot labels\n textPlot = plot_text(None, scores, mask=_attr['validation'],\n cLass=_attr['cl'], text=_attr['labels'],\n col1=0, col2=1, usemask=_attr['xval'], tit='',\n xL='', yL='')\n for each in textPlot:\n plot.append(each)\n \n if _attr['pconf']:\n # 95% confidence interval\n plot.append(PolyEllipse(mScores, colour='black', width=1,\n dim=(2.15 * 2, 2.15 * 2),\n style=wx.PENSTYLE_SOLID))\n # 95% confidence about the mean\n plot.append(PolyEllipse(mScores, colour='blue', width=1,\n dim=((1.95 / np.sqrt(len(nCl)) * 2),\n (1.95 / np.sqrt(len(nCl)) * 2)),\n style=wx.PENSTYLE_SOLID))\n # class centroids\n plot.append(PolyMarker(mScores[:, 0:2], colour='black',\n size=2, marker='plus'))\n # force boundary\n plot.append(PolyMarker([[min(mScores[:, 0] - 2.15),\n min(mScores[:, 1] - 2.15)],\n [max(mScores[:, 0] + 2.15),\n max(mScores[:, 1] + 2.15)]],\n colour='white', size=1, marker='circle'))\n\n # class centroid label\n if (_attr['symb'] is False) & (_attr['text'] is False):\n uC, centLab, centLabOrds = np.unique(_attr['cl']), [], []\n for gC in range(len(uC)):\n Idx = _index(_attr['cl'], uC[gC])[0]\n centLab.append(_attr['labels'][Idx])\n centLabOrds.append(np.reshape(mScores[gC, :], (scores.shape[1], )).tolist())\n \n # print centroid labels\n centPlot = plot_text(None, np.array(centLabOrds),\n cLass=np.arange(1, len(centLab) + 1),\n text=centLab, col1=0, col2=1,\n tit='', xL='', yL='', usemask=False)\n for each in centPlot:\n plot.append(each)\n \n canvas.draw(PlotGraphics(plot, _attr['title'],\n _attr['xLabel'], _attr['yLabel']))\n \n else:\n canvas.xSpec = 'none'\n if _attr['text']:\n # plot labels\n textPlot = plot_text(None, scores, mask=_attr['validation'],\n cLass=_attr['cl'], text=_attr['labels'],\n col1=_attr['col1'], col2=_attr['col1'],\n tit=_attr['title'], xL='Arbitrary',\n yL=_attr['yLabel'], usemask=_attr['xval'])\n # each are PolyMarkers\n\n for each in textPlot:\n print('each legend', each.getLegend())\n print('each tp labels attributes:', each.attributes['labels'])\n # noinspection PyProtectedMember\n print('each tp labels _attributes:', each._attributes['labels'])\n # print('each tp labels _attr:', each._attr['labels'])\n plot.append(each)\n \n if _attr['symb']:\n # plot symbols\n sym_plot, output = plot_symbols(None, scores,\n mask=_attr['validation'],\n cLass=_attr['cl'],\n text=_attr['labels'],\n usemask=_attr['xval'],\n col1=_attr['col1'],\n col2=_attr['col1'], tit='', xL='',\n yL='', usecol=_attr['usecol'],\n usesym=_attr['usesym'])\n \n # create window in background for changing symbols/colours\n create_sym_col_select(canvas, output)\n \n for each in sym_plot:\n print('each sp labels attributes:', each, each.attributes['labels'])\n # noinspection PyProtectedMember\n print('each sp labels _attributes:', each._attributes['labels'])\n plot.append(each)\n \n if _attr['text'] or _attr['symb']:\n graphic = PlotGraphics(plot, _attr['title'], '', _attr['yLabel'])\n print('legen names: ', graphic.getLegendNames())\n canvas.draw(graphic)\n \n \nclass SymColSelectTool(wx.Dialog):\n def __init__(self, parent):\n wx.Dialog.__init__(self, parent=parent, style=0)\n\n self.parent = parent\n self.SetSize((300, 0))\n self.SetAutoLayout(True)\n\n def on_btn_close(self, _):\n self.Show(False)\n\n # noinspection PyUnresolvedReferences\n def on_btn_apply(self, _):\n # get list of new colours\n collist = []\n for each in self.colctrls:\n exec('collist.append(self.' + each + '.GetColour())')\n # get list of new symbols\n symlist = []\n for each in self.symctrls:\n exec('symlist.append(self.' + each + '.symname)')\n # plot loadings\n self.parent.do_plot(loadType=3, symcolours=collist, symsymbols=symlist)\n self.parent.load_idx = 3\n \n def on_btn_symbol(self, evt):\n # symbol select dialog\n btn = evt.GetEventObject()\n dlg = SymDialog(self, btn)\n pos = btn.ClientToScreen((0, 0))\n sz = btn.GetSize()\n dlg.SetPosition((pos[0]-155, pos[1] + sz[1]))\n dlg.ShowModal()\n \nclass SymDialog(wx.Dialog):\n\n def __init__(self, parent, btn):\n wx.Dialog.__init__(self, id=-1, name=u'SymDialog', parent=parent,\n pos=(589, 316), size=(156, 155),\n style=wx.DEFAULT_DIALOG_STYLE,\n title=u'Select Symbol')\n\n self._init_ctrls()\n self.btn = btn\n\n def _init_ctrls(self):\n # generated method, don't edit\n\n self.SetClientSize((140, 119))\n self.SetToolTip(u'')\n\n bmp = wx.Bitmap(op_join('bmp', 'square.bmp'), wx.BITMAP_TYPE_BMP)\n self.tbSquare = wx.BitmapButton(bitmap=bmp, id=-1, name=u'tbSquare',\n parent=self, pos=(0, 0),\n size=(69, 38), style=0)\n self.tbSquare.Bind(wx.EVT_BUTTON, self.on_tb_square)\n\n bmp = wx.Bitmap(op_join('bmp', 'circle.bmp'), wx.BITMAP_TYPE_BMP)\n self.tbCircle = wx.BitmapButton(bitmap=bmp, id=-1, name=u'tbCircle',\n parent=self, pos=(71, 0),\n size=(69, 38), style=0)\n self.tbCircle.Bind(wx.EVT_BUTTON, self.on_tb_circle)\n\n bmp = wx.Bitmap(op_join('bmp', 'plus.bmp'), wx.BITMAP_TYPE_BMP)\n self.tbPlus = wx.BitmapButton(bitmap=bmp, id=-1, name=u'tbPlus',\n parent=self, pos=(0, 40),\n size=(69, 38), style=0)\n self.tbPlus.Bind(wx.EVT_BUTTON, self.on_tb_plus)\n\n bmp = wx.Bitmap(op_join('bmp', 'triangle.bmp'), wx.BITMAP_TYPE_BMP)\n self.tbTriangleUp = wx.BitmapButton(bitmap=bmp, id=-1,\n name=u'tbTriangleUp', parent=self,\n pos=(71, 40),\n size=(69, 38), style=0)\n self.tbTriangleUp.Bind(wx.EVT_BUTTON, self.on_tb_triangle_up)\n\n bmp = wx.Bitmap(op_join('bmp', 'triangle_down.bmp'), wx.BITMAP_TYPE_BMP)\n self.tbTriangleDown = wx.BitmapButton(bitmap=bmp, id=-1,\n name=u'tbTriangleDown',\n parent=self, pos=(0, 80),\n size=(69, 38), style=0)\n self.tbTriangleDown.Bind(wx.EVT_BUTTON, self.on_tb_triangle_down)\n\n bmp = wx.Bitmap(op_join('bmp', 'cross.bmp'), wx.BITMAP_TYPE_BMP)\n self.tbCross = wx.BitmapButton(bitmap=bmp, id=-1, name=u'tbCross',\n parent=self, pos=(71, 80),\n size=(69, 38), style=0)\n self.tbCross.Bind(wx.EVT_BUTTON, self.on_tb_cross)\n\n self._init_sizers()\n\n def _init_sizers(self):\n # generated method, don't edit\n self.grs_symdialog = wx.GridSizer(cols=2, hgap=2, rows=3, vgap=2)\n\n self._init_coll_grs_symdialog_items(self.grs_symdialog)\n\n self.SetSizer(self.grs_symdialog)\n\n def _init_coll_grs_symdialog_items(self, parent):\n \"\"\"\"\"\"\n parent.Add(self.tbSquare, 0, border=0, flag=wx.EXPAND)\n parent.Add(self.tbCircle, 0, border=0, flag=wx.EXPAND)\n parent.Add(self.tbPlus, 0, border=0, flag=wx.EXPAND)\n parent.Add(self.tbTriangleUp, 0, border=0, flag=wx.EXPAND)\n parent.Add(self.tbTriangleDown, 0, border=0, flag=wx.EXPAND)\n parent.Add(self.tbCross, 0, border=0, flag=wx.EXPAND)\n \n def on_tb_square(self, _):\n self.btn.SetBitmapLabel(wx.Bitmap(op_join('bmp', 'square.bmp')))\n self.btn.symname = 'square'\n self.Destroy()\n\n def on_tb_circle(self, _):\n self.btn.SetBitmapLabel(wx.Bitmap(op_join('bmp', 'circle.bmp')))\n self.btn.symname = 'circle'\n self.Destroy()\n\n def on_tb_plus(self, _):\n self.btn.SetBitmapLabel(wx.Bitmap(op_join('bmp', 'plus.bmp')))\n self.btn.symname = 'plus'\n self.Destroy()\n\n def on_tb_triangle_up(self, _):\n self.btn.SetBitmapLabel(wx.Bitmap(op_join('bmp', 'triangle.bmp')))\n self.btn.symname = 'triangle'\n self.Destroy()\n\n def on_tb_triangle_down(self, _):\n bmp = wx.Bitmap(op_join('bmp', 'triangle_down.bmp'))\n self.btn.SetBitmapLabel(bmp)\n self.btn.symname = 'triangle_down'\n self.Destroy()\n\n def on_tb_cross(self, _):\n bmp = wx.Bitmap(op_join('bmp', 'cross.bmp'))\n self.btn.SetBitmapLabel(bmp)\n self.btn.symname = 'cross'\n self.Destroy()\n\nclass MyPlotCanvas(wlpc.PlotCanvas):\n def __init__(self, parent, id_, pos, size, style, name, toolbar):\n wlpc.PlotCanvas.__init__(self, parent, id_, pos, size, style, name)\n\n wlpc.PlotCanvas._interEnabled = False\n wlpc.PlotCanvas._justDragged = False\n self._interEnabled = False\n\n self.xSpec = 'min'\n self.ySpec = 'min'\n\n self.minXrange = 0\n self.maxXrange = 0\n self.minYrange = 0\n self.maxYrange = 0\n\n self.Bind(wx.EVT_RIGHT_DOWN, self.OnMouseRightDown)\n self.Bind(wx.EVT_LEFT_DOWN, self.OnMouseLeftDown)\n\n self._init_utils()\n self.parent = parent\n self.tbMain = toolbar\n\n def _init_utils(self):\n self.plotMenu = wx.Menu(title='')\n\n self._init_plot_menu(self.plotMenu)\n\n def _init_plot_menu(self, parent):\n \n parent.Append(helpString='', id=MNUPLOTCOPY, kind=wx.ITEM_NORMAL,\n item='Copy Figure')\n parent.Append(helpString='', id=MNUPLOTCOORDS, kind=wx.ITEM_NORMAL,\n item='Copy Coordinates')\n parent.Append(helpString='', id=MNUPLOTPRINT, kind=wx.ITEM_NORMAL,\n item='Print')\n parent.Append(helpString='', id=MNUPLOTSAVE, kind=wx.ITEM_NORMAL,\n item='Save')\n\n self.Bind(wx.EVT_MENU, self.on_mnu_plot_copy, id=MNUPLOTCOPY)\n self.Bind(wx.EVT_MENU, self.on_mnu_plot_print, id=MNUPLOTPRINT)\n self.Bind(wx.EVT_MENU, self.on_mnu_plot_save, id=MNUPLOTSAVE)\n self.Bind(wx.EVT_MENU, self.on_mnu_plot_coords, id=MNUPLOTCOORDS)\n \n def on_mnu_plot_copy(self, _):\n # for windows\n self.Redraw(wx.MetafileDC()).SetClipboard()\n\n # for linux\n # wx.TheClipboard.Open()\n # wx.TheClipboard.SetData(self.Copy())\n # wx.TheClipboard.Close()\n \n def on_mnu_plot_print(self, _):\n self.Printout()\n \n def on_mnu_plot_save(self, _):\n self.SaveFile()\n \n # def OnMnuPlotProperties(self, event):\n # dlg = plotProperties(self)\n # dlg.SetSize((450, 350))\n # dlg.Center(wx.BOTH)\n #\n # # Set up dialog for specific cases\n # # dfa & pca score plots\n # if self.GetName() in ['plcDFAscores',\n # 'plcPCAscore', 'plcGaFeatPlot']:\n # dlg.scoreSets.Enable(True)\n # # pca score plots minus conf intervals\n # if self.GetName() in ['plcPCAscore', 'plcGaFeatPlot']:\n # dlg.tbConf.Enable(False)\n # dlg.tbConf.SetValue(False)\n # # ga-dfa score plots\n # if self.GetName() in ['plcGaModelPlot1']:\n # if self.prnt.prnt.splitPrnt.type in ['DFA']:\n # dlg.scoreSets.Enable(True)\n # if self.GetName() in ['plcPcaLoadsV', 'plcDfaLoadsV',\n # 'plcGaSpecLoad', 'plcPLSloading']:\n # dlg.loadSets.Enable(True)\n # dlg.Iconize(False)\n # dlg.ShowModal()\n \n def on_mnu_plot_coords(self, _):\n # send coords to clipboard\n getPoints = self.last_draw[0].objects\n coords = []\n for each in getPoints:\n # noinspection PyProtectedMember\n coords.extend(each._points.tolist())\n \n data = np.array2string(coords, separator='\\t')\n wx.TheClipboard.Open()\n wx.TheClipboard.SetData(wx.TextDataObject('X\\tY\\n' + data))\n wx.TheClipboard.Close()\n \n def OnMouseRightDown(self, event):\n pt = event.GetPosition()\n self.PopupMenu(self.plotMenu, pt) \n \n def OnMouseLeftDown(self, event):\n # put info in tb\n self.populate_toolbar()\n # get coords for zoom centre\n self._zoomCorner1[0], self._zoomCorner1[1] = self._getXY(event)\n self._screenCoordinates = np.array(event.GetPosition())\n if self._dragEnabled:\n self.SetCursor(self.GrabHandCursor)\n self.tbMain.canvas.CaptureMouse()\n if self._interEnabled:\n if self.last_draw is not None:\n graphics, xAxis, yAxis = self.last_draw\n xy = self.PositionScreenToUser(self._screenCoordinates)\n graphics.objects.append(\n PolyLine([[xy[0], yAxis[0]], [xy[0], yAxis[1]]],\n colour='red'))\n self._Draw(graphics, xAxis, yAxis)\n \n def populate_toolbar(self):\n # enable plot toolbar\n self.tbMain.Enable(True)\n self.tbMain.Refresh()\n \n # populate plot toolbar\n self.tbMain.canvas = self\n self.tbMain.graph = self.last_draw[0]\n \n self.tbMain.txtPlot.SetValue(self.tbMain.graph.title)\n self.tbMain.txtXlabel.SetValue(self.tbMain.graph.xLabel)\n self.tbMain.txtYlabel.SetValue(self.tbMain.graph.yLabel)\n \n self.tbMain.spnAxesFont.SetValue(self.fontSizeAxis)\n self.tbMain.spn_title.SetValue(self.fontSizeTitle)\n \n self.minXrange = self.xCurrentRange[0]\n self.maxXrange = self.xCurrentRange[1]\n self.minYrange = self.yCurrentRange[0]\n self.maxYrange = self.yCurrentRange[1]\n \n self.tbMain.Increment = (self.maxXrange - self.minXrange)/100\n \n self.tbMain.txtXmin.SetValue('%.2f' % self.minXrange)\n self.tbMain.txtXmax.SetValue('%.2f' % self.maxXrange)\n self.tbMain.txtYmax.SetValue('%.2f' % self.maxYrange)\n self.tbMain.txtYmin.SetValue('%.2f' % self.minYrange)\n \n # enable controls\n names = ['plcPcaLoadsV', 'plcDfaLoadsV', 'plcGaSpecLoad',\n 'plcPLSloading', 'plcGaModelPlot1', 'plcDFAscores',\n 'plcGaFeatPlot']\n\n if self.GetName() not in names:\n # disable for general case\n self.tbMain.tbConf.Enable(False)\n self.tbMain.tbPoints.Enable(False)\n self.tbMain.tbSymbols.Enable(False)\n \n if self.GetName() in ['plcPCAscore', 'plcGaFeatPlot']:\n self.tbMain.tbPoints.Enable(True)\n self.tbMain.tbSymbols.Enable(True)\n \n if len(self.GetName().split('plcPredPls')) > 1:\n if self.parent.prnt.prnt.prnt.data['plstype'] == 1:\n self.tbMain.tbPoints.Enable(True)\n self.tbMain.tbSymbols.Enable(True)\n else:\n self.tbMain.tbSymbols.Enable(True)\n \n if self.GetName() in ['plcDFAscores']:\n # dfa score plots\n self.tbMain.tbConf.Enable(True)\n self.tbMain.tbPoints.Enable(True)\n self.tbMain.tbSymbols.Enable(True)\n else:\n self.tbMain.tbConf.Enable(False)\n \n if len(self.GetName().split('plcGaModelPlot')) > 1:\n # ga-dfa score plots\n if self.parent.prnt.prnt.splitPrnt.dtype in ['DFA']:\n self.tbMain.tbConf.Enable(True)\n self.tbMain.tbPoints.Enable(True)\n self.tbMain.tbSymbols.Enable(True)\n else:\n self.tbMain.tbConf.Enable(False)\n self.tbMain.tbPoints.Enable(False)\n self.tbMain.tbSymbols.Enable(True)\n \n if self.GetName() in ['plcPcaLoadsV', 'plcDfaLoadsV', 'plcPLSloading']:\n self.tbMain.tbLoadLabels.Enable(True)\n self.tbMain.tbLoadLabStd1.Enable(True)\n self.tbMain.tbLoadLabStd2.Enable(True)\n self.tbMain.tbLoadSymStd2.Enable(True)\n self.tbMain.tbPoints.Enable(False)\n self.tbMain.tbSymbols.Enable(False)\n else:\n self.tbMain.tbLoadLabels.Enable(False)\n self.tbMain.tbLoadLabStd1.Enable(False)\n self.tbMain.tbLoadLabStd2.Enable(False)\n self.tbMain.tbLoadSymStd2.Enable(False)\n\n def enable_drag(self, value):\n \"\"\"Set True to enable drag.\"\"\"\n if value not in [True, False]:\n raise TypeError(\"Value should be True or False\")\n if value:\n if self.GetEnableZoom():\n self.enableZoom = False\n if self.get_enable_interactive():\n self.enable_interactive(False)\n self.SetCursor(self.HandCursor)\n else:\n self.SetCursor(wx.CROSS_CURSOR)\n self._dragEnabled = value\n\n def enable_interactive(self, value):\n \"\"\"Set True to enable interactive mode - RMJ 03/2008.\"\"\"\n if value not in [True, False]:\n raise TypeError(\"Value should be True or False\")\n if value:\n if self.GetEnableZoom():\n self.enableZoom = False\n if self.GetEnableDrag():\n self.enable_drag(False)\n self.SetCursor(wx.Cursor(wx.CURSOR_PENCIL))\n else:\n self.SetCursor(wx.CROSS_CURSOR)\n self._interEnabled = value\n\n def get_enable_interactive(self):\n return self._interEnabled\n \n\nclass Pca(wx.Panel):\n \"\"\"principal component analysis\n\n \"\"\"\n def __init__(self, parent, id_, pos, size, style, name):\n \"\"\"\"\"\"\n wx.Panel.__init__(self, id=wxID_PCA, name='Pca', parent=parent,\n pos=(-12, 22), size=(1024, 599),\n style=wx.TAB_TRAVERSAL)\n\n _, _, _, _, _ = id_, pos, size, style, name\n\n self.parent = parent\n self._init_ctrls()\n self._init_sizers()\n\n def _init_ctrls(self):\n \"\"\"\"\"\"\n self.SetClientSize((1016, 565))\n self.SetAutoLayout(True)\n self.SetToolTip('')\n\n self.plcPCeigs = MyPlotCanvas(id_=-1, name='plcPCeigs',\n parent=self, pos=(589, 283),\n size=(200, 200), style=0,\n toolbar=self.parent.parent.tbMain)\n self.plcPCeigs.SetToolTip('')\n self.plcPCeigs.fontSizeTitle = 10\n self.plcPCeigs.enableZoom = True\n self.plcPCeigs.fontSizeAxis = 8\n self.plcPCeigs.SetConstraints(\n LayoutAnchors(self.plcPCeigs, False, True, False, True))\n self.plcPCeigs.fontSizeLegend = 8\n\n self.plcPCvar = MyPlotCanvas(id_=-1, name='plcPCvar', parent=self,\n pos=(176, 283),\n size=(200, 200), style=0,\n toolbar=self.parent.parent.tbMain)\n self.plcPCvar.fontSizeAxis = 8\n self.plcPCvar.fontSizeTitle = 10\n self.plcPCvar.enableZoom = True\n self.plcPCvar.SetToolTip('')\n self.plcPCvar.fontSizeLegend = 8\n\n self.plcPCAscore = MyPlotCanvas(\n parent=self, id_=-1, name='plcPCAscore', pos=(0, 24),\n size=(200, 200), style=0, toolbar=self.parent.parent.tbMain)\n self.plcPCAscore.fontSizeTitle = 10\n self.plcPCAscore.fontSizeAxis = 8\n self.plcPCAscore.enableZoom = True\n self.plcPCAscore.enableLegend = True\n self.plcPCAscore.SetToolTip('')\n self.plcPCAscore.fontSizeLegend = 8\n\n self.plcPcaLoadsV = MyPlotCanvas(\n id_=-1, name='plcPcaLoadsV', parent=self, pos=(0, 24),\n size=(200, 200), style=0, toolbar=self.parent.parent.tbMain)\n self.plcPcaLoadsV.SetToolTip('')\n self.plcPcaLoadsV.fontSizeTitle = 10\n self.plcPcaLoadsV.enableZoom = True\n self.plcPcaLoadsV.fontSizeAxis = 8\n self.plcPcaLoadsV.enableLegend = True\n self.plcPcaLoadsV.fontSizeLegend = 8\n\n self.titleBar = TitleBar(\n self, id_=-1, text=\"Principal Component Analysis\",\n style=bp.BP_USE_GRADIENT, alignment=bp.BP_ALIGN_LEFT)\n\n def _init_sizers(self):\n \"\"\"\"\"\"\n self.bxsPca1 = wx.BoxSizer(orient=wx.HORIZONTAL)\n self.bxsPca2 = wx.BoxSizer(orient=wx.VERTICAL)\n self.grsPca1 = wx.GridSizer(cols=2, hgap=2, rows=2, vgap=2)\n\n self.bxsPca1.Add(self.bxsPca2, 1, border=0, flag=wx.EXPAND)\n\n self.bxsPca2.Add(self.titleBar, 0, border=0, flag=wx.EXPAND)\n self.bxsPca2.Add(self.grsPca1, 1, border=0, flag=wx.EXPAND)\n\n self.grsPca1.Add(self.plcPCAscore, 0, border=0, flag=wx.EXPAND)\n self.grsPca1.Add(self.plcPcaLoadsV, 0, border=0, flag=wx.EXPAND)\n self.grsPca1.Add(self.plcPCvar, 0, border=0, flag=wx.EXPAND)\n self.grsPca1.Add(self.plcPCeigs, 0, border=0, flag=wx.EXPAND)\n \n self.SetSizer(self.bxsPca1)\n\n def reset(self):\n self.titleBar.spnNumPcs1.Enable(0)\n self.titleBar.spnNumPcs2.Enable(0)\n self.titleBar.spnNumPcs1.SetValue(1)\n self.titleBar.spnNumPcs2.SetValue(2)\n \n objects = {'plcPCeigs': ['Eigenvalues', 'Principal Component', 'Eigenvalue'],\n 'plcPCvar': ['Percentage Explained Variance',\n 'Principal Component', 'Cumulative % Variance'],\n 'plcPCAscore': ['PCA Scores', 't[1]', 't[2]'],\n 'plcPcaLoadsV': ['PCA Loading', 'w[1]', 'w[2]']}\n\n # noinspection PyUnusedLocal, PyTypeChecker\n curve = PolyLine([[0, 0], [1, 1]], colour='white', width=1,\n style=wx.PENSTYLE_TRANSPARENT)\n \n for each in objects.keys():\n exec('self.' + each + '.Draw(PlotGraphics([curve], ' +\n 'objects[\"' + each + '\"][0], ' + 'objects[\"' + each +\n '\"][1], ' + 'objects[\"' + each + '\"][2]))')\n\n\nclass TitleBar(bp.ButtonPanel):\n def __init__(self, parent, id_, text, style, alignment):\n \"\"\"\"\"\"\n bp.ButtonPanel.__init__(self, parent=parent, id=-1,\n text=\"Principal Component Analysis\",\n agwStyle=bp.BP_USE_GRADIENT,\n alignment=bp.BP_ALIGN_LEFT)\n\n _, _, _, _ = id_, text, style, alignment\n\n self.data = None\n self.parent = parent\n self._init_btnpanel_ctrls()\n self.create_buttons()\n\n def _init_btnpanel_ctrls(self):\n \"\"\"\"\"\"\n self.Bind(wx.EVT_PAINT, self.on_btnpanel_paint)\n \n bmp = wx.Bitmap(op_join('bmp', 'run.png'), wx.BITMAP_TYPE_PNG)\n self.btnRunPCA = bp.ButtonInfo(self, -1, bmp, kind=wx.ITEM_NORMAL,\n shortHelp='Run PCA',\n longHelp='Run Principal Component Analysis')\n self.btnRunPCA.Enable(False)\n self.Bind(wx.EVT_BUTTON, self.on_btn_run_pca, id=self.btnRunPCA.GetId())\n\n bmp = wx.Bitmap(op_join('bmp', 'export.png'), wx.BITMAP_TYPE_PNG)\n self.btnExportPcaResults = bp.ButtonInfo(self, -1, bmp,\n kind=wx.ITEM_NORMAL,\n shortHelp='Export PCA Results',\n longHelp='Export PCA Results')\n self.btnExportPcaResults.Enable(False)\n self.Bind(wx.EVT_BUTTON, self.on_btn_export_pca_results,\n id=self.btnExportPcaResults.GetId())\n\n choices = ['Raw spectra', 'Processed spectra']\n self.cbxData = wx.Choice(choices=choices, id=-1, name='cbxData',\n parent=self, pos=(118, 23),\n size=(100, 23), style=0)\n self.cbxData.SetSelection(0)\n \n self.cbxPcaType = wx.Choice(choices=['NIPALS', 'SVD'], id=-1,\n name='cbxPcaType', parent=self,\n pos=(56, 23),\n size=(64, 23), style=0)\n self.cbxPcaType.Bind(wx.EVT_COMBOBOX, self.on_cbx_pca_type, id=ID_PCATYPE)\n self.cbxPcaType.SetSelection(0)\n \n choices = ['Correlation matrix', 'Covariance matrix']\n self.cbxPreprocType = wx.Choice(choices=choices, id=-1,\n name='cbxPreprocType', parent=self,\n pos=(118, 23),\n size=(110, 23), style=0, )\n self.cbxPreprocType.SetSelection(0)\n \n self.spnPCAnum = wx.SpinCtrl(id=ID_SPNPCS, initial=3, max=100,\n min=3, name='spnPCAnum', parent=self,\n pos=(112, 158),\n size=(46, 23),\n style=wx.SP_ARROW_KEYS)\n self.spnPCAnum.SetToolTip('')\n self.spnPCAnum.SetValue(3)\n \n self.spnNumPcs1 = wx.SpinCtrl(id=ID_NUMPCS1, initial=1,\n max=100, min=1, name='spnNumPcs1',\n parent=self, pos=(240, 184),\n size=(46, 23),\n style=wx.SP_ARROW_KEYS)\n self.spnNumPcs1.Enable(0)\n self.spnNumPcs1.Bind(wx.EVT_SPINCTRL, self.on_spn_num_pcs1, id=-1)\n \n self.spnNumPcs2 = wx.SpinCtrl(id=ID_NUMPCS2, initial=1, max=100, min=1,\n name='spnNumPcs2', parent=self,\n pos=(240, 184),\n size=(46, 23),\n style=wx.SP_ARROW_KEYS)\n self.spnNumPcs2.Enable(0)\n self.spnNumPcs2.Bind(wx.EVT_SPINCTRL, self.on_spn_num_pcs2, id=-1)\n\n def create_buttons(self):\n self.Freeze()\n self.set_properties()\n style = wx.TRANSPARENT_WINDOW\n\n self.AddControl(self.cbxData)\n self.AddControl(self.cbxPreprocType)\n self.AddControl(self.cbxPcaType)\n self.AddControl(GenStaticText(self, -1, 'No. PCs:', style=style))\n self.AddControl(self.spnPCAnum)\n self.AddSeparator()\n self.AddControl(GenStaticText(self, -1, 'PC', style=style))\n self.AddControl(self.spnNumPcs1)\n self.AddControl(GenStaticText(self, -1, ' vs. ', style=style))\n self.AddControl(self.spnNumPcs2)\n self.AddSeparator()\n self.AddButton(self.btnRunPCA)\n self.AddSeparator()\n self.AddButton(self.btnExportPcaResults)\n \n self.Thaw()\n self.DoLayout()\n\n # noinspection PyMethodMayBeStatic\n def on_btnpanel_paint(self, event):\n event.Skip()\n \n def set_properties(self):\n\n # Sets the colours for the two demos: called only if the user didn't\n # modify the colours and sizes using the Settings Panel\n bpArt = self.GetBPArt()\n \n # set the color the text is drawn with\n bpArt.SetColour(bp.BP_TEXT_COLOUR, wx.WHITE)\n\n background = self.GetBackgroundColour() \n bpArt.SetColour(bp.BP_TEXT_COLOUR, wx.BLUE)\n bpArt.SetColour(bp.BP_BORDER_COLOUR,\n bp.BrightenColour(background, 0.85))\n bpArt.SetColour(bp.BP_SEPARATOR_COLOUR,\n bp.BrightenColour(background, 0.85))\n bpArt.SetColour(bp.BP_BUTTONTEXT_COLOUR, wx.BLACK)\n bpArt.SetColour(bp.BP_SELECTION_BRUSH_COLOUR, wx.Colour(242, 242, 235))\n bpArt.SetColour(bp.BP_SELECTION_PEN_COLOUR, wx.Colour(206, 206, 195))\n\n def on_btn_run_pca(self, _):\n self.run_pca()\n \n def on_btn_export_pca_results(self, _):\n dlg = wx.FileDialog(self, \"Choose a file\", \".\", \"\", \n \"Any files (*.*)|*.*\", wx.FD_SAVE)\n try:\n if dlg.ShowModal() == wx.ID_OK:\n saveFile = dlg.GetPath()\n scrs = np.array2string(self.data['pcscores'], separator='\\t')\n lods = np.array2string(self.data['pcloads'], separator='\\t')\n eign = np.array2string(self.data['pceigs'], separator='\\t')\n pcex = np.array2string(self.data['pcpervar'], separator='\\t')\n out = '#PRINCIPAL_COMPONENT_SCORES\\n' + scrs + '\\n' +\\\n '#PRINCIPAL_COMPONENT_LOADINGS\\n' + lods + '\\n' +\\\n '#EIGENVALUES\\n' + eign + '\\n' +\\\n '#CUMULATIVE_PERCENTAGE_EXPLAINED_VARIANCE\\n' + pcex + '\\n'\n\n with open(saveFile, 'w') as f:\n f.write(out)\n\n finally:\n dlg.Destroy()\n \n def on_cbx_pca_type(self, _):\n if self.cbxPcaType.GetSelection() == 1:\n self.spnPCAnum.Enable(0)\n else:\n self.spnPCAnum.Enable(1)\n \n def get_data(self, data):\n self.data = data\n \n def run_pca(self):\n \"\"\"Run principal component analysis\n\n \"\"\"\n xdata = None\n\n try:\n self.spnNumPcs1.Enable(1)\n self.spnNumPcs2.Enable(1)\n self.spnNumPcs1.SetValue(1)\n self.spnNumPcs2.SetValue(2)\n \n if self.cbxData.GetSelection() == 0:\n xdata = self.data['rawtrunc']\n elif self.cbxData.GetSelection() == 1:\n xdata = self.data['proctrunc']\n \n if self.cbxPreprocType.GetSelection() == 0:\n self.data['pcatype'] = 'covar'\n elif self.cbxPreprocType.GetSelection() == 1:\n self.data['pcatype'] = 'corr'\n \n if self.cbxPcaType.GetSelection() == 1:\n # run PCA using SVD\n scores, loads, self.data['pcpervar'], eigs = \\\n chemtrics.pca_svd(xdata, self.data['pcatype'])\n \n self.data['pcscores'] = scores[:, 0: len(eigs)]\n self.data['pcloads'] = loads[0: len(eigs), :]\n self.data['pceigs'] = eigs\n self.data['niporsvd'] = 'svd'\n \n elif self.cbxPcaType.GetSelection() == 0:\n # run PCA using NIPALS\n self.data['pcscores'], self.data['pcloads'], self.data['pcpervar'], self.data['pceigs'] = \\\n chemtrics.pca_nipals(xdata, self.spnPCAnum.GetValue(),\n self.data['pcatype'],\n self.parent.parent.parent.sbMain)\n \n self.data['niporsvd'] = 'nip'\n \n # Enable ctrls\n self.btnExportPcaResults.Enable(1)\n self.spnNumPcs1.SetRange(1, len(self.data['pceigs']))\n self.spnNumPcs1.SetValue(1)\n self.spnNumPcs2.SetRange(1, len(self.data['pceigs']))\n self.spnNumPcs2.SetValue(2)\n \n # check for metadata & setup limits for dfa\n tbar = self.parent.parent.parent.plDfa.titleBar\n klass = self.data['class']\n if (sum(klass[:, 0]) != 0) and (klass is not None):\n tbar.cbxData.SetSelection(0)\n tbar.spnDfaPcs.SetRange(2, len(self.data['pceigs']))\n tbar.spnDfaDfs.SetRange(1, len(np.unique(klass[:, 0])) - 1)\n \n # plot results\n self.plot_pca()\n \n except Exception as error:\n error_box(self, '%s' % str(error))\n raise\n \n def plot_pca(self):\n # Plot pca scores and loadings\n pc1 = self.spnNumPcs1.GetValue()\n pc2 = self.spnNumPcs2.GetValue()\n\n xL = 't[' + str(pc1) + '] (' + \\\n '%.2f' % (self.data['pcpervar'][pc1] -\n self.data['pcpervar'][pc1 - 1]) + '%)'\n \n yL = 't[' + str(pc2) + '] (' + \\\n '%.2f' % (self.data['pcpervar'][pc2] -\n self.data['pcpervar'][pc2-1]) + '%)'\n \n plot_scores(self.parent.plcPCAscore, self.data['pcscores'],\n cl=self.data['class'][:, 0],\n labels=self.data['label'],\n validation=self.data['validation'],\n col1=pc1-1, col2=pc2-1, pconf=False,\n title='PCA Scores', xLabel=xL, yLabel=yL, xval=False,\n symb=self.parent.parent.parent.tbMain.tbSymbols.GetValue(),\n text=self.parent.parent.parent.tbMain.tbPoints.GetValue(),\n usecol=[], usesym=[])\n \n # Plot loadings\n if pc1 != pc2:\n plot_loads(self.parent.plcPcaLoadsV,\n np.transpose(self.data['pcloads']),\n xaxis=self.data['indlabels'], col1=pc1-1,\n col2=pc2-1, title='PCA Loadings',\n xLabel='w[' + str(pc1) + ']',\n yLabel='w[' + str(pc2) + ']',\n type=self.parent.prnt.parent.tbMain.get_load_plot_idx(),\n usecol=[], usesym=[])\n else:\n idx = pc1-1\n plot_line(self.parent.plcPcaLoadsV,\n self.data['pcloads'],\n xaxis=self.data['xaxis'], rownum=idx, tit='PCA Loadings',\n type='single', xLabel='Variable',\n yLabel='w['+str(idx+1)+']', wdth=1, ledge=[])\n \n # Plot % variance\n plot_line(self.parent.plcPCvar, np.transpose(self.data['pcpervar']),\n xaxis=np.arange(0, len(self.data['pcpervar']))[:, nax],\n rownum=0, tit='Percentage Explained Variance', type='single',\n xLabel='Principal Component', yLabel='Cumulative % Variance',\n wdth=3, ledge=[])\n \n # Plot eigenvalues\n plot_line(self.parent.plcPCeigs, np.transpose(self.data['pceigs']),\n xaxis=np.arange(1, len(self.data['pceigs']) + 1)[:, nax],\n rownum=0, tit='Eigenvalues', xLabel='Principal Component',\n yLabel='Eigenvalue', wdth=3, type='single', ledge=[])\n \n # make sure ctrls enabled\n self.spnNumPcs1.Enable(True)\n self.spnNumPcs2.Enable(True)\n self.btnExportPcaResults.Enable(True)\n \n def on_spn_num_pcs1(self, _):\n pc1 = self.spnNumPcs1.GetValue()\n pc2 = self.spnNumPcs2.GetValue()\n self.plot_pca()\n set_btn_state(pc1, pc2, self.parent.prnt.prnt.tbMain)\n \n def on_spn_num_pcs2(self, _):\n pc1 = self.spnNumPcs1.GetValue()\n pc2 = self.spnNumPcs2.GetValue()\n self.plot_pca()\n set_btn_state(pc1, pc2, self.parent.prnt.prnt.tbMain)\n\n\nclass PlotProperties(wx.Dialog):\n \"\"\"\"\"\"\n def __init__(self, parent):\n \"\"\"\"\"\"\n wx.Dialog.__init__(self, id=-1, name='', parent=parent,\n pos=(0, 0), size=(530, 480),\n style=wx.MAXIMIZE_BOX | wx.DEFAULT_DIALOG_STYLE,\n title='Plot Properties')\n\n self._init_plot_prop_ctrls()\n self._init_plot_prop_sizers()\n self._init_grs_df_scores()\n self._init_grs_loads()\n\n self.foldPnl.Expand(self.genSets)\n self.foldPnl.Collapse(self.scoreSets)\n self.foldPnl.Collapse(self.loadSets)\n\n self.graph = parent.last_draw[0]\n self.canvas = parent\n\n self.minXrange = parent.get_x_current_range()[0]\n self.maxXrange = parent.get_x_current_range()[1]\n self.minYrange = parent.get_y_current_range()[0]\n self.maxYrange = parent.get_y_current_range()[1]\n\n self.Increment = (self.maxXrange - self.minXrange) / 100\n\n self.txtXmin.SetValue('%.3f' % self.minXrange)\n self.txtXmax.SetValue('%.3f' % self.maxXrange)\n self.txtYmin.SetValue('%.3f' % self.minYrange)\n self.txtYmax.SetValue('%.3f' % self.maxYrange)\n\n self.txtTitle.SetValue(self.graph.get_title())\n self.txtXlabel.SetValue(self.graph.get_xlabel())\n self.txtYlabel.SetValue(self.graph.get_ylabel())\n\n self.spnFontSizeAxes.SetValue(parent.get_font_size_axis())\n self.spnFontSizeTitle.SetValue(parent.get_font_size_title())\n\n if self.canvas.get_enable_grid():\n self.tbGrid.SetValue(1)\n if self.canvas.get_enable_zoom():\n self.tbZoom.SetValue(1)\n if self.canvas.get_enable_drag():\n self.tbDrag.SetValue(1)\n if self.canvas.get_enable_point_label():\n self.tbPointLabel.SetValue(1)\n\n def _init_grs_df_scores(self):\n\n self.grsDfScores = wx.GridSizer(cols=2, hgap=4, rows=2, vgap=4)\n\n self._init_coll_grs_df_scores(self.grsDfScores)\n\n self.scorePnl.SetSizer(self.grsDfScores)\n\n def _init_coll_grs_df_scores(self, parent):\n # generated method, don't edit\n\n parent.Add(self.tbConf, 0, border=0, flag=wx.EXPAND)\n parent.Add(self.tbPoints, 0, border=0, flag=wx.EXPAND)\n parent.Add(self.tbSymbols, 0, border=0, flag=wx.EXPAND)\n \n def _init_grs_loads(self):\n # generated method, don't edit\n self.grsLoadings = wx.GridSizer(cols=2, hgap=4, rows=2, vgap=4)\n\n self._init_coll_grs_loads(self.grsLoadings)\n\n self.loadPnl.SetSizer(self.grsLoadings)\n \n def _init_coll_grs_loads(self, parent):\n # generated method, don't edit\n parent.Add(self.tbLoadLabels, 0, border=0, flag=wx.EXPAND)\n parent.Add(self.tbLoadLabStd1, 0, border=0, flag=wx.EXPAND)\n parent.Add(self.tbLoadLabStd2, 0, border=0, flag=wx.EXPAND)\n parent.Add(self.tbLoadSymStd2, 0, border=0, flag=wx.EXPAND)\n \n def _init_coll_gbs_plot_props(self, parent):\n # generated method, don't edit\n flag = wx.EXPAND\n parent.Add(self.stTitle, (0, 0), border=4, flag=flag, span=(1, 1))\n parent.Add(self.txtTitle, (0, 1), border=4, flag=flag, span=(1, 5))\n parent.Add(wx.StaticText(self.genPnl, -1, 'Axes font',\n style=wx.ALIGN_LEFT),\n (1, 0), flag=flag, span=(1, 1))\n parent.Add(self.spnFontSizeAxes, (1, 1), border=4, flag=flag,\n span=(1, 1))\n parent.Add(wx.StaticText(self.genPnl, -1, 'Title font',\n style=wx.ALIGN_LEFT),\n (1, 2), flag=flag, span=(1, 1))\n parent.Add(self.spnFontSizeTitle, (1, 3), border=4, flag=flag,\n span=(1, 1))\n parent.Add(self.stXlabel, (2, 0), border=4, flag=flag, span=(1, 1))\n parent.Add(self.txtXlabel, (2, 1), border=4, flag=flag, span=(1, 5))\n parent.Add(self.stYlabel, (3, 0), border=4, flag=flag, span=(1, 1))\n parent.Add(self.txtYlabel, (3, 1), border=4, flag=flag, span=(1, 5))\n parent.Add(self.stXfrom, (4, 0), border=4, flag=flag, span=(1, 1))\n parent.Add(self.txtXmin, (4, 1), border=4, flag=flag, span=(1, 1))\n parent.Add(self.spnXmin, (4, 2), border=4, flag=flag, span=(1, 1))\n parent.Add(self.stXto, (4, 3), border=4, flag=flag, span=(1, 1))\n parent.Add(self.txtXmax, (4, 4), border=4, flag=flag, span=(1, 1))\n parent.Add(self.spnXmax, (4, 5), border=4, flag=flag, span=(1, 1))\n parent.Add(self.stYfrom, (5, 0), border=4, flag=flag, span=(1, 1))\n parent.Add(self.txtYmin, (5, 1), border=4, flag=flag, span=(1, 1))\n parent.Add(self.spnYmin, (5, 2), border=4, flag=flag, span=(1, 1))\n parent.Add(self.stYto, (5, 3), border=4, flag=flag, span=(1, 1))\n parent.Add(self.txtYmax, (5, 4), border=4, flag=flag, span=(1, 1))\n parent.Add(self.spnYmax, (5, 5), border=4, flag=flag, span=(1, 1))\n parent.Add(self.tbDrag, (6, 1), border=4, flag=flag, span=(1, 1))\n parent.Add(self.tbGrid, (6, 2), border=4, flag=flag, span=(1, 1))\n parent.Add(self.tbPointLabel, (6, 3), border=4, flag=flag, span=(1, 1))\n parent.Add(self.tbZoom, (6, 4), border=4, flag=flag, span=(1, 1))\n parent.Add(self.cbApply, (7, 0), border=4, flag=flag, span=(1, 1))\n parent.Add(self.btnApply, (7, 1), border=4, flag=flag, span=(1, 5))\n # parent.AddSpacer((8, 8), (8, 0), flag=flag, span=(2, 6))\n\n # noinspection PyMethodMayBeStatic\n def _init_coll_gbs_plot_props_growables(self, parent):\n # generated method, don't edit\n for col in range(6):\n parent.AddGrowableCol(col)\n\n def _init_plot_prop_sizers(self):\n # generated method, don't edit\n self.gbsPlotProps = wx.GridBagSizer(8, 8)\n self.gbsPlotProps.SetCols(6)\n self.gbsPlotProps.SetRows(6)\n self.gbsPlotProps.SetNonFlexibleGrowMode(wx.FLEX_GROWMODE_SPECIFIED)\n self.gbsPlotProps.SetMinSize((250, 439))\n self.gbsPlotProps.SetEmptyCellSize((0, 0))\n self.gbsPlotProps.SetFlexibleDirection(wx.HORIZONTAL)\n\n self._init_coll_gbs_plot_props(self.gbsPlotProps)\n self._init_coll_gbs_plot_props_growables(self.gbsPlotProps)\n\n self.genPnl.SetSizer(self.gbsPlotProps)\n\n def _init_plot_prop_ctrls(self):\n\n self.SetAutoLayout(True)\n \n self.foldPnl = fpb.FoldPanelBar(self, -1, wx.DefaultPosition,\n (525, 450),\n fpb.FPB_EXCLUSIVE_FOLD)\n self.foldPnl.SetConstraints(\n LayoutAnchors(self.foldPnl, True, True, True, True))\n self.foldPnl.SetAutoLayout(True)\n \n icons = wx.ImageList(16, 16)\n icons.Add(wx.Bitmap(op_join('bmp', 'arrown.png'), wx.BITMAP_TYPE_PNG))\n icons.Add(wx.Bitmap(op_join('bmp', 'arrows.png'), wx.BITMAP_TYPE_PNG))\n \n self.genSets = self.foldPnl.AddFoldPanel(\"General properties\", \n collapsed=True,\n foldIcons=icons)\n \n self.scoreSets = self.foldPnl.AddFoldPanel(\"Score plots\", \n collapsed=True,\n foldIcons=icons)\n self.scoreSets.Enable(False)\n \n self.loadSets = self.foldPnl.AddFoldPanel(\"Loadings plots\", \n collapsed=True,\n foldIcons=icons)\n self.loadSets.Enable(False)\n \n self.genPnl = wx.Panel(id=-1, name='genPnl', parent=self.genSets,\n pos=(0, 0), size=(20, 250),\n style=wx.TAB_TRAVERSAL)\n self.genPnl.SetToolTip('')\n \n self.scorePnl = wx.Panel(id=-1, name='scorePnl', parent=self.scoreSets,\n pos=(0, 0), size=(20, 100),\n style=wx.TAB_TRAVERSAL)\n self.scorePnl.SetToolTip('') \n \n self.loadPnl = wx.Panel(id=-1, name='loadPnl', parent=self.loadSets,\n pos=(0, 0), size=(20, 100),\n style=wx.TAB_TRAVERSAL)\n self.loadPnl.SetToolTip('') \n \n self.stTitle = wx.StaticText(id=-1, label='Title', name=u'stTitle',\n parent=self.genPnl, pos=(0, 0),\n size=(21, 24), style=0)\n self.stTitle.SetToolTip('')\n\n self.stYfrom = wx.StaticText(id=-1, label=u'Y-Axis From:',\n name=u'stYfrom', parent=self.genPnl,\n pos=(0, 131), size=(42, 24),\n style=0)\n self.stYfrom.SetToolTip('')\n\n self.stYto = wx.StaticText(id=-1, label='To:', name=u'stYto',\n parent=self.genPnl, pos=(144, 131),\n size=(40, 21), style=0)\n self.stYto.SetToolTip('')\n\n self.stXfrom = wx.StaticText(id=-1, label=u'X-Axis From:',\n name=u'stXfrom', parent=self.genPnl,\n pos=(0, 103), size=(40, 21),\n style=0)\n self.stXfrom.SetToolTip('')\n\n self.stXto = wx.StaticText(id=-1, label='To:', name=u'stXto',\n parent=self.genPnl, pos=(144, 103),\n size=(40, 21), style=0)\n self.stXto.SetToolTip('')\n\n self.stXlabel = wx.StaticText(id=-1, label='X label', name=u'stXlabel',\n parent=self.genPnl, pos=(0, 53),\n size=(40, 21), style=0)\n self.stXlabel.SetToolTip('')\n\n self.stYlabel = wx.StaticText(id=-1, label='Y label', name=u'stYlabel',\n parent=self.genPnl, pos=(0, 78),\n size=(40, 21), style=0)\n self.stYlabel.SetToolTip('')\n\n self.txtTitle = wx.TextCtrl(id=-1, name='txtTitle', parent=self.genPnl,\n pos=(15, 0), size=(40, 21),\n style=0, value='')\n self.txtTitle.SetToolTip('')\n self.txtTitle.Bind(wx.EVT_TEXT, self.on_txt_title)\n\n self.txtYlabel = wx.TextCtrl(id=-1, name='txtYlabel', parent=self.genPnl,\n pos=(15, 78), size=(40, 21),\n style=0, value='')\n self.txtYlabel.SetToolTip('')\n\n self.txtXlabel = wx.TextCtrl(id=-1, name='txtXlabel',\n parent=self.genPnl, pos=(15, 53),\n size=(40, 21), style=0, value='')\n self.txtXlabel.SetToolTip('')\n\n self.txtXmin = wx.TextCtrl(id=-1, name='txtXmin',\n parent=self.genPnl, pos=(15, 103),\n size=(40, 21), style=0, value='')\n self.txtXmin.SetToolTip('')\n\n self.spnXmin = wx.SpinButton(id=-1, name='spnXmin',\n parent=self.genPnl, pos=(96, 103),\n size=(15, 21), style=wx.SP_VERTICAL)\n self.spnXmin.SetToolTip('')\n self.spnXmin.Bind(wx.EVT_SPIN_DOWN, self.on_spn_xmin_down)\n self.spnXmin.Bind(wx.EVT_SPIN_UP, self.on_spn_xmin_up)\n self.spnXmin.Bind(wx.EVT_SPIN, self.on_spn_xmin)\n\n self.spnXmax = wx.SpinButton(id=-1, name='spnXmax',\n parent=self.genPnl, pos=(240, 103),\n size=(15, 21), style=wx.SP_VERTICAL)\n self.spnXmax.SetToolTip('')\n self.spnXmax.Bind(wx.EVT_SPIN_DOWN, self.on_spn_xmax_down)\n self.spnXmax.Bind(wx.EVT_SPIN_UP, self.on_spn_xmax_up)\n self.spnXmax.Bind(wx.EVT_SPIN, self.on_spn_xmax)\n\n self.spnYmax = wx.SpinButton(id=-1, name='spnYmax',\n parent=self.genPnl, pos=(240, 131),\n size=(15, 21), style=wx.SP_VERTICAL)\n self.spnYmax.SetToolTip('')\n self.spnYmax.Bind(wx.EVT_SPIN_DOWN, self.on_spn_ymax_down)\n self.spnYmax.Bind(wx.EVT_SPIN_UP, self.on_spn_ymax_up)\n self.spnYmax.Bind(wx.EVT_SPIN, self.on_spn_ymax)\n\n self.spnYmin = wx.SpinButton(id=-1, name='spnYmin',\n parent=self.genPnl, pos=(96, 131),\n size=(15, 21), style=wx.SP_VERTICAL)\n self.spnYmin.SetToolTip('')\n self.spnYmin.Bind(wx.EVT_SPIN_DOWN, self.on_spn_ymin_down)\n self.spnYmin.Bind(wx.EVT_SPIN_UP, self.on_spn_ymin_up)\n self.spnYmin.Bind(wx.EVT_SPIN, self.on_spn_ymin)\n\n self.txtXmax = wx.TextCtrl(id=-1, name='txtXmax', parent=self.genPnl,\n pos=(192, 103), size=(40, 21),\n style=0, value='')\n self.txtXmax.SetToolTip('')\n\n self.txtYmax = wx.TextCtrl(id=-1, name='txtYmax', parent=self.genPnl,\n pos=(192, 131), size=(40, 21),\n style=0, value='')\n self.txtYmax.SetToolTip('')\n\n self.txtYmin = wx.TextCtrl(id=-1, name='txtYmin',\n parent=self.genPnl, pos=(15, 131),\n size=(40, 21), style=0, value='')\n self.txtYmin.SetToolTip('')\n\n self.stFont = wx.StaticText(id=-1, name=u'stFont',\n label='Font size axes and title (pt)',\n parent=self.genPnl, pos=(0, 28),\n size=(40, 21), style=0)\n self.stFont.SetToolTip('')\n\n self.spnFontSizeAxes = wx.SpinCtrl(id=-1, name='spnFontSizeAxes',\n initial=8, max=76, min=4,\n parent=self.genPnl,\n pos=(15, 28),\n size=(40, 21),\n style=wx.SP_ARROW_KEYS)\n self.spnFontSizeAxes.SetToolTip('')\n self.spnFontSizeAxes.SetValue(8)\n self.spnFontSizeAxes.SetRange(4, 76)\n self.spnFontSizeAxes.Bind(wx.EVT_SPIN, self.on_spn_font_size_axes)\n \n self.spnFontSizeTitle = wx.SpinCtrl(id=-1, initial=8, max=76, min=4,\n name='spnFontSizeTitle',\n parent=self.genPnl,\n pos=(15, 28),\n size=(40, 21),\n style=wx.SP_ARROW_KEYS)\n self.spnFontSizeTitle.SetToolTip('')\n self.spnFontSizeTitle.SetValue(8)\n self.spnFontSizeTitle.SetRange(4, 76)\n self.spnFontSizeTitle.Bind(wx.EVT_SPIN, self.on_spn_font_size_title)\n \n self.tbGrid = wxTogBut(id=-1, name='tbGrid', label='Grid',\n parent=self.genPnl, pos=(248, 48),\n size=(40, 21), style=0)\n self.tbGrid.SetValue(False)\n self.tbGrid.SetToolTip('')\n self.tbGrid.Bind(wx.EVT_BUTTON, self.on_tb_grid)\n \n self.tbDrag = wxTogBut(id=-1, name='tbDrag', label='Drag',\n parent=self.genPnl, pos=(248, 48),\n size=(40, 21), style=0)\n self.tbDrag.SetValue(False)\n self.tbDrag.SetToolTip('')\n self.tbDrag.Bind(wx.EVT_BUTTON, self.on_tb_drag_button)\n \n self.tbPointLabel = wxTogBut(id=-1, label='Points',\n name='tbPointLabel', parent=self.genPnl,\n pos=(248, 48),\n size=(40, 21), style=0)\n self.tbPointLabel.SetValue(False)\n self.tbPointLabel.SetToolTip('')\n self.tbPointLabel.Bind(wx.EVT_BUTTON, self.on_tb_point_label)\n \n self.tbZoom = wxTogBut(id=-1, label='zoom', name='tbZoom',\n parent=self.genPnl, pos=(248, 48),\n size=(40, 21), style=0)\n self.tbZoom.SetValue(True)\n self.tbZoom.SetToolTip('')\n self.tbZoom.Bind(wx.EVT_BUTTON, self.on_tb_zoom_button)\n \n self.cbApply = wx.CheckBox(id=-1, name='cbApply',\n label='Immediate Apply',\n parent=self.genPnl, pos=(48, 96),\n size=(70, 13), style=0)\n \n self.btnApply = wx.Button(id=-1, label='Apply & Close',\n name='btnApply', parent=self.genPnl,\n pos=(192, 136),\n size=(40, 21), style=0)\n self.btnApply.Bind(wx.EVT_BUTTON, self.on_btn_apply)\n \n self.tbConf = wxTogBut(id=-1, name='tbConf',\n label='95% Confidence Circles',\n parent=self.scorePnl, pos=(248, 48),\n size=(40, 21))\n self.tbConf.SetValue(True)\n self.tbConf.SetToolTip('')\n self.tbConf.Bind(wx.EVT_BUTTON, self.on_tb_conf)\n \n self.tbPoints = wxTogBut(id=-1, label='Labels',\n name='tbPoints', parent=self.scorePnl,\n pos=(248, 48), size=(40, 21))\n self.tbPoints.SetValue(True)\n self.tbPoints.SetToolTip('')\n self.tbPoints.Bind(wx.EVT_BUTTON, self.on_tb_points)\n \n self.tbSymbols = wxTogBut(id=-1, name='tbSymbols', label='Symbols',\n parent=self.scorePnl, pos=(248, 48),\n size=(40, 21))\n self.tbSymbols.SetValue(False)\n self.tbSymbols.SetToolTip('')\n self.tbSymbols.Bind(wx.EVT_BUTTON, self.on_tb_symbols)\n \n self.tbLoadLabels = wx.Button(id=-1, name='tbLoadLabels',\n label='Labels', parent=self.loadPnl,\n pos=(248, 48),\n size=(40, 21))\n self.tbLoadLabels.SetToolTip('')\n self.tbLoadLabels.Bind(wx.EVT_BUTTON, self.on_tb_load_labels)\n \n self.tbLoadLabStd1 = wx.Button(id=-1, name='tbLoadLabStd1',\n label='Labels & 1 Std',\n parent=self.loadPnl,\n pos=(248, 48),\n size=(40, 21))\n self.tbLoadLabStd1.SetToolTip('')\n self.tbLoadLabStd1.Bind(wx.EVT_BUTTON, self.on_tb_load_lab_std1)\n \n self.tbLoadLabStd2 = wx.Button(id=-1, name='tbLoadLabStd2',\n label='Labels & 2 Std',\n parent=self.loadPnl,\n pos=(248, 48),\n size=(40, 21))\n self.tbLoadLabStd2.SetToolTip('')\n self.tbLoadLabStd2.Bind(wx.EVT_BUTTON, self.on_tb_load_lab_std2)\n \n self.tbLoadSymStd2 = wx.Button(id=-1, label='Symbols & 2 Std',\n name='tbLoadSymStd2',\n parent=self.loadPnl,\n pos=(248, 48),\n size=(40, 21))\n self.tbLoadSymStd2.SetToolTip('')\n self.tbLoadSymStd2.Bind(wx.EVT_BUTTON, self.on_tb_load_sym_std2)\n \n style = fpb.FPB_ALIGN_WIDTH\n self.foldPnl.AddFoldPanelWindow(self.genSets, self.genPnl, style)\n self.foldPnl.AddFoldPanelWindow(self.scoreSets, self.scorePnl, style)\n self.foldPnl.AddFoldPanelWindow(self.loadSets, self.loadPnl, style)\n \n # self.btnFont = wx.Button(id_=-1, label='Font',\n # name='btnFont', parent=self.genSets, pos=(192, 136),\n # size=(40, 21), style=0)\n # self.btnFont.Bind(wx.EVT_BUTTON, self.OnBtnFont)\n \n def on_tb_load_labels(self, _):\n # plot loadings\n self.do_plot(loadType=0)\n \n def on_tb_load_lab_std1(self, _):\n # plot loadings\n self.do_plot(loadType=1)\n \n def on_tb_load_lab_std2(self, _):\n # plot loadings\n self.do_plot(loadType=2)\n \n def on_tb_load_sym_std2(self, _):\n # plot loadings\n self.do_plot(loadType=3)\n \n def on_tb_conf(self, _):\n if (self.tbPoints.GetValue() is False) & \\\n (self.tbConf.GetValue() is False) & \\\n (self.tbSymbols.GetValue() is False) is False:\n # plot scores\n self.do_plot()\n \n def on_tb_points(self, _):\n if (self.tbPoints.GetValue() is False) & \\\n (self.tbConf.GetValue() is False) & \\\n (self.tbSymbols.GetValue() is False) is False:\n # plot scores\n self.do_plot()\n \n def on_tb_symbols(self, _):\n if (self.tbPoints.GetValue() is False) & \\\n (self.tbConf.GetValue() is False) & \\\n (self.tbSymbols.GetValue() is False) is False:\n # plot scores\n self.do_plot()\n \n def do_plot(self, loadType=0):\n tbar = self.canvas.prnt.titleBar\n ptbar = self.canvas.prnt.prnt.splitPrnt.titleBar\n pprnt = self.canvas.prnt.prnt.prnt.splitPrnt\n if self.canvas.GetName() in ['plcDFAscores']:\n plot_scores(self.canvas, tbar.data['dfscores'],\n cl=tbar.data['class'][:, 0],\n labels=tbar.data['label'],\n validation=tbar.data['validation'],\n col1=tbar.spnDfaScore1.GetValue() - 1,\n col2=tbar.spnDfaScore2.GetValue() - 1,\n title=self.graph.title, xLabel=self.graph.xLabel,\n yLabel=self.graph.yLabel,\n xval=tbar.cbDfaXval.GetValue(),\n text=self.tbPoints.GetValue(),\n pconf=self.tbConf.GetValue(),\n symb=self.tbSymbols.GetValue(), usecol=[], usesym=[])\n \n elif self.canvas.GetName() in ['plcPCAscore']:\n plot_scores(self.canvas, tbar.data['pcscores'],\n cl=tbar.data['class'][:, 0],\n labels=tbar.data['label'],\n validation=tbar.data['validation'],\n col1=tbar.spnNumPcs1.GetValue() - 1,\n col2=tbar.spnNumPcs2.GetValue() - 1,\n title=self.graph.title, xLabel=self.graph.xLabel,\n yLabel=self.graph.yLabel, xval=False,\n text=self.tbPoints.GetValue(), pconf=False,\n symb=self.tbSymbols.GetValue(), usecol=[], usesym=[])\n \n elif len(self.GetName().split('plcPredPls')) > 1:\n self.canvas = plot_pls_model(self.canvas, model='full',\n tbar=self.canvas.prnt.prnt.prnt.prnt.tbMain,\n cL=tbar.data['class'][:, nax],\n label=tbar.data['label'],\n scores=tbar.data['plst'],\n predictions=tbar.data['plspred'],\n validation=np.array(tbar.data['validation'], 'i')[:, nax],\n RMSEPT=tbar.data['RMSEPT'],\n factors=tbar.data['plsfactors'],\n type=tbar.data['plstype'],\n col1=tbar.spnPLSfactor1.GetValue() - 1,\n col2=tbar.spnPLSfactor2.GetValue() - 1,\n symbols=self.tbSymbols.GetValue(),\n usetxt=self.tbPoints.GetValue(),\n plScL=tbar.data['pls_class'])\n \n elif self.canvas.GetName() in ['plcGaFeatPlot']:\n plot_scores(self.canvas, ptbar.data['gavarcoords'],\n cl=ptbar.data['class'][:, 0],\n labels=ptbar.data['label'],\n validation=ptbar.data['validation'],\n col1=0, col2=1, title=self.graph.title,\n xLabel=self.graph.xLabel, yLabel=self.graph.yLabel,\n xval=True, text=self.tbPoints.GetValue(), pconf=False,\n symb=self.tbSymbols.GetValue(), usecol=[], usesym=[])\n \n elif len(self.GetName().split('plcGaModelPlot')) > 1:\n if self.canvas.prnt.prnt.splitPrnt.type in ['DFA']:\n plot_scores(self.canvas, ptbar.data['gadfadfscores'],\n cl=ptbar.data['class'][:, 0],\n labels=ptbar.data['label'],\n validation=ptbar.data['validation'],\n col1=ptbar.spnGaScoreFrom.GetValue() - 1,\n col2=ptbar.spnGaScoreTo.GetValue() - 1,\n title=self.graph.title, xLabel=self.graph.xLabel,\n yLabel=self.graph.yLabel, xval=True,\n text=self.tbPoints.GetValue(),\n pconf=self.tbConf.GetValue(),\n symb=self.tbSymbols.GetValue(), usecol=[], usesym=[])\n else:\n self.canvas = plot_pls_model(self.canvas, model='ga',\n tbar=self.canvas.prnt.prnt.splitPrnt.prnt.prnt.tbMain,\n cL=ptbar.data['class'][:, 0],\n scores=None,\n label=ptbar.data['label'],\n predictions=ptbar.data['gaplsscores'],\n validation=ptbar.data['validation'],\n RMSEPT=ptbar.data['gaplsrmsept'],\n factors=ptbar.data['gaplsfactors'],\n type=0, col1=ptbar.spnGaScoreFrom.GetValue()-1,\n col2=ptbar.spnGaScoreTo.GetValue()-1,\n symbols=self.tbSymbols.GetValue(),\n usetxt=self.tbPoints.GetValue(),\n usecol=[], usesym=[],\n plScL=ptbar.data['pls_class'])\n \n elif self.canvas.GetName() in ['plcPcaLoadsV']:\n plot_loads(self.canvas, np.transpose(tbar.data['pcloads']),\n xaxis=tbar.data['indlabels'],\n col1=tbar.spnNumPcs1.GetValue()-1,\n col2=tbar.spnNumPcs2.GetValue()-1,\n title=self.graph.title, xLabel=self.graph.xLabel,\n yLabel=self.graph.yLabel, type=loadType,\n usecol=[], usesym=[])\n \n elif self.canvas.GetName() in ['plcPLSloading']:\n plot_loads(self.canvas, tbar.data['plsloads'],\n xaxis=tbar.data['indlabels'],\n col1=tbar.spnPLSfactor1.GetValue()-1,\n col2=tbar.spnPLSfactor2.GetValue()-1,\n title=self.graph.title, xLabel=self.graph.xLabel,\n yLabel=self.graph.yLabel, type=loadType,\n usecol=[], usesym=[])\n \n elif self.canvas.GetName() in ['plcDfaLoadsV']:\n plot_loads(self.canvas, tbar.data['dfloads'],\n xaxis=tbar.data['indlabels'],\n col1=tbar.spnDfaScore1.GetValue() - 1,\n col2=tbar.spnDfaScore2.GetValue() - 1,\n title=self.graph.title, xLabel=self.graph.xLabel,\n yLabel=self.graph.yLabel, type=loadType,\n usecol=[], usesym=[])\n \n elif self.canvas.GetName() in ['plcGaSpecLoad']:\n if pprnt.type in ['DFA']:\n labels = []\n for each in pprnt.titleBar.data['gacurrentchrom']:\n labels.append(pprnt.titleBar.data['indlabels'][int(each)])\n\n plot_loads(self.canvas, pprnt.titleBar.data['gadfadfaloads'],\n xaxis=labels, title=self.graph.title,\n xLabel=self.graph.xLabel, yLabel=self.graph.yLabel,\n type=loadType, usecol=[], usesym=[])\n \n elif self.canvas.GetName() in ['plcGaSpecLoad']:\n if self.canvas.prnt.prnt.splitPrnt.type in ['PLS']:\n labels = []\n for each in pprnt.titleBar.data['gacurrentchrom']:\n labels.append(pprnt.titleBar.data['indlabels'][int(each)])\n\n plot_loads(self.canvas, ptbar.data['gaplsplsloads'],\n xaxis=labels, title=self.graph.title,\n xLabel=self.graph.xLabel, yLabel=self.graph.yLabel,\n type=loadType, usecol=[], usesym=[])\n \n # def OnBtnFont(self, event):\n # data = wx.FontData()\n # data.EnableEffects(True)\n # data.SetColour(self.canvas.GetForegroundColour())\n # data.SetInitialFont(self.canvas.GetFont())\n #\n # dlg = wx.FontDialog(self, data)\n # if dlg.ShowModal() == wx.ID_OK:\n # self.font = dlg.GetFontData().GetChosenFont()\n # self.colour = dlg.GetFontData().GetColour()\n #\n # if self.cbApply.GetValue() is True:\n # self.canvas.SetFont(self.font)\n # self.canvas.SetForegroundColour(self.colour)\n # self.canvas.redraw()\n \n def on_txt_title(self, _):\n if self.cbApply.GetValue() is True:\n self.graph.set_title(self.txtTitle.GetValue())\n self.canvas.redraw()\n \n def on_tb_grid(self, _):\n self.canvas.enable_grid(self.tbGrid.GetValue())\n \n def on_tb_drag_button(self, _):\n self.canvas.enable_drag(self.tbDrag.GetValue())\n \n def on_tb_point_label(self, _):\n self.canvas.enable_point_label(self.tbPointLabel.GetValue())\n \n def on_tb_zoom_button(self, _):\n self.canvas.enable_zoom(self.tbZoom.GetValue())\n \n def on_btn_apply(self, _):\n self.canvas.font_size_axis = self.spnFontSizeAxes.GetValue()\n self.canvas.font_size_title = self.spnFontSizeTitle.GetValue()\n \n self.graph.set_title(self.txtTitle.GetValue())\n self.graph.set_xlabel(self.txtXlabel.GetValue())\n self.graph.set_ylabel(self.txtYlabel.GetValue())\n \n xmin = float(self.txtXmin.GetValue())\n xmax = float(self.txtXmax.GetValue())\n ymin = float(self.txtYmin.GetValue())\n ymax = float(self.txtYmax.GetValue())\n\n if (xmin < xmax) and (ymin < ymax):\n self.canvas.last_draw = [self.canvas.last_draw[0],\n np.array([xmin, xmax]),\n np.array([ymin, ymax])]\n self.canvas.redraw()\n self.Close()\n \n def on_spn_font_size_axes(self, _):\n if self.cbApply.GetValue() is True:\n self.canvas.font_size_axis = self.spnFontSizeAxes.GetValue()\n self.canvas.redraw()\n \n def on_spn_font_size_title(self, _):\n if self.cbApply.GetValue() is True:\n self.canvas.font_size_title = self.spnFontSizeTitle.GetValue()\n self.canvas.redraw()\n \n def resize_axes(self):\n xmin = float(self.txtXmin.GetValue())\n xmax = float(self.txtXmax.GetValue())\n ymin = float(self.txtYmin.GetValue())\n ymax = float(self.txtYmax.GetValue())\n\n if (xmin < xmax) and (ymin < ymax) and self.cbApply.GetValue():\n self.canvas.last_draw = [self.canvas.last_draw[0],\n np.array([xmin, xmax]),\n np.array([ymin, ymax])]\n self.canvas.redraw()\n \n def on_spn_xmin(self, _):\n self.resize_axes()\n \n def on_spn_xmax(self, _):\n self.resize_axes()\n \n def on_spn_ymin(self, _):\n self.resize_axes()\n \n def on_spn_ymax(self, _):\n self.resize_axes()\n \n def on_spn_xmin_up(self, _):\n curr = float(self.txtXmin.GetValue())\n curr = curr + self.Increment\n self.txtXmin.SetValue('%.3f' % curr)\n\n def on_spn_xmin_down(self, _):\n curr = float(self.txtXmin.GetValue())\n curr = curr - self.Increment\n self.txtXmin.SetValue('%.3f' % curr)\n\n def on_spn_xmax_up(self, _):\n curr = float(self.txtXmax.GetValue())\n curr = curr + self.Increment\n self.txtXmax.SetValue('%.3f' % curr)\n\n def on_spn_xmax_down(self, _):\n curr = float(self.txtXmax.GetValue())\n curr = curr - self.Increment\n self.txtXmax.SetValue('%.3f' % curr)\n\n def on_spn_ymax_up(self, _):\n curr = float(self.txtYmax.GetValue())\n curr = curr + self.Increment\n self.txtYmax.SetValue('%.3f' % curr)\n\n def on_spn_ymax_down(self, _):\n curr = float(self.txtYmax.GetValue())\n curr = curr - self.Increment\n self.txtYmax.SetValue('%.3f' % curr)\n\n def on_spn_ymin_up(self, _):\n curr = float(self.txtYmin.GetValue())\n curr = curr + self.Increment\n self.txtYmin.SetValue('%.3f' % curr)\n\n def on_spn_ymin_down(self, _):\n curr = float(self.txtYmin.GetValue())\n curr = curr - self.Increment\n self.txtYmin.SetValue('%.3f' % curr)\n\n\nif __name__ == '__main__':\n app = wx.App()\n # This needs parameters to work\n # noinspection PyTypeChecker\n props = PlotProperties(None)\n app.MainLoop()\n", "repo_name": "joaquinabian/PyChem3K", "sub_path": "pca.py", "file_name": "pca.py", "file_ext": "py", "file_size_in_byte": 110910, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "20", "api": [{"api_name": "wx.NewId", "line_number": 27, "usage_type": "call"}, {"api_name": "wx.NewId", "line_number": 31, "usage_type": "call"}, {"api_name": "wx.NewId", "line_number": 42, "usage_type": "call"}, {"api_name": "wx.NewId", "line_number": 45, "usage_type": "call"}, {"api_name": "wx.Button", "line_number": 74, "usage_type": "call"}, {"api_name": "wx.NewId", "line_number": 74, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 75, "usage_type": "attribute"}, {"api_name": "wx.Button", "line_number": 77, "usage_type": "call"}, {"api_name": "wx.NewId", "line_number": 77, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 78, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 80, "usage_type": "call"}, {"api_name": "wx.GridSizer", "line_number": 102, "usage_type": "call"}, {"api_name": "wx.EXPAND", "line_number": 104, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 105, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 106, "usage_type": "attribute"}, {"api_name": "numpy.unique", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 137, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 149, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 150, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 152, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 154, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 156, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 158, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 160, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 162, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 164, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 166, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 170, "usage_type": "name"}, {"api_name": "commons.PolyMarker", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 174, "usage_type": "name"}, {"api_name": "wx.lib.plot.polyobjects.PlotGraphics", "line_number": 181, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 215, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 218, "usage_type": "attribute"}, {"api_name": "numpy.take", "line_number": 223, "usage_type": "call"}, {"api_name": "mva.chemometrics._index", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 225, "usage_type": "call"}, {"api_name": "mva.chemometrics._index", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 227, "usage_type": "call"}, {"api_name": "mva.chemometrics._index", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 228, "usage_type": "call"}, {"api_name": "mva.chemometrics._index", "line_number": 228, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 231, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 232, "usage_type": "name"}, {"api_name": "wx.BRUSHSTYLE_SOLID", "line_number": 235, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 243, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 245, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 249, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 249, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 251, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 254, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 256, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PlotGraphics", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 321, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 322, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 324, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 325, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 381, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 382, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 383, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 386, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 387, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 388, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 389, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 389, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 390, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 391, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 392, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 394, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 395, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 396, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 397, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 397, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 403, "usage_type": "call"}, {"api_name": "wx.BRUSHSTYLE_TRANSPARENT", "line_number": 406, "usage_type": "attribute"}, {"api_name": "commons.PolyMarker", "line_number": 408, "usage_type": "call"}, {"api_name": "wx.BRUSHSTYLE_TRANSPARENT", "line_number": 411, "usage_type": "attribute"}, {"api_name": "commons.PolyMarker", "line_number": 413, "usage_type": "call"}, {"api_name": "wx.BRUSHSTYLE_TRANSPARENT", "line_number": 416, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 419, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 419, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 422, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PlotGraphics", "line_number": 424, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 433, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 462, "usage_type": "name"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 463, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 463, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 465, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PlotGraphics", "line_number": 466, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 473, "usage_type": "name"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 475, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 475, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 479, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 481, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 481, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 484, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PlotGraphics", "line_number": 488, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 506, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 507, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 508, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 510, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 511, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 514, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PlotGraphics", "line_number": 516, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 541, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 564, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 567, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 569, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 571, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 571, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 571, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 572, "usage_type": "name"}, {"api_name": "commons.PolyMarker", "line_number": 581, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 587, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 591, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 598, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 607, "usage_type": "call"}, {"api_name": "wx.Colour", "line_number": 609, "usage_type": "call"}, {"api_name": "wx.Colour", "line_number": 610, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 614, "usage_type": "call"}, {"api_name": "wx.Colour", "line_number": 616, "usage_type": "call"}, {"api_name": "wx.Colour", "line_number": 617, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PlotGraphics", "line_number": 623, "usage_type": "call"}, {"api_name": "mva.chemometrics._index", "line_number": 665, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 668, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 669, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 671, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 675, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 676, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 679, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 680, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 682, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 685, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 685, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 685, "usage_type": "name"}, {"api_name": "commons.PolyMarker", "line_number": 689, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 692, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 693, "usage_type": "call"}, {"api_name": "mva.chemometrics._index", "line_number": 696, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 698, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 698, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 700, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PlotGraphics", "line_number": 704, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PlotGraphics", "line_number": 707, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 736, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 736, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 737, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 738, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 738, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 739, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 739, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 747, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 750, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 752, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 757, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 767, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 773, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 783, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 790, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 794, "usage_type": "call"}, {"api_name": "numpy.mod", "line_number": 800, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 806, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 813, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 824, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 827, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 827, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 833, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 833, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 836, "usage_type": "call"}, {"api_name": "plot.PolyEllipse", "line_number": 836, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 838, "usage_type": "attribute"}, {"api_name": "plot.append", "line_number": 840, "usage_type": "call"}, {"api_name": "plot.PolyEllipse", "line_number": 840, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 842, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PlotGraphics", "line_number": 845, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 883, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 889, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 889, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 890, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 892, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 894, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 895, "usage_type": "call"}, {"api_name": "numpy.take", "line_number": 895, "usage_type": "call"}, {"api_name": "mva.chemometrics._index", "line_number": 895, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 896, "usage_type": "name"}, {"api_name": "plot.append", "line_number": 914, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 923, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 927, "usage_type": "call"}, {"api_name": "plot.PolyEllipse", "line_number": 927, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 929, "usage_type": "attribute"}, {"api_name": "plot.append", "line_number": 931, "usage_type": "call"}, {"api_name": "plot.PolyEllipse", "line_number": 931, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 932, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 933, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_SOLID", "line_number": 934, "usage_type": "attribute"}, {"api_name": "plot.append", "line_number": 936, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 936, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 939, "usage_type": "call"}, {"api_name": "commons.PolyMarker", "line_number": 939, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 947, "usage_type": "call"}, {"api_name": "mva.chemometrics._index", "line_number": 949, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 951, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 954, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 955, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 959, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PlotGraphics", "line_number": 961, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 981, "usage_type": "call"}, {"api_name": "plot.append", "line_number": 1002, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PlotGraphics", "line_number": 1005, "usage_type": "call"}, {"api_name": "wx.Dialog", "line_number": 1010, "usage_type": "attribute"}, {"api_name": "wx.Dialog.__init__", "line_number": 1012, "usage_type": "call"}, {"api_name": "wx.Dialog", "line_number": 1012, "usage_type": "attribute"}, {"api_name": "wx.Dialog", "line_number": 1044, "usage_type": "attribute"}, {"api_name": "wx.Dialog.__init__", "line_number": 1047, "usage_type": "call"}, {"api_name": "wx.Dialog", "line_number": 1047, "usage_type": "attribute"}, {"api_name": "wx.DEFAULT_DIALOG_STYLE", "line_number": 1049, "usage_type": "attribute"}, {"api_name": "wx.Bitmap", "line_number": 1061, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1061, "usage_type": "call"}, {"api_name": "wx.BITMAP_TYPE_BMP", "line_number": 1061, "usage_type": "attribute"}, {"api_name": "wx.BitmapButton", "line_number": 1062, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 1065, "usage_type": "attribute"}, {"api_name": "wx.Bitmap", "line_number": 1067, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1067, "usage_type": "call"}, {"api_name": "wx.BITMAP_TYPE_BMP", "line_number": 1067, "usage_type": "attribute"}, {"api_name": "wx.BitmapButton", "line_number": 1068, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 1071, "usage_type": "attribute"}, {"api_name": "wx.Bitmap", "line_number": 1073, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1073, "usage_type": "call"}, {"api_name": "wx.BITMAP_TYPE_BMP", "line_number": 1073, "usage_type": "attribute"}, {"api_name": "wx.BitmapButton", "line_number": 1074, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 1077, "usage_type": "attribute"}, {"api_name": "wx.Bitmap", "line_number": 1079, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1079, "usage_type": "call"}, {"api_name": "wx.BITMAP_TYPE_BMP", "line_number": 1079, "usage_type": "attribute"}, {"api_name": "wx.BitmapButton", "line_number": 1080, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 1084, "usage_type": "attribute"}, {"api_name": "wx.Bitmap", "line_number": 1086, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1086, "usage_type": "call"}, {"api_name": "wx.BITMAP_TYPE_BMP", "line_number": 1086, "usage_type": "attribute"}, {"api_name": "wx.BitmapButton", "line_number": 1087, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 1091, "usage_type": "attribute"}, {"api_name": "wx.Bitmap", "line_number": 1093, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1093, "usage_type": "call"}, {"api_name": "wx.BITMAP_TYPE_BMP", "line_number": 1093, "usage_type": "attribute"}, {"api_name": "wx.BitmapButton", "line_number": 1094, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 1097, "usage_type": "attribute"}, {"api_name": "wx.GridSizer", "line_number": 1103, "usage_type": "call"}, {"api_name": "wx.EXPAND", "line_number": 1111, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1112, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1113, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1114, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1115, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1116, "usage_type": "attribute"}, {"api_name": "wx.Bitmap", "line_number": 1119, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1119, "usage_type": "call"}, {"api_name": "wx.Bitmap", "line_number": 1124, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1124, "usage_type": "call"}, {"api_name": "wx.Bitmap", "line_number": 1129, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1129, "usage_type": "call"}, {"api_name": "wx.Bitmap", "line_number": 1134, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1134, "usage_type": "call"}, {"api_name": "wx.Bitmap", "line_number": 1139, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1139, "usage_type": "call"}, {"api_name": "wx.Bitmap", "line_number": 1145, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1145, "usage_type": "call"}, {"api_name": "wx.lib.plot.plotcanvas.PlotCanvas", "line_number": 1150, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.plotcanvas", "line_number": 1150, "usage_type": "name"}, {"api_name": "wx.lib.plot.plotcanvas.PlotCanvas.__init__", "line_number": 1152, "usage_type": "call"}, {"api_name": "wx.lib.plot.plotcanvas.PlotCanvas", "line_number": 1152, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.plotcanvas", "line_number": 1152, "usage_type": "name"}, {"api_name": "wx.lib.plot.plotcanvas.PlotCanvas", "line_number": 1154, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.plotcanvas", "line_number": 1154, "usage_type": "name"}, {"api_name": "wx.lib.plot.plotcanvas.PlotCanvas", "line_number": 1155, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.plotcanvas", "line_number": 1155, "usage_type": "name"}, {"api_name": "wx.EVT_RIGHT_DOWN", "line_number": 1166, "usage_type": "attribute"}, {"api_name": "wx.EVT_LEFT_DOWN", "line_number": 1167, "usage_type": "attribute"}, {"api_name": "wx.Menu", "line_number": 1174, "usage_type": "call"}, {"api_name": "wx.ITEM_NORMAL", "line_number": 1180, "usage_type": "attribute"}, {"api_name": "wx.ITEM_NORMAL", "line_number": 1182, "usage_type": "attribute"}, {"api_name": "wx.ITEM_NORMAL", "line_number": 1184, "usage_type": "attribute"}, {"api_name": "wx.ITEM_NORMAL", "line_number": 1186, "usage_type": "attribute"}, {"api_name": "wx.EVT_MENU", "line_number": 1189, "usage_type": "attribute"}, {"api_name": "wx.EVT_MENU", "line_number": 1190, "usage_type": "attribute"}, {"api_name": "wx.EVT_MENU", "line_number": 1191, "usage_type": "attribute"}, {"api_name": "wx.EVT_MENU", "line_number": 1192, "usage_type": "attribute"}, {"api_name": "wx.MetafileDC", "line_number": 1196, "usage_type": "call"}, {"api_name": "numpy.array2string", "line_number": 1241, "usage_type": "call"}, {"api_name": "wx.TheClipboard.Open", "line_number": 1242, "usage_type": "call"}, {"api_name": "wx.TheClipboard", "line_number": 1242, "usage_type": "attribute"}, {"api_name": "wx.TheClipboard.SetData", "line_number": 1243, "usage_type": "call"}, {"api_name": "wx.TheClipboard", "line_number": 1243, "usage_type": "attribute"}, {"api_name": "wx.TextDataObject", "line_number": 1243, "usage_type": "call"}, {"api_name": "wx.TheClipboard.Close", "line_number": 1244, "usage_type": "call"}, {"api_name": "wx.TheClipboard", "line_number": 1244, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 1255, "usage_type": "call"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 1264, "usage_type": "call"}, {"api_name": "wx.CROSS_CURSOR", "line_number": 1361, "usage_type": "attribute"}, {"api_name": "wx.Cursor", "line_number": 1373, "usage_type": "call"}, {"api_name": "wx.CURSOR_PENCIL", "line_number": 1373, "usage_type": "attribute"}, {"api_name": "wx.CROSS_CURSOR", "line_number": 1375, "usage_type": "attribute"}, {"api_name": "wx.Panel", "line_number": 1382, "usage_type": "attribute"}, {"api_name": "wx.Panel.__init__", "line_number": 1388, "usage_type": "call"}, {"api_name": "wx.Panel", "line_number": 1388, "usage_type": "attribute"}, {"api_name": "wx.TAB_TRAVERSAL", "line_number": 1390, "usage_type": "attribute"}, {"api_name": "wx.lib.anchors.LayoutAnchors", "line_number": 1413, "usage_type": "call"}, {"api_name": "wx.lib.agw.buttonpanel.BP_USE_GRADIENT", "line_number": 1448, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1448, "usage_type": "name"}, {"api_name": "wx.lib.agw.buttonpanel.BP_ALIGN_LEFT", "line_number": 1448, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 1452, "usage_type": "call"}, {"api_name": "wx.HORIZONTAL", "line_number": 1452, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 1453, "usage_type": "call"}, {"api_name": "wx.VERTICAL", "line_number": 1453, "usage_type": "attribute"}, {"api_name": "wx.GridSizer", "line_number": 1454, "usage_type": "call"}, {"api_name": "wx.EXPAND", "line_number": 1456, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1458, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1459, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1461, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1462, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1463, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1464, "usage_type": "attribute"}, {"api_name": "wx.lib.plot.polyobjects.PolyLine", "line_number": 1481, "usage_type": "call"}, {"api_name": "wx.PENSTYLE_TRANSPARENT", "line_number": 1482, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel.ButtonPanel", "line_number": 1490, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1490, "usage_type": "name"}, {"api_name": "wx.lib.agw.buttonpanel.ButtonPanel.__init__", "line_number": 1493, "usage_type": "call"}, {"api_name": "wx.lib.agw.buttonpanel.ButtonPanel", "line_number": 1493, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1493, "usage_type": "name"}, {"api_name": "wx.lib.agw.buttonpanel.BP_USE_GRADIENT", "line_number": 1495, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1495, "usage_type": "name"}, {"api_name": "wx.lib.agw.buttonpanel.BP_ALIGN_LEFT", "line_number": 1496, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1496, "usage_type": "name"}, {"api_name": "wx.EVT_PAINT", "line_number": 1507, "usage_type": "attribute"}, {"api_name": "wx.Bitmap", "line_number": 1509, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1509, "usage_type": "call"}, {"api_name": "wx.BITMAP_TYPE_PNG", "line_number": 1509, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel.ButtonInfo", "line_number": 1510, "usage_type": "call"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1510, "usage_type": "name"}, {"api_name": "wx.ITEM_NORMAL", "line_number": 1510, "usage_type": "attribute"}, {"api_name": "wx.EVT_BUTTON", "line_number": 1514, "usage_type": "attribute"}, {"api_name": "wx.Bitmap", "line_number": 1516, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1516, "usage_type": "call"}, {"api_name": "wx.BITMAP_TYPE_PNG", "line_number": 1516, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel.ButtonInfo", "line_number": 1517, "usage_type": "call"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1517, "usage_type": "name"}, {"api_name": "wx.ITEM_NORMAL", "line_number": 1518, "usage_type": "attribute"}, {"api_name": "wx.EVT_BUTTON", "line_number": 1522, "usage_type": "attribute"}, {"api_name": "wx.Choice", "line_number": 1526, "usage_type": "call"}, {"api_name": "wx.Choice", "line_number": 1531, "usage_type": "call"}, {"api_name": "wx.EVT_COMBOBOX", "line_number": 1535, "usage_type": "attribute"}, {"api_name": "wx.Choice", "line_number": 1539, "usage_type": "call"}, {"api_name": "wx.SpinCtrl", "line_number": 1545, "usage_type": "call"}, {"api_name": "wx.SP_ARROW_KEYS", "line_number": 1549, "usage_type": "attribute"}, {"api_name": "wx.SpinCtrl", "line_number": 1553, "usage_type": "call"}, {"api_name": "wx.SP_ARROW_KEYS", "line_number": 1557, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPINCTRL", "line_number": 1559, "usage_type": "attribute"}, {"api_name": "wx.SpinCtrl", "line_number": 1561, "usage_type": "call"}, {"api_name": "wx.SP_ARROW_KEYS", "line_number": 1565, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPINCTRL", "line_number": 1567, "usage_type": "attribute"}, {"api_name": "wx.TRANSPARENT_WINDOW", "line_number": 1572, "usage_type": "attribute"}, {"api_name": "wx.lib.stattext.GenStaticText", "line_number": 1577, "usage_type": "call"}, {"api_name": "wx.lib.stattext.GenStaticText", "line_number": 1580, "usage_type": "call"}, {"api_name": "wx.lib.stattext.GenStaticText", "line_number": 1582, "usage_type": "call"}, {"api_name": "wx.lib.agw.buttonpanel.BP_TEXT_COLOUR", "line_number": 1603, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1603, "usage_type": "name"}, {"api_name": "wx.WHITE", "line_number": 1603, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel.BP_TEXT_COLOUR", "line_number": 1606, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1606, "usage_type": "name"}, {"api_name": "wx.BLUE", "line_number": 1606, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel.BP_BORDER_COLOUR", "line_number": 1607, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1607, "usage_type": "name"}, {"api_name": "wx.lib.agw.buttonpanel.BrightenColour", "line_number": 1608, "usage_type": "call"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1608, "usage_type": "name"}, {"api_name": "wx.lib.agw.buttonpanel.BP_SEPARATOR_COLOUR", "line_number": 1609, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1609, "usage_type": "name"}, {"api_name": "wx.lib.agw.buttonpanel.BrightenColour", "line_number": 1610, "usage_type": "call"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1610, "usage_type": "name"}, {"api_name": "wx.lib.agw.buttonpanel.BP_BUTTONTEXT_COLOUR", "line_number": 1611, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1611, "usage_type": "name"}, {"api_name": "wx.BLACK", "line_number": 1611, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel.BP_SELECTION_BRUSH_COLOUR", "line_number": 1612, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1612, "usage_type": "name"}, {"api_name": "wx.Colour", "line_number": 1612, "usage_type": "call"}, {"api_name": "wx.lib.agw.buttonpanel.BP_SELECTION_PEN_COLOUR", "line_number": 1613, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.buttonpanel", "line_number": 1613, "usage_type": "name"}, {"api_name": "wx.Colour", "line_number": 1613, "usage_type": "call"}, {"api_name": "wx.FileDialog", "line_number": 1619, "usage_type": "call"}, {"api_name": "wx.FD_SAVE", "line_number": 1620, "usage_type": "attribute"}, {"api_name": "wx.ID_OK", "line_number": 1622, "usage_type": "attribute"}, {"api_name": "numpy.array2string", "line_number": 1624, "usage_type": "call"}, {"api_name": "numpy.array2string", "line_number": 1625, "usage_type": "call"}, {"api_name": "numpy.array2string", "line_number": 1626, "usage_type": "call"}, {"api_name": "numpy.array2string", "line_number": 1627, "usage_type": "call"}, {"api_name": "mva.chemometrics.pca_svd", "line_number": 1673, "usage_type": "call"}, {"api_name": "mva.chemometrics", "line_number": 1673, "usage_type": "name"}, {"api_name": "mva.chemometrics.pca_nipals", "line_number": 1683, "usage_type": "call"}, {"api_name": "mva.chemometrics", "line_number": 1683, "usage_type": "name"}, {"api_name": "numpy.unique", "line_number": 1702, "usage_type": "call"}, {"api_name": "commons.error_box", "line_number": 1708, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 1737, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 1753, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 1754, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 1754, "usage_type": "name"}, {"api_name": "numpy.transpose", "line_number": 1760, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 1761, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 1761, "usage_type": "name"}, {"api_name": "wx.Dialog", "line_number": 1783, "usage_type": "attribute"}, {"api_name": "wx.Dialog.__init__", "line_number": 1787, "usage_type": "call"}, {"api_name": "wx.Dialog", "line_number": 1787, "usage_type": "attribute"}, {"api_name": "wx.MAXIMIZE_BOX", "line_number": 1789, "usage_type": "attribute"}, {"api_name": "wx.DEFAULT_DIALOG_STYLE", "line_number": 1789, "usage_type": "attribute"}, {"api_name": "wx.GridSizer", "line_number": 1834, "usage_type": "call"}, {"api_name": "wx.EXPAND", "line_number": 1843, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1844, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1845, "usage_type": "attribute"}, {"api_name": "wx.GridSizer", "line_number": 1849, "usage_type": "call"}, {"api_name": "wx.EXPAND", "line_number": 1857, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1858, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1859, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1860, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 1864, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 1867, "usage_type": "call"}, {"api_name": "wx.ALIGN_LEFT", "line_number": 1868, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 1872, "usage_type": "call"}, {"api_name": "wx.ALIGN_LEFT", "line_number": 1873, "usage_type": "attribute"}, {"api_name": "wx.GridBagSizer", "line_number": 1909, "usage_type": "call"}, {"api_name": "wx.FLEX_GROWMODE_SPECIFIED", "line_number": 1912, "usage_type": "attribute"}, {"api_name": "wx.HORIZONTAL", "line_number": 1915, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.foldpanelbar.FoldPanelBar", "line_number": 1926, "usage_type": "call"}, {"api_name": "wx.lib.agw.foldpanelbar", "line_number": 1926, "usage_type": "name"}, {"api_name": "wx.DefaultPosition", "line_number": 1926, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.foldpanelbar.FPB_EXCLUSIVE_FOLD", "line_number": 1928, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.foldpanelbar", "line_number": 1928, "usage_type": "name"}, {"api_name": "wx.lib.anchors.LayoutAnchors", "line_number": 1930, "usage_type": "call"}, {"api_name": "wx.ImageList", "line_number": 1933, "usage_type": "call"}, {"api_name": "wx.Bitmap", "line_number": 1934, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1934, "usage_type": "call"}, {"api_name": "wx.BITMAP_TYPE_PNG", "line_number": 1934, "usage_type": "attribute"}, {"api_name": "wx.Bitmap", "line_number": 1935, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1935, "usage_type": "call"}, {"api_name": "wx.BITMAP_TYPE_PNG", "line_number": 1935, "usage_type": "attribute"}, {"api_name": "wx.Panel", "line_number": 1951, "usage_type": "call"}, {"api_name": "wx.TAB_TRAVERSAL", "line_number": 1953, "usage_type": "attribute"}, {"api_name": "wx.Panel", "line_number": 1956, "usage_type": "call"}, {"api_name": "wx.TAB_TRAVERSAL", "line_number": 1958, "usage_type": "attribute"}, {"api_name": "wx.Panel", "line_number": 1961, "usage_type": "call"}, {"api_name": "wx.TAB_TRAVERSAL", "line_number": 1963, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 1966, "usage_type": "call"}, {"api_name": "wx.StaticText", "line_number": 1971, "usage_type": "call"}, {"api_name": "wx.StaticText", "line_number": 1977, "usage_type": "call"}, {"api_name": "wx.StaticText", "line_number": 1982, "usage_type": "call"}, {"api_name": "wx.StaticText", "line_number": 1988, "usage_type": "call"}, {"api_name": "wx.StaticText", "line_number": 1993, "usage_type": "call"}, {"api_name": "wx.StaticText", "line_number": 1998, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 2003, "usage_type": "call"}, {"api_name": "wx.EVT_TEXT", "line_number": 2007, "usage_type": "attribute"}, {"api_name": "wx.TextCtrl", "line_number": 2009, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 2014, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 2019, "usage_type": "call"}, {"api_name": "wx.SpinButton", "line_number": 2024, "usage_type": "call"}, {"api_name": "wx.SP_VERTICAL", "line_number": 2026, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN_DOWN", "line_number": 2028, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN_UP", "line_number": 2029, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN", "line_number": 2030, "usage_type": "attribute"}, {"api_name": "wx.SpinButton", "line_number": 2032, "usage_type": "call"}, {"api_name": "wx.SP_VERTICAL", "line_number": 2034, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN_DOWN", "line_number": 2036, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN_UP", "line_number": 2037, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN", "line_number": 2038, "usage_type": "attribute"}, {"api_name": "wx.SpinButton", "line_number": 2040, "usage_type": "call"}, {"api_name": "wx.SP_VERTICAL", "line_number": 2042, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN_DOWN", "line_number": 2044, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN_UP", "line_number": 2045, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN", "line_number": 2046, "usage_type": "attribute"}, {"api_name": "wx.SpinButton", "line_number": 2048, "usage_type": "call"}, {"api_name": "wx.SP_VERTICAL", "line_number": 2050, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN_DOWN", "line_number": 2052, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN_UP", "line_number": 2053, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN", "line_number": 2054, "usage_type": "attribute"}, {"api_name": "wx.TextCtrl", "line_number": 2056, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 2061, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 2066, "usage_type": "call"}, {"api_name": "wx.StaticText", "line_number": 2071, "usage_type": "call"}, {"api_name": "wx.SpinCtrl", "line_number": 2077, "usage_type": "call"}, {"api_name": "wx.SP_ARROW_KEYS", "line_number": 2082, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN", "line_number": 2086, "usage_type": "attribute"}, {"api_name": "wx.SpinCtrl", "line_number": 2088, "usage_type": "call"}, {"api_name": "wx.SP_ARROW_KEYS", "line_number": 2093, "usage_type": "attribute"}, {"api_name": "wx.EVT_SPIN", "line_number": 2097, "usage_type": "attribute"}, {"api_name": "wx.lib.buttons.GenToggleButton", "line_number": 2099, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 2104, "usage_type": "attribute"}, {"api_name": "wx.lib.buttons.GenToggleButton", "line_number": 2106, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 2111, "usage_type": "attribute"}, {"api_name": "wx.lib.buttons.GenToggleButton", "line_number": 2113, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 2119, "usage_type": "attribute"}, {"api_name": "wx.lib.buttons.GenToggleButton", "line_number": 2121, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 2126, "usage_type": "attribute"}, {"api_name": "wx.CheckBox", "line_number": 2128, "usage_type": "call"}, {"api_name": "wx.Button", "line_number": 2133, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 2137, "usage_type": "attribute"}, {"api_name": "wx.lib.buttons.GenToggleButton", "line_number": 2139, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 2145, "usage_type": "attribute"}, {"api_name": "wx.lib.buttons.GenToggleButton", "line_number": 2147, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 2152, "usage_type": "attribute"}, {"api_name": "wx.lib.buttons.GenToggleButton", "line_number": 2154, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 2159, "usage_type": "attribute"}, {"api_name": "wx.Button", "line_number": 2161, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 2166, "usage_type": "attribute"}, {"api_name": "wx.Button", "line_number": 2168, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 2174, "usage_type": "attribute"}, {"api_name": "wx.Button", "line_number": 2176, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 2182, "usage_type": "attribute"}, {"api_name": "wx.Button", "line_number": 2184, "usage_type": "call"}, {"api_name": "wx.EVT_BUTTON", "line_number": 2190, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.foldpanelbar.FPB_ALIGN_WIDTH", "line_number": 2192, "usage_type": "attribute"}, {"api_name": "wx.lib.agw.foldpanelbar", "line_number": 2192, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 2272, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 2276, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 2276, "usage_type": "name"}, {"api_name": "numpy.transpose", "line_number": 2327, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2423, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2424, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2446, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2447, "usage_type": "call"}, {"api_name": "wx.App", "line_number": 2504, "usage_type": "call"}]}
+{"seq_id": "37614464044", "text": "import json\nimport urllib.parse\nimport urllib.request\n\n# Google Knowledge Graph API key\napi_key = \"AIzaSyDPZceqCgLVGytRa14EOvYfcYarjfqMLm0\"\n\nquery = 'university'\nservice_url = 'https://kgsearch.googleapis.com/v1/entities:search'\nparams = {\n 'query': query,\n # 'ids': 'kg:/m/0gg594v',\n 'limit': 3,\n 'indent': True,\n 'key': api_key,\n}\n\nurl = service_url + '?' + urllib.parse.urlencode(params)\n\nwith urllib.request.urlopen(url) as response:\n bb = response.read()\n obj = bb.decode('utf-8')\n\nobj = json.loads(obj)\nitems = obj['itemListElement']\nfor element in items:\n print(element['result']['name'] + ' (' + str(element['resultScore']) + ')')\n\n", "repo_name": "amirhdrz/nlp_experiments", "sub_path": "knowledge_graph.py", "file_name": "knowledge_graph.py", "file_ext": "py", "file_size_in_byte": 664, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "25", "api": [{"api_name": "urllib.parse.parse.urlencode", "line_number": 18, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 18, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 18, "usage_type": "name"}, {"api_name": "urllib.parse.request.urlopen", "line_number": 20, "usage_type": "call"}, {"api_name": "urllib.parse.request", "line_number": 20, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 20, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 24, "usage_type": "call"}]}
+{"seq_id": "13310558836", "text": "\"\"\"basenodemaps.py\nBase maps for HDF5 objects with node methods.\n\"\"\"\n# Package Header #\nfrom ...header import *\n\n# Header #\n__author__ = __author__\n__credits__ = __credits__\n__maintainer__ = __maintainer__\n__email__ = __email__\n\n\n# Imports #\n# Standard Libraries #\n\n# Third-Party Packages #\nimport h5py\n\n# Local Packages #\nfrom ...hdf5bases import HDF5Map, DatasetMap\nfrom ...dataset import ObjectReferenceComponent\nfrom ..datasetcomponents import NodeDatasetComponent\nfrom ..groupcomponents import NodeGroupComponent\n\n\n# Definitions #\n# Classes #\nclass BaseNodeDatasetMap(DatasetMap):\n \"\"\"A dataset map which outlines a dataset with basic node methods.\"\"\"\n\n default_dtype = ((\"Node\", h5py.ref_dtype),)\n default_component_types = {\n \"object_reference\": (\n ObjectReferenceComponent,\n {\n \"reference_fields\": {\"dataset\": \"Dataset\"},\n \"primary_reference_field\": \"dataset\",\n },\n ),\n \"tree_node\": (NodeDatasetComponent, {}),\n }\n\n\nclass BaseNodeGroupMap(HDF5Map):\n \"\"\"A group map which outlines a group with basic node methods.\"\"\"\n\n default_attribute_names = {\"tree_type\": \"tree_type\"}\n default_attributes = {\"tree_type\": \"Node\"}\n default_map_names = {\"node_map\": \"node_map\"}\n default_maps = {\"node_map\": BaseNodeDatasetMap()}\n default_component_types = {\n \"tree_node\": (NodeGroupComponent, {}),\n }\n", "repo_name": "FongAnthonyM/python-hdf5objects", "sub_path": "src/hdf5objects/treehierarchy/maps/basenodemaps.py", "file_name": "basenodemaps.py", "file_ext": "py", "file_size_in_byte": 1416, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "hdf5bases.DatasetMap", "line_number": 29, "usage_type": "name"}, {"api_name": "h5py.ref_dtype", "line_number": 32, "usage_type": "attribute"}, {"api_name": "dataset.ObjectReferenceComponent", "line_number": 35, "usage_type": "name"}, {"api_name": "datasetcomponents.NodeDatasetComponent", "line_number": 41, "usage_type": "name"}, {"api_name": "hdf5bases.HDF5Map", "line_number": 45, "usage_type": "name"}, {"api_name": "groupcomponents.NodeGroupComponent", "line_number": 53, "usage_type": "name"}]}
+{"seq_id": "38082556492", "text": "import sys\r\nimport importlib\r\n#from data_esod import ESOD_Test\r\nimport torch\r\nimport time\r\nfrom progress.bar import Bar\r\nimport os\r\nfrom collections import OrderedDict\r\nimport cv2\r\nfrom PIL import Image\r\nimport numpy as np\r\n\r\nfrom base.framework_factory import load_framework\r\nfrom base.data import Test_Dataset\r\nfrom base.metric import *\r\nfrom base.util import *\r\n\r\n\r\ndef test_model(model, test_sets, config, saver=None):\r\n model.eval()\r\n st = time.time()\r\n for set_name, test_set in test_sets.items():\r\n save_folder = os.path.join(config['save_path'], set_name)\r\n check_path(save_folder)\r\n \r\n titer = test_set.size\r\n MR = MetricRecorder(titer)\r\n scores = []\r\n \r\n test_bar = Bar('Dataset {:10}:'.format(set_name), max=titer)\r\n for j in range(titer):\r\n image, gt, name = test_set.load_data(j)\r\n Y = model(image.cuda())\r\n pred = Y['final'][0, 0].sigmoid_().cpu().data.numpy()\r\n \r\n out_shape = gt.shape\r\n \r\n #pred = np.array(Image.fromarray(pred).resize((out_shape[::-1]), resample=0))\r\n pred = cv2.resize(pred, (out_shape[::-1]), interpolation=cv2.INTER_LINEAR)\r\n \r\n pred, gt = normalize_pil(pred, gt)\r\n pred = np.clip(np.round(pred * 255) / 255., 0, 1)\r\n MR.update(pre=pred, gt=gt)\r\n \r\n #scores.append(get_scores(pred, gt))\r\n #print(get_scores(pred, gt))\r\n \r\n # save predictions\r\n if config['save']:\r\n fnl_folder = os.path.join(save_folder, 'final')\r\n check_path(fnl_folder)\r\n im_path = os.path.join(fnl_folder, name + '.png')\r\n Image.fromarray((pred * 255)).convert('L').save(im_path)\r\n \r\n if saver is not None:\r\n saver(Y, gt, name, save_folder, config)\r\n pass\r\n \r\n Bar.suffix = '{}/{}'.format(j, titer)\r\n test_bar.next()\r\n \r\n #scores = np.array(scores)\r\n #print(np.mean(scores, axis=0))\r\n \r\n mae, (maxf, meanf, *_), sm, em, wfm = MR.show(bit_num=3)\r\n #print(' MAE: {}, Max-F: {}, Maen-F: {}, SM: {}, EM: {}, Fbw: {}.'.format(mae, maxf, meanf, sm, em, wfm))\r\n print(' Max-F: {:.3f}, Maen-F: {:.3f}, Fbw: {:.3f}, MAE: {:.3f}, SM: {:.3f}, EM: {:.3f}.'.format(maxf, meanf, wfm, mae, sm, em))\r\n \r\n print('Test using time: {}.'.format(round(time.time() - st, 3)))\r\n\r\ndef main():\r\n if len(sys.argv) > 1:\r\n net_name = sys.argv[1]\r\n else:\r\n print('Need model name!')\r\n return\r\n \r\n config, model, _, _, _, saver = load_framework(net_name)\r\n print(config)\r\n \r\n #model.load_state_dict(torch.load(config['weight'], map_location='cpu'))\r\n saved_model = torch.load(config['weight'], map_location='cpu')\r\n new_name = {}\r\n for k, v in saved_model.items():\r\n if k.startswith('model'):\r\n new_name[k[6:]] = v\r\n else:\r\n new_name[k] = v\r\n model.load_state_dict(new_name)\r\n\r\n test_sets = OrderedDict()\r\n for set_name in config['vals']:\r\n test_sets[set_name] = Test_Dataset(name=set_name, config=config)\r\n \r\n model = model.cuda()\r\n \r\n test_model(model, test_sets, config, saver=saver)\r\n \r\nif __name__ == \"__main__\":\r\n main()", "repo_name": "moothes/SALOD", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 3423, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 38, "dataset": "github-code", "pt": "27", "api": [{"api_name": "time.time", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "progress.bar.Bar", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.clip", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 53, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 53, "usage_type": "name"}, {"api_name": "progress.bar.Bar.suffix", "line_number": 59, "usage_type": "attribute"}, {"api_name": "progress.bar.Bar", "line_number": 59, "usage_type": "name"}, {"api_name": "time.time", "line_number": 69, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 72, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 73, "usage_type": "attribute"}, {"api_name": "base.framework_factory.load_framework", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 82, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 91, "usage_type": "call"}, {"api_name": "base.data.Test_Dataset", "line_number": 93, "usage_type": "call"}]}
+{"seq_id": "73091392705", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[24]:\n\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\nfrom matplotlib import cm\nimport tov_solver as tov\nfrom scipy import interpolate\n\ndef modified_wave_equation(rho_c = 1e-3):\n \n def M(x):\n ## interpolation of M(r) from TOV solution\n sol = tov.TOV_solver(rho_c)\n r = np.linspace(0,sol.t[-1],len(sol.y[0]))\n Mr = interpolate.interp1d(r, sol.y[0])\n return Mr(x)\n def v(x):\n ## interpolation of v(r) from TOV solution\n sol = tov.TOV_solver(rho_c)\n r = np.linspace(0,sol.t[-1],len(sol.y[0]))\n vr = interpolate.interp1d(r, sol.y[1])\n return vr(x)\n def p(x):\n ## interpolation of p(r) from TOV solution\n sol = tov.TOV_solver(rho_c)\n r = np.linspace(0,sol.t[-1],len(sol.y[0]))\n pr = interpolate.interp1d(r, sol.y[2])\n return pr(x)\n\n def rho(x):\n ## interpolation of rho(r) from TOV solution\n return (p(x)/50)**(1/2)\n\n def func(r):\n ## function f(r) in the wave equation\n sol = tov.TOV_solver(rho_c)\n return np.exp(M(r)/2)*(1 - 2*v(r)/r)**(1/2)\n\n def gfunc(r):\n ## defined function g for wave equation\n return 1/r**2\n\n def init(r):\n ## inital condition for the wave\n return -2*1e-3*np.exp(-r**2/2)*r\n def init2(r):\n ## inital condition for the wave\n return 1e-3*np.exp(-r**2/2)\n\n def hfunc(r):\n return rho(r)-3*p(r)\n\n sol = tov.TOV_solver(rho_c)\n c = 1 #speed of light\n N = 300 # mesh element number\n r = np.linspace(0,7,N) #space\n t = np.linspace(0,5,N) #time\n dr = r[1] - r[0] #differential space element\n dt = t[1] - t[0] # differential time element\n lam = c*dt/dr # lambda\n\n #discretize functions\n g = np.zeros(N)\n g = gfunc(r)\n f = np.zeros(N)\n f = func(r)\n h = np.zeros(N)\n h = hfunc(r)\n f[0] = 1\n g[0] = 1e4\n Psi = np.zeros((N,N))\n Pi = np.zeros((N,N))\n psi = np.zeros((N,N))\n #initial condition\n for i in range(N):\n Psi[0,i] = init(r[i])\n psi[0,i] = init2(r[i])\n\n #BCs\n Psi[:,0] = 0\n Pi [:,0] = 0\n psi[:,0] = 1e-3\n\n Psi[:,-1] = 0\n Pi [:,-1] = 0\n\n #solving PDE\n T = 250\n for n in range(1,T): ##time\n for j in range(1,N-1): ##space\n if (n == 0):\n Psi[1,j] = 0.75*Psi[0,j] + 0.25*Psi[2,j] + lam*0.5*((f[j+1]*Pi[0,j+1]) - (f[j-1]*Pi[0,j-1]))\n Pi[1,j] = 0.75*Pi[0,j] + 0.25*Pi[2,j] + lam*0.5*g[j]*(((1/g[j+1])*f[j+1]*Psi[0,j+1]) - ((1/g[j-1])*f[j-1]*Psi[0,j-1]))\n +48*np.pi*np.exp(-12*psi[0,j]**2)*psi[0,j]*(h[j])*dt\n psi[n,j+1] = psi[n,j] + dr*Psi[n,j+1]\n\n Psi[n+1,j] = Psi[n-1,j] + lam*((f[j+1]*Pi[n,j+1]) - (f[j+1]*Pi[n,j-1]))\n Pi[n+1,j] = Pi[n-1,j] + lam*g[j]*(((1/g[j+1])*f[j+1]*Psi[n,j+1]) - ((1/g[j-1])*f[j-1]*Psi[n,j-1])) \n +48*np.pi*np.exp(-12*psi[n,j]**2)*psi[n,j]*(h[j])*dt\n psi[n,j+1] = psi[n,j] + dr*Psi[n,j+1]\n #plotting space vs time\n fig = plt.figure()\n ax = fig.gca(projection='3d')\n X, Y = np.meshgrid(r[2:], t[0:T:20])\n Z = psi[0:T:20,2:]\n ax.set_xlabel('r')\n ax.set_ylabel('time')\n ax.set_zlabel('psi')\n surf = ax.scatter3D(X, Y, Z, cmap='viridis')\n ax.set_title('Solution for psi');\n ax.set_xlim3d(0, sol.t[-1])\n #ax.set_ylim3d(0,0.2)\n ax.set_zlim3d(-0.001,0.000)\n\n", "repo_name": "ekremdemirboga/PHYS-414-514-Final-Project", "sub_path": "Beyond Einstein/wave_equation_with_added_term.py", "file_name": "wave_equation_with_added_term.py", "file_ext": "py", "file_size_in_byte": 3458, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "25", "api": [{"api_name": "tov_solver.TOV_solver", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 19, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 20, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 20, "usage_type": "name"}, {"api_name": "tov_solver.TOV_solver", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 25, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 26, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 26, "usage_type": "name"}, {"api_name": "tov_solver.TOV_solver", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 31, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 32, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 32, "usage_type": "name"}, {"api_name": "tov_solver.TOV_solver", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 53, "usage_type": "call"}, {"api_name": "tov_solver.TOV_solver", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 99, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 104, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "numpy.meshgrid", "line_number": 109, "usage_type": "call"}]}
+{"seq_id": "28865489309", "text": "#!/usr/bin/env python\n# -*- coding:utf-8 -*-\nimport re\nfrom libs.pyaml import configure\nfrom modules.interaction.metric import AbstractMetric\n\n\ndef _set_metric(metric=None):\n if not metric:\n metric = AbstractMetric()\n for key, value in configure['metric'].items():\n setattr(metric, key, value)\n if hasattr(metric, 'expend_time'):\n seconds = 0\n items = re.findall(r'\\d+', metric.expend_time)\n bases = [1, 60, 3600, 86400]\n for i, item in enumerate(reversed(items)):\n seconds += int(item) * bases[i]\n print(seconds, '秒')\n print(metric)\n\n\nif __name__ == '__main__':\n _set_metric()\n", "repo_name": "beikejinmiao/HawkSec", "sub_path": "tests/metric_test.py", "file_name": "metric_test.py", "file_ext": "py", "file_size_in_byte": 663, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "modules.interaction.metric.AbstractMetric", "line_number": 10, "usage_type": "call"}, {"api_name": "libs.pyaml.configure", "line_number": 11, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 15, "usage_type": "call"}]}
+{"seq_id": "17966669232", "text": "import logging\n\nfrom flask_restx import Api\nfrom web import settings\n\n# from sqlalchemy.orm.exc import NoResultFound\n\nlog = logging.getLogger(__name__)\n\napi = Api(\n version='1.0',\n title='GoodAttitude 결제 구현 API.',\n description='GA 유저가 텀블러를 사용하기 위해, 카드를 선 등록 후 -> 대여 기간에 늦거나 반납을 하지 않았을 시, 해당 결제 금액만큼 자동결제(빌링)\\n'\n '카드사1: toss payments\\n'\n '카드사1: kakaopay (구현 예정)\\n'\n '기능 1: 일반 카드 결제\\n'\n '기능 2: 카드 선 등록 -> 자동 결제 (빌링)\\n'\n '기능 2-1: 대여기간이 늦어질 시, status code: 1\\n'\n '기능 2-2: 제품에 하자가 생길 시, status code:2\\n'\n '기능 2-3: 반납이 안될 시(대여기간이 늦어진 상태에서 1차 결제를 하고, 그 후에도 반납이 안되면 추가 결제), status code:3\\n'\n '기능 3: 결제 취소 (구현 예정)\\n'\n '기능 4: 결제 조회 (구현 예정)\\n'\n '기능 5: 결제 승인 및 취소 조회 (구현 예정)\\n'\n '기능 6: 현금영수증 발급 및 취소 (구현 예정)\\n'\n '기능 7: 수동 정산 (구현 예정)\\n'\n '기능 8: 카드사 혜택 조회 (구현 예정)\\n'\n '기능 9: 세금 처리 (구현 예정)\\n',\n terms_url=\"/\",\n contact_email=\"tyty587587@gmail.com\",\n contact=\"010-7704-1961\",\n)\n\n\n@api.errorhandler\ndef default_error_handler(e):\n message = 'An unhandled exception occurred.'\n log.exception(message)\n\n if not settings.FLASK_DEBUG:\n return {'message': message}, 500\n\n\n# @api.errorhandler(NoResultFound)\n# def database_not_found_error_handler(e):\n# \"\"\"No results found in database\"\"\"\n# log.warning(traceback.format_exc())\n# return {'message': 'A database result was required but none was found.'}, 404", "repo_name": "KYUSEONGHAN/payment", "sub_path": "web/restapi.py", "file_name": "restapi.py", "file_ext": "py", "file_size_in_byte": 2001, "program_lang": "python", "lang": "ko", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "flask_restx.Api", "line_number": 10, "usage_type": "call"}, {"api_name": "web.settings.FLASK_DEBUG", "line_number": 39, "usage_type": "attribute"}, {"api_name": "web.settings", "line_number": 39, "usage_type": "name"}]}
+{"seq_id": "73337791744", "text": "from PIL import Image\nimport pytesseract\nimport os\n\ndef getPhotoToText():\n pytesseract.pytesseract.tesseract_cmd = r'C:\\Program Files\\Tesseract-OCR\\tesseract.exe'\n\n arr = os.listdir(\"images\")\n image_path = f\"images/{arr[0]}\"\n image = Image.open(image_path)\n\n text = pytesseract.image_to_string(image, lang=\"rus\")\n return text", "repo_name": "KirillhacT/projects", "sub_path": "PhotoToTextAPI/photoToText.py", "file_name": "photoToText.py", "file_ext": "py", "file_size_in_byte": 343, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "pytesseract.pytesseract", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 8, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 10, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 10, "usage_type": "name"}, {"api_name": "pytesseract.image_to_string", "line_number": 12, "usage_type": "call"}]}
+{"seq_id": "31916554562", "text": "import copy\nimport json\nimport os\n\nDIR_1 = '/Users/hsor001/Projects/musculoskeletal/workflows/sparc/data/argon_viewer_out'\nDIR_2 = '/Users/hsor001/Projects/musculoskeletal/data/vickie_shim_6f1/'\n\nVERSION_INFO = {\n \"OpenCMISS-Argon Version\": [\n \"0\",\n \"2\",\n \"2\"\n ],\n}\n\nSCENE_GRAPHICS = {\n \"Scene\": {\n \"Graphics\": [\n {\n \"BoundaryMode\": \"ALL\",\n \"CoordinateField\": \"coordinates\",\n \"ElementFaceType\": \"ALL\",\n \"FieldDomainType\": \"MESH2D\",\n \"Material\": \"black\",\n \"RenderLineWidth\": 1,\n \"RenderPointSize\": 1,\n \"RenderPolygonMode\": \"SHADED\",\n \"Scenecoordinatesystem\": \"LOCAL\",\n \"SelectMode\": \"ON\",\n \"SelectedMaterial\": \"default_selected\",\n \"Surfaces\": {},\n \"Tessellation\": \"default\",\n \"Type\": \"SURFACES\",\n \"VisibilityFlag\": True\n }\n ],\n \"VisibilityFlag\": True\n }\n}\n\nFIELD_MODULE = {\n \"Fieldmodule\": {\n \"Fields\": [\n {\n \"CoordinateSystemType\": \"RECTANGULAR_CARTESIAN\",\n \"FieldFiniteElement\": {\n \"ComponentNames\": [\n \"x\",\n \"y\",\n \"z\"\n ],\n \"NumberOfComponents\": 3\n },\n \"IsManaged\": True,\n \"IsTypeCoordinate\": True,\n \"Name\": \"coordinates\"\n },\n {\n \"CoordinateSystemType\": \"FIBRE\",\n \"FieldFiniteElement\": {\n \"ComponentNames\": [\n \"fibre angle\",\n \"imbrication angle\",\n \"sheet angle\"\n ],\n \"NumberOfComponents\": 3\n },\n \"IsManaged\": True,\n \"IsTypeCoordinate\": False,\n \"Name\": \"fibres\"\n }\n ]\n },\n}\n\nEMPTY_REGION = {\n \"Fieldmodule\": None,\n \"Scene\": {\n \"Graphics\": None,\n \"VisibilityFlag\": True\n }\n}\n\nSKIP_REGIONS = ['maxilla', ]\n\n\ndef main():\n os.walk(DIR_2)\n data_files = []\n for root, dirs, files in os.walk(DIR_2, topdown=True):\n current_dir = {\n 'node_files': [],\n 'elem_files': []\n }\n for file in files:\n # print(file)\n if file.endswith('.EXNODE'):\n current_dir['node_files'].append(os.path.join(root, file))\n if file.endswith('.EXELEM'):\n current_dir['elem_files'].append(os.path.join(root, file))\n\n if len(current_dir[\"node_files\"]):\n data_files.append(current_dir)\n\n common_path = os.path.commonpath([d[\"node_files\"][0] for d in data_files])\n\n argon_document = {\n **VERSION_INFO\n }\n\n root_region = copy.deepcopy(EMPTY_REGION)\n\n bits = []\n for index, data in enumerate(data_files):\n\n exnode_file = data[\"node_files\"][0]\n region_path = exnode_file.replace(common_path, '')\n\n region_parts = region_path.split('/')\n region_parts.pop(0)\n base_region = root_region\n for i in range(len(region_parts) - 1):\n current_region = region_parts[i].lower()\n\n if \"ChildRegions\" not in base_region:\n base_region[\"ChildRegions\"] = []\n\n child_region_names = []\n for region_info in base_region[\"ChildRegions\"]:\n child_region_names.append(region_info[\"Name\"])\n\n if current_region not in child_region_names:\n new_child = copy.deepcopy(EMPTY_REGION)\n new_child['Name'] = current_region\n base_region[\"ChildRegions\"].append(new_child)\n child_region_names.append(current_region)\n\n j = child_region_names.index(current_region)\n\n base_region = base_region[\"ChildRegions\"][j]\n\n # base_region[\"Fieldmodule\"] = copy.deepcopy(FIELD_MODULE[\"Fieldmodule\"])\n base_region[\"Scene\"] = copy.deepcopy(SCENE_GRAPHICS[\"Scene\"])\n bit = f\"'{region_parts[-2].lower()}',\"\n # if bit not in bits:\n # bits.append(bit)\n if region_parts[-2].lower() in SKIP_REGIONS:\n continue\n\n if \"Model\" not in base_region:\n base_region[\"Model\"] = {\"Sources\": []}\n\n for node_file in data['node_files']:\n exnode_path = node_file # .replace(common_path, '')[1:]\n base_region[\"Model\"][\"Sources\"].insert(\n 0,\n {\n \"FileName\": exnode_path,\n \"RegionName\": os.path.dirname(region_path).lower(),\n \"Type\": \"FILE\"\n }\n )\n for elem_file in data['elem_files']:\n exelem_path = elem_file # .replace(common_path, '')[1:]\n base_region[\"Model\"][\"Sources\"].append(\n {\n \"FileName\": exelem_path,\n \"RegionName\": os.path.dirname(region_path).lower(),\n \"Type\": \"FILE\"\n }\n )\n\n if 'MUSCLES' in region_path or 'NECK' in region_path:\n base_region[\"Scene\"][\"Graphics\"][0][\"Material\"] = \"muscle\"\n if 'BONE' in region_path:\n base_region[\"Scene\"][\"Graphics\"][0][\"Material\"] = \"bone\"\n if 'LIGAMENT' in region_path:\n base_region[\"Scene\"][\"Graphics\"][0][\"Material\"] = \"white\"\n if 'SKIN' in region_path:\n base_region[\"Scene\"][\"Graphics\"][0][\"Material\"] = \"brown\"\n\n argon_document[\"RootRegion\"] = root_region\n\n print('\\n'.join(bits))\n with open(os.path.join(DIR_2, 'test_file.json'), 'w') as f:\n f.write(json.dumps(argon_document, default=lambda o: o.__dict__, sort_keys=True, indent=2))\n\n\nif __name__ == \"__main__\":\n main()\n\nmodel_sources = {\n \"Model\": {\n \"Sources\": [\n {\n \"FileName\": \"FEMUR.EXNODE\",\n \"RegionName\": \"/left_lower_limb/bones/femur\",\n \"Type\": \"FILE\"\n },\n {\n \"FileName\": \"FEMUR.EXELEM\",\n \"RegionName\": \"/left_lower_limb/bones/femur\",\n \"Type\": \"FILE\"\n }\n ]\n },\n}\n", "repo_name": "hsorby/shell-scripts", "sub_path": "opencmiss/generate_argon_document.py", "file_name": "generate_argon_document.py", "file_ext": "py", "file_size_in_byte": 6332, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "os.walk", "line_number": 88, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path", "line_number": 98, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "os.path.commonpath", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}, {"api_name": "copy.deepcopy", "line_number": 111, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 133, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path", "line_number": 159, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path", "line_number": 168, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path", "line_number": 185, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 186, "usage_type": "call"}]}
+{"seq_id": "14915590730", "text": "from PyQt5 import QtCore, QtGui, QtWidgets\nfrom PyQt5.QtWidgets import QMainWindow, QApplication\nimport sys\nimport boj\n\nfrom matplotlib.backends.backend_qt5agg import FigureCanvasQTAgg as FigureCanvas\nfrom matplotlib.backends.backend_qt5agg import NavigationToolbar2QT as NavigationToolbar\nimport matplotlib.pyplot as plt\n\nimport random\n\nclass MainWindow(object):\n def setupUI(self, MainWindow):\n MainWindow.setObjectName(\"MainWindow\")\n MainWindow.setWindowModality(QtCore.Qt.NonModal)\n MainWindow.resize(640, 320)\n self.centralwidget = QtWidgets.QWidget(MainWindow)\n self.centralwidget.setObjectName(\"centralwidget\")\n self.pushButton = QtWidgets.QPushButton(self.centralwidget)\n self.pushButton.setGeometry(QtCore.QRect(270, 240, 100, 40))\n self.pushButton.setObjectName(\"pushButton\")\n self.label = QtWidgets.QLabel(self.centralwidget)\n self.label.setEnabled(True)\n self.label.setGeometry(QtCore.QRect(0, 30, 640, 130))\n font = QtGui.QFont()\n font.setFamily(\"Arial Rounded MT Bold\")\n font.setPointSize(128)\n font.setBold(True)\n font.setWeight(75)\n self.label.setFont(font)\n self.label.setFrameShape(QtWidgets.QFrame.NoFrame)\n self.label.setFrameShadow(QtWidgets.QFrame.Raised)\n self.label.setTextFormat(QtCore.Qt.RichText)\n self.label.setAlignment(QtCore.Qt.AlignCenter)\n self.label.setOpenExternalLinks(False)\n self.label.setObjectName(\"label\")\n self.label_2 = QtWidgets.QLabel(self.centralwidget)\n self.label_2.setGeometry(QtCore.QRect(0, 160, 640, 21))\n font = QtGui.QFont()\n font.setFamily(\"Arial Rounded MT Bold\")\n font.setPointSize(16)\n font.setBold(True)\n font.setWeight(75)\n self.label_2.setFont(font)\n self.label_2.setFrameShape(QtWidgets.QFrame.NoFrame)\n self.label_2.setFrameShadow(QtWidgets.QFrame.Raised)\n self.label_2.setTextFormat(QtCore.Qt.RichText)\n self.label_2.setAlignment(QtCore.Qt.AlignCenter)\n self.label_2.setOpenExternalLinks(False)\n self.label_2.setObjectName(\"label_2\")\n self.lineEdit = QtWidgets.QLineEdit(self.centralwidget)\n self.lineEdit.setEnabled(True)\n self.lineEdit.setGeometry(QtCore.QRect(230, 200, 180, 30))\n self.lineEdit.setText(\"\")\n self.lineEdit.setObjectName(\"lineEdit\")\n MainWindow.setCentralWidget(self.centralwidget)\n self.statusbar = QtWidgets.QStatusBar(MainWindow)\n self.statusbar.setObjectName(\"statusbar\")\n MainWindow.setStatusBar(self.statusbar)\n self.actionsddjd = QtWidgets.QAction(MainWindow)\n self.actionsddjd.setObjectName(\"actionsddjd\")\n self.actionvdmns = QtWidgets.QAction(MainWindow)\n self.actionvdmns.setObjectName(\"actionvdmns\")\n\n self.retranslateUI(MainWindow)\n QtCore.QMetaObject.connectSlotsByName(MainWindow)\n\n def retranslateUI(self, MainWindow):\n _translate = QtCore.QCoreApplication.translate\n MainWindow.setWindowTitle(_translate(\"MainWindow\", \"BOJ.GG\"))\n self.pushButton.setText(_translate(\"MainWindow\", \"검색\"))\n self.label.setText(_translate(\"MainWindow\", \"BOJ.GG\"))\n self.label_2.setText(_translate(\"MainWindow\", \"Beakjoon Online Judge 전적검색 프로그램\"))\n self.actionsddjd.setText(_translate(\"MainWindow\", \"sddjd\"))\n self.actionvdmns.setText(_translate(\"MainWindow\", \"vdmns\"))\n\n\nclass UserInfoWindow(object):\n def setupUI(self, MainWindow, userID):\n _translate = QtCore.QCoreApplication.translate\n\n MainWindow.setObjectName(\"MainWindow\")\n MainWindow.resize(840, 640)\n\n self.userID = userID\n\n self.centralwidget = QtWidgets.QWidget(MainWindow)\n self.centralwidget.setObjectName(\"centralwidget\")\n\n self.statusTable = QtWidgets.QTableWidget(self.centralwidget)\n self.statusTable.setGeometry(QtCore.QRect(220, 40, 611, 331))\n self.statusTable.setShowGrid(False)\n self.statusTable.setObjectName(\"statusTable\")\n\n self.statusTableTitle = [\"문제 번호\", \"문제 이름\", \"결과\", \"메모리\", \"시간\", \"언어\", \"코드 길이\", \"제출한 시간\"]\n self.statusTable.setColumnCount(len(self.statusTableTitle))\n for i in range(len(self.statusTableTitle)):\n item = QtWidgets.QTableWidgetItem()\n item.setText(_translate(\"MainWindow\", self.statusTableTitle[i]))\n self.statusTable.setHorizontalHeaderItem(i, item)\n self.statusTable.setSortingEnabled(False)\n self.statusTable.verticalHeader().setVisible(False)\n\n self.infoTable = QtWidgets.QTableWidget(self.centralwidget)\n self.infoTable.setGeometry(QtCore.QRect(10, 40, 201, 331))\n self.infoTable.setShowGrid(False)\n self.infoTable.setObjectName(\"infoTable\")\n self.infoTable.setColumnCount(2)\n self.infoTable.setColumnWidth(0, 70)\n self.infoTable.setColumnWidth(1, 130)\n self.infoTable.setRowCount(12)\n for i in range(self.infoTable.rowCount()):\n for j in range(self.infoTable.columnCount()):\n item = QtWidgets.QTableWidgetItem()\n self.infoTable.setItem(i, j, item)\n self.infoTable.horizontalHeader().setVisible(False)\n self.infoTable.verticalHeader().setVisible(False)\n\n self.verticalLayoutWidget = QtWidgets.QWidget(self.centralwidget)\n self.verticalLayoutWidget.setGeometry(QtCore.QRect(10, 380, 821, 251))\n self.verticalLayoutWidget.setObjectName(\"verticalLayoutWidget\")\n self.verticalLayout = QtWidgets.QVBoxLayout(self.verticalLayoutWidget)\n self.verticalLayout.setContentsMargins(10, 10, 10, 10)\n self.verticalLayout.setObjectName(\"verticalLayout\")\n self.figure = plt.figure()\n self.graphView = FigureCanvas(self.figure)\n self.verticalLayout.addWidget(self.graphView)\n\n self.refreshButton = QtWidgets.QPushButton(self.centralwidget)\n self.refreshButton.setGeometry(QtCore.QRect(740, 580, 50, 50))\n self.refreshButton.setText(\"\")\n icon = QtGui.QIcon()\n icon.addPixmap(QtGui.QPixmap(\"refresh.png\"), QtGui.QIcon.Normal,\n QtGui.QIcon.Off)\n self.refreshButton.setIcon(icon)\n self.refreshButton.setIconSize(QtCore.QSize(32, 32))\n self.refreshButton.setObjectName(\"refreshButton\")\n\n self.infoLabel = QtWidgets.QLabel(self.centralwidget)\n self.infoLabel.setGeometry(QtCore.QRect(10, 10, 101, 21))\n self.infoLabel.setObjectName(\"infoLabel\")\n\n self.statusLabel = QtWidgets.QLabel(self.centralwidget)\n self.statusLabel.setGeometry(QtCore.QRect(220, 10, 101, 21))\n self.statusLabel.setObjectName(\"statusLabel\")\n\n self.exitButton = QtWidgets.QPushButton(self.centralwidget)\n self.exitButton.setGeometry(QtCore.QRect(780, 580, 50, 50))\n icon1 = QtGui.QIcon()\n icon1.addPixmap(QtGui.QPixmap(\"close.png\"), QtGui.QIcon.Normal, QtGui.QIcon.Off)\n self.exitButton.setIcon(icon1)\n self.exitButton.setIconSize(QtCore.QSize(32, 32))\n self.exitButton.setObjectName(\"exitButton\")\n MainWindow.setCentralWidget(self.centralwidget)\n\n MainWindow.setWindowTitle(_translate(\"MainWindow\", \"BOJ.GG\"))\n self.infoLabel.setText(_translate(\"MainWindow\", self.userID + \" 정보\"))\n self.statusLabel.setText(_translate(\"MainWindow\", \"채점 현황\"))\n\n QtCore.QMetaObject.connectSlotsByName(MainWindow)\n\n self.setupUserInfo(MainWindow)\n self.setupStatus(MainWindow)\n self.setupGraph(MainWindow)\n\n\n def setupUserInfo(self, MainWindow):\n _translate = QtCore.QCoreApplication.translate\n __sortingEnabled = self.infoTable.isSortingEnabled()\n self.infoTable.setSortingEnabled(False)\n data = boj.getUserInfo(self.userID)\n for i in range(len(data)):\n for j in range(len(data[i])):\n item = self.infoTable.item(i, j)\n item.setText(_translate(\"MainWindow\", data[i][j]))\n self.infoTable.setSortingEnabled(__sortingEnabled)\n\n def setupStatus(self, MainWindow):\n _translate = QtCore.QCoreApplication.translate\n __sortingEnabled = self.statusTable.isSortingEnabled()\n data = boj.getStatus(self.userID)\n self.statusTable.setRowCount(len(data))\n for i in range(len(data)):\n for j in range(len(data[i])):\n item = QtWidgets.QTableWidgetItem()\n\n if j == 2:\n font = QtGui.QFont()\n font.setBold(True)\n font.setWeight(75)\n item.setFont(font)\n brush = QtGui.QBrush(QtGui.QColor(0, 0, 0))\n brush.setStyle(QtCore.Qt.NoBrush)\n item.setBackground(brush)\n if data[i][j] == \"맞았습니다!!\":\n brush = QtGui.QBrush(QtGui.QColor(0, 255, 0))\n else:\n brush = QtGui.QBrush(QtGui.QColor(255, 0, 0))\n brush.setStyle(QtCore.Qt.NoBrush)\n item.setForeground(brush)\n\n item.setText(_translate(\"MainWindow\", data[i][j]))\n self.statusTable.setItem(i, j, item)\n self.statusTable.setSortingEnabled(__sortingEnabled)\n\n def setupGraph(self, MainWindow):\n data = boj.getAcceptedData(self.userID)\n self.figure.clear()\n ax = self.figure.add_subplot(111)\n # ax.bar(reversed(list(data.keys())), reversed(list(data.values())), color='g')\n ax.bar(list(reversed(list(data.keys()))), list(reversed(list(data.values()))), color='g')\n # ax.hist(list(data.keys()), data.values())\n self.graphView.draw()\n\nclass Window(QMainWindow):\n def __init__(self, parent=None):\n super(Window, self).__init__(parent)\n self.mainWindow = MainWindow()\n self.userInfoWindow = UserInfoWindow()\n self.startMainWindow()\n # self.startUserInfoWindow('sunjbs98')\n\n def startMainWindow(self):\n self.mainWindow.setupUI(self)\n self.mainWindow.pushButton.clicked.connect(lambda: self.startUserInfoWindow(self.mainWindow.lineEdit.text()))\n self.show()\n\n def startUserInfoWindow(self, userID):\n if boj.isBOJUser(userID):\n self.userInfoWindow.setupUI(self, userID)\n self.userInfoWindow.exitButton.clicked.connect(self.startMainWindow)\n self.show()\n else:\n self.mainWindow.statusbar.showMessage(\"존재하지 않는 아이디입니다.\")\n\nif __name__ == '__main__':\n app = QApplication(sys.argv)\n w = Window()\n sys.exit(app.exec_())\n", "repo_name": "ParkJH1/BOJ.GG", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 10751, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "27", "api": [{"api_name": "PyQt5.QtCore.Qt", "line_number": 15, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 15, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 17, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 17, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 19, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 19, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 20, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 20, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 22, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 22, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 24, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 24, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 25, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 25, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFrame", "line_number": 31, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 31, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFrame", "line_number": 32, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 32, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 33, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 33, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 34, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 34, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 37, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 37, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 38, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 38, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 39, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 39, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFrame", "line_number": 45, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 45, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFrame", "line_number": 46, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 46, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 47, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 47, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 48, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 48, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 51, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 51, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 53, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 53, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QStatusBar", "line_number": 57, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 57, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAction", "line_number": 60, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 60, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAction", "line_number": 62, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 62, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QMetaObject.connectSlotsByName", "line_number": 66, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QMetaObject", "line_number": 66, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 66, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QCoreApplication", "line_number": 69, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 69, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QCoreApplication", "line_number": 80, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 80, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 87, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 87, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTableWidget", "line_number": 90, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 90, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 91, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 91, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 98, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 98, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTableWidget", "line_number": 104, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 104, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 105, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 105, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 114, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 114, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 119, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 119, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 120, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 120, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 122, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.backends.backend_qt5agg.FigureCanvasQTAgg", "line_number": 126, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 129, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 129, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 130, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 130, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 132, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 132, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 133, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 133, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 133, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 134, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 134, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QSize", "line_number": 136, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 136, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 139, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 139, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 140, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 140, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 143, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 143, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 144, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 144, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 147, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 147, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 148, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 148, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 149, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 149, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 150, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 150, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 150, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QSize", "line_number": 152, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 152, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QMetaObject.connectSlotsByName", "line_number": 160, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QMetaObject", "line_number": 160, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 160, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QCoreApplication", "line_number": 168, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 168, "usage_type": "name"}, {"api_name": "boj.getUserInfo", "line_number": 171, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QCoreApplication", "line_number": 179, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 179, "usage_type": "name"}, {"api_name": "boj.getStatus", "line_number": 181, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 185, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 185, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 188, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 188, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QBrush", "line_number": 192, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 192, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QColor", "line_number": 192, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 193, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 193, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QBrush", "line_number": 196, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 196, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QColor", "line_number": 196, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QBrush", "line_number": 198, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 198, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QColor", "line_number": 198, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 199, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 199, "usage_type": "name"}, {"api_name": "boj.getAcceptedData", "line_number": 207, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 215, "usage_type": "name"}, {"api_name": "boj.isBOJUser", "line_number": 229, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 237, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 237, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 239, "usage_type": "call"}]}
+{"seq_id": "31525616767", "text": "import numpy as np\nfrom sklearn.ensemble import RandomForestRegressor\nfrom sklearn.ensemble import RandomForestClassifier\nfrom typing import Tuple\n\nfrom simulation.dgps import outcomes_not_treated, outcomes_treated\n\n\ndef regression_prediction(x_train: np.ndarray, y_train: np.ndarray, x_test: np.ndarray, **kwargs) -> np.ndarray:\n # fit random forest regression\n model = RandomForestRegressor(**kwargs)\n model.fit(x_train, y_train)\n # predict outcomes\n y_pred = model.predict(x_test)\n return y_pred\n\ndef classification_prediction(x_train: np.ndarray, w_train: np.ndarray, x_test: np.ndarray, **kwargs) -> np.ndarray:\n # fit random forest classification\n model = RandomForestClassifier(**kwargs)\n model.fit(x_train, w_train)\n # predict treatment probabilities\n w_pred = model.predict_proba(x_test)[:,1]\n return w_pred\n\ndef estimate_ate(y: np.ndarray, w: np.ndarray, x: np.ndarray, x_policy: np.ndarray, true_ate: float, cate_type: str, \n nfolds: int=2, under_sample_train: bool=False, under_sample_test: bool=False, **kwargs) -> float:\n # if both train and test sets are under-sampled, under-sample the entire dataset\n if under_sample_train and under_sample_test:\n y, w, x = _under_sample_majority_treatment(y, w, x)\n # compute pseudo-outcomes\n tau = _estimate_pseudo_outcomes(y, w, x, nfolds, under_sample_fitting=under_sample_train and not under_sample_test,**kwargs)\n # estimate ATE using doubly robust estimator\n ate = np.mean(tau)\n # compute optimal policy and its regret\n w_opt = _compute_optimal_policy(tau-true_ate, x, x_policy, **kwargs) # use the true ATE as the cost for implementing the policy\n regret = _compute_regret(w_opt,x_policy, true_ate, cate_type)\n return ate, regret\n\n\ndef _estimate_pseudo_outcomes(y: np.ndarray, w: np.ndarray, x: np.ndarray, nfolds: int=2, under_sample_fitting: bool=False, **kwargs) -> np.ndarray:\n # function to estimate pseudo-outcomes using cross-fitting\n # split sample into folds\n n = x.shape[0]\n idx = np.random.choice(np.arange(n), size=n, replace=False)\n idx = np.array_split(idx, nfolds)\n # estimate ration of treated to non-treated\n ratio_treated = np.sum(w)/np.sum(1-w)\n # initialize pseudo-outcomes\n tau = np.zeros(n)\n # loop over folds\n for i in range(nfolds):\n # split sample into train and test\n idx_test = idx[i]\n idx_train = np.concatenate(idx[:i] + idx[(i+1):])\n x_train = x[idx_train,:]\n y_train = y[idx_train]\n w_train = w[idx_train]\n x_test = x[idx_test,:]\n y_test = y[idx_test]\n w_test = w[idx_test]\n # if train and/or test sample have no treated or no non-treated, set tau to nan\n if (np.sum(w_train==1)==0) or (np.sum(w_train==0)==0) or (np.sum(w_test==1)==0) or (np.sum(w_test==0)==0):\n tau[idx_test] = np.nan\n continue\n # under-sample fitting folds if specified\n if under_sample_fitting:\n y_train, w_train, x_train=_under_sample_majority_treatment(y_train, w_train, x_train)\n # predict outcomes using data on the treated\n y_pred_treated = regression_prediction(x_train[w_train==1,:], y_train[w_train==1], x_test, **kwargs)\n # predict outcomes using data on the non-treated\n y_pred_not_treated = regression_prediction(x_train[w_train==0,:], y_train[w_train==0], x_test, **kwargs)\n # predict treatment probabilities\n w_pred = classification_prediction(x_train, w_train, x_test, **kwargs)\n # correct predicted probabilities for under-sampling (Dal Pozzolo et al., 2015)\n if under_sample_fitting:\n if ratio_treated < 1:\n # correct for under-sampling of the treated\n w_pred = ratio_treated*w_pred/(ratio_treated*w_pred - w_pred + 1)\n else:\n # correct for under-sampling of the non-treated\n w_pred = w_pred/((1-w_pred)/ratio_treated - w_pred)\n # compute pseudo-outcomes on test set\n tau[idx_test] = y_pred_treated-y_pred_not_treated + w_test*(y_test-y_pred_treated)/(w_pred+1e-10) - (1-w_test)*(y_test-y_pred_not_treated)/(1-w_pred+1e-10)\n return tau\n\n\n\ndef _under_sample_majority_treatment(y: np.ndarray, w: np.ndarray, x: np.ndarray) -> Tuple[np.ndarray, np.ndarray, np.ndarray]:\n n = x.shape[0]\n # under-sample the majority class\n n_treated = np.sum(w)\n n_not_treated = n - n_treated\n if n_treated > n_not_treated:\n # under-sample treated\n idx = np.where(w == 1)[0]\n idx = np.random.choice(idx, size=n_not_treated, replace=False)\n idx = np.concatenate((idx, np.where(w == 0)[0]))\n else:\n # under-sample not treated\n idx = np.where(w == 0)[0]\n idx = np.random.choice(idx, size=n_treated, replace=False)\n idx = np.concatenate((idx, np.where(w == 1)[0]))\n x = x[idx,:]\n y = y[idx]\n w = w[idx]\n return y, w, x\n\n\ndef _compute_optimal_policy(pseudo_outcome: np.ndarray, x_test: np.ndarray, x_policy: np.ndarray, **kwargs) -> np.ndarray:\n # define classification target:\n # 1 if pseudo-outcome is positive, 0 otherwise\n pseudo_outcome_sign = pseudo_outcome > 0\n # define weights for random forest classification\n weight = np.abs(pseudo_outcome)\n # fit random forest classification\n model = RandomForestClassifier(**kwargs)\n model.fit(x_test, pseudo_outcome_sign, sample_weight=weight)\n # predict optimal policy\n w_opt = model.predict(x_policy)\n return w_opt\n\ndef _compute_regret(w_opt: np.ndarray, x_policy: np.ndarray, true_ate: bool, cate_type: str) -> float:\n y_treated = outcomes_treated(x_policy, true_ate, cate_type)\n y_not_treated = outcomes_not_treated(x_policy)\n # determine oracle policy (i.e. policy that maximizes the expected outcome)\n w_oracle = y_treated - y_not_treated > true_ate # treat only if individual CATE is larger than true ATE\n # define outcome based on policy\n y_policy = y_treated*w_opt + y_not_treated*(1-w_opt) - w_opt*true_ate\n y_oracle = y_treated*w_oracle + y_not_treated*(1-w_oracle) - w_oracle*true_ate\n # compute regret as the average outcome of the oracle policy minus the average outcome of the optimal policy\n regret = np.mean(y_oracle) - np.mean(y_policy)\n return regret\n", "repo_name": "dballinari/simulation-unbalanced-treatment", "sub_path": "simulation/estimator.py", "file_name": "estimator.py", "file_ext": "py", "file_size_in_byte": 6318, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "numpy.ndarray", "line_number": 9, "usage_type": "attribute"}, {"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.array_split", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 63, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 96, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 102, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 88, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 109, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 114, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 122, "usage_type": "attribute"}, {"api_name": "simulation.dgps.outcomes_treated", "line_number": 123, "usage_type": "call"}, {"api_name": "simulation.dgps.outcomes_not_treated", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 131, "usage_type": "call"}]}
+{"seq_id": "5245815186", "text": "from django.shortcuts import render, get_object_or_404\nfrom django.views import View\nfrom news.models import News\nfrom core.models import Index\nfrom contacts.models import *\nfrom pages.models import Page\nfrom django.http import Http404\n\n\nclass IndexView(View):\n def get(self, request):\n news = News.objects.filter(is_active=True)[:3]\n index = Index.objects.first()\n\n addresses = Address.objects.all()\n ptypes = PhoneType.objects.all()\n emails = Email.objects.all()\n schedule = Schedule.objects.all()\n map_code = MapCode.objects.filter(map_type='contacts').first()\n\n areas = ActivityArea.objects.all()\n area_code = MapCode.objects.filter(map_type='area').first()\n\n context = {\n 'news': news,\n 'index': index,\n 'addresses': addresses,\n 'ptypes': ptypes,\n 'emails': emails,\n 'schedule': schedule,\n 'map_code': map_code,\n 'areas': areas,\n 'area_code': area_code,\n }\n return render(request, 'core/index.html', context)\n\n\nclass CalcView(View):\n def get(self, request):\n index = Index.objects.first()\n koef = str(index.koef).replace(',', '.')\n\n context = {\n 'koef': koef,\n }\n return render(request, 'core/calc.html', context)\n\n\nclass DropMenuView(View):\n def get(self, request, drop_menu):\n drop_page = get_object_or_404(Page, slug=drop_menu)\n\n context = {\n 'drop_page': drop_page,\n }\n return render(request, 'core/drop_menu.html', context)", "repo_name": "gurgenXD/maykoptec", "sub_path": "core/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1610, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "django.views.View", "line_number": 10, "usage_type": "name"}, {"api_name": "news.models", "line_number": 12, "usage_type": "name"}, {"api_name": "news.models.News.objects.filter", "line_number": 12, "usage_type": "call"}, {"api_name": "news.models.News.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "news.models.News", "line_number": 12, "usage_type": "name"}, {"api_name": "core.models.Index.objects.first", "line_number": 13, "usage_type": "call"}, {"api_name": "core.models.Index.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "core.models.Index", "line_number": 13, "usage_type": "name"}, {"api_name": "news.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 35, "usage_type": "call"}, {"api_name": "django.views.View", "line_number": 38, "usage_type": "name"}, {"api_name": "core.models.Index.objects.first", "line_number": 40, "usage_type": "call"}, {"api_name": "core.models.Index.objects", "line_number": 40, "usage_type": "attribute"}, {"api_name": "core.models.Index", "line_number": 40, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 46, "usage_type": "call"}, {"api_name": "django.views.View", "line_number": 49, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 51, "usage_type": "call"}, {"api_name": "pages.models.Page", "line_number": 51, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 56, "usage_type": "call"}]}
+{"seq_id": "40969445493", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Wed May 20 13:22:29 2015\r\n\r\n@author: Don\r\n\"\"\"\r\n\r\n#! /usr/bin/python\r\n\r\nimport datetime\r\nimport time\r\nimport csv\r\nimport smtplib\r\nimport os, sys\r\nimport itertools\r\n\r\nfrom email.mime.multipart import MIMEMultipart\r\nfrom email.mime.text import MIMEText\r\n\r\n \r\ndef processData():\r\n \r\n #partialfilename = \"C:\\myFolder\\StockChange\"\r\n \r\n ordersDataList = []\r\n \r\n ordersDataList.append(getListingCount()) #HourlyListingCount\r\n \r\n emailSubject = constructEmail(\"S\",ordersDataList)\r\n emailBody = constructEmail(\"B\",ordersDataList)\r\n \r\n sendEmail(emailSubject,emailBody)\r\n \r\n return\r\n\r\ndef cleanData():\r\n \r\n partialfilename = \"C:\\myFolder\\StockChange\"\r\n lstngTimeStampPrev = \" \"\r\n \r\n fib1 = open(partialfilename+\".csv\") \r\n fi_b1 = csv.reader(fib1)\r\n \r\n fob1 = open(partialfilename+\"Clean.csv\",'w')\r\n fo_b1 = csv.writer(fob1,lineterminator='\\n')\r\n\r\n for row1 in fi_b1:\r\n \r\n if row1[3] == lstngTimeStampPrev:\r\n x = 0\r\n else:\r\n lstngTimeStampPrev = row1[3]\r\n fo_b1.writerow(row1)\r\n\r\n fib1.close()\r\n fob1.close()\r\n \r\n return \r\n \r\ndef getListingCount():\r\n \r\n partialfilename = \"C:\\myFolder\\StockChange\"\r\n columnSKU = 0\r\n columnStkNow = 1\r\n columnLvlChg = 2 \r\n columnDate = 3\r\n columnLstr = 4\r\n columnImgId = 5\r\n actDay = 7\r\n formatAs = \" - \"\r\n customerReturn = 0\r\n \r\n today = datetime.datetime.today()\r\n thisHour = datetime.datetime.now().strftime('%I') \r\n \r\n two_hours_ago = datetime.datetime.now() - datetime.timedelta(hours=2)\r\n d = modification_date(partialfilename+\".csv\")\r\n\r\n if d < two_hours_ago :\r\n return \" Restart Linnworks ---------> \" + partialfilename + \".csv is not current \\n\\n\" \r\n \r\n lines = \"\"\r\n\r\n partialfilename = \"C:\\myFolder\\PosInv.csv\"\r\n \r\n totValue = 0\r\n\r\n fib1 = open(partialfilename,encoding='UTF8') \r\n fi_b1 = csv.reader(fib1)\r\n next (fi_b1)\r\n \r\n fob1 = open(\"C:\\myFolder\\Tmp.csv\",'w')\r\n fo_b1 = csv.writer(fob1,lineterminator='\\n')\r\n\r\n for row1 in fi_b1:\r\n departmnt = row1[0]\r\n totValue = int(float(row1[1]))\r\n \r\n if departmnt == \"TILE\":\r\n fo_b1.writerow([departmnt,totValue,200000,totValue-200000])\r\n elif departmnt == \"LAMINATE\":\r\n fo_b1.writerow([departmnt,totValue,65000,totValue-65000])\r\n \r\n fib1.close()\r\n fob1.close()\r\n\r\n sortLinesInFile(\"C:\\myFolder\\Tmp.csv\")\r\n \r\n f = open(\"C:\\myFolder\\Tmp.csv\", \"r\")\r\n for i, line in enumerate(f):\r\n lines = lines + \" \" + line.split(\",\")[0].zfill(15) + formatAs + line.split(\",\")[1].zfill(3) + formatAs + line.split(\",\")[2].zfill(3) + formatAs + line.split(\",\")[3].zfill(3) + \"\\n\"\r\n f.close()\r\n \r\n return lines\r\n\r\n \r\ndef sortLinesInFile(fileName):\r\n f = open(fileName, \"r\")\r\n lines = [line for line in f if line.strip()]\r\n f.close()\r\n lines.sort()\r\n \r\n f = open(fileName, 'w')\r\n f.writelines(lines)\r\n f.close() \r\n\r\ndef modification_date(filename):\r\n t = os.path.getmtime(filename)\r\n return datetime.datetime.fromtimestamp(t)\r\n \r\ndef constructEmail(emailSubjectOrBody,ordersDataList): \r\n today = datetime.date.today().strftime('%x')\r\n thisHour = datetime.datetime.now().strftime('%I:%M:%S %p')\r\n now_time = datetime.datetime.now()\r\n \r\n if emailSubjectOrBody == \"S\" and now_time > now_time.replace(hour=9, minute=0, second=0, microsecond=0) :\r\n return \"Inventory Targets : \" + str(today) + \" \" + str(thisHour) \r\n \r\n elif emailSubjectOrBody == \"B\" and now_time > now_time.replace(hour=9, minute=0, second=0, microsecond=0) :\r\n body = \"PAYLESS COMPONENTS : \\n******************\\n\" \r\n body = body + \" Listings by the Hour today: \\n\" + ordersDataList[0] + \"\\n\"\r\n return body \r\n \r\ndef sendEmail(emailSubject,emailBody):\r\n me = \"reports@paylesscomponents.com\"\r\n you = \"don@paylesscomponents.com\"\r\n #you = [\"don@paylesscomponents.com\", \"sales@paylesscomponents.com\"]\r\n #you = [\"kyle@paylesscomponents.com\", \"don@paylesscomponents.com\", \"nick@paylesscomponents.com\"]\r\n #you = [\"kyle@paylesscomponents.com\", \"jhamamy@factory-surplus.com\", \"nick@paylesscomponents.com\", \"desirae@paylesscomponents.com\", \"don@paylesscomponents.com\"]\r\n \r\n COMMASPACE = ', '\r\n\r\n msg = MIMEMultipart('alternative')\r\n msg['Subject'] = emailSubject\r\n msg['From'] = me\r\n msg['To'] = COMMASPACE.join(you)\r\n\r\n part1 = MIMEText(emailBody, 'plain') \r\n msg.attach(part1)\r\n \r\n s = smtplib.SMTP_SSL('smtpout.secureserver.net',465) \r\n s.login(\"don@paylesscomponents.com\", \"donpay123\")\r\n \r\n s.sendmail(me, you, msg.as_string())\r\n s.quit()\r\n\r\n#cleanData()\r\n\r\nprocessData()", "repo_name": "paylessc/pointOfSaleAutomation", "sub_path": "QBPOS%20Custom%20Reports/PosInventoryTargets.py", "file_name": "PosInventoryTargets.py", "file_ext": "py", "file_size_in_byte": 5006, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "csv.reader", "line_number": 42, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 73, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 73, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 74, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 74, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 76, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 76, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 76, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 89, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path.getmtime", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path", "line_number": 128, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 129, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 129, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 132, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 132, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 133, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 133, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 134, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 134, "usage_type": "attribute"}, {"api_name": "email.mime.multipart.MIMEMultipart", "line_number": 153, "usage_type": "call"}, {"api_name": "email.mime.text.MIMEText", "line_number": 158, "usage_type": "call"}, {"api_name": "smtplib.SMTP_SSL", "line_number": 161, "usage_type": "call"}]}
+{"seq_id": "24385020187", "text": "\"\"\"\nAbstraction of the deployment functionality for processors.\n\nThe Processing Server provides the configuration parameters to the Deployer agent.\nThe Deployer agent runs the RabbitMQ Server, MongoDB and the Processing Hosts.\nEach Processing Host may have several Processing Workers.\nEach Processing Worker is an instance of an OCR-D processor.\n\"\"\"\nfrom __future__ import annotations\nfrom typing import Dict, List, Union\nfrom re import search as re_search\nfrom pathlib import Path\nimport subprocess\nfrom time import sleep\n\nfrom ocrd_utils import config, getLogger, safe_filename\n\nfrom .deployment_utils import (\n create_docker_client,\n DeployType,\n verify_mongodb_available,\n verify_rabbitmq_available,\n)\nfrom .logging import get_mets_server_logging_file_path\nfrom .runtime_data import (\n DataHost,\n DataMongoDB,\n DataProcessingWorker,\n DataProcessorServer,\n DataRabbitMQ\n)\nfrom .utils import (\n is_mets_server_running,\n stop_mets_server,\n validate_and_load_config\n)\n\n\nclass Deployer:\n def __init__(self, config_path: str) -> None:\n self.log = getLogger('ocrd_network.deployer')\n config = validate_and_load_config(config_path)\n\n self.data_mongo: DataMongoDB = DataMongoDB(config['database'])\n self.data_queue: DataRabbitMQ = DataRabbitMQ(config['process_queue'])\n self.data_hosts: List[DataHost] = []\n self.internal_callback_url = config.get('internal_callback_url', None)\n for config_host in config['hosts']:\n self.data_hosts.append(DataHost(config_host))\n self.mets_servers: Dict = {} # {\"mets_server_url\": \"mets_server_pid\"}\n\n # TODO: Reconsider this.\n def find_matching_processors(\n self,\n worker_only: bool = False,\n server_only: bool = False,\n docker_only: bool = False,\n native_only: bool = False,\n str_names_only: bool = False,\n unique_only: bool = False\n ) -> Union[List[str], List[object]]:\n \"\"\"Finds and returns a list of matching data objects of type:\n `DataProcessingWorker` and `DataProcessorServer`.\n\n :py:attr:`worker_only` match only processors with worker status\n :py:attr:`server_only` match only processors with server status\n :py:attr:`docker_only` match only docker processors\n :py:attr:`native_only` match only native processors\n :py:attr:`str_only` returns the processor_name instead of data object\n :py:attr:`unique_only` remove duplicates from the matches\n\n `worker_only` and `server_only` are mutually exclusive to each other\n `docker_only` and `native_only` are mutually exclusive to each other\n `unique_only` is allowed only together with `str_names_only`\n \"\"\"\n\n if worker_only and server_only:\n raise ValueError(f\"Only 'worker_only' or 'server_only' is allowed, not both.\")\n if docker_only and native_only:\n raise ValueError(f\"Only 'docker_only' or 'native_only' is allowed, not both.\")\n if not str_names_only and unique_only:\n raise ValueError(f\"Value 'unique_only' is allowed only together with 'str_names_only'\")\n\n # Find all matching objects of type:\n # DataProcessingWorker or DataProcessorServer\n matched_objects = []\n for data_host in self.data_hosts:\n if not server_only:\n for data_worker in data_host.data_workers:\n if data_worker.deploy_type == DeployType.NATIVE and docker_only:\n continue\n if data_worker.deploy_type == DeployType.DOCKER and native_only:\n continue\n matched_objects.append(data_worker)\n if not worker_only:\n for data_server in data_host.data_servers:\n if data_server.deploy_type == DeployType.NATIVE and docker_only:\n continue\n if data_server.deploy_type == DeployType.DOCKER and native_only:\n continue\n matched_objects.append(data_server)\n if str_names_only:\n # gets only the processor names of the matched objects\n name_list = [match.processor_name for match in matched_objects]\n if unique_only:\n # removes the duplicates, if any\n return list(dict.fromkeys(name_list))\n return name_list\n return matched_objects\n\n def resolve_processor_server_url(self, processor_name) -> str:\n processor_server_url = ''\n for data_host in self.data_hosts:\n for data_server in data_host.data_servers:\n if data_server.processor_name == processor_name:\n processor_server_url = f'http://{data_host.address}:{data_server.port}/'\n return processor_server_url\n\n def kill_all(self) -> None:\n \"\"\" kill all started services: hosts, database, queue\n\n The order of killing is important to optimize graceful shutdown in the future. If RabbitMQ\n server is killed before killing Processing Workers, that may have bad outcome and leave\n Processing Workers in an unpredictable state\n \"\"\"\n self.kill_hosts()\n self.kill_mongodb()\n self.kill_rabbitmq()\n\n def deploy_hosts(\n self,\n mongodb_url: str,\n rabbitmq_url: str\n ) -> None:\n for host_data in self.data_hosts:\n if host_data.needs_ssh:\n host_data.create_client(client_type='ssh')\n assert host_data.ssh_client\n if host_data.needs_docker:\n host_data.create_client(client_type='docker')\n assert host_data.docker_client\n\n self.log.debug(f'Deploying processing workers on host: {host_data.address}')\n for data_worker in host_data.data_workers:\n self._deploy_processing_worker(\n mongodb_url,\n rabbitmq_url,\n host_data,\n data_worker\n )\n\n self.log.debug(f'Deploying processor servers on host: {host_data.address}')\n for data_server in host_data.data_servers:\n self._deploy_processor_server(\n mongodb_url,\n host_data,\n data_server\n )\n\n if host_data.ssh_client:\n host_data.ssh_client.close()\n host_data.ssh_client = None\n if host_data.docker_client:\n host_data.docker_client.close()\n host_data.docker_client = None\n\n def _deploy_processing_worker(\n self,\n mongodb_url: str,\n rabbitmq_url: str,\n host_data: DataHost,\n data_worker: DataProcessingWorker\n ) -> None:\n self.log.debug(f\"Deploying processing worker, \"\n f\"environment: '{data_worker.deploy_type}', \"\n f\"name: '{data_worker.processor_name}', \"\n f\"address: '{host_data.address}'\")\n\n if data_worker.deploy_type == DeployType.NATIVE:\n assert host_data.ssh_client # to satisfy mypy\n pid = self.start_native_processor(\n ssh_client=host_data.ssh_client,\n processor_name=data_worker.processor_name,\n queue_url=rabbitmq_url,\n database_url=mongodb_url,\n )\n data_worker.pid = pid\n elif data_worker.deploy_type == DeployType.DOCKER:\n assert host_data.docker_client # to satisfy mypy\n pid = self.start_docker_processor(\n docker_client=host_data.docker_client,\n processor_name=data_worker.processor_name,\n _queue_url=rabbitmq_url,\n _database_url=mongodb_url\n )\n data_worker.pid = pid\n sleep(0.2)\n\n # TODO: Revisit this to remove code duplications of deploy_* methods\n def _deploy_processor_server(\n self,\n mongodb_url: str,\n host_data: DataHost,\n data_server: DataProcessorServer,\n ) -> None:\n self.log.debug(f\"Deploying processing worker, \"\n f\"environment: '{data_server.deploy_type}', \"\n f\"name: '{data_server.processor_name}', \"\n f\"address: '{data_server.host}:{data_server.port}'\")\n\n if data_server.deploy_type == DeployType.NATIVE:\n assert host_data.ssh_client\n pid = self.start_native_processor_server(\n ssh_client=host_data.ssh_client,\n processor_name=data_server.processor_name,\n agent_address=f'{data_server.host}:{data_server.port}',\n database_url=mongodb_url,\n )\n data_server.pid = pid\n\n if data_server.processor_name in host_data.server_ports:\n name = data_server.processor_name\n port = data_server.port\n if host_data.server_ports[name]:\n host_data.server_ports[name] = host_data.server_ports[name].append(port)\n else:\n host_data.server_ports[name] = [port]\n else:\n host_data.server_ports[data_server.processor_name] = [data_server.port]\n elif data_server.deploy_type == DeployType.DOCKER:\n raise Exception(\"Deploying docker processor server is not supported yet!\")\n\n def deploy_rabbitmq(\n self,\n image: str,\n detach: bool,\n remove: bool,\n ports_mapping: Union[Dict, None] = None\n ) -> str:\n if self.data_queue.skip_deployment:\n self.log.debug(f\"RabbitMQ is externaly managed. Skipping deployment\")\n verify_rabbitmq_available(\n self.data_queue.address,\n self.data_queue.port,\n self.data_queue.vhost,\n self.data_queue.username,\n self.data_queue.password\n )\n return self.data_queue.url\n self.log.debug(f\"Trying to deploy '{image}', with modes: \"\n f\"detach='{detach}', remove='{remove}'\")\n\n if not self.data_queue or not self.data_queue.address:\n raise ValueError('Deploying RabbitMQ has failed - missing configuration.')\n\n client = create_docker_client(\n self.data_queue.address,\n self.data_queue.ssh_username,\n self.data_queue.ssh_password,\n self.data_queue.ssh_keypath\n )\n if not ports_mapping:\n # 5672, 5671 - used by AMQP 0-9-1 and AMQP 1.0 clients without and with TLS\n # 15672, 15671: HTTP API clients, management UI and rabbitmq admin, without and with TLS\n # 25672: used for internode and CLI tools communication and is allocated from\n # a dynamic range (limited to a single port by default, computed as AMQP port + 20000)\n ports_mapping = {\n 5672: self.data_queue.port,\n 15672: 15672,\n 25672: 25672\n }\n res = client.containers.run(\n image=image,\n detach=detach,\n remove=remove,\n ports=ports_mapping,\n # The default credentials to be used by the processing workers\n environment=[\n f'RABBITMQ_DEFAULT_USER={self.data_queue.username}',\n f'RABBITMQ_DEFAULT_PASS={self.data_queue.password}'\n ]\n )\n assert res and res.id, \\\n f'Failed to start RabbitMQ docker container on host: {self.data_queue.address}'\n self.data_queue.pid = res.id\n client.close()\n\n rmq_host = self.data_queue.address\n rmq_port = int(self.data_queue.port)\n rmq_vhost = '/'\n\n verify_rabbitmq_available(\n host=rmq_host,\n port=rmq_port,\n vhost=rmq_vhost,\n username=self.data_queue.username,\n password=self.data_queue.password\n )\n self.log.info(f'The RabbitMQ server was deployed on URL: '\n f'{rmq_host}:{rmq_port}{rmq_vhost}')\n return self.data_queue.url\n\n def deploy_mongodb(\n self,\n image: str,\n detach: bool,\n remove: bool,\n ports_mapping: Union[Dict, None] = None\n ) -> str:\n if self.data_mongo.skip_deployment:\n self.log.debug('MongoDB is externaly managed. Skipping deployment')\n verify_mongodb_available(self.data_mongo.url)\n return self.data_mongo.url\n\n self.log.debug(f\"Trying to deploy '{image}', with modes: \"\n f\"detach='{detach}', remove='{remove}'\")\n\n if not self.data_mongo or not self.data_mongo.address:\n raise ValueError('Deploying MongoDB has failed - missing configuration.')\n\n client = create_docker_client(\n self.data_mongo.address,\n self.data_mongo.ssh_username,\n self.data_mongo.ssh_password,\n self.data_mongo.ssh_keypath\n )\n if not ports_mapping:\n ports_mapping = {\n 27017: self.data_mongo.port\n }\n if self.data_mongo.username:\n environment = [\n f'MONGO_INITDB_ROOT_USERNAME={self.data_mongo.username}',\n f'MONGO_INITDB_ROOT_PASSWORD={self.data_mongo.password}'\n ]\n else:\n environment = []\n\n res = client.containers.run(\n image=image,\n detach=detach,\n remove=remove,\n ports=ports_mapping,\n environment=environment\n )\n if not res or not res.id:\n raise RuntimeError('Failed to start MongoDB docker container on host: '\n f'{self.data_mongo.address}')\n self.data_mongo.pid = res.id\n client.close()\n\n mongodb_hostinfo = f'{self.data_mongo.address}:{self.data_mongo.port}'\n self.log.info(f'The MongoDB was deployed on host: {mongodb_hostinfo}')\n return self.data_mongo.url\n\n def kill_rabbitmq(self) -> None:\n if self.data_queue.skip_deployment:\n return\n elif not self.data_queue.pid:\n self.log.warning('No running RabbitMQ instance found')\n return\n client = create_docker_client(\n self.data_queue.address,\n self.data_queue.ssh_username,\n self.data_queue.ssh_password,\n self.data_queue.ssh_keypath\n )\n client.containers.get(self.data_queue.pid).stop()\n self.data_queue.pid = None\n client.close()\n self.log.info('The RabbitMQ is stopped')\n\n def kill_mongodb(self) -> None:\n if self.data_mongo.skip_deployment:\n return\n elif not self.data_mongo.pid:\n self.log.warning('No running MongoDB instance found')\n return\n client = create_docker_client(\n self.data_mongo.address,\n self.data_mongo.ssh_username,\n self.data_mongo.ssh_password,\n self.data_mongo.ssh_keypath\n )\n client.containers.get(self.data_mongo.pid).stop()\n self.data_mongo.pid = None\n client.close()\n self.log.info('The MongoDB is stopped')\n\n def kill_hosts(self) -> None:\n self.log.debug('Starting to kill/stop hosts')\n # Kill processing hosts\n for host_data in self.data_hosts:\n if host_data.needs_ssh:\n host_data.create_client(client_type='ssh')\n assert host_data.ssh_client\n if host_data.needs_docker:\n host_data.create_client(client_type='docker')\n assert host_data.docker_client\n\n self.log.debug(f'Killing/Stopping processing workers on host: {host_data.address}')\n self.kill_processing_workers(host_data)\n\n self.log.debug(f'Killing/Stopping processor servers on host: {host_data.address}')\n self.kill_processor_servers(host_data)\n\n if host_data.ssh_client:\n host_data.ssh_client.close()\n host_data.ssh_client = None\n if host_data.docker_client:\n host_data.docker_client.close()\n host_data.docker_client = None\n\n # TODO: Optimize the code duplication from start_* and kill_* methods\n def kill_processing_workers(self, host_data: DataHost) -> None:\n amount = len(host_data.data_workers)\n if not amount:\n self.log.info(f'No active processing workers to be stopped.')\n return\n self.log.info(f\"Trying to stop {amount} processing workers:\")\n for worker in host_data.data_workers:\n if not worker.pid:\n continue\n if worker.deploy_type == DeployType.NATIVE:\n host_data.ssh_client.exec_command(f'kill {worker.pid}')\n self.log.info(f\"Stopped native worker with pid: '{worker.pid}'\")\n elif worker.deploy_type == DeployType.DOCKER:\n host_data.docker_client.containers.get(worker.pid).stop()\n self.log.info(f\"Stopped docker worker with container id: '{worker.pid}'\")\n host_data.data_workers = []\n\n def kill_processor_servers(self, host_data: DataHost) -> None:\n amount = len(host_data.data_servers)\n if not amount:\n self.log.info(f'No active processor servers to be stopped.')\n return\n self.log.info(f\"Trying to stop {amount} processing workers:\")\n for server in host_data.data_servers:\n if not server.pid:\n continue\n if server.deploy_type == DeployType.NATIVE:\n host_data.ssh_client.exec_command(f'kill {server.pid}')\n self.log.info(f\"Stopped native server with pid: '{server.pid}'\")\n elif server.deploy_type == DeployType.DOCKER:\n host_data.docker_client.containers.get(server.pid).stop()\n self.log.info(f\"Stopped docker server with container id: '{server.pid}'\")\n host_data.data_servers = []\n\n def start_native_processor(\n self,\n ssh_client,\n processor_name: str,\n queue_url: str,\n database_url: str\n ) -> str:\n \"\"\" start a processor natively on a host via ssh\n\n Args:\n ssh_client: paramiko SSHClient to execute commands on a host\n processor_name: name of processor to run\n queue_url: url to rabbitmq\n database_url: url to database\n\n Returns:\n str: pid of running process\n \"\"\"\n self.log.info(f'Starting native processing worker: {processor_name}')\n channel = ssh_client.invoke_shell()\n stdin, stdout = channel.makefile('wb'), channel.makefile('rb')\n cmd = f'{processor_name} worker --database {database_url} --queue {queue_url} &'\n # the only way (I could find) to make it work to start a process in the background and\n # return early is this construction. The pid of the last started background process is\n # printed with `echo $!` but it is printed inbetween other output. Because of that I added\n # `xyz` before and after the code to easily be able to filter out the pid via regex when\n # returning from the function\n\n self.log.debug(f'About to execute command: {cmd}')\n stdin.write(f'{cmd}\\n')\n stdin.write('echo xyz$!xyz \\n exit \\n')\n output = stdout.read().decode('utf-8')\n stdout.close()\n stdin.close()\n return re_search(r'xyz([0-9]+)xyz', output).group(1) # type: ignore\n\n def start_docker_processor(\n self,\n docker_client,\n processor_name: str,\n _queue_url: str,\n _database_url: str\n ) -> str:\n # TODO: Raise an exception here as well?\n # raise Exception(\"Deploying docker processing worker is not supported yet!\")\n\n self.log.info(f'Starting docker container processor: {processor_name}')\n # TODO: add real command here to start processing server in docker here\n res = docker_client.containers.run('debian', 'sleep 500s', detach=True, remove=True)\n assert res and res.id, f'Running processor: {processor_name} in docker-container failed'\n return res.id\n\n # TODO: Just a copy of the above start_native_processor() method.\n # Far from being great... But should be good as a starting point\n def start_native_processor_server(\n self,\n ssh_client,\n processor_name: str,\n agent_address: str,\n database_url: str\n ) -> str:\n self.log.info(f\"Starting native processor server: {processor_name} on {agent_address}\")\n channel = ssh_client.invoke_shell()\n stdin, stdout = channel.makefile('wb'), channel.makefile('rb')\n cmd = f'{processor_name} server --address {agent_address} --database {database_url} &'\n self.log.debug(f'About to execute command: {cmd}')\n stdin.write(f'{cmd}\\n')\n stdin.write('echo xyz$!xyz \\n exit \\n')\n output = stdout.read().decode('utf-8')\n stdout.close()\n stdin.close()\n return re_search(r'xyz([0-9]+)xyz', output).group(1) # type: ignore\n\n # TODO: No support for TCP version yet\n def start_unix_mets_server(self, mets_path: str) -> Path:\n log_file = get_mets_server_logging_file_path(mets_path=mets_path)\n mets_server_url = Path(config.OCRD_NETWORK_SOCKETS_ROOT_DIR, f\"{safe_filename(mets_path)}.sock\")\n\n if is_mets_server_running(mets_server_url=str(mets_server_url)):\n self.log.info(f\"The mets server is already started: {mets_server_url}\")\n return mets_server_url\n\n cwd = Path(mets_path).parent\n self.log.info(f'Starting UDS mets server: {mets_server_url}')\n sub_process = subprocess.Popen(\n args=['nohup', 'ocrd', 'workspace', '--mets-server-url', f'{mets_server_url}',\n '-d', f'{cwd}', 'server', 'start'],\n shell=False,\n stdout=open(file=log_file, mode='w'),\n stderr=open(file=log_file, mode='a'),\n cwd=cwd,\n universal_newlines=True\n )\n # Wait for the mets server to start\n sleep(2)\n self.mets_servers[mets_server_url] = sub_process.pid\n return mets_server_url\n\n def stop_unix_mets_server(self, mets_server_url: str) -> None:\n self.log.info(f'Stopping UDS mets server: {mets_server_url}')\n if Path(mets_server_url) in self.mets_servers:\n mets_server_pid = self.mets_servers[Path(mets_server_url)]\n else:\n raise Exception(f\"Mets server not found: {mets_server_url}\")\n\n '''\n subprocess.run(\n args=['kill', '-s', 'SIGINT', f'{mets_server_pid}'],\n shell=False,\n universal_newlines=True\n )\n '''\n\n # TODO: Reconsider this again\n # Not having this sleep here causes connection errors\n # on the last request processed by the processing worker.\n # Sometimes 3 seconds is enough, sometimes not.\n sleep(5)\n stop_mets_server(mets_server_url=mets_server_url)\n", "repo_name": "OCR-D/core", "sub_path": "ocrd_network/ocrd_network/deployer.py", "file_name": "deployer.py", "file_ext": "py", "file_size_in_byte": 23362, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 117, "dataset": "github-code", "pt": "27", "api": [{"api_name": "ocrd_utils.getLogger", "line_number": 41, "usage_type": "call"}, {"api_name": "ocrd_utils.config", "line_number": 42, "usage_type": "name"}, {"api_name": "utils.validate_and_load_config", "line_number": 42, "usage_type": "call"}, {"api_name": "runtime_data.DataMongoDB", "line_number": 44, "usage_type": "name"}, {"api_name": "ocrd_utils.config", "line_number": 44, "usage_type": "name"}, {"api_name": "runtime_data.DataRabbitMQ", "line_number": 45, "usage_type": "name"}, {"api_name": "ocrd_utils.config", "line_number": 45, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 46, "usage_type": "name"}, {"api_name": "runtime_data.DataHost", "line_number": 46, "usage_type": "name"}, {"api_name": "ocrd_utils.config.get", "line_number": 47, "usage_type": "call"}, {"api_name": "ocrd_utils.config", "line_number": 47, "usage_type": "name"}, {"api_name": "ocrd_utils.config", "line_number": 48, "usage_type": "name"}, {"api_name": "runtime_data.DataHost", "line_number": 49, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 50, "usage_type": "name"}, {"api_name": "deployment_utils.DeployType.NATIVE", "line_number": 90, "usage_type": "attribute"}, {"api_name": "deployment_utils.DeployType", "line_number": 90, "usage_type": "name"}, {"api_name": "deployment_utils.DeployType.DOCKER", "line_number": 92, "usage_type": "attribute"}, {"api_name": "deployment_utils.DeployType", "line_number": 92, "usage_type": "name"}, {"api_name": "deployment_utils.DeployType.NATIVE", "line_number": 97, "usage_type": "attribute"}, {"api_name": "deployment_utils.DeployType", "line_number": 97, "usage_type": "name"}, {"api_name": "deployment_utils.DeployType.DOCKER", "line_number": 99, "usage_type": "attribute"}, {"api_name": "deployment_utils.DeployType", "line_number": 99, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 61, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 61, "usage_type": "name"}, {"api_name": "runtime_data.DataHost", "line_number": 171, "usage_type": "name"}, {"api_name": "runtime_data.DataProcessingWorker", "line_number": 172, "usage_type": "name"}, {"api_name": "deployment_utils.DeployType.NATIVE", "line_number": 179, "usage_type": "attribute"}, {"api_name": "deployment_utils.DeployType", "line_number": 179, "usage_type": "name"}, {"api_name": "deployment_utils.DeployType.DOCKER", "line_number": 188, "usage_type": "attribute"}, {"api_name": "deployment_utils.DeployType", "line_number": 188, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 197, "usage_type": "call"}, {"api_name": "runtime_data.DataHost", "line_number": 203, "usage_type": "name"}, {"api_name": "runtime_data.DataProcessorServer", "line_number": 204, "usage_type": "name"}, {"api_name": "deployment_utils.DeployType.NATIVE", "line_number": 211, "usage_type": "attribute"}, {"api_name": "deployment_utils.DeployType", "line_number": 211, "usage_type": "name"}, {"api_name": "deployment_utils.DeployType.DOCKER", "line_number": 230, "usage_type": "attribute"}, {"api_name": "deployment_utils.DeployType", "line_number": 230, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 238, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 238, "usage_type": "name"}, {"api_name": "deployment_utils.verify_rabbitmq_available", "line_number": 242, "usage_type": "call"}, {"api_name": "deployment_utils.create_docker_client", "line_number": 256, "usage_type": "call"}, {"api_name": "deployment_utils.verify_rabbitmq_available", "line_number": 292, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 308, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 308, "usage_type": "name"}, {"api_name": "deployment_utils.verify_mongodb_available", "line_number": 312, "usage_type": "call"}, {"api_name": "deployment_utils.create_docker_client", "line_number": 321, "usage_type": "call"}, {"api_name": "deployment_utils.create_docker_client", "line_number": 362, "usage_type": "call"}, {"api_name": "deployment_utils.create_docker_client", "line_number": 379, "usage_type": "call"}, {"api_name": "runtime_data.DataHost", "line_number": 415, "usage_type": "name"}, {"api_name": "deployment_utils.DeployType.NATIVE", "line_number": 424, "usage_type": "attribute"}, {"api_name": "deployment_utils.DeployType", "line_number": 424, "usage_type": "name"}, {"api_name": "deployment_utils.DeployType.DOCKER", "line_number": 427, "usage_type": "attribute"}, {"api_name": "deployment_utils.DeployType", "line_number": 427, "usage_type": "name"}, {"api_name": "runtime_data.DataHost", "line_number": 432, "usage_type": "name"}, {"api_name": "deployment_utils.DeployType.NATIVE", "line_number": 441, "usage_type": "attribute"}, {"api_name": "deployment_utils.DeployType", "line_number": 441, "usage_type": "name"}, {"api_name": "deployment_utils.DeployType.DOCKER", "line_number": 444, "usage_type": "attribute"}, {"api_name": "deployment_utils.DeployType", "line_number": 444, "usage_type": "name"}, {"api_name": "re.search", "line_number": 483, "usage_type": "call"}, {"api_name": "re.search", "line_number": 520, "usage_type": "call"}, {"api_name": "logging.get_mets_server_logging_file_path", "line_number": 524, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 525, "usage_type": "call"}, {"api_name": "ocrd_utils.config.OCRD_NETWORK_SOCKETS_ROOT_DIR", "line_number": 525, "usage_type": "attribute"}, {"api_name": "ocrd_utils.config", "line_number": 525, "usage_type": "name"}, {"api_name": "ocrd_utils.safe_filename", "line_number": 525, "usage_type": "call"}, {"api_name": "utils.is_mets_server_running", "line_number": 527, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 531, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 533, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 543, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 523, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 549, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 550, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 566, "usage_type": "call"}, {"api_name": "utils.stop_mets_server", "line_number": 567, "usage_type": "call"}]}
+{"seq_id": "72709351104", "text": "from flask import Flask\nimport feedparser\nimport csv\nimport json\nimport numpy as np\nimport pandas as pd\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.svm import SVC\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.metrics import classification_report, confusion_matrix\n\napp = Flask(__name__)\ncont =0\n@app.route(\"/knn\")\ndef main():\n\turl_test = 'test.csv'\n\turl_train = 'train.csv'\n\ttraining_data = []\n\ttest_data = []\n\tprint(\"datos totales\")\n\tprint(test_data)\n\tprint(\"----------------------------------\")\n\tdf_test = pd.read_csv(url_test)\n\tdf_train = pd.read_csv(url_train)\n\tprint('\\nEstadísticas del dataset:')\n\tprint(df_train.describe())\n\tprint(df_test.describe())\n\tprint('\\nMedia')\n\tprint(df_test.mean())\n\tprint(df_train.mean())\n\tprint(\"\\nPrecisión\")\n\tX = np.array(df_train.drop(['species'], 1))\n\ty = np.array(df_train['species'])\n\tprint(\"train\",np.array(df_train['species']))\n\tprint(X)\n\tprint(y)\n\tX_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.20)\n\tknn = KNeighborsClassifier(n_neighbors = 3)\n\tknn.fit(X_train, y_train)\n\tY_pred = knn.predict(X_test)\n\tprediccion_knn = knn.predict(df_test)\n\tprint(\"Aqui esta el df_test\", df_test.iloc[:,:-1].values)\n\tprint(\"prediccion del KNN\")\n\tprint(prediccion_knn)\n\tprint('Precisión Vecinos más Cercanos:')\n\tprint(knn.score(X_train, y_train))\n\tprint(\"Matriz de confusion\")\n\tprint(confusion_matrix(y_test, Y_pred))\n\tprint(\"Reporte de clasificacion\")\n\tprint(classification_report(y_test, Y_pred))\n\tjst = {str(i):prediccion_knn[i] for i in range(len(prediccion_knn))}\n\tjst[\"saludo\"] =\"hola\"\n\treturn jst #dict({\"informacion_knn\":[{i:prediccion_knn[i] for i in range(len(prediccion_knn))},{\"hola\":\"holin\"}]})\nif __name__ == \"__main__\":\n app.run(debug=True)", "repo_name": "SebastianTT/algoritmo_knn", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1852, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "flask.Flask", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 39, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 50, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 52, "usage_type": "call"}]}
+{"seq_id": "9166796731", "text": "from selenium import webdriver\n\n#driver = webdriver.Chrome(executable_path=r'C:\\Users\\Southville\\Downloads\\chromedriver_win32\\chromedriver')\n#driver = webdriver.Firefox(executable_path=r'C:\\Users\\Southville\\Downloads\\geckodriver-v0.27.0-win64\\geckodriver')\ndriver = webdriver.Ie(executable_path=r'C:\\Users\\Southville\\Downloads\\IEDriverServer_x64_3.150.1\\IEDriverServer')\ndriver.get('https://rahulshettyacademy.com/#/index') # http://206.189.237.25/\ndriver.maximize_window()\nprint(driver.title)\nprint(driver.current_url)\ndriver.get('https://rahulshettyacademy.com/#/index')\nprint(driver.title)\nprint(driver.current_url)\ndriver.back()\ndriver.refresh()\ndriver.minimize_window()\ndriver.close() # only the current window will close\n#driver.quit() # all the windows which are open will close simultanuously", "repo_name": "nadiamehmood/GitDemo", "sub_path": "PythonSelenium/PythonDemo.py", "file_name": "PythonDemo.py", "file_ext": "py", "file_size_in_byte": 813, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "selenium.webdriver.Ie", "line_number": 5, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 5, "usage_type": "name"}]}
+{"seq_id": "28663830837", "text": "\"\"\"Script for parsing emails. By default finds all ips (v4 and v6) and domains in letter,\r\nwrites it into db and prints it in console. Can also search headers by pattern or substring\r\n(start script with parameters -hs --header-string to search by string, and -hp --header-pattern\r\nto use regex pattern).\r\nWrites logs into main.log. It's possible to change level of logging (by adding -cl\r\n--change-level argument).\"\"\"\r\n\r\n\r\nimport re\r\nimport json\r\nimport sys\r\nimport argparse\r\nimport logging\r\nimport mysql.connector\r\n\r\nlogging.basicConfig(level=logging.DEBUG, format='%(asctime)s %(name)s %(levelname)s: %(message)s',\r\n filename='main.log')\r\nLOGGER = logging.getLogger()\r\nDB_CONF_FILE = 'db_conf.json'\r\nLEVELS = {'DEBUG': logging.DEBUG, 'INFO': logging.INFO, 'WARNING': logging.WARNING,\r\n 'ERROR': logging.ERROR, 'CRITICAL': logging.CRITICAL}\r\n\r\n\r\nclass DbConnection:\r\n \"\"\"Class, based on singleton pattern for creating database connection instance.\r\n Uses MySQL connector.\"\"\"\r\n\r\n _instance = None\r\n\r\n def __new__(cls, *ars, **kwars):\r\n \"\"\"Singleton pattern realization\"\"\"\r\n\r\n if not isinstance(cls._instance, DbConnection):\r\n cls._instance = object.__new__(cls)\r\n return cls._instance\r\n\r\n def __init__(self, host='127.0.0.1', port=3306, user='root', password='', data_base=''):\r\n \"\"\"Initializes a class and creates db connection instance.\r\n\r\n :param host: MySQL server host\r\n :param port: MySQL server port\r\n :param user: username of user with needed privileges\r\n :param password: user's password\r\n :param data_base: name of already created schema\r\n \"\"\"\r\n\r\n self.host = host\r\n self.port = port\r\n self.user = user\r\n\r\n if password != '':\r\n self.password = password\r\n else:\r\n logging.error('Password could not be an empty string. Check db configuration.')\r\n sys.exit(2)\r\n\r\n if data_base != '':\r\n self.data_base = data_base\r\n else:\r\n logging.error('You have to select database to connect to. Check db configuration.')\r\n sys.exit(2)\r\n\r\n try:\r\n self.conn = mysql.connector.connect(host=self.host, database=self.data_base,\r\n user=self.user, password=self.password,\r\n port=self.port)\r\n if self.conn.is_connected():\r\n logging.info(\"Database connected.\")\r\n self.recreate_tables()\r\n else:\r\n logging.info(\"Connection failed.\")\r\n self.conn = None\r\n except mysql.connector.Error:\r\n logging.info(\"Can not connect to database!\")\r\n self.conn = None\r\n\r\n def recreate_tables(self):\r\n \"\"\"Method for cleaning database. It drops old tables and creates new empty\r\n tables 'ips' and 'domains\"\"\"\r\n\r\n cursor = self.conn.cursor()\r\n\r\n try:\r\n cursor.execute(\"DROP TABLE ips\")\r\n logging.debug(\"Old table ips dropped\")\r\n except mysql.connector.errors.ProgrammingError:\r\n pass\r\n finally:\r\n cursor.execute(\"CREATE TABLE ips (id INT PRIMARY KEY AUTO_INCREMENT UNIQUE, ip \"\r\n \"VARCHAR(255) NOT NULL)\")\r\n logging.info(\"Table ips created\")\r\n try:\r\n cursor.execute(\"DROP TABLE domains\")\r\n logging.debug(\"Old table domains dropped\")\r\n except mysql.connector.errors.ProgrammingError:\r\n pass\r\n finally:\r\n cursor.execute(\r\n \"CREATE TABLE domains (id INT PRIMARY KEY AUTO_INCREMENT UNIQUE, \"\r\n \"domain VARCHAR(255) NOT NULL)\")\r\n logging.info(\"Table domains created\")\r\n\r\n\r\ndef start_db_connection():\r\n \"\"\"Function for loading db configuration json file and create DbConnection instance.\r\n :returns DbConnection instance\"\"\"\r\n\r\n try:\r\n db_conf = json.load(open(DB_CONF_FILE))\r\n logging.debug(\"Db configuration loaded\")\r\n except FileNotFoundError:\r\n logging.error('DB configuration file not found')\r\n sys.exit(2)\r\n db_instance = DbConnection(host=db_conf['host'], data_base=db_conf['db'],\r\n user=db_conf['user'], password=db_conf['password'],\r\n port=db_conf['port'])\r\n logging.debug(\"DbConnection instance created\")\r\n if not db_instance.conn:\r\n logging.error('DB connection failed. Check configuration file and db server status.')\r\n sys.exit(2)\r\n\r\n return db_instance\r\n\r\n\r\ndef open_mail(path_to_mail: str):\r\n \"\"\"\r\n Function to correctly open and read email file. If there are some problems with file\r\n (wrong extension, file not found) func logs message and exits\r\n\r\n :param path_to_mail: path to email file (str) with extension .eml\r\n :return: content of email (str)\r\n \"\"\"\r\n\r\n if not isinstance(path_to_mail, str):\r\n logging.error('Wrong datatype input. Path to mail have to be string')\r\n sys.exit(2)\r\n\r\n if path_to_mail.find('.eml') == -1:\r\n logging.error('File has wrong extension. You have to choose \\\".eml\\\" file.')\r\n sys.exit(2)\r\n\r\n try:\r\n with open(path_to_mail, 'r') as file:\r\n file_text = file.read()\r\n logging.debug(\"Email file opened\")\r\n except FileNotFoundError:\r\n logging.error('File not found. Check your path or choose another file. Entered path: %s'\r\n % path_to_mail)\r\n sys.exit(2)\r\n return file_text\r\n\r\n\r\ndef tld_config():\r\n \"\"\"Func loads TLD configuration file and makes a regex pattern from all domains from it\r\n :returns regex pattern of TLD\"\"\"\r\n\r\n try:\r\n tld_list = open('TLD.conf').read()\r\n except FileNotFoundError:\r\n logging.info('No TLD config found. Setting a default TLD configuration...')\r\n tld_list = 'com|org|net|int|edu|gov|mil|arpa'\r\n last_n = tld_list.rfind('\\n')\r\n if len(tld_list) - last_n == 1:\r\n tld_list = tld_list[:last_n]\r\n tld_list = tld_list.replace('\\n', '|')\r\n return tld_list\r\n\r\n\r\ndef parse_mail(email_content: str, db_conn: DbConnection):\r\n \"\"\"\r\n Function parses email text to find IPs (v4 and v6) and domains, and writes it into db.\r\n\r\n :param db_conn: MySQL db connection\r\n :param email_content: text content of email (str)\r\n \"\"\"\r\n\r\n if not isinstance(email_content, str):\r\n logging.debug('Email content variable should be a string.')\r\n logging.error('Email content has incorrect type')\r\n sys.exit(2)\r\n\r\n ip_v4 = re.findall(r\"(?\", methods=[\"GET\",\"POST\"]) #buton will get the id & post request \ndef update(id):\n form= TeamForm()\n team= Teams.query.filter_by(id=id).first()\n\n if request.method == 'GET':\n form.form_name.data = team.name # put data in the input box \n form.form_city.data = team.city\n form.form_conference.data = team.conference\n form.form_rank.data = team.rank\n\n elif request.method ==\"POST\": # id the form has been posted \n team.name = form.form_name.data # refere to model.py && send to the data base \n team.city = form.form_city.data\n team.conference = form.form_conference.data\n team.rank = form.form_rank.data\n db.session.commit()\n return redirect(url_for(\"home\")) \n return render_template(\"update.html\", form=form, title=\"Update Team\", team=team)\n \n@app.route(\"/delete/\")\ndef delete(id):\n team=Teams.query.filter_by(id=id).first()\n for player in team.players: \n db.session.delete(player)\n db.session.delete(team)\n db.session.commit()\n return redirect(url_for(\"home\"))\n\n\n@app.route(\"/player/\",methods=[\"GET\", \"POST\"]) \ndef player(id):\n form=PlayerForm()\n if request.method == \"POST\": \n if form.validate_on_submit():\n new_player = Players(\n pl_name = form.form_name.data, # filling in each column \n pl_position = form.form_position.data,\n teamid = id \n ) \n db.session.add(new_player) # add the new team to the route \n db.session.commit() # commit to the data base itself\n return redirect(url_for(\"home\")) #redirect back to the home page\n return render_template(\"add-Player.html\", title = \"Add Players to your Teams\", form=form) # display ftml file", "repo_name": "PhilipL1/project-bball", "sub_path": "application/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 3184, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "application.models.Teams.query.all", "line_number": 11, "usage_type": "call"}, {"api_name": "application.models.Teams.query", "line_number": 11, "usage_type": "attribute"}, {"api_name": "application.models.Teams", "line_number": 11, "usage_type": "name"}, {"api_name": "application.models.Players.query.all", "line_number": 12, "usage_type": "call"}, {"api_name": "application.models.Players.query", "line_number": 12, "usage_type": "attribute"}, {"api_name": "application.models.Players", "line_number": 12, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 14, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 8, "usage_type": "call"}, {"api_name": "application.app", "line_number": 8, "usage_type": "name"}, {"api_name": "application.app.route", "line_number": 9, "usage_type": "call"}, {"api_name": "application.app", "line_number": 9, "usage_type": "name"}, {"api_name": "application.forms.TeamForm", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 19, "usage_type": "name"}, {"api_name": "application.models.Teams", "line_number": 21, "usage_type": "call"}, {"api_name": "application.db.session.add", "line_number": 27, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 27, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 27, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 28, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 28, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 30, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 16, "usage_type": "call"}, {"api_name": "application.app", "line_number": 16, "usage_type": "name"}, {"api_name": "application.forms.TeamForm", "line_number": 34, "usage_type": "call"}, {"api_name": "application.models.Teams.query.filter_by", "line_number": 35, "usage_type": "call"}, {"api_name": "application.models.Teams.query", "line_number": 35, "usage_type": "attribute"}, {"api_name": "application.models.Teams", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 43, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 43, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 48, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 48, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 48, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 50, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 32, "usage_type": "call"}, {"api_name": "application.app", "line_number": 32, "usage_type": "name"}, {"api_name": "application.models.Teams.query.filter_by", "line_number": 54, "usage_type": "call"}, {"api_name": "application.models.Teams.query", "line_number": 54, "usage_type": "attribute"}, {"api_name": "application.models.Teams", "line_number": 54, "usage_type": "name"}, {"api_name": "application.db.session.delete", "line_number": 56, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 56, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 56, "usage_type": "name"}, {"api_name": "application.db.session.delete", "line_number": 57, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 57, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 57, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 58, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 58, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 59, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 52, "usage_type": "call"}, {"api_name": "application.app", "line_number": 52, "usage_type": "name"}, {"api_name": "application.forms.PlayerForm", "line_number": 64, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 65, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 65, "usage_type": "name"}, {"api_name": "application.models.Players", "line_number": 67, "usage_type": "call"}, {"api_name": "application.db.session.add", "line_number": 72, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 72, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 72, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 73, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 73, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 73, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 75, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 62, "usage_type": "call"}, {"api_name": "application.app", "line_number": 62, "usage_type": "name"}]}
+{"seq_id": "71075521342", "text": "# -- coding:utf-8--\nimport multiprocessing\nfrom urllib import request\nimport os\nimport shutil\nimport zipfile\nimport time\nimport datetime\nimport json\nimport sys\nimport socket\n\n\ndef rm():\n\tprint(\"one\")\n\ttime.sleep(2)\n\tprint(\"two\")\n\tos.system(r\"rm.py\")\n\ndef jc_update():\n\tjg = True\n\tupdate_url = \"http://lucyx.cn/zzz/update.json\"\n\trequest.urlretrieve(update_url, r\"update.json\")\n\twith open(\"update.json\") as zx_1:\n\t\tedition = json.load(zx_1)\n\tif edition == 2 :\n\t\tjg = False\n\telse:\n\t\tjg = True\n\treturn jg\n\nif __name__ == \"__main__\":\n\t#防止程序打包无限循环\n\tmultiprocessing.freeze_support()\n\tif jc_update():\n\t\tprint(\"yes\")\n\telse:\n\t\tprint(\"no\")\n\n\tp = multiprocessing.Process(target=rm)\n\t#运行脚本\n\tp.start()\n\t#主进程同时打开V2RAY\n\ts.remove(r\"hi.py\")", "repo_name": "ZiJie-Duan/lcv2", "sub_path": "src/test/hi.py", "file_name": "hi.py", "file_ext": "py", "file_size_in_byte": 764, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "time.sleep", "line_number": 16, "usage_type": "call"}, {"api_name": "os.system", "line_number": 18, "usage_type": "call"}, {"api_name": "urllib.request.urlretrieve", "line_number": 23, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 23, "usage_type": "name"}, {"api_name": "json.load", "line_number": 25, "usage_type": "call"}, {"api_name": "multiprocessing.freeze_support", "line_number": 34, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 40, "usage_type": "call"}]}
+{"seq_id": "23924056205", "text": "import argparse\nimport contextlib\nimport logging\nimport os\nimport string\nimport sys\nimport traceback\n\nfrom typing import Dict\nfrom typing import List\nfrom typing import Tuple\n\nfrom sampletester import convention\nfrom sampletester import environment_registry\nfrom sampletester import inputs\nfrom sampletester import runner\nfrom sampletester import summary\nfrom sampletester import testplan\nfrom sampletester import xunit\n\nVERSION = '0.16.3'\nEXITCODE_SUCCESS = 0\nEXITCODE_TEST_FAILURE = 1\nEXITCODE_FLAG_ERROR = 2\nEXITCODE_SETUP_ERROR = 3\nEXITCODE_USER_ABORT = 4\n\n# Set this to True to get a backtrace for debugging, or enable debug-level\n# logging from the command line.\nDEBUGME=False\n\ndef main():\n args, usage = parse_cli()\n if not args:\n exit(EXITCODE_SETUP_ERROR)\n\n if args.version:\n print(\"sampletester version {}\".format(VERSION))\n exit(EXITCODE_SUCCESS)\n\n log_level = LOG_LEVELS[args.logging]\n logging.getLogger().setLevel(log_level)\n logging.debug(\"argv: {}\".format(sys.argv))\n\n global DEBUGME\n DEBUGME = DEBUGME or (log_level == logging.DEBUG)\n\n try:\n indexed_docs = inputs.index_docs(*args.files)\n\n registry = environment_registry.new(args.convention, indexed_docs)\n test_suites = testplan.suites_from(indexed_docs, args.suites, args.cases)\n\n if len(test_suites) == 0:\n exit(EXITCODE_SUCCESS)\n manager = testplan.Manager(registry, test_suites, args.envs)\n\n except Exception as e:\n logging.error(f'fatal error: {repr(e)}')\n print(f'\\nERROR: could not run tests because {e}\\n')\n if DEBUGME:\n traceback.print_exc(file=sys.stdout)\n else:\n print(usage)\n exit(EXITCODE_SETUP_ERROR)\n\n verbosity = VERBOSITY_LEVELS[args.verbosity]\n quiet = verbosity == summary.Detail.NONE\n visitor = testplan.MultiVisitor(runner.Visitor(args.fail_fast),\n summary.SummaryVisitor(verbosity,\n not args.suppress_failures,\n debug=DEBUGME))\n try:\n success = manager.accept(visitor)\n except KeyboardInterrupt:\n print('\\nkeyboard interrupt; aborting')\n exit(EXITCODE_USER_ABORT)\n\n if not quiet or (not success and not args.suppress_failures):\n print()\n if success:\n print(\"Tests passed\")\n else:\n print(\"Tests failed\")\n\n if args.xunit:\n try:\n with smart_open(args.xunit) as xunit_output:\n xunit_output.write(manager.accept(xunit.Visitor()))\n if not quiet:\n print('xUnit output written to \"{}\"'.format(args.xunit))\n except Exception as e:\n print(\"could not write xunit output to {}: {}\".format(args.xunit, e))\n if DEBUGME:\n traceback.print_exc(file=sys.stdout)\n exit(EXITCODE_FLAG_ERROR)\n\n exit(EXITCODE_SUCCESS if success else EXITCODE_TEST_FAILURE)\n\n\nLOG_LEVELS = {\"none\": logging.CRITICAL, \"info\": logging.INFO, \"debug\": logging.DEBUG}\nDEFAULT_LOG_LEVEL = \"none\"\n\nVERBOSITY_LEVELS = {\"quiet\": summary.Detail.NONE, \"summary\": summary.Detail.BRIEF, \"detailed\": summary.Detail.FULL}\nDEFAULT_VERBOSITY_LEVEL = \"summary\"\n\ndef parse_cli():\n epilog = \"\"\"CONFIGS consists of any number of the following, in any order:\n\n TEST.yaml files: these are the test plans to execute against the CONVENTION\n\n arbitrary files/paths, depending on CONVENTION. For `tag` (the\n default), these should be paths to `MANIFEST.manifest.yaml` files.\n \"\"\"\n\n parser = argparse.ArgumentParser(\n description=\"A tool to run tests on equivalent samples in different languages\",\n epilog=epilog,\n formatter_class=argparse.RawDescriptionHelpFormatter)\n\n parser.add_argument(\n \"-c\",\n \"--convention\",\n metavar=\"CONVENTION:ARG,ARG,...\",\n help=('name of the convention to use in resolving artifact names in ' +\n 'specific languages, and a comma-separated list of arguments to ' +\n 'that convention (default: \"{}\")'.format(convention.DEFAULT)),\n default=convention.DEFAULT\n )\n\n parser.add_argument(\n \"--xunit\", metavar=\"FILE\", help=\"xunit output file (use `-` for stdout)\")\n\n parser.add_argument(\n \"-v\", \"--verbosity\",\n help=('how much output to show for passing tests (default: \"{}\")'\n .format(DEFAULT_VERBOSITY_LEVEL)),\n choices=list(VERBOSITY_LEVELS.keys()),\n default=\"summary\"\n )\n\n parser.add_argument(\n \"-f\", \"--suppress_failures\",\n help=\"suppress showing output for failing cases\",\n action='store_true')\n\n parser.add_argument(\n \"-l\",\n \"--logging\",\n help=('show logs at the specified level (default: \"{}\")'\n .format(DEFAULT_LOG_LEVEL)),\n choices=list(LOG_LEVELS.keys()),\n default=\"none\")\n\n parser.add_argument(\n \"--envs\",\n metavar=\"TESTENV_FILTER\",\n help=\"regex filtering test environments to execute\"\n )\n\n parser.add_argument(\n \"--suites\",\n metavar=\"SUITE_FILTER\",\n help=\"regex filtering test suites to execute\"\n )\n\n parser.add_argument(\n \"--cases\",\n metavar=\"CASE_FILTER\",\n help=\"regex filtering test cases to execute\"\n )\n\n parser.add_argument(\n \"--version\",\n help=\"print version number\",\n action=\"store_true\")\n\n parser.add_argument(\n \"--fail-fast\",\n help=(\"stop execution as soon as any test case fails, preempting \" +\n \"additional test cases/suites/environments from running\"),\n action=\"store_true\")\n\n parser.add_argument(\"files\", metavar=\"CONFIGS\", nargs=argparse.REMAINDER)\n return parser.parse_args(), parser.format_usage()\n\n\n# from https://stackoverflow.com/a/17603000\n@contextlib.contextmanager\ndef smart_open(filename: str=None):\n if filename and filename != \"-\":\n fh = open(filename, \"w\")\n else:\n fh = sys.stdout\n\n try:\n yield fh\n finally:\n if fh is not sys.stdout:\n fh.close()\n\n\nif __name__ == \"__main__\":\n main()\n", "repo_name": "googleapis/sample-tester", "sub_path": "sampletester/cli.py", "file_name": "cli.py", "file_ext": "py", "file_size_in_byte": 5847, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 10, "dataset": "github-code", "pt": "24", "api": [{"api_name": "logging.getLogger", "line_number": 42, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 43, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 43, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 46, "usage_type": "attribute"}, {"api_name": "sampletester.inputs.index_docs", "line_number": 49, "usage_type": "call"}, {"api_name": "sampletester.inputs", "line_number": 49, "usage_type": "name"}, {"api_name": "sampletester.environment_registry.new", "line_number": 51, "usage_type": "call"}, {"api_name": "sampletester.environment_registry", "line_number": 51, "usage_type": "name"}, {"api_name": "sampletester.testplan.suites_from", "line_number": 52, "usage_type": "call"}, {"api_name": "sampletester.testplan", "line_number": 52, "usage_type": "name"}, {"api_name": "sampletester.testplan.Manager", "line_number": 56, "usage_type": "call"}, {"api_name": "sampletester.testplan", "line_number": 56, "usage_type": "name"}, {"api_name": "logging.error", "line_number": 59, "usage_type": "call"}, {"api_name": "traceback.print_exc", "line_number": 62, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 62, "usage_type": "attribute"}, {"api_name": "sampletester.summary.Detail", "line_number": 68, "usage_type": "attribute"}, {"api_name": "sampletester.summary", "line_number": 68, "usage_type": "name"}, {"api_name": "sampletester.testplan.MultiVisitor", "line_number": 69, "usage_type": "call"}, {"api_name": "sampletester.testplan", "line_number": 69, "usage_type": "name"}, {"api_name": "sampletester.runner.Visitor", "line_number": 69, "usage_type": "call"}, {"api_name": "sampletester.runner", "line_number": 69, "usage_type": "name"}, {"api_name": "sampletester.summary.SummaryVisitor", "line_number": 70, "usage_type": "call"}, {"api_name": "sampletester.summary", "line_number": 70, "usage_type": "name"}, {"api_name": "sampletester.xunit.Visitor", "line_number": 89, "usage_type": "call"}, {"api_name": "sampletester.xunit", "line_number": 89, "usage_type": "name"}, {"api_name": "traceback.print_exc", "line_number": 95, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 95, "usage_type": "attribute"}, {"api_name": "logging.CRITICAL", "line_number": 101, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 101, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 101, "usage_type": "attribute"}, {"api_name": "sampletester.summary.Detail", "line_number": 104, "usage_type": "attribute"}, {"api_name": "sampletester.summary", "line_number": 104, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 116, "usage_type": "call"}, {"api_name": "argparse.RawDescriptionHelpFormatter", "line_number": 119, "usage_type": "attribute"}, {"api_name": "sampletester.convention.DEFAULT", "line_number": 127, "usage_type": "attribute"}, {"api_name": "sampletester.convention", "line_number": 127, "usage_type": "name"}, {"api_name": "sampletester.convention.DEFAULT", "line_number": 128, "usage_type": "attribute"}, {"api_name": "sampletester.convention", "line_number": 128, "usage_type": "name"}, {"api_name": "argparse.REMAINDER", "line_number": 184, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 194, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 199, "usage_type": "attribute"}, {"api_name": "contextlib.contextmanager", "line_number": 189, "usage_type": "attribute"}]}
+{"seq_id": "22684095149", "text": "from tornado.web import RequestHandler\n\nfrom handlers.http_handler.mongo_conn import MongoModel, JSONEncoder\n\n\nclass SewageHandler(RequestHandler):\n\n def get(self, *args, **kwargs):\n point_id = self.get_argument('PointID')\n count = self.get_argument('count')\n my_set = MongoModel().get_collection(name='datas')\n mn = point_id + '~32_2011'\n result_json = JSONEncoder().encode(my_set.find_one({'code': mn}))\n\n self.write('查到数据:{}'.format(result_json))\n", "repo_name": "MarnoonFeng/my_project", "sub_path": "src/handlers/http_handler/sewage_handler.py", "file_name": "sewage_handler.py", "file_ext": "py", "file_size_in_byte": 503, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "tornado.web.RequestHandler", "line_number": 6, "usage_type": "name"}, {"api_name": "handlers.http_handler.mongo_conn.MongoModel", "line_number": 11, "usage_type": "call"}, {"api_name": "handlers.http_handler.mongo_conn.JSONEncoder", "line_number": 13, "usage_type": "call"}]}
+{"seq_id": "40875865840", "text": "from keras.layers import Bidirectional, Concatenate, Permute, Dot, Input, LSTM, Multiply\nfrom keras.layers import RepeatVector, Dense, Activation, Lambda\nimport tensor as tf\n\nx = tf.Variable([[1.,2.,3.],[3.,4.,5.]])\ny = tf.Variable([[4.,5.,6.],[6.,7.,8.]])\nz = Dot(axes=1)([x,y])\n\ninit = tf.initialize_all_variables()\nsess = tf.Session()\nsess.run(init)\nsess.run(x)\nsess.run(y)\nsess.run(z)\n", "repo_name": "david-d-an/Chapter5Week3", "sub_path": "NeuralMachineTranslation/KerasNotes.py", "file_name": "KerasNotes.py", "file_ext": "py", "file_size_in_byte": 389, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "tensor.Variable", "line_number": 5, "usage_type": "call"}, {"api_name": "tensor.Variable", "line_number": 6, "usage_type": "call"}, {"api_name": "keras.layers.Dot", "line_number": 7, "usage_type": "call"}, {"api_name": "tensor.initialize_all_variables", "line_number": 9, "usage_type": "call"}, {"api_name": "tensor.Session", "line_number": 10, "usage_type": "call"}]}
+{"seq_id": "74346546945", "text": "from airflow import DAG\nfrom airflow.operators import LoadDimensionOperator\n\ndef load_dimension_table_dag(\n parent_dag_name,\n child_dag_name,\n args,\n tables_and_queries,\n redshift_conn_id):\n\n subdag = DAG(\n dag_id='{0}.{1}'.format(parent_dag_name, child_dag_name),\n default_args=args\n )\n with subdag:\n for table, query in tables_and_queries.items():\n load_dimension_table = LoadDimensionOperator(\n task_id=f'Load_{table}_dim_table',\n redshift_conn_id=redshift_conn_id,\n table={table},\n query=query,\n dag=subdag\n )\n\n return subdag", "repo_name": "bellabellahuang/Udacity-DEND", "sub_path": "data-pipeline/airflow/dags/subdag.py", "file_name": "subdag.py", "file_ext": "py", "file_size_in_byte": 674, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "airflow.DAG", "line_number": 11, "usage_type": "call"}, {"api_name": "airflow.operators.LoadDimensionOperator", "line_number": 17, "usage_type": "call"}]}
+{"seq_id": "71170619261", "text": "from uuid import uuid4\n\nfrom . import simple_storage\nfrom . import simple_network\nfrom . import resource\nfrom .templates.redfish_computer_system import REDFISH_TEMPLATE\nfrom api_emulator.exceptions import CreatePooledNodeError\nfrom api_emulator import utils\n\nfrom .processor import Processors, Processor\n\nfrom threading import Thread\nfrom time import sleep\n\nimport collections\n\nclass ResetWorker(Thread):\n def __init__(self, states, cs):\n super(ResetWorker, self).__init__()\n self.states = states\n self.cs = cs\n\n def run(self):\n self.cs.config['PowerState'] = self.states[0]\n sleep(10)\n self.cs.config['PowerState'] = self.states[1]\n\n\nclass ComputerSystem(object):\n \"\"\"\n Pooled node based on the ComputerSystem.1.00.0.ComputerSystem\n \"\"\"\n def __init__(self, config, cs_puid, rest_base, suffix):\n \"\"\"\n ComputerSystem Constructor\n\n Note: If OVF is set to True, then config must be of type \n \n \"\"\"\n self.suffix = suffix\n self.cs_puid = cs_puid\n self.rb = rest_base\n self.config = {}\n self.processor_count = 0\n self.total_memory_gb = 0\n\n self.SimpleNetwork = simple_network.EthernetNetworkInterfaceCollection(self.rb)\n self.EthernetInterfaces= simple_network.EthernetNetworkInterfaceCollection(self.rb)\n self.SimpleStorage = simple_storage.SimpleStorageCollection(self.rb, suffix)\n self.Processors = Processors(rest_base, suffix, cs_puid)\n\n self.SimpleNetwork.init_config(self.cs_puid, suffix)\n self.EthernetInterfaces.init_config(self.cs_puid,suffix)\n self.SimpleStorage.init_config(self.cs_puid)\n\n self.configure(config)\n\n @property\n def configuration(self):\n config = self.config.copy()\n\n if self.SimpleNetwork.members:\n config['Links']['SimpleNetwork'] = \\\n {'@odata.id': self.SimpleNetwork.odata_id}\n if self.EthernetInterfaces.members:\n config['Links']['EthernetInterfaces'] = \\\n {'@odata.id': self.EthernetInterfaces.odata_id}\n if self.SimpleStorage.members:\n config['Links']['SimpleStorage'] = \\\n {'@odata.id': self.SimpleStorage.odata_id}\n\n return self.config\n\n @property\n def storage_gb(self):\n \"\"\"\n Return the amount of storage the system has in GB\n \"\"\"\n return self.SimpleStorage.storage_gb\n\n @property\n def network_ports(self):\n \"\"\"\n Return the number of VLAN network ports in the system\n \"\"\"\n return self.SimpleNetwork.port_count\n\n def configure(self, config):\n \"\"\"\n Overridden configure() method\n \"\"\"\n self._base_configure()\n self.config['Processors']['Count'] = int(config['NumberOfProcessors'])\n self.config['Links']['Processors'] = {'@odata.id': self.Processors.odata_id}\n self.processor_count = int(config['NumberOfProcessors'])\n self._create_processors(self.processor_count)\n\n self.config['Memory']['TotalSystemMemoryGB'] = int(config['TotalSystemMemoryGB'])\n self.total_memory_gb = self.config['Memory']['TotalSystemMemoryGB']\n\n if 'Boot' in config:\n self.config['Boot'] = config['Boot']\n\n self.add_network_ports(int(config['NumberOfNetworkPorts']))\n self.add_storage(config['Devices'])\n\n def reboot(self, config):\n state = config['PowerState']\n\n if state == 'On':\n states = [ 'On', 'Off' ]\n elif state == 'Off':\n states = [ 'Off', 'On' ]\n else:\n raise CreatePooledNodeError('Incorrect PowerState value.')\n\n ResetWorker(states, self).start()\n\n def _replace_config(self, config, change):\n for key, value in change.iteritems():\n if isinstance(value, collections.Mapping):\n ret = self._replace_config(config.get(key, {}), value)\n config[key] = ret\n else:\n config[key] = change[key]\n return config\n\n def update_config(self,change):\n self.config = self._replace_config(self.config, change)\n\n def add_network_ports(self, amount):\n \"\"\"\n Add network ports to the pooled node\n\n Arguments:\n amount - Number of ports to add\n \"\"\"\n status = resource.Status(resource.StateEnum.ENABLED, resource.HealthEnum.OK)\n network_objs = []\n for i in range(amount):\n config = simple_network.EthernetNetworkInterface.create_config(\n self.rb, self.suffix, self.cs_puid, i + 1, status, None, 100)\n eth = simple_network.EthernetNetworkInterface(config, None)\n network_objs.append(eth)\n self.SimpleNetwork.add_network_objects(network_objs)\n self.EthernetInterfaces.add_network_objects(network_objs)\n\n def add_storage(self, device_list):\n \"\"\"\n Add the given devices to the storage for the pooled node.\n\n NOTE: The device_list should be a list of dictionaries that can be\n turned into SimpleStorage objects\n \"\"\"\n devices = []\n\n # Creating a SimpleStorage object\n for dev in device_list:\n status = resource.Status(\n state=resource.StateEnum.ENABLED,\n health=resource.HealthEnum.OK)\n device = simple_storage.Device(\n name=dev['Name'],\n manufacturer='N/A',\n model='N/A',\n size=dev['Size'],\n status=status,\n oem={})\n devices.append(device)\n\n # Creating SimpleStorage objects\n ss = simple_storage.SimpleStorage()\n ss.init_config(\n item_id=1,\n parent_cs_puid=self.cs_puid,\n status=resource.Status(\n resource.StateEnum.ENABLED,\n resource.HealthEnum.OK,\n resource.HealthEnum.OK),\n devices=devices,\n rest_base=self.rb,\n system_suffix=self.suffix)\n storage_objects = [ss]\n self.SimpleStorage.append(storage_objects)\n\n def _create_processors(self, count):\n \"\"\"\n Populates the Processors attribute\n \"\"\"\n status = resource.Status(resource.StateEnum.ENABLED,\n resource.HealthEnum.OK)\n for i in range(count):\n p = Processor(self.rb, self.suffix, self.cs_puid,\n i + 1, i + 1, 3700, 8, 4, 4, 2, status)\n self.Processors.add_processor(p)\n\n def _base_configure(self):\n \"\"\"\n Private method to do a base configuration of the pooled node\n \"\"\"\n try:\n self.config = REDFISH_TEMPLATE.copy()\n self.odata_id = self.config['@odata.id'].format(\n rest_base=self.rb, cs_puid=self.cs_puid)\n self.config['@odata.context'] = self.config['@odata.context'].format(\n rest_base=self.rb, cs_puid=self.cs_puid)\n self.config['@odata.id'] = self.odata_id\n self.config['Actions']['#ComputerSystem.Reset']['target'] = \\\n self.config['Actions']['#ComputerSystem.Reset']['target'].format(\n cs_puid=self.cs_puid)\n self.config['Id'] = self.cs_puid\n\n except KeyError as e:\n raise CreatePooledNodeError(\n 'Incorrect configuration, missing key: ' + e.message)\n", "repo_name": "facebook/openbmc", "sub_path": "tools/Redfish-Interface-Emulator/api_emulator/redfish/computer_system.py", "file_name": "computer_system.py", "file_ext": "py", "file_size_in_byte": 7479, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 601, "dataset": "github-code", "pt": "24", "api": [{"api_name": "threading.Thread", "line_number": 17, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 25, "usage_type": "call"}, {"api_name": "processor.Processors", "line_number": 50, "usage_type": "call"}, {"api_name": "api_emulator.exceptions.CreatePooledNodeError", "line_number": 115, "usage_type": "call"}, {"api_name": "collections.Mapping", "line_number": 121, "usage_type": "attribute"}, {"api_name": "processor.Processor", "line_number": 193, "usage_type": "call"}, {"api_name": "templates.redfish_computer_system.REDFISH_TEMPLATE.copy", "line_number": 202, "usage_type": "call"}, {"api_name": "templates.redfish_computer_system.REDFISH_TEMPLATE", "line_number": 202, "usage_type": "name"}, {"api_name": "api_emulator.exceptions.CreatePooledNodeError", "line_number": 214, "usage_type": "call"}]}
+{"seq_id": "15445848706", "text": "# -*- coding: utf-8 -*-\n\nfrom odoo import models, fields, api\nimport json\nfrom odoo.addons import decimal_precision as dp\n\nclass Product(models.Model):\n _inherit = 'product.product'\n attribute_value_ids_json = fields.Char()\n attrs_jsonb_show = fields.Char(compute='_attrs_jsonb_show')\n pname = fields.Char(related='name', store=True)\n lst_price = fields.Float(\n 'Sale Price', compute='_compute_product_lst_price', store=True,\n digits=dp.get_precision('Product Price'), inverse='_set_product_lst_price',\n help=\"The sale price is managed from the product template. Click on the 'Configure Variants' button to set the extra attribute prices.\")\n # attrs_json = fields.Json()\n\n def _attrs_jsonb_show(self):\n for r in self:\n query = '''SELECT attrs_jsonb from product_product\n WHERE id = %s'''%(r.id)\n self.env.cr.execute(query)\n rs = self.env.cr.dictfetchall()\n if rs:\n rs = rs[0]['attrs_jsonb']\n r. attrs_jsonb_show = rs\n\n def thay_doi_attribute_value_ids(self,obj, vals):\n if 'attribute_value_ids' in vals:\n for r in obj:\n adict = {}\n for attr in r.attribute_value_ids:\n attribute_id = attr.attribute_id.name\n adict[attribute_id] = attr.id\n # r.attribute_value_ids_json = adict\n query = '''UPDATE product_product\n SET attrs_jsonb = '%s'::jsonb\n WHERE id = %s'''%(json.dumps(adict), r.id)\n self.env.cr.execute(query)\n # rs = self.env.cr.fetchall()\n # print ('***rs***',rs)\n\n\n @api.model\n def create(self,vals):\n # self.clear_caches()\n rs = super(Product, self.sudo()).create(vals)\n self.thay_doi_attribute_value_ids(rs, vals)\n return rs\n\n \n @api.multi\n def write(self, vals):\n # self.clear_caches()\n rs = super(Product, self.sudo()).write(vals)\n self.thay_doi_attribute_value_ids(self, vals)\n return rs", "repo_name": "tu95ctv/tglteam", "sub_path": "attrs_json/models/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 2124, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "odoo.models.Model", "line_number": 7, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 7, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 9, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 9, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 10, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 10, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 11, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 11, "usage_type": "name"}, {"api_name": "odoo.fields.Float", "line_number": 12, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 12, "usage_type": "name"}, {"api_name": "odoo.addons.decimal_precision.get_precision", "line_number": 14, "usage_type": "call"}, {"api_name": "odoo.addons.decimal_precision", "line_number": 14, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 38, "usage_type": "call"}, {"api_name": "odoo.api.model", "line_number": 44, "usage_type": "attribute"}, {"api_name": "odoo.api", "line_number": 44, "usage_type": "name"}, {"api_name": "odoo.api.multi", "line_number": 52, "usage_type": "attribute"}, {"api_name": "odoo.api", "line_number": 52, "usage_type": "name"}]}
+{"seq_id": "40409449554", "text": "import streamlit as st\nimport pandas as pd\nimport numpy as np\nimport viz_utility as vutil\nimport altair as alt\n\nst.title('MH URS Parsinator - Data Review')\nst.header('Single-State Data Explorer')\n\nst.text('To view data, select a state.')\nst.text('Data is currently only available for NOMS domain, but ACCESS tab is visible for testing.')\n\n#Initialize\nstate_select = 'Select State'\n\n@st.cache_data\ndef load_data():\n data = vutil.viz_data_get()\n return data\n\n@st.cache_data\ndef load_states():\n data = vutil.url_data_get()\n return data\n\nstates = load_states()\n\nstate_select = st.selectbox(\n \"Which state's data would you like to review?\",\n ('Select State',)+tuple(states['state_name'].unique()))\n\ndata = load_data()\n\ndomains = [\"NOMS\", \"ACCESS\"]\n\nif state_select != 'Select State':\n data = data[data.state_name == state_select]\n \n tab1, tab2= st.tabs(domains)\n tabs = [tab1,tab2]\n\n for i in range(0,len(domains),1):\n domain_data = data[data.domain == domains[i]]\n\n with tabs[i]:\n for table in domain_data['table_name'].unique().tolist(): \n table_data = domain_data[domain_data.table_name == table]\n with st.expander(table):\n for metric in table_data['metric_name'].unique().tolist():\n st.header(metric)\n ch = alt.Chart(table_data[table_data['metric_name']==metric]).mark_line().encode(x='year',y='metric_result',text='state_name')\n st.altair_chart(ch, use_container_width=True)\n st.divider()\n\n\n\n", "repo_name": "PoofyOddish/mh-urs-parsinator", "sub_path": "pages/2_Single_State_Data_Explorer.py", "file_name": "2_Single_State_Data_Explorer.py", "file_ext": "py", "file_size_in_byte": 1599, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "26", "api": [{"api_name": "streamlit.title", "line_number": 7, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 8, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 10, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 11, "usage_type": "call"}, {"api_name": "viz_utility.viz_data_get", "line_number": 18, "usage_type": "call"}, {"api_name": "streamlit.cache_data", "line_number": 16, "usage_type": "attribute"}, {"api_name": "viz_utility.url_data_get", "line_number": 23, "usage_type": "call"}, {"api_name": "streamlit.cache_data", "line_number": 21, "usage_type": "attribute"}, {"api_name": "streamlit.selectbox", "line_number": 28, "usage_type": "call"}, {"api_name": "streamlit.tabs", "line_number": 39, "usage_type": "call"}, {"api_name": "streamlit.expander", "line_number": 48, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 50, "usage_type": "call"}, {"api_name": "altair.Chart", "line_number": 51, "usage_type": "call"}, {"api_name": "streamlit.altair_chart", "line_number": 52, "usage_type": "call"}, {"api_name": "streamlit.divider", "line_number": 53, "usage_type": "call"}]}
+{"seq_id": "17200223586", "text": "from django.db import models\nfrom inventario.models import Semilla\nfrom usuario.models import Agricultor\n# Create your models here.\n\n\nclass EntregaAAgricultor(models.Model):\n FechaEntrega = models.DateField(\n verbose_name=\"Fecha de movimiento\",\n auto_now=True\n )\n HectareasAPlantar = models.PositiveIntegerField(verbose_name=\"Hectareas a plantar\")\n SemillasDadas = models.PositiveIntegerField(\n verbose_name=\"KG de semillas dadas\",\n default=1,\n )\n TipoSemilla = models.ForeignKey(\n Semilla,\n verbose_name=\"Tipo de semillas que se darán\",\n null=True,\n blank=True,\n on_delete=models.CASCADE\n )\n NombreAgricultor = models.ForeignKey(\n Agricultor,\n default=2,\n verbose_name=\"Agricultor al que se le darán las semillas\",\n null=True,\n blank=True,\n on_delete=models.CASCADE\n )\n\n def __str__(self):\n return str(self.SemillasDadas) + \" KG de \" + str(self.TipoSemilla) + \" - \" + str(self.NombreAgricultor)\n\n class Meta:\n ordering = [\"-FechaEntrega\"]\n verbose_name_plural = \"Entregas a agricultores\"\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "repo_name": "miguelzetina/sistema", "sub_path": "movimiento/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1164, "program_lang": "python", "lang": "es", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.db.models.Model", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 17, "usage_type": "call"}, {"api_name": "inventario.models.Semilla", "line_number": 18, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 24, "usage_type": "call"}, {"api_name": "usuario.models.Agricultor", "line_number": 25, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 30, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}]}
+{"seq_id": "37488531762", "text": "#!/usr/local/bin/python3\n# -*- coding: utf-8 -*-\n\n\"\"\"\n@File : ui_stdf.py\n@Author : Link\n@Time : 2022/5/2 10:44\n@Mark : \n\"\"\"\nimport math\nimport sys\nimport datetime as dt\nfrom typing import Union, List\nfrom pydoc import help\n\nimport numpy as np\nimport pandas as pd\nfrom PySide2.QtCore import Slot, QTimer, Qt, Signal\nfrom PySide2.QtGui import QCloseEvent\nfrom PySide2.QtWidgets import QMainWindow, QApplication, QWidget, QMessageBox\n\nfrom pyqtgraph.dockarea import *\n\nfrom chart_core.chart_pyqtgraph.core.mixin import ChartType\nfrom chart_core.chart_pyqtgraph.poll import ChartDockWindow\nfrom common.li import Li, SummaryCore\nfrom ui_component.ui_analysis_stdf.ui_components.ui_data_group import DataGroupWidget\nfrom ui_component.ui_common.ui_console import ConsoleWidget\nfrom ui_component.ui_analysis_stdf.ui_designer.ui_home_load import Ui_MainWindow\nfrom ui_component.ui_common.ui_utils import QTableUtils\nfrom ui_component.ui_analysis_stdf.ui_components.ui_file_load_widget import FileLoadWidget # 文件选取\nfrom ui_component.ui_analysis_stdf.ui_components.ui_tree_load_widget import TreeLoadWidget # 载入的数据选取\nfrom ui_component.ui_analysis_stdf.ui_components.ui_table_load_widget import TableLoadWidget # 测试项选取\n\n\nclass StdfLoadUi(QMainWindow, Ui_MainWindow):\n \"\"\"\n 懒加载界面\n \"\"\"\n closeSignal = Signal(int)\n parent = None\n\n def __init__(self, parent=None, space_nm=1, path_select=False, select=True, is_web=False):\n super(StdfLoadUi, self).__init__(parent)\n self.setupUi(self)\n self.space_nm = space_nm\n self.li = Li()\n self.summary = SummaryCore()\n self.title = \"STDF数据载入空间: {}\".format(space_nm)\n self.setWindowTitle(self.title)\n\n self.area = DockArea()\n self.setCentralWidget(self.area)\n \" FileLoadWidget用来载入数据文件,内部用到多线程 \"\n self.stdf_select_widget = FileLoadWidget(self.summary, self, space_nm=space_nm)\n self.dock_stdf_load = Dock(\"STDF File Select\", size=(400, 100))\n self.dock_stdf_load.addWidget(self.stdf_select_widget)\n self.area.addDock(self.dock_stdf_load)\n\n \" TreeLoadWidget用来载入和配对内部数据,会占用到大量的IO资源 \"\n self.tree_load_widget = TreeLoadWidget(self.li, self.summary, self)\n dock_tree_load = Dock(\"Data Tree Select & Config\", size=(200, 300))\n dock_tree_load.addWidget(self.tree_load_widget)\n self.area.addDock(dock_tree_load, \"bottom\", self.dock_stdf_load)\n # self.area.moveDock(self.dock_stdf_load, 'above', dock_tree_load)\n\n \" TableLoadWidget用来载入和配对内部数据,会占用到大量的IO资源, 并且是作为主要的分析界面 \"\n self.table_load_widget = TableLoadWidget(self.li, self.summary, self)\n dock_table_load = Dock(\"Data TEST NO&ITEM Analysis\", size=(200, 300))\n dock_table_load.addWidget(self.table_load_widget)\n self.area.addDock(dock_table_load, \"right\", dock_tree_load)\n # self.area.moveDock(dock_tree_load, 'above', dock_table_load)\n\n \" DataGroupWidget用来对数据进行简单的分组和筛选 \"\n self.group_table_widget = DataGroupWidget(self.li)\n dock_group_load = Dock(\"Data Group\", size=(50, 300))\n dock_group_load.addWidget(self.group_table_widget)\n self.area.addDock(dock_group_load)\n\n text = \"\"\"\n 载入的功能包: np(numpy), pd(pandas), math\n 载入的数据:\n 点击RUN后会被刷新 \n li: 数据空间的集合数据, 通过help(li)查看\n RUN.\n \"\"\"\n self.namespace = {\n \"np\": np,\n \"pd\": pd,\n \"math\": math,\n \"help\": help,\n \"li\": self.li,\n \"summary\": self.summary\n }\n self.console = ConsoleWidget(parent=self, namespace=self.namespace, text=text)\n self.dock_console_load = Dock(\"Python Console\", size=(400, 100))\n self.dock_console_load.addWidget(self.console)\n self.area.addDock(self.dock_console_load, \"bottom\")\n # self.area.moveDock(self.dock_stdf_load, 'above', dock_console_load)\n \"------------------------------------------------------------------------------\"\n self.area.restoreState(\n {\n 'main': (\n 'horizontal',\n [\n ('dock', 'Data Group', {}),\n ('vertical',\n [('horizontal', [\n (\n 'vertical',\n [\n (\n 'dock',\n 'STDF File Select',\n {}\n ),\n ('dock',\n 'Data Tree Select & Config',\n {}\n )\n ],\n {\n 'sizes': [147, 198]\n }\n ),\n (\n 'dock',\n 'Python Console',\n {})],\n {'sizes': [737, 429]}),\n ('dock',\n 'Data TEST NO&ITEM Analysis',\n {})],\n {'sizes': [350, 439]})],\n {'sizes': [372, 1171]}), 'float': []})\n self.dock_console_load.hide() # 可以显示和隐藏\n self.init_signal()\n\n \" 用来存放chart的 \"\n self.chart_ui = ChartDockWindow(self.li, None, icon=None, space_nm=space_nm) # type:ChartDockWindow\n\n if select and is_web is False:\n if path_select:\n self.stdf_select_widget.first_directory_select()\n else:\n self.stdf_select_widget.first_select()\n\n def init_signal(self):\n self.stdf_select_widget.finished.connect(self.tree_load_widget.set_tree)\n self.li.QCalculation.connect(self.q_calculation)\n self.li.QMessage.connect(self.message_show)\n self.li.QStatusMessage.connect(self.mdi_space_message_emit)\n\n def q_calculation(self):\n self.table_load_widget.cal_table()\n self.group_table_widget.checkbox_changed()\n\n @Slot(SummaryCore)\n def merge_data_emit(self, data: SummaryCore):\n self.summary = data\n self.tree_load_widget.set_tree()\n\n @Slot()\n def on_action_dock_structure_triggered(self):\n \"\"\" 将 area 布局保存在泡菜中 \"\"\"\n state = self.area.saveState()\n print(state)\n\n @Slot()\n def on_action_sava_data_triggered(self):\n \"\"\" 将数据保存在csv文件中 \"\"\"\n\n @Slot()\n def on_action_console_triggered(self):\n if self.action_console.isChecked():\n self.dock_console_load.show()\n else:\n self.dock_console_load.hide()\n\n @Slot()\n def on_action_limit_triggered(self):\n self.li.show_limit_diff()\n\n @Slot(str)\n def mdi_space_message_emit(self, message: str):\n \"\"\"\n append message\n :param message:\n :return:\n \"\"\"\n self.statusbar.showMessage(\"==={}==={}===\".format(dt.datetime.now().strftime(\"%H:%M:%S\"), message))\n\n def message_show(self, text: str) -> bool:\n res = QMessageBox.question(self, '待确认', text,\n QMessageBox.Yes | QMessageBox.No,\n QMessageBox.Yes)\n if res == QMessageBox.Yes:\n return True\n else:\n return False\n\n def get_test_id_column(self) -> Union[List[int], None]:\n \"\"\"\n\n :return:\n \"\"\"\n return QTableUtils.get_table_widget_test_id(self.table_load_widget.cpk_info_table)\n\n def get_text_column(self) -> Union[List[str], None]:\n \"\"\"\n\n :return:\n \"\"\"\n test_id_column = self.get_test_id_column()\n if not test_id_column:\n return None\n text_column = []\n for each in test_id_column:\n table_row = self.li.capability_key_dict[each]\n text = str(table_row[\"TEST_NUM\"]) + \":\" + table_row[\"TEST_TXT\"]\n text_column.append(text)\n return text_column\n\n @Slot()\n def on_action_qt_distribution_trans_triggered(self):\n \"\"\" 使用PYQT来拉出横向柱状分布图 \"\"\"\n test_id_column: List[int] = self.get_test_id_column()\n self.chart_ui.add_chart_dock(test_id_column, ChartType.TransBar)\n self.chart_ui.show()\n self.chart_ui.raise_()\n\n @Slot()\n def on_action_qt_scatter_triggered(self):\n \"\"\" 使用PYQT来拉出线性散点图 \"\"\"\n test_id_column: List[int] = self.get_test_id_column()\n self.chart_ui.add_chart_dock(test_id_column, ChartType.TransScatter)\n self.chart_ui.show()\n self.chart_ui.raise_()\n\n @Slot()\n def on_action_qt_mapping_triggered(self):\n \"\"\" 使用PYQT来拉出Mapping图 \"\"\"\n pass\n\n @Slot()\n def on_action_qt_visual_map_triggered(self):\n \"\"\" 使用PYQT来拉出Visual Map图 \"\"\"\n test_id_column = self.get_test_id_column()\n self.chart_ui.add_chart_dock(test_id_column, ChartType.VisualMap)\n self.chart_ui.show()\n self.chart_ui.raise_()\n\n def closeEvent(self, a0: QCloseEvent) -> None:\n \"\"\"\n 删除mdi时, 需要将与其对应的chart也删除\n \"\"\"\n self.closeSignal.emit(self.space_nm)\n return super(StdfLoadUi, self).closeEvent(a0)\n", "repo_name": "Crazytommy90/ATE_STDF_ANALYSIS", "sub_path": "ui_component/ui_analysis_stdf/ui_stdf.py", "file_name": "ui_stdf.py", "file_ext": "py", "file_size_in_byte": 9798, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 2, "dataset": "github-code", "pt": "24", "api": [{"api_name": "PySide2.QtWidgets.QMainWindow", "line_number": 36, "usage_type": "name"}, {"api_name": "ui_component.ui_analysis_stdf.ui_designer.ui_home_load.Ui_MainWindow", "line_number": 36, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Signal", "line_number": 40, "usage_type": "call"}, {"api_name": "common.li.Li", "line_number": 47, "usage_type": "call"}, {"api_name": "common.li.SummaryCore", "line_number": 48, "usage_type": "call"}, {"api_name": "ui_component.ui_analysis_stdf.ui_components.ui_file_load_widget.FileLoadWidget", "line_number": 55, "usage_type": "call"}, {"api_name": "ui_component.ui_analysis_stdf.ui_components.ui_tree_load_widget.TreeLoadWidget", "line_number": 61, "usage_type": "call"}, {"api_name": "ui_component.ui_analysis_stdf.ui_components.ui_table_load_widget.TableLoadWidget", "line_number": 68, "usage_type": "call"}, {"api_name": "ui_component.ui_analysis_stdf.ui_components.ui_data_group.DataGroupWidget", "line_number": 75, "usage_type": "call"}, {"api_name": "pydoc.help", "line_number": 91, "usage_type": "name"}, {"api_name": "ui_component.ui_common.ui_console.ConsoleWidget", "line_number": 95, "usage_type": "call"}, {"api_name": "chart_core.chart_pyqtgraph.poll.ChartDockWindow", "line_number": 140, "usage_type": "call"}, {"api_name": "common.li.SummaryCore", "line_number": 159, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 158, "usage_type": "call"}, {"api_name": "common.li.SummaryCore", "line_number": 158, "usage_type": "argument"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 163, "usage_type": "call"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 169, "usage_type": "call"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 173, "usage_type": "call"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 180, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 191, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 191, "usage_type": "attribute"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 184, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets.QMessageBox.question", "line_number": 194, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets.QMessageBox", "line_number": 194, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QMessageBox.Yes", "line_number": 195, "usage_type": "attribute"}, {"api_name": "PySide2.QtWidgets.QMessageBox", "line_number": 195, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QMessageBox.No", "line_number": 195, "usage_type": "attribute"}, {"api_name": "PySide2.QtWidgets.QMessageBox.Yes", "line_number": 196, "usage_type": "attribute"}, {"api_name": "PySide2.QtWidgets.QMessageBox", "line_number": 196, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QMessageBox.Yes", "line_number": 197, "usage_type": "attribute"}, {"api_name": "PySide2.QtWidgets.QMessageBox", "line_number": 197, "usage_type": "name"}, {"api_name": "ui_component.ui_common.ui_utils.QTableUtils.get_table_widget_test_id", "line_number": 207, "usage_type": "call"}, {"api_name": "ui_component.ui_common.ui_utils.QTableUtils", "line_number": 207, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 202, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 202, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 209, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 209, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 227, "usage_type": "name"}, {"api_name": "chart_core.chart_pyqtgraph.core.mixin.ChartType.TransBar", "line_number": 228, "usage_type": "attribute"}, {"api_name": "chart_core.chart_pyqtgraph.core.mixin.ChartType", "line_number": 228, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 224, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 235, "usage_type": "name"}, {"api_name": "chart_core.chart_pyqtgraph.core.mixin.ChartType.TransScatter", "line_number": 236, "usage_type": "attribute"}, {"api_name": "chart_core.chart_pyqtgraph.core.mixin.ChartType", "line_number": 236, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 232, "usage_type": "call"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 240, "usage_type": "call"}, {"api_name": "chart_core.chart_pyqtgraph.core.mixin.ChartType.VisualMap", "line_number": 249, "usage_type": "attribute"}, {"api_name": "chart_core.chart_pyqtgraph.core.mixin.ChartType", "line_number": 249, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 245, "usage_type": "call"}, {"api_name": "PySide2.QtGui.QCloseEvent", "line_number": 253, "usage_type": "name"}]}
+{"seq_id": "71521177342", "text": "import collections\nimport itertools\n\nimport numpy as np\n\nfrom ... import draw, mesh\nfrom ...draw import internal\n\nLightInfo = collections.namedtuple(\n 'LightInfo', ['normal', 'magnitude'])\n\n@internal.ShapeDecorator\nclass Disks(draw.Disks):\n __doc__ = draw.Disks.__doc__\n\n def render(self, rotation=(1, 0, 0, 0), name_suffix='', illo_id='illo',\n **kwargs):\n # in the zdog coordinate system, x is to the right, y is down,\n # and z is toward you\n lines = []\n\n particles = zip(*mesh.unfoldProperties([\n self.positions*(1, -1), self.diameters, self.colors*255]))\n for i, (position, (diameter,), color) in enumerate(particles):\n group_index = 'disk_{}_{}'.format(name_suffix, i)\n\n (r, g, b) = map(int, color[:3])\n color_str = '\"rgba({}, {}, {}, {})\"'.format(\n r, g, b, color[3]/255)\n\n lines.append(\"\"\"\n new Zdog.Shape({{\n addTo: {illo_id},\n translate: {{x: {pos[0]}, y: {pos[1]}}},\n stroke: {diameter},\n color: {color},\n }});\"\"\".format(\n group_index=group_index, illo_id=illo_id, pos=position,\n diameter=diameter, color=color_str))\n\n return lines\n", "repo_name": "glotzerlab/plato", "sub_path": "plato/draw/zdog/Disks.py", "file_name": "Disks.py", "file_ext": "py", "file_size_in_byte": 1288, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 11, "dataset": "github-code", "pt": "24", "api": [{"api_name": "collections.namedtuple", "line_number": 9, "usage_type": "call"}, {"api_name": "draw.Disks", "line_number": 13, "usage_type": "attribute"}, {"api_name": "draw.Disks", "line_number": 14, "usage_type": "attribute"}, {"api_name": "draw.internal.ShapeDecorator", "line_number": 12, "usage_type": "attribute"}, {"api_name": "draw.internal", "line_number": 12, "usage_type": "name"}]}
+{"seq_id": "43150134416", "text": "import asyncio\nimport random\n\nimport discord\nimport json\n\nfrom discord.ext import commands\n\n\nclass Shop(commands.Cog):\n def __init__(self, bot: commands.Bot):\n self.bot = bot\n self.config = json.load(open(\"config.json\", \"r\"))\n self.colors = [0xFFE4E1, 0x00FF7F, 0xD8BFD8, 0xDC143C, 0xFF4500, 0xDEB887, 0xADFF2F, 0x800000, 0x4682B4, 0x006400, 0x808080, 0xA0522D, 0xF08080, 0xC71585, 0xFFB6C1, 0x00CED1]\n self.cache = {}\n self.messages = {}\n self.embeds = {}\n\n @commands.command()\n @commands.has_permissions(administrator=True)\n async def new(self, ctx: commands.Context):\n questions = [\n \"What will be the product name ?\",\n \"What will be the product prize ?\",\n \"Please send the product image here.\",\n \"What will be the destination channel ?\"\n ]\n answers = []\n\n def check(m):\n return m.author == ctx.author and m.channel == ctx.channel\n i = 1\n for q in questions:\n xx = await ctx.send(embed=discord.Embed(title=f\"Question {i}\", color=random.choice(self.colors), description=q))\n msg = await self.bot.wait_for(\"message\", check=check)\n answers.append(msg.content)\n i+=1\n await xx.delete()\n await msg.delete()\n try:\n int(answers[3][2:-1])\n float(answers[1])\n except ValueError:\n return await ctx.send(\"The channel you mentionned or the prize are not correct.\")\n\n embed = discord.Embed(title=answers[0], color=random.choice(self.colors), description=f\"{answers[1]}€\\nStock!\")\n embed.set_image(url=answers[2])\n channel = self.bot.get_channel(int(answers[3][2:-1]))\n x = await channel.send(embed=embed)\n await x.add_reaction(\"🛒\")\n await x.add_reaction(\"❌\")\n\n products = json.load(open(\"products.json\", \"r\"))\n products[str(x.id)] = {\n \"name\": answers[0],\n \"prize\": float(answers[1])\n }\n json.dump(products, open(\"products.json\", \"w\"), indent=4)\n\n @commands.Cog.listener()\n async def on_raw_reaction_add(self, payload):\n member = payload.member\n user = self.bot.get_user(payload.user_id)\n emoji = payload.emoji\n message_id = payload.message_id\n if self.bot.user in [member, user] : return\n is_dm = isinstance(self.bot.get_channel(payload.channel_id), discord.DMChannel)\n products = json.load(open(\"products.json\", \"r\"))\n if str(message_id) in products.keys():\n channel = self.bot.get_channel(payload.channel_id)\n message = await channel.fetch_message(message_id)\n await message.remove_reaction(emoji, member)\n\n async def edit_cart():\n p = []\n total_prize = 0\n for product in cart_data.keys():\n if self.cache[member.id][product] == 0:\n continue\n else:\n p.append(\n f\"**{products[str(product)]['name']}** x **{self.cache[member.id][product]}**= **{round(products[str(product)]['prize'] * self.cache[member.id][product], 2)}€**\\n\")\n total_prize += round(products[str(product)][\"prize\"], 2) * self.cache[member.id][product]\n\n products_string = \"\".join(p)\n delemitation = \"~--------------------------------------~\\nReact with ✔️ to confirm the order and create a ticket.\\nReact with ❌ to cancel the order\"\n due_embed = discord.Embed(title=f\"Ticket for {member.name}\", color=random.choice(self.colors),\n description=f\"{products_string}\\n**Total Prize: __{total_prize}__**\\n{delemitation}\")\n\n if member.id not in self.messages.keys():\n dm = await member.send(embed=due_embed)\n await dm.add_reaction(\"✔️\")\n await dm.add_reaction(\"❌\")\n self.messages[member.id] = dm.id\n self.embeds[member.id] = due_embed.to_dict()\n else:\n chann_user = self.bot.get_user(member.id)\n priv_chann = chann_user.dm_channel\n temp_msg = await priv_chann.fetch_message(self.messages[member.id])\n self.embeds[member.id][\n \"description\"] = f\"{products_string}\\n**Total Prize: __{round(total_prize, 2)}€__**\\n{delemitation}\"\n edited_embed = discord.Embed.from_dict(self.embeds[member.id])\n await temp_msg.edit(embed=edited_embed)\n\n if str(emoji) == \"🛒\":\n if member.id not in self.cache.keys():\n self.cache[member.id] = {}\n if message_id not in self.cache[member.id].keys():\n self.cache[member.id][message_id] = 0\n self.cache[member.id][message_id] += 1\n cart_data = self.cache[member.id]\n await edit_cart()\n\n if str(emoji) == \"❌\":\n if member.id not in self.cache.keys() or message_id not in self.cache[member.id].keys() or self.cache[member.id][message_id] <= 0: return\n self.cache[member.id][message_id] -= 1\n cart_data = self.cache[member.id]\n await edit_cart()\n\n elif is_dm:\n if str(emoji) == \"✔️\":\n guild = self.bot.get_guild(self.config[\"guild_id\"])\n private_channel = user.dm_channel\n msg = await private_channel.fetch_message(self.messages[user.id])\n tickets_catergory = discord.utils.get(guild.categories, id=self.config[\"tickets_category\"])\n await tickets_catergory.set_permissions(user, read_messages=True)\n command_channel = await guild.create_text_channel(f\"ticket-{user.name}\", category=tickets_catergory)\n embed = discord.Embed(title=f\"Ticket {user.name.capitalize()}\", color=random.choice(self.colors), description=\n f\"**Order Summary ->**\\n\\n{self.embeds[user.id]['description'].split('~')[0]}\\n\"\n f\"Use the command `!close` to close this ticket!\")\n await command_channel.send(embed=embed)\n del self.embeds[user.id]\n del self.cache[user.id]\n del self.messages[user.id]\n await msg.delete()\n\n tickets = json.load(open(\"tickets.json\", \"r\"))\n tickets[str(command_channel.id)] = user.id\n json.dump(tickets, open(\"tickets.json\", \"w\"), indent=4)\n\n if str(emoji) == \"❌\":\n private_channel = user.dm_channel\n msg = await private_channel.fetch_message(self.messages[user.id])\n del self.embeds[user.id]\n del self.cache[user.id]\n del self.messages[user.id]\n await msg.edit(embed=discord.Embed(description=\"Order Canceled.\", color=random.choice(self.colors)))\n for reaction in msg.reactions:\n await msg.remove_reaction(reaction, self.bot.user)\n await asyncio.sleep(5)\n await msg.delete()\n\n @commands.command()\n @commands.has_permissions(administrator=True)\n async def delete(self, ctx, msg_id):\n data = json.load(open(\"products.json\", \"r\"))\n if not msg_id in data.keys():\n return await ctx.send(embed=discord.Embed(color=random.choice(self.colors), description=\"No product found with this id\"))\n name = data[msg_id][\"name\"]\n try:\n msg = await ctx.channel.fetch_message(int(msg_id))\n await msg.delete()\n except Exception:\n pass\n await ctx.send(embed=discord.Embed(color=random.choice(self.colors), description=f\"The product `{name}` was removed from the shop.\"))\n del data[msg_id]\n json.dump(data, open(\"products.json\", \"w\"), indent=4)\n\n @commands.command()\n async def close(self, ctx):\n tickets = json.load(open(\"tickets.json\", \"r\"))\n if str(ctx.channel.id) in tickets.keys():\n channel = self.bot.get_channel(ctx.channel.id)\n embed = discord.Embed(title=\"Ticket closed.\", color=random.choice(self.colors),\n description=\"This ticket channel will be closed in 5 seconds.\")\n await channel.send(embed=embed)\n await asyncio.sleep(5)\n await channel.delete()\n del tickets[str(ctx.channel.id)]\n json.dump(tickets, open(\"tickets.json\", \"w\"), indent=4)\n\ndef setup(bot): bot.add_cog(Shop(bot))", "repo_name": "AirReaper/Bot-Discord-Shop-Keeper", "sub_path": "cogs/shop.py", "file_name": "shop.py", "file_ext": "py", "file_size_in_byte": 8771, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 10, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 10, "usage_type": "name"}, {"api_name": "discord.ext.commands.Bot", "line_number": 11, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 11, "usage_type": "name"}, {"api_name": "json.load", "line_number": 13, "usage_type": "call"}, {"api_name": "discord.ext.commands.Context", "line_number": 21, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 21, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 34, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 34, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 46, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 46, "usage_type": "call"}, {"api_name": "json.load", "line_number": 53, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 58, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 19, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 19, "usage_type": "name"}, {"api_name": "discord.ext.commands.has_permissions", "line_number": 20, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 20, "usage_type": "name"}, {"api_name": "discord.DMChannel", "line_number": 67, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 68, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 87, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 87, "usage_type": "call"}, {"api_name": "discord.Embed.from_dict", "line_number": 102, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 102, "usage_type": "attribute"}, {"api_name": "discord.utils.get", "line_number": 125, "usage_type": "call"}, {"api_name": "discord.utils", "line_number": 125, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 128, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 128, "usage_type": "call"}, {"api_name": "json.load", "line_number": 137, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 139, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 147, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 147, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 150, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 60, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 60, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 60, "usage_type": "name"}, {"api_name": "json.load", "line_number": 156, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 158, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 158, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 165, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 165, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 167, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 153, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 153, "usage_type": "name"}, {"api_name": "discord.ext.commands.has_permissions", "line_number": 154, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 154, "usage_type": "name"}, {"api_name": "json.load", "line_number": 171, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 174, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 174, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 177, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 180, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 169, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 169, "usage_type": "name"}]}
+{"seq_id": "72048417346", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nfrom sklearn.decomposition import PCA\nfrom sklearn import datasets\nfrom sklearn.cluster import KMeans\n\n# load the dataset\niris_df = datasets.load_iris()\npca = PCA(2)\n\n# print y-values names\nprint(iris_df.target_names)\n\n# split the dataset\nX, y = iris_df.data, iris_df.target\nX_proj = pca.fit_transform(X)\n\n# plot the dataset\nplt.scatter(X_proj[:, 0], X_proj[:, 1], c=y)\nplt.show()\n\n# set the size of the plot\nplt.figure(figsize=(10, 4))\n\n# create color map\ncolormap = np.array(['red', 'lime', 'black', 'blue', 'yellow', 'green', 'red'])\n\nk = 3 # running kmeans clustering into two\nkmeans = KMeans(n_clusters=k, random_state=0).fit(X_proj) # train the classifier\n\n# set the classifiers y-values to labels\nlabels = kmeans.labels_\n\n# plot the original classifier\nplt.subplot(1, 2, 1)\nplt.scatter(X_proj[:, 0], X_proj[:, 1], c=colormap[y], s=40)\nplt.title('Real Classification')\n\n# plot the model classifier\nplt.subplot(1, 2, 2)\nplt.scatter(X_proj[:, 0], X_proj[:, 1], c=colormap[labels], s=40)\nplt.title('K-Mean Classification')\n\nplt.show()\n", "repo_name": "alex-marquardt/MachineLearningExam2019", "sub_path": "Exercise 3/iris_dataset_kmeans_pca.py", "file_name": "iris_dataset_kmeans_pca.py", "file_ext": "py", "file_size_in_byte": 1091, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "sklearn.datasets.load_iris", "line_number": 8, "usage_type": "call"}, {"api_name": "sklearn.datasets", "line_number": 8, "usage_type": "name"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}]}
+{"seq_id": "33317312855", "text": "from pymongo import MongoClient\nfrom exceptions.StudentExistsException import StudentExistsException\nfrom email_sender import send_email\nimport pickle\nimport os\n\nclient = MongoClient(os.environ['MONGODB_HOST_NAME'])\ndb = client[os.environ['MONGODB_DB_NAME']]\ncollection = db[os.environ['MONGODB_COLLECTION_NAME']]\n\n\ndef add_student_to_course(course, student):\n # serialization of objects\n student_binary = pickle.dumps(student)\n course_binary = pickle.dumps(course)\n\n if not student in course.students:\n collection.update_one({'_id': course.unique_id},\n {'$set':\n {'course': course_binary},\n '$push':\n {'students': student_binary},\n },\n upsert=True)\n course.add_student(student)\n send_email(student, course, 'verification')\n else:\n raise StudentExistsException()\n\n\ndef remove_course(course):\n collection.delete_one({'_id': course.unique_id})\n\n\ndef get_courses():\n documents = list(collection.find({}))\n courses = []\n for course_binary in documents:\n # deserializes objects\n course = pickle.loads(course_binary['course'])\n\n for student_binary in course_binary['students']:\n student = pickle.loads(student_binary)\n course.add_student(student)\n courses.append(course)\n return courses\n\n\ndef get_number_of_courses():\n return collection.estimated_document_count()\n", "repo_name": "mertbarutcuoglu/Seat-Checker", "sub_path": "database.py", "file_name": "database.py", "file_ext": "py", "file_size_in_byte": 1555, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "pymongo.MongoClient", "line_number": 7, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pickle.dumps", "line_number": 14, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 15, "usage_type": "call"}, {"api_name": "email_sender.send_email", "line_number": 26, "usage_type": "call"}, {"api_name": "exceptions.StudentExistsException.StudentExistsException", "line_number": 28, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 40, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 43, "usage_type": "call"}]}
+{"seq_id": "3925888555", "text": "import numpy\nimport spidev # import the spidev module\nimport time\nimport OSC\n\nnumChannels = 6;\nchannelOffset = 2;\nnumSamples = 4;\n\n# values = numpy.empty((numChannels, numSamples), dtype=numpy.uint16)\nsampleIndex = 0\nvalues = [-1, -1, -1, -1, -1, -1]\n# lastValues = [-1, -1, -1, -1, -1, -1, -1, -1]\n\nspi = spidev.SpiDev() # create a new spidev object\nspi.open(0, 1) # open bus 0, chip enable 1\nspi.max_speed_hz = 500000 # 1MHz\n\nsc = OSC.OSCClient()\nsc.connect(('127.0.0.1', 57120)) #send locally to sc\n# sc.connect(('192.168.1.35', 57120)) #send to external sc\n\ndef init():\n values = values * 0\n lastSendValue = lastSensValue * 0\n\ndef sendOSC(_name, _values):\n msg = OSC.OSCMessage()\n msg.setAddress(_name)\n for value in _values:\n msg.append(4095 - value)\n try:\n sc.send(msg)\n except:\n 1+1 # dummy\n #print msg\n\ndef readADC(_channel):\n if _channel > 7 or _channel < 0:\n return -1\n\n input_mode = 1; # single ended = 1, differential = 0\n command = 0x04; # start flag\n command |= (input_mode<<1);\n command |= (_channel>>2) & 0x01; # add msb of channel in our first command byte\n bytes = spi.xfer2([command, _channel<<6, 0x00])\n result = (bytes[1] & 0x0f)<<8 | bytes[2]\n return result;\n\ntry:\n while True:\n for channel in range(numChannels):\n # values[channel, sampleIndex] = readADC(channel)\n values[channel] = readADC(channel + channelOffset)\n\n # medianValues = numpy.median(values, axis=1)\n # medianValues = 1 - (medianValues/4095.0)\n # print \"%4d\" % value,\n sendOSC(\"/adc\", values)\n # sampleIndex += 1\n # sampleIndex %= numSamples\n\n time.sleep(0.05)\n # print;\nexcept KeyboardInterrupt:\n # Ctrl+C pressed, so...\n spi.close()\n", "repo_name": "constantin3000/PiCollider", "sub_path": "adc2osc_bundle.py", "file_name": "adc2osc_bundle.py", "file_ext": "py", "file_size_in_byte": 1708, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 8, "dataset": "github-code", "pt": "24", "api": [{"api_name": "spidev.SpiDev", "line_number": 15, "usage_type": "call"}, {"api_name": "OSC.OSCClient", "line_number": 19, "usage_type": "call"}, {"api_name": "OSC.OSCMessage", "line_number": 28, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 63, "usage_type": "call"}]}
+{"seq_id": "32893567818", "text": "from django.urls import path\nfrom . import views\n\nurlpatterns = [\n\tpath('', views.homepage, name='homepage'),\n\tpath('quiz_start', views.quiz_start, name='quiz_start'),\n\tpath('quiz', views.quiz, name='quiz'),\n\tpath('feedback',views.feedback, name = 'feedback'),\n\tpath('other', views.other, name = 'other'),\n\tpath('serene',views.serene, name = 'serene'),\n]", "repo_name": "Mouzier/Mindpal_WebProject", "sub_path": "MindPal/mindreader/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 354, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}]}
+{"seq_id": "46191537092", "text": "from setuptools import setup, find_packages\n\ntry:\n with open('requirements.txt') as f:\n requires = f.read().splitlines()\nexcept IOError:\n with open('codeyoloco.egg-info/requires.txt') as f:\n requires = f.read().splitlines()\n \nwith open('VERSION') as f:\n version = f.read().strip()\n\nsetup(\n name = \"codeyoloco\",\n version = version,\n packages = find_packages(),\n package_dir = {'codeyoloco':'codeyoloco'},\n author = 'codeyoloco',\n author_email = 'surajshah525@gmail.com',\n description = 'iCTF and Project Package',\n license = \"PSF\",\n include_package_data = True,\n ) \n", "repo_name": "theskullcrusher/codeyoloco", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 645, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "setuptools.setup", "line_number": 13, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 16, "usage_type": "call"}]}
+{"seq_id": "25873324262", "text": "import json\nfrom twisted.internet.defer import inlineCallbacks\nfrom labrad.wrappers import connectAsync\nfrom lib.helpers import sleep\n\nfrom conductor_device.conductor_parameter import ConductorParameter\n\nclass State(ConductorParameter):\n priority = 1\n\n @inlineCallbacks\n def initialize(self):\n self.cxn = yield connectAsync(name=self.name)\n yield self.cxn.power_supply.select_device('quadrant_coils')\n \n @inlineCallbacks\n def update(self):\n parameter_values = yield self.cxn.conductor.get_parameter_values()\n sequence = json.loads(parameter_values)['sequencer']['sequence']\n yield self.cxn.power_supply.state(False)\n if 'evaporate' in sequence:\n yield sleep(6)\n yield self.cxn.power_supply.state(self._value)\n", "repo_name": "yesrgang/labrad_tools", "sub_path": "conductor/devices/quadrant_coils/seperate_settings/state.py", "file_name": "state.py", "file_ext": "py", "file_size_in_byte": 791, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3, "dataset": "github-code", "pt": "26", "api": [{"api_name": "conductor_device.conductor_parameter.ConductorParameter", "line_number": 8, "usage_type": "name"}, {"api_name": "labrad.wrappers.connectAsync", "line_number": 13, "usage_type": "call"}, {"api_name": "twisted.internet.defer.inlineCallbacks", "line_number": 11, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 19, "usage_type": "call"}, {"api_name": "lib.helpers.sleep", "line_number": 22, "usage_type": "call"}, {"api_name": "twisted.internet.defer.inlineCallbacks", "line_number": 16, "usage_type": "name"}]}
+{"seq_id": "2286078012", "text": "import torch\nfrom torch import nn\n\n\nclass Lenet(nn.Module):\n def __init__(self, num_classes=10, grayscale=False):\n super(Lenet, self).__init__()\n\n self.grayscale = grayscale\n self.num_classes = num_classes\n\n if self.grayscale:\n in_channels = 1\n else:\n in_channels = 3\n\n self.features = nn.Sequential(\n nn.Conv2d(in_channels, 20, kernel_size=5, stride=1),\n nn.ReLU(inplace=True),\n nn.MaxPool2d(kernel_size=2),\n nn.Conv2d(20, 20, kernel_size=5, stride=1),\n nn.ReLU(inplace=True),\n nn.MaxPool2d(kernel_size=2),\n )\n\n self.classifier = nn.Sequential(\n nn.Linear(20 * 4 * 4, 120),\n nn.ReLU(inplace=True),\n nn.Linear(120, 84),\n # nn.Sigmoid(),\n nn.Linear(84, num_classes)\n )\n\n def forward(self, x):\n x = self.features(x)\n # print(x.shape)\n x = torch.flatten(x, 1)\n # print(x.shape)\n logits = self.classifier(x)\n # probas = F.softmax(logits, dim=1)\n\n return logits", "repo_name": "silidada/Lenet", "sub_path": "python/model/Lenet.py", "file_name": "Lenet.py", "file_ext": "py", "file_size_in_byte": 1117, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "torch.nn.Module", "line_number": 5, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 5, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.flatten", "line_number": 37, "usage_type": "call"}]}
+{"seq_id": "17350435876", "text": "import gym\nimport roomba_env\n\nimport time\n\nimport random, numpy, math\n\nfrom keras.models import Sequential\nfrom keras.layers import *\nfrom keras.optimizers import *\n\nclass Model:\n def __init__(self, stateCnt, actionCnt):\n\n self.stateCnt = stateCnt\n self.actionCnt = actionCnt\n\n self.model = self._createModel()\n # self.model.load_weights(\"cartpole-basic.h5\")\n\n def _createModel(self):\n self.model = Sequential()\n self.model.add(Dense(24, input_shape=(5,), activation=\"relu\"))\n self.model.add(Dense(24, activation=\"relu\"))\n self.model.add(Dense(5, activation=\"softmax\"))\n self.model.compile(loss=\"mse\", optimizer=Adam(lr=0.00025))\n\n return self.model\n\n def train(self, x, y, epoch=1, verbose=0):\n self.model.fit(x, y, batch_size=64, nb_epoch=epoch, verbose=verbose)\n\n def predict(self, s):\n return self.model.predict(s)\n\n def predictOne(self, s):\n return self.predict(s.reshape(1, self.stateCnt)).flatten()\n\n\n\nclass Memory: # stored as ( s, a, r, s_ )\n samples = []\n\n def __init__(self, capacity):\n self.capacity = capacity\n\n def add(self, sample):\n self.samples.append(sample)\n\n if len(self.samples) > self.capacity:\n self.samples.pop(0)\n\n def sample(self, n):\n n = min(n, len(self.samples))\n return random.sample(self.samples, n)\n\n\nMEMORY_CAPACITY = 100000\nBATCH_SIZE = 64\n\nGAMMA = 0.99\n\n#exploration rate\nMAX_EPSILON = 1\nMIN_EPSILON = 0.01\n\nLAMBDA = 0.001 # speed of decay\n\n\nclass Agent:\n steps = 0\n epsilon = MAX_EPSILON\n\n def __init__(self, stateCnt, actionCnt):\n self.stateCnt = stateCnt\n self.actionCnt = actionCnt\n\n self.brain = Model(stateCnt, actionCnt)\n self.memory = Memory(MEMORY_CAPACITY)\n\n def act(self, s):\n if random.random() < self.epsilon:\n return random.randint(0, self.actionCnt - 1)\n else:\n return numpy.argmax(self.brain.predictOne(s))\n\n def observe(self, sample): # in (s, a, r, s_) format\n self.memory.add(sample)\n\n # slowly decrease Epsilon based on our experience\n self.steps += 1\n self.epsilon = MIN_EPSILON + (MAX_EPSILON - MIN_EPSILON) * math.exp(-LAMBDA * self.steps)\n\n def replay(self):\n batch = self.memory.sample(BATCH_SIZE)\n print(batch)\n batchLen = len(batch)\n\n no_state = numpy.zeros(self.stateCnt)\n\n states = numpy.array([o for o in batch])\n print(states)\n states_ = numpy.array([no_state if o is None else o for o in batch])\n print(states_)\n p = self.brain.predict(states)\n p_ = self.brain.predict(states_)\n\n x = numpy.zeros((batchLen, self.stateCnt))\n y = numpy.zeros((batchLen, self.actionCnt))\n\n for i in range(batchLen):\n o = batch[i]\n print(i)\n s = o[0]\n a = o[1]\n r = o[2]\n s_ = o[3]\n\n t = p[i]\n print(t)\n if s_ is None:\n t[a] = r\n else:\n t[a] = r + GAMMA * numpy.amax(p_[i])\n\n x[i] = s\n y[i] = t\n\n self.brain.train(x, y)\n\n\n\n\n\nenv = gym.make('roomba-v0')\n\nACTIONS = [\"F\", \"B\", \"L\", \"R\", \"S\"]\n\nenemy = \"enemy\"\nreward_enemy = 0.0\n\nfriendly = \"friendly\"\nreward_friendly = 0.0\n\nstate_enemy, state_friendly = env.reset()\n\n\nagent_enemy = Agent(5,5)\nagent_friendly = Agent(5,5)\n\nfor i in range(1,100000000000001):\n\n done = False\n\n step = 0\n\n initial_input_enemy = []\n\n initial_input_enemy.append(state_enemy[0])\n initial_input_enemy.append(state_enemy[1])\n initial_input_enemy.append(state_friendly[0])\n initial_input_enemy.append(state_friendly[1])\n initial_input_enemy.append(reward_enemy)\n input_array_enemy = np.asarray(initial_input_enemy)\n\n initial_input_friendly = []\n\n initial_input_friendly.append(state_enemy[0])\n initial_input_friendly.append(state_enemy[1])\n initial_input_friendly.append(state_friendly[0])\n initial_input_friendly.append(state_friendly[1])\n initial_input_friendly.append(reward_friendly)\n input_array_friendly = np.asarray(initial_input_friendly)\n\n while done == False:\n\n if(step %2 ==0):\n\n\n\n action = agent_enemy.act(input_array_enemy)\n\n action_code = ACTIONS[action]\n\n state_next_enemy, state_friendly, reward_enemy, reward_friendly, done, info_enemy, info_friendly = env.step(action_code, enemy)\n\n env.render()\n\n #klopt dit?!\n agent_enemy.observe([state_next_enemy[0],state_next_enemy[1], state_friendly[0], state_friendly[1], reward_enemy])\n agent_enemy.replay()\n\n input_next_enemy = []\n input_next_enemy.append(state_next_enemy[0])\n input_next_enemy.append(state_next_enemy[1])\n input_next_enemy.append(state_friendly[0])\n input_next_enemy.append(state_friendly[1])\n input_next_enemy.append(reward_enemy)\n\n input_array_next_enemy = np.asarray(input_next_enemy)\n print(input_array_next_enemy)\n\n\n input_array_enemy = input_array_next_enemy\n\n\n\n # print(\"Game: \" + str(i) + \", exploration: \" + str(dqn_solver_enemy.exploration_rate) + \", score: \" + str(\"nog niet geimplementeerd\"))\n\n\n else:\n action = agent_enemy.act(input_array_friendly)\n\n action_code = ACTIONS[action]\n\n state, state_next_friendly, reward_enemy, reward_friendly, done, info_enemy, info_friendly = env.step(\n action_code, friendly)\n\n env.render()\n\n agent_enemy.observe((input_array_friendly[0], action, reward_friendly, state_next_friendly))\n agent_enemy.replay()\n\n input_next_friendly = []\n input_next_friendly.append(state_next_friendly[0])\n input_next_friendly.append(state_next_friendly[1])\n input_next_friendly.append(state_enemy[0])\n input_next_friendly.append(state_enemy[1])\n input_next_friendly.append(reward_friendly)\n\n input_array_next_friendly = np.asarray(input_next_friendly)\n\n input_array_friendly = input_array_next_friendly\n\n\n\n\n\n time.sleep(0.1)\n step +=1\n print(step)\n\n env.reset()\n env.render()\n time.sleep(3)\n\n\n\n\n", "repo_name": "VanbecelaereVincent/Q_Learning_roomba-env", "sub_path": "Q-learning/test_files/DQN2.py", "file_name": "DQN2.py", "file_ext": "py", "file_size_in_byte": 6358, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "keras.models.Sequential", "line_number": 22, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 55, "usage_type": "call"}, {"api_name": "random.random", "line_number": 82, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 85, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 124, "usage_type": "call"}, {"api_name": "gym.make", "line_number": 135, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 239, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 245, "usage_type": "call"}]}
+{"seq_id": "27716224445", "text": "\"\"\"\nWalk Agent MIB (SNMPv1)\n+++++++++++++++++++++++\n\nPerform SNMP GETNEXT operation with the following options:\n\n* with SNMPv1, community 'public'\n* over IPv4/UDP\n* to an Agent at demo.snmplabs.com:161\n* for OID in tuple form\n\nThis script performs similar to the following Net-SNMP command:\n\n| $ snmpwalk -v1 -c public -ObentU demo.snmplabs.com 1.3.6\n\n\"\"\"#\nfrom pysnmp.carrier.asyncore.dispatch import AsyncoreDispatcher\nfrom pysnmp.carrier.asyncore.dgram import udp\nfrom pyasn1.codec.ber import encoder, decoder\nfrom pysnmp.proto import api\nfrom time import time\n\n# Protocol version to use\npMod = api.PROTOCOL_MODULES[api.SNMP_VERSION_1]\n# pMod = api.protoModules[api.protoVersion2c]\n\n# SNMP table header\nheadVars = [pMod.ObjectIdentifier((1, 3, 6))]\n\n# Build PDU\nreqPDU = pMod.GetNextRequestPDU()\npMod.apiPDU.setDefaults(reqPDU)\npMod.apiPDU.setVarBinds(reqPDU, [(x, pMod.null) for x in headVars])\n\n# Build message\nreqMsg = pMod.Message()\npMod.apiMessage.setDefaults(reqMsg)\npMod.apiMessage.setCommunity(reqMsg, 'public')\npMod.apiMessage.setPDU(reqMsg, reqPDU)\n\nstartedAt = time()\n\n\ndef cbTimerFun(timeNow):\n if timeNow - startedAt > 3:\n raise Exception(\"Request timed out\")\n\n\n# noinspection PyUnusedLocal\ndef cbRecvFun(transportDispatcher, transportDomain, transportAddress,\n wholeMsg, reqPDU=reqPDU, headVars=headVars):\n\n while wholeMsg:\n rspMsg, wholeMsg = decoder.decode(wholeMsg, asn1Spec=pMod.Message())\n rspPDU = pMod.apiMessage.getPDU(rspMsg)\n\n # Match response to request\n if pMod.apiPDU.getRequestID(reqPDU) == pMod.apiPDU.getRequestID(rspPDU):\n\n # Check for SNMP errors reported\n errorStatus = pMod.apiPDU.getErrorStatus(rspPDU)\n if errorStatus and errorStatus != 2:\n raise Exception(errorStatus)\n\n # Format var-binds table\n varBindTable = pMod.apiPDU.getVarBindTable(reqPDU, rspPDU)\n\n # Report SNMP table\n for tableRow in varBindTable:\n for name, val in tableRow:\n print('from: %s, %s = %s' % (transportAddress,\n name.prettyPrint(),\n val.prettyPrint()))\n\n # Stop on EOM\n for oid, val in varBindTable[-1]:\n if not isinstance(val, pMod.Null):\n break\n\n else:\n transportDispatcher.jobFinished(1)\n continue\n\n # Generate request for next row\n pMod.apiPDU.setVarBinds(\n reqPDU, [(x, pMod.null) for x, y in varBindTable[-1]]\n )\n\n pMod.apiPDU.setRequestID(reqPDU, pMod.getNextRequestID())\n\n transportDispatcher.sendMessage(\n encoder.encode(reqMsg), transportDomain, transportAddress\n )\n\n global startedAt\n\n if time() - startedAt > 3:\n raise Exception('Request timed out')\n\n startedAt = time()\n\n return wholeMsg\n\n\ntransportDispatcher = AsyncoreDispatcher()\n\ntransportDispatcher.registerRecvCbFun(cbRecvFun)\ntransportDispatcher.registerTimerCbFun(cbTimerFun)\n\ntransportDispatcher.registerTransport(\n udp.DOMAIN_NAME, udp.UdpSocketTransport().openClientMode()\n)\n\ntransportDispatcher.sendMessage(\n encoder.encode(reqMsg), udp.DOMAIN_NAME, ('demo.snmplabs.com', 161)\n)\n\ntransportDispatcher.jobStarted(1)\n\ntransportDispatcher.runDispatcher()\n\ntransportDispatcher.closeDispatcher()\n", "repo_name": "etingof/pysnmp", "sub_path": "examples/v1arch/asyncore/manager/cmdgen/getnext-pull-whole-mib.py", "file_name": "getnext-pull-whole-mib.py", "file_ext": "py", "file_size_in_byte": 3505, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 544, "dataset": "github-code", "pt": "24", "api": [{"api_name": "pysnmp.proto.api.PROTOCOL_MODULES", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pysnmp.proto.api", "line_number": 24, "usage_type": "name"}, {"api_name": "pysnmp.proto.api.SNMP_VERSION_1", "line_number": 24, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 41, "usage_type": "call"}, {"api_name": "pyasn1.codec.ber.decoder.decode", "line_number": 54, "usage_type": "call"}, {"api_name": "pyasn1.codec.ber.decoder", "line_number": 54, "usage_type": "name"}, {"api_name": "pyasn1.codec.ber.encoder.encode", "line_number": 92, "usage_type": "call"}, {"api_name": "pyasn1.codec.ber.encoder", "line_number": 92, "usage_type": "name"}, {"api_name": "time.time", "line_number": 97, "usage_type": "call"}, {"api_name": "time.time", "line_number": 100, "usage_type": "call"}, {"api_name": "pysnmp.carrier.asyncore.dispatch.AsyncoreDispatcher", "line_number": 105, "usage_type": "call"}, {"api_name": "pysnmp.carrier.asyncore.dgram.udp.DOMAIN_NAME", "line_number": 111, "usage_type": "attribute"}, {"api_name": "pysnmp.carrier.asyncore.dgram.udp", "line_number": 111, "usage_type": "name"}, {"api_name": "pysnmp.carrier.asyncore.dgram.udp.UdpSocketTransport", "line_number": 111, "usage_type": "call"}, {"api_name": "pyasn1.codec.ber.encoder.encode", "line_number": 115, "usage_type": "call"}, {"api_name": "pyasn1.codec.ber.encoder", "line_number": 115, "usage_type": "name"}, {"api_name": "pysnmp.carrier.asyncore.dgram.udp.DOMAIN_NAME", "line_number": 115, "usage_type": "attribute"}, {"api_name": "pysnmp.carrier.asyncore.dgram.udp", "line_number": 115, "usage_type": "name"}]}
+{"seq_id": "71171030782", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\n\ndef get_points(ds, show_images_count: int):\n r_up = np.array([])\n r_down = np.array([])\n c = np.array([])\n i = 0\n for d_s in ['train', 'val', 'test']:\n for image, target in ds[d_s]:\n my_mask = image[0].detach()\n my_mask[my_mask < 1e-18] = 0\n my_mask[my_mask > 1e-18] = 1\n np_im = my_mask.numpy()\n rows = np.argwhere(np.sum(np_im, axis=1) > 10)\n cols = np.argwhere(np.sum(np_im, axis=0) > 10)\n col1 = np.argmax(cols)\n row1 = rows[0]\n row2 = rows[-1]\n if col1 < 2700:\n r_up = np.append(r_up, row1)\n r_down = np.append(r_down, row2)\n c = np.append(c, col1)\n if i < show_images_count:\n if col1 >= 2700:\n # new_image = np_im[row1[0]-50:row2[0]+50, 0:col1+50]\n print(target['file'])\n new_image = np_im[:, 0:2590]\n print(col1)\n plt.imshow(new_image)\n plt.show()\n # plt.imshow(image[0])\n # plt.title(str(i))\n # plt.show()\n i += 1\n print(np.min(r_up), np.max(r_down), np.max(c))\n return np.min(r_up), np.max(r_down), np.max(c)\n", "repo_name": "xkuubix/Breast-Cancer-Classification-with-MIL-models", "sub_path": "get_points_to_crop.py", "file_name": "get_points_to_crop.py", "file_ext": "py", "file_size_in_byte": 1370, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "numpy.array", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.min", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 38, "usage_type": "call"}]}
+{"seq_id": "71873973821", "text": "import time\nfrom multiprocessing import cpu_count\nfrom typing import Union, NamedTuple\n# from torchsummary import summary\n\nimport torch\nimport torch.backends.cudnn # Backend for using NVIDIA CUDA\nimport numpy as np\nimport pickle\n\nfrom torch import nn, optim\nfrom torch.nn import functional as F\nfrom torch.optim .optimizer import Optimizer\nfrom torch.utils.tensorboard import SummaryWriter\nfrom torch.utils.data import DataLoader\nfrom torch.utils import data\nfrom random import randint\n\nimport argparse\nfrom pathlib import Path\n\n# Enable benchmark mode on CUDNN since the input sizes do not vary. This finds the best algorithm to implement the convolutions given the layout.\ntorch.backends.cudnn.benchmark = True\n\n# Add argument parser\nparser = argparse.ArgumentParser(\n description=\"Training a 4-conv-layer CNN on UrbanSound8K\",\n formatter_class=argparse.ArgumentDefaultsHelpFormatter,\n)\n\n# Add arguments to parse\nparser.add_argument(\n \"--log-dir\",\n default=Path(\"logs\"),\n type=Path\n )\nparser.add_argument(\n \"--learning-rate\",\n default=1e-3,\n type=float,\n help=\"Learning rate\"\n )\nparser.add_argument(\n \"--batch-size\",\n default=32,\n type=int,\n help=\"Number of images within each mini-batch\",\n)\nparser.add_argument(\n \"--epochs\",\n default=50,\n type=int,\n help=\"Number of epochs (passes through the entire dataset) to train for\",\n)\nparser.add_argument(\n \"--val-frequency\",\n default=5,\n type=int,\n help=\"How frequently to test the model on the validation set in number of epochs\",\n)\nparser.add_argument(\n \"--log-frequency\",\n default=10,\n type=int,\n help=\"How frequently to save logs to tensorboard in number of steps\",\n)\nparser.add_argument(\n \"--print-frequency\",\n default=300,\n type=int,\n help=\"How frequently to print progress to the command line in number of steps\",\n)\nparser.add_argument(\n \"-j\",\n \"--worker-count\",\n default=cpu_count(),\n type=int,\n help=\"Number of worker processes used to load data.\",\n)\nparser.add_argument(\n \"--momentum\",\n default=0.9,\n type=float,\n)\nparser.add_argument(\n \"--dropout\",\n default=0.5,\n type=float,\n)\nparser.add_argument(\n \"--mode\",\n default=\"LMC\",\n type=str,\n help=\"The type of data to train the network on (LMC, MC, MLMC)\"\n)\nparser.add_argument(\n \"--optimiser\",\n default=\"SGD\",\n type=str,\n help=\"The optimiser used (SGD, Adam, AdamW)\"\n)\nparser.add_argument(\n \"--weight-decay\",\n default=0.01,\n type=float,\n help=\"The L2 regularisation decay parameter\"\n)\nparser.add_argument(\n \"--TSCNN\",\n action = 'store_true',\n help=\"Parameter for dealing with TSCNN combining of logits\"\n)\nparser.add_argument(\n \"--improvements\",\n action = 'store_true',\n help=\"Parameter for adding improvements to the architecture CNN\"\n)\n\nclass DataShape(NamedTuple):\n height: int\n width: int\n channels: int\n\n# Use GPU if cuda is available\nif torch.cuda.is_available():\n DEVICE = torch.device(\"cuda\")\n print (\"Using CUDA...\")\nelse:\n DEVICE = torch.device(\"cpu\")\n print (\"Using CPU...\")\n\n\n# Main function loop for training and testing the data\ndef main(args):\n\n # Load and prepare the data\n istrain = True\n train_dataset = UrbanSound8KDataset(\"./UrbanSound8K_train.pkl\", istrain, args.mode, args.improvements)\n test_dataset = UrbanSound8KDataset(\"./UrbanSound8K_test.pkl\", not istrain, args.mode, args.improvements)\n train_loader = torch.utils.data.DataLoader(\n train_dataset,\n shuffle=True,\n batch_size=args.batch_size,\n pin_memory=True,\n num_workers=args.worker_count,\n )\n test_loader = torch.utils.data.DataLoader(\n test_dataset,\n shuffle=False,\n batch_size=args.batch_size,\n num_workers=args.worker_count,\n pin_memory=True,\n )\n\n # Get the dimensions of the data\n data_channels = train_dataset.__getitem__(0)[0].shape[0]\n data_height = train_dataset.__getitem__(0)[0].shape[1]\n data_width = train_dataset.__getitem__(0)[0].shape[2]\n\n # Define the CNN model\n model = CNN(height=data_height, width=data_width, channels=data_channels, class_count=10, dropout=args.dropout, mode=args.mode, improvements=args.improvements)\n\n # Running Torch Summary to check the architecture\n # summary(model, (data_channels,data_height,data_width))\n\n # Define the unbalanced class weight of the data and move it to the appropriate device (hardcoded from analysis of the dataset)\n data_weight = torch.Tensor(6299/(np.array([6295,1825,6248,5121,5682,6282,1112,5886,5819,6299]))).to(DEVICE)\n\n # Define the criterion to be softmax cross entropy\n criterion = nn.CrossEntropyLoss(weight=data_weight)\n\n # Define the optimizer based on parsed arguments\n if args.optimiser == \"Adam\":\n optimizer = optim.Adam(model.parameters(), lr = args.learning_rate, betas = (args.momentum, 0.999), weight_decay=args.weight_decay)\n elif args.optimiser == \"AdamW\":\n optimizer = optim.AdamW(model.parameters(), lr = args.learning_rate, betas = (args.momentum, 0.999), weight_decay=args.weight_decay)\n elif args.optimiser == \"SGD\":\n optimizer = optim.SGD(model.parameters(), lr = args.learning_rate, momentum = args.momentum, weight_decay=args.weight_decay)\n else:\n print(\"Error: Invalid optimiser argument, defaulting to SGD...\")\n optimizer = optim.SGD(model.parameters(), lr = args.learning_rate, momentum = args.momentum, weight_decay=args.weight_decay)\n\n # Setup directory for the logs\n log_dir = get_summary_writer_log_dir(args)\n print(f\"Writing logs to {log_dir}\")\n\n # Define the summary writer for logging\n summary_writer = SummaryWriter(\n str(log_dir),\n flush_secs=5\n )\n\n # Prep notes file for reference\n f = open(\"logs/notes-sbatch.md\", \"a\")\n f.write(\"Logged to: \" + log_dir + (\" - TSCNN\" if args.TSCNN else f\" - storing {args.mode}\") + \"\\n\")\n f.close()\n\n # Define the model trainer\n trainer = Trainer(\n model, train_loader, test_loader, criterion, optimizer, summary_writer, DEVICE, log_dir, args.TSCNN, args.mode\n )\n\n # Use the trainer to train the model\n trainer.train(\n args.epochs,\n args.val_frequency,\n print_frequency=args.print_frequency,\n log_frequency=args.log_frequency,\n )\n\n # Close the summary writer at the end of the training\n summary_writer.close()\n\n# The Dataset class\nclass UrbanSound8KDataset(data.Dataset):\n def __init__(self, dataset_path, istrain, mode, improvements):\n\n # Load the dataset\n self.dataset = pickle.load(open(dataset_path, 'rb'))\n self.mode = mode\n self.improvements = improvements\n self.istrain = istrain\n\n def __getitem__(self, index):\n\n # Extract the necessary features from the loaded dataset\n LM = self.dataset[index][\"features\"][\"logmelspec\"]\n MFCC = self.dataset[index][\"features\"][\"mfcc\"]\n C = self.dataset[index][\"features\"][\"chroma\"]\n SC = self.dataset[index][\"features\"][\"spectral_contrast\"]\n T = self.dataset[index][\"features\"][\"tonnetz\"]\n\n # Appropriately prepare the data given the selected mode, based on the specifications of the paper\n if self.mode == 'LMC':\n LMC = np.concatenate((LM, C, SC, T), axis=0)\n if self.improvements and self.istrain:\n LMC = LMC*randint(1,4) # Random augmentation of the training data\n feature = torch.from_numpy(LMC.astype(np.float32)).unsqueeze(0)\n elif self.mode == 'MC':\n MC = np.concatenate((MFCC, C, SC, T), axis=0)\n if self.improvements and self.istrain:\n MC = MC*randint(1,4) # Random augmentation of the training data\n feature = torch.from_numpy(MC.astype(np.float32)).unsqueeze(0)\n elif self.mode == 'MLMC':\n MLMC = np.concatenate((MFCC, LM, C, SC, T), axis=0)\n if self.improvements and self.istrain:\n MLMC = MLMC*randint(1,4) # Random augmentation of the training data\n feature = torch.from_numpy(MLMC.astype(np.float32)).unsqueeze(0)\n label = self.dataset[index]['classID']\n fname = self.dataset[index]['filename']\n\n return feature, label, fname, index\n\n def __len__(self):\n return len(self.dataset)\n\n# The architecture class\nclass CNN(nn.Module):\n def __init__(self, height: int, width: int, channels: int, class_count: int, dropout: float, mode: str, improvements: bool):\n super().__init__()\n\n # Define some global class variables\n self.input_shape = DataShape(height=height, width=width, channels=channels)\n self.class_count = class_count\n self.mode = mode\n self.improvements = improvements\n\n # Defining the first convolutional layer & initialising its weights using Kaiming\n self.conv1 = nn.Conv2d(\n in_channels=self.input_shape.channels,\n out_channels=32,\n bias=False,\n kernel_size=(3,3),\n padding=(1,1),\n # padding=(43,21),\n # stride=(2,2)\n )\n self.initialise_layer(self.conv1)\n\n # Defining batch normalisation of the outputs of the first conv layer\n self.bnorm1 = nn.BatchNorm2d(\n num_features=32\n )\n\n # Defining the second convolutional layer & initialising its weights using Kaiming\n self.conv2 = nn.Conv2d(\n in_channels = 32,\n out_channels = 32,\n kernel_size = (3, 3),\n bias=False,\n padding=(1,1),\n # padding = (43, 21),\n # stride=(2,2)\n )\n self.initialise_layer(self.conv2)\n\n # Defining batch normalisation of the outputs of the second conv layer\n self.bnorm2 = nn.BatchNorm2d(\n num_features = 32\n )\n\n # Defining the pooling layer for the batch normalised 2nd conv output\n self.pool2 = nn.MaxPool2d(\n kernel_size=(2, 2),\n padding=(1,1),\n stride=(2, 2)\n )\n\n # Defining the third convolutional layer & initialising its weights using Kaiming\n self.conv3 = nn.Conv2d(\n in_channels=32,\n out_channels=64,\n kernel_size=(3,3),\n bias=False,\n padding=(1,1),\n # padding = (22,11),\n # stride=(2,2)\n )\n self.initialise_layer(self.conv3)\n\n # Defining batch normalisation of the outputs of the third conv layer\n self.bnorm3 = nn.BatchNorm2d(\n num_features=64\n )\n\n # Defining the fourth convolutional layer & initialising its weights using Kaiming\n # Could use Max Pooling for the last layer, but probably more likely to be stride (based on the paper)\n self.conv4 = nn.Conv2d(\n in_channels=64,\n out_channels=64,\n kernel_size=(3,3),\n padding=(1,1),\n stride=(2,2),\n bias=False,\n )\n self.initialise_layer(self.conv4)\n\n # Adding a pooling layer with stride instead of conv4 with stride\n # self.pool4 = nn.MaxPool2d(\n # kernel_size=(2, 2),\n # padding=(1,1),\n # stride=(2, 2)\n # )\n\n # Defining batch normalisation of the outputs of the fourth conv layer\n self.bnorm4 = nn.BatchNorm2d(\n num_features=64\n )\n\n # Defining the first fully connected layer & initialising the weights using Kaiming\n # The size of the data for MLMC is larger and requires a larger fully connected layer\n if self.mode == \"MLMC\":\n self.fc1 = nn.Linear(26048, 1024)\n else:\n self.fc1 = nn.Linear(15488, 1024)\n self.initialise_layer(self.fc1)\n\n # Defining batch normalisation of the outputs of the first fully connected layer\n self.bnormfc1 = nn.BatchNorm1d(\n num_features = 1024\n )\n\n # Defining the final fully connected layer to 10 classes & initialising the weights using Kaiming\n self.fc2 = nn.Linear (1024, 10)\n self.initialise_layer(self.fc2)\n\n # Defining the dropout used in the CNN\n self.dropout = nn.Dropout2d(p=dropout)\n\n def forward(self, input_data: torch.Tensor) -> torch.Tensor:\n\n # Implementing the first conv hidden layer\n x = self.conv1(input_data)\n x = self.bnorm1(x)\n x = F.relu(x)\n\n # Implementing the second conv hidden layer\n x = self.conv2(self.dropout(x))\n x = self.bnorm2(x)\n x = F.relu(x)\n\n # Implementing a pooling stage to the outputs of the first layer\n x = self.pool2(x)\n\n # Implementing the third conv hidden layer\n # if self.improvements:\n # x = self.dropout(x)\n x = self.conv3(x)\n x = self.bnorm3(x)\n x = F.relu(x)\n\n # Implementing the fourth conv hidden layer\n x = self.conv4(self.dropout(x))\n x = self.bnorm4(x)\n x = F.relu(x)\n\n # Use pooling with stride instead of conv4 with stride\n # x = self.pool4(x)\n\n # Flattening the output of the fourth conv layer for the first fc layer\n x = torch.flatten(x, start_dim = 1)\n\n # Implementing the first fully connected hidden layer\n x = self.fc1(self.dropout(x))\n # x = self.bnormfc1(x) # This was not in the paper\n x = torch.sigmoid(x)\n\n # Implementing the final fully connected hidden layer\n x = self.fc2(x)\n return x\n\n @staticmethod\n def initialise_layer(layer):\n if hasattr(layer, \"bias\"):\n if layer.bias is not None:\n nn.init.zeros_(layer.bias)\n if hasattr(layer, \"weight\"):\n nn.init.kaiming_normal_(layer.weight)\n\n# Class for the execution of the main training loop\nclass Trainer:\n def __init__(\n self,\n model: nn.Module,\n train_loader: DataLoader,\n val_loader: DataLoader,\n criterion: nn.Module,\n optimizer: Optimizer,\n summary_writer: SummaryWriter,\n device: torch.device,\n log_dir: str,\n TSCNN: bool,\n mode: str,\n ):\n self.model = model.to(device)\n self.device = device\n self.train_loader = train_loader\n self.val_loader = val_loader\n self.criterion = criterion\n self.optimizer = optimizer\n self.summary_writer = summary_writer\n self.step = 0\n self.log_dir = log_dir\n self.TSCNN = TSCNN\n self.mode = mode\n\n def train(\n self,\n epochs: int,\n val_frequency: int,\n print_frequency: int = 20,\n log_frequency: int = 5,\n start_epoch: int = 0,\n ):\n # Setting model to training mode\n self.model.train()\n\n # Defining list of results for each epoch\n results_epoch = {}\n\n # Main training loop\n for epoch in range(start_epoch, epochs):\n self.model.train()\n\n # Extracting required data from loader\n data_load_start_time = time.time()\n for batch, labels, fname, index in self.train_loader:\n batch = batch.to(self.device)\n labels = labels.to(self.device)\n data_load_end_time = time.time()\n\n # Compute the forward pass of the model\n logits = self.model.forward(batch)\n\n # Calculate the loss of the forward pass\n loss = self.criterion(logits, labels)\n\n # Implement backpropogation\n loss.backward()\n\n # Update the optimiser parameters and set the update grads to zero again\n self.optimizer.step()\n self.optimizer.zero_grad()\n\n # Disabling autograd when calculationg the accuracy\n with torch.no_grad():\n preds = logits.argmax(-1)\n accuracy = compute_accuracy(labels, preds)\n\n # Writing to logs and printing out the progress\n data_load_time = data_load_end_time - data_load_start_time\n step_time = time.time() - data_load_end_time\n if ((self.step + 1) % log_frequency) == 0:\n self.log_metrics(epoch, accuracy, loss, data_load_time, step_time)\n if ((self.step + 1) % print_frequency) == 0:\n self.print_metrics(epoch, accuracy, loss, data_load_time, step_time)\n\n # Update loop params for next batch\n self.step += 1\n data_load_start_time = time.time()\n\n # Write to summary writer at the end of each epoch\n self.summary_writer.add_scalar(\"epoch\", epoch, self.step)\n if ((epoch + 1) % val_frequency) == 0:\n results_epoch[epoch] = self.validate(epoch, epochs, self.log_dir)\n\n # Exporting data\n if not self.TSCNN:\n pickle.dump(results_epoch, open(\"TSCNN_store_\" + self.mode + \".pkl\", \"wb\"))\n else:\n pickle.dump(results_epoch, open(\"TSCNN_store_\" + \"TSCNN\" + \".pkl\", \"wb\"))\n\n # Function used to print the progress\n def print_metrics(self, epoch, accuracy, loss, data_load_time, step_time):\n epoch_step = self.step % len(self.train_loader)\n print(\n f\"epoch: [{epoch}], \"\n f\"step: [{epoch_step}/{len(self.train_loader)}], \"\n f\"batch loss: {loss:.5f}, \"\n f\"batch accuracy: {accuracy * 100:2.2f}, \"\n f\"data load time: \"\n f\"{data_load_time:.5f}, \"\n f\"step time: {step_time:.5f},\"\n\n )\n\n # Function used to log the progress\n def log_metrics(self, epoch, accuracy, loss, data_load_time, step_time):\n self.summary_writer.add_scalar(\"epoch\", epoch, self.step)\n self.summary_writer.add_scalars(\n \"accuracy\",\n {\"train\": accuracy},\n self.step\n )\n self.summary_writer.add_scalars(\n \"loss\",\n {\"train\": float(loss.item())},\n self.step\n )\n self.summary_writer.add_scalar(\n \"time/data\", data_load_time, self.step\n )\n self.summary_writer.add_scalar(\n \"time/data\", step_time, self.step\n )\n\n # Function used to validate the model\n def validate(self, epoch, epochs, log_dir):\n results = {\"preds\": [], \"labels\": [], \"logits\": [], \"indices\": []}\n total_loss = 0\n\n # Loading data from previous runs and defining softmax to combine for TSCNN\n if self.TSCNN:\n results_epoch_LMC = pickle.load(open(\"TSCNN_store_LMC.pkl\", \"rb\"))\n results_epoch_MC = pickle.load(open(\"TSCNN_store_MC.pkl\", \"rb\"))\n smax = nn.Softmax(dim=-1)\n counter = 0 # used to load the appropriate logits from the stored data\n\n # Put model in validation mode\n self.model.eval()\n\n # No need to track gradients for validation, we're not optimizing.\n with torch.no_grad():\n for batch, labels, fname, index in self.val_loader:\n\n # Shifting batch and labels to appropriate device for efficiency\n batch = batch.to(self.device)\n labels = labels.to(self.device)\n\n # Calculating the logits of the testing batch\n logits = self.model(batch)\n\n # Averaging the logits by fname and making the new labels\n fname_logits, fname_labels, fname_indices = orderbyfname(labels,fname,logits, index)\n fname_logits = fname_logits.to(self.device)\n fname_labels = fname_labels.to(self.device)\n\n # Combining saved LMC and current MC logits for TSCNN, including some sanity checks\n if self.TSCNN:\n combined_logits = []\n if len(fname_logits) != len(fname_labels):\n print(\"ERROR: Incorrect lengths of logit and label arrays, sanity check failed!\") # Sanity check\n for fname_logit, fname_label in zip(fname_logits, fname_labels):\n if fname_label != results_epoch_LMC[epoch][\"labels\"][counter] or fname_label != results_epoch_MC[epoch][\"labels\"][counter]:\n print(\"ERROR: Incorrect label, sanity check failed!\") # Sanity check\n combined_logits.append(np.array(smax(results_epoch_LMC[epoch][\"logits\"][counter]).cpu() + smax(results_epoch_MC[epoch][\"logits\"][counter]).cpu()))\n counter += 1\n\n # Overwriting logits with new combined logits and preparing the tensor\n fname_logits = (torch.Tensor(combined_logits).type(torch.float)).to(self.device)\n\n\n # Calculating loss with new logits and labels\n loss = self.criterion(fname_logits, fname_labels)\n total_loss += loss.item()\n\n # Getting predictions of new logits\n preds = fname_logits.argmax(dim=-1).cpu().numpy()\n\n # Appending results\n results[\"logits\"].extend(list(fname_logits))\n results[\"preds\"].extend(list(preds))\n results[\"labels\"].extend(list(fname_labels.cpu().numpy()))\n results[\"indices\"].extend(list(fname_indices))\n\n # Find the overall accuracy of the model\n accuracy = compute_accuracy(\n np.array(results[\"labels\"]), np.array(results[\"preds\"])\n )\n\n # Find the class accuracy of the model\n class_accuracy = compute_class_accuracy(\n np.array(results[\"labels\"]), np.array(results[\"preds\"])\n )\n\n # Find the average class accuracy\n class_accuracy_avg = sum(class_accuracy)/100*len(class_accuracy)\n\n # Compute the average loss\n average_loss = total_loss / len(self.val_loader)\n\n # Write the progress to the logs\n self.summary_writer.add_scalars(\n \"accuracy\",\n {\"test\": accuracy},\n self.step\n )\n self.summary_writer.add_scalars(\n \"average_class\",\n {\"test\": class_accuracy_avg},\n self.step\n )\n self.summary_writer.add_scalars(\n \"loss\",\n {\"test\": average_loss},\n self.step\n )\n\n # Switch model back to evaluation mode\n self.model.train()\n\n # Print the progress & exporting the softmaxed logits and labels\n display_text = f\"validation loss: {average_loss:.5f}, accuracy: {accuracy * 100:2.2f}, class_accuracy: {class_accuracy}\\nclass_avg: {class_accuracy_avg}\"\n if (epoch+1) == epochs:\n f = open(\"logs/accuracy.md\", \"a\")\n f.write(log_dir + \"\\n\")\n f.write(display_text + \"\\n\\n\")\n f.close()\n print(display_text)\n return results\n\n# Function for averaging the logits and proucing new labels from old\ndef orderbyfname(labels,fname,logits, index):\n fname_set = sorted(set(fname))\n new_logits = []\n new_labels = []\n new_indices = []\n for iter,name in enumerate(fname_set):\n\n # Determining the indices of the batch which are from the same filename\n indices = np.where(np.array(fname)==name)[0]\n sum = np.zeros(10)\n index_store_temp = []\n\n # Using the indices to average the logits\n for i in indices:\n sum += np.array(logits[i].cpu())\n index_store_temp.append(index[i]) # qualitative analysis\n sum = sum/len(indices)\n\n # appending new data\n new_logits.append(sum)\n new_labels.append(labels[indices[0]])\n\n # Storing the actual test data indices for the qualitative analysis\n new_indices.append(index_store_temp)\n\n return torch.Tensor(new_logits).type(torch.float), torch.Tensor(new_labels).type(torch.long), new_indices\n\n# Function for computing the overall accuracy of the model\ndef compute_accuracy(\n labels: Union[torch.Tensor, np.ndarray], preds: Union[torch.Tensor, np.ndarray]\n) -> float:\n \"\"\"\n Args:\n labels: ``(batch_size, class_count)`` tensor or array containing example labels\n preds: ``(batch_size, class_count)`` tensor or array containing model prediction\n \"\"\"\n assert len(labels) == len(preds)\n return float((labels == preds).sum()) / len(labels)\n\n# Function for computing the class accuracy of the model\ndef compute_class_accuracy(labels: Union[torch.Tensor, np.ndarray], preds: Union[torch.Tensor, np.ndarray], class_count: int = 10) -> float:\n assert len(labels) == len(preds)\n class_accuracy = []\n for class_label in range(0,class_count):\n class_labels = np.where(labels == class_label, class_label, class_label)\n class_accuracy.append(float(np.logical_and((preds == class_labels),(labels == class_labels)).sum())*100 / np.array(labels == class_labels).sum())\n return class_accuracy\n\n\n\n# Function for handling the directory for writing logs to\ndef get_summary_writer_log_dir(args: argparse.Namespace) -> str:\n \"\"\"Get a unique directory that hasn't been logged to before for use with a TB\n SummaryWriter.\n\n Args:\n args: CLI Arguments\n\n Returns:\n Subdirectory of log_dir with unique subdirectory name to prevent multiple runs\n from getting logged to the same TB log directory (which you can't easily\n untangle in TB).\n \"\"\"\n tb_log_dir_prefix = (f'CNN_bn_epochs={args.epochs}_dropout={args.dropout}_bs={args.batch_size}_optim={args.optimiser}_decay={args.weight_decay}_lr={args.learning_rate}_momentum={args.momentum}_mode=' + (\"TSCNN\" if args.TSCNN else args.mode) + (\"_improvements_\" if args.improvements else \"\") +'_run_')\n i = 0\n while i < 1000:\n tb_log_dir = args.log_dir / (tb_log_dir_prefix + str(i))\n if not tb_log_dir.exists():\n return str(tb_log_dir)\n i += 1\n return str(tb_log_dir)\n\n# Running the Progamme\nif __name__ == \"__main__\":\n start = time.time()\n main(parser.parse_args())\n print (\"Total time taken: {}\".format(time.time() - start))\n", "repo_name": "fz16336/Applied-Deep-Learning", "sub_path": "code/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 26170, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3, "dataset": "github-code", "pt": "24", "api": [{"api_name": "torch.backends", "line_number": 23, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 26, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 34, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 35, "usage_type": "name"}, {"api_name": "multiprocessing.cpu_count", "line_number": 76, "usage_type": "call"}, {"api_name": "typing.NamedTuple", "line_number": 119, "usage_type": "name"}, {"api_name": "torch.cuda.is_available", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 125, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 126, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 140, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 147, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 167, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 170, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 170, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 174, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 174, "usage_type": "name"}, {"api_name": "torch.optim.AdamW", "line_number": 176, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 176, "usage_type": "name"}, {"api_name": "torch.optim.SGD", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 178, "usage_type": "name"}, {"api_name": "torch.optim.SGD", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 181, "usage_type": "name"}, {"api_name": "torch.utils.tensorboard.SummaryWriter", "line_number": 188, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 215, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 215, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 235, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 237, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 238, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 240, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 242, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 243, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 245, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 247, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 248, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 258, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 258, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 269, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 269, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 281, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 281, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 286, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 286, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 298, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 298, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 303, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 303, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 310, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 310, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 322, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 322, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 328, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 328, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 346, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 346, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 353, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 353, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 355, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 355, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 359, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 359, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 364, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 364, "usage_type": "name"}, {"api_name": "torch.nn.Dropout2d", "line_number": 368, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 368, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 370, "usage_type": "attribute"}, {"api_name": "torch.nn.functional.relu", "line_number": 375, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 375, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 380, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 380, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 390, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 390, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 395, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 395, "usage_type": "name"}, {"api_name": "torch.flatten", "line_number": 401, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 406, "usage_type": "call"}, {"api_name": "torch.nn.init.zeros_", "line_number": 416, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 416, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 416, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_normal_", "line_number": 418, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 418, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 418, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 424, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 424, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 425, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 426, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 427, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 427, "usage_type": "name"}, {"api_name": "torch.optim.optimizer.Optimizer", "line_number": 428, "usage_type": "name"}, {"api_name": "torch.utils.tensorboard.SummaryWriter", "line_number": 429, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 430, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 466, "usage_type": "call"}, {"api_name": "time.time", "line_number": 470, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 486, "usage_type": "call"}, {"api_name": "time.time", "line_number": 492, "usage_type": "call"}, {"api_name": "time.time", "line_number": 500, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 509, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 511, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 554, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 555, "usage_type": "call"}, {"api_name": "torch.nn.Softmax", "line_number": 556, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 556, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 563, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 586, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 590, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 590, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 608, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 613, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 661, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 661, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 662, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 667, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 678, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 678, "usage_type": "attribute"}, {"api_name": "torch.long", "line_number": 678, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 682, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 682, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 682, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 693, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 693, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 693, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 697, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 698, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 698, "usage_type": "call"}, {"api_name": "argparse.Namespace", "line_number": 704, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 727, "usage_type": "call"}, {"api_name": "time.time", "line_number": 729, "usage_type": "call"}]}
+{"seq_id": "73775317181", "text": "__author__ = \"ContraxSuite, LLC; LexPredict, LLC\"\n__copyright__ = \"Copyright 2015-2021, ContraxSuite, LLC\"\n__license__ = \"https://github.com/LexPredict/lexpredict-lexnlp/blob/2.3.0/LICENSE\"\n__version__ = \"2.3.0\"\n__maintainer__ = \"LexPredict, LLC\"\n__email__ = \"support@contraxsuite.com\"\n\n\nimport re\nfrom typing import List, Generator\nfrom lexnlp.extract.common.annotations.definition_annotation import DefinitionAnnotation\nfrom lexnlp.extract.common.definitions.common_definition_patterns import CommonDefinitionPatterns\nfrom lexnlp.extract.common.definitions.universal_definition_parser import UniversalDefinitionsParser\nfrom lexnlp.extract.common.pattern_found import PatternFound\nfrom lexnlp.extract.es.language_tokens import EsLanguageTokens\nfrom lexnlp.utils.lines_processing.line_processor import LineSplitParams\n\n\nclass SpanishParsingMethods:\n \"\"\"\n the class contains methods with the same signature:\n def method_name(phrase: str) -> List[DefinitionMatch]:\n the methods are used for finding definition \"candidates\"\n \"\"\"\n reg_hereafter = re.compile(\"(?<=(en adelante[,\\\\s]))[\\\\w\\\\s*\\\\\\\"*]+\", re.UNICODE)\n reg_reffered = re.compile(\"^.+(?=se refiere)\", re.UNICODE)\n reg_first_word_is = re.compile(r\"^.+?(?=es\\s+\\w+\\W+\\w+|está\\s+\\w+\\W+\\w+)\", re.UNICODE)\n\n @staticmethod\n def match_es_def_by_hereafter(phrase: str) -> List[PatternFound]:\n \"\"\"\n :param phrase: las instrucciones de uso o instalación del software o todas las descripciones\n de uso del mismo (de aquí en adelante, la \"Documentación\");\n :return: {name: 'Documentación', probability: 100, ...}\n \"\"\"\n reg = SpanishParsingMethods.reg_hereafter\n dfs = CommonDefinitionPatterns. \\\n collect_regex_matches_with_quoted_chunks(phrase, reg, 100,\n lambda p, m, e: 0,\n lambda p, m, e: m.start() + e.end(),\n lambda p, m: 0,\n lambda p, m: m.end())\n return dfs\n\n @staticmethod\n def match_es_def_by_reffered(phrase: str) -> List[PatternFound]:\n \"\"\"\n :param phrase: En este acuerdo, el término \"Software\" se refiere a: (i) el programa informático\n que acompaña a este Acuerdo y todos sus componentes;\n :return: definitions (objects)\n \"\"\"\n reg = SpanishParsingMethods.reg_reffered\n dfs = CommonDefinitionPatterns. \\\n collect_regex_matches_with_quoted_chunks(phrase, reg, 100,\n lambda p, m, e: m.start() + e.start(),\n lambda p, m, e: len(phrase),\n lambda p, m: m.start(),\n lambda p, m: len(p))\n return dfs\n\n @staticmethod\n def match_first_word_is(phrase: str) -> List[PatternFound]:\n \"\"\"\n :param phrase: El tabaquismo es la adicción al tabaco, provocada principalmente.\n :return: definitions (objects)\n \"\"\"\n reg = SpanishParsingMethods.reg_first_word_is\n dfs = CommonDefinitionPatterns.\\\n collect_regex_matches_with_quoted_chunks(phrase, reg, 65,\n lambda p, m, e: m.start() + e.start(),\n lambda p, m, e: len(phrase),\n lambda p, m: m.start(),\n lambda p, m: len(p))\n return dfs\n\n\ndef make_es_definitions_parser():\n split_params = LineSplitParams()\n split_params.line_breaks = {'\\n', '.', ';', '!', '?'}\n split_params.abbreviations = EsLanguageTokens.abbreviations\n split_params.abbr_ignore_case = True\n\n functions = [CommonDefinitionPatterns.match_es_def_by_semicolon,\n CommonDefinitionPatterns.match_acronyms,\n SpanishParsingMethods.match_es_def_by_hereafter,\n SpanishParsingMethods.match_es_def_by_reffered,\n SpanishParsingMethods.match_first_word_is]\n\n return UniversalDefinitionsParser(functions, split_params)\n\n\nparser = make_es_definitions_parser()\n\n\ndef get_definition_annotations(text: str, language: str = 'es') -> Generator[DefinitionAnnotation, None, None]:\n yield from parser.parse(text, language)\n\n\ndef get_definition_annotation_list(text: str, language: str = 'es') -> List[DefinitionAnnotation]:\n return list(get_definition_annotations(text, language))\n\n\ndef get_definitions(text: str, language: str = 'es') -> Generator[dict, None, None]:\n for annotation in parser.parse(text, language):\n yield annotation.to_dictionary()\n\n\ndef get_definition_list(text: str, language: str = 'es') -> List[dict]:\n return list(get_definitions(text, language))\n", "repo_name": "LexPredict/lexpredict-lexnlp", "sub_path": "lexnlp/extract/es/definitions.py", "file_name": "definitions.py", "file_ext": "py", "file_size_in_byte": 5005, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 654, "dataset": "github-code", "pt": "24", "api": [{"api_name": "re.compile", "line_number": 25, "usage_type": "call"}, {"api_name": "re.UNICODE", "line_number": 25, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 26, "usage_type": "call"}, {"api_name": "re.UNICODE", "line_number": 26, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 27, "usage_type": "call"}, {"api_name": "re.UNICODE", "line_number": 27, "usage_type": "attribute"}, {"api_name": "lexnlp.extract.common.definitions.common_definition_patterns.CommonDefinitionPatterns.collect_regex_matches_with_quoted_chunks", "line_number": 37, "usage_type": "call"}, {"api_name": "lexnlp.extract.common.definitions.common_definition_patterns.CommonDefinitionPatterns", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 30, "usage_type": "name"}, {"api_name": "lexnlp.extract.common.pattern_found.PatternFound", "line_number": 30, "usage_type": "name"}, {"api_name": "lexnlp.extract.common.definitions.common_definition_patterns.CommonDefinitionPatterns.collect_regex_matches_with_quoted_chunks", "line_number": 53, "usage_type": "call"}, {"api_name": "lexnlp.extract.common.definitions.common_definition_patterns.CommonDefinitionPatterns", "line_number": 53, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 46, "usage_type": "name"}, {"api_name": "lexnlp.extract.common.pattern_found.PatternFound", "line_number": 46, "usage_type": "name"}, {"api_name": "lexnlp.extract.common.definitions.common_definition_patterns.CommonDefinitionPatterns.collect_regex_matches_with_quoted_chunks", "line_number": 68, "usage_type": "call"}, {"api_name": "lexnlp.extract.common.definitions.common_definition_patterns.CommonDefinitionPatterns", "line_number": 68, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 62, "usage_type": "name"}, {"api_name": "lexnlp.extract.common.pattern_found.PatternFound", "line_number": 62, "usage_type": "name"}, {"api_name": "lexnlp.utils.lines_processing.line_processor.LineSplitParams", "line_number": 78, "usage_type": "call"}, {"api_name": "lexnlp.extract.es.language_tokens.EsLanguageTokens.abbreviations", "line_number": 80, "usage_type": "attribute"}, {"api_name": "lexnlp.extract.es.language_tokens.EsLanguageTokens", "line_number": 80, "usage_type": "name"}, {"api_name": "lexnlp.extract.common.definitions.common_definition_patterns.CommonDefinitionPatterns.match_es_def_by_semicolon", "line_number": 83, "usage_type": "attribute"}, {"api_name": "lexnlp.extract.common.definitions.common_definition_patterns.CommonDefinitionPatterns", "line_number": 83, "usage_type": "name"}, {"api_name": "lexnlp.extract.common.definitions.common_definition_patterns.CommonDefinitionPatterns.match_acronyms", "line_number": 84, "usage_type": "attribute"}, {"api_name": "lexnlp.extract.common.definitions.common_definition_patterns.CommonDefinitionPatterns", "line_number": 84, "usage_type": "name"}, {"api_name": "lexnlp.extract.common.definitions.universal_definition_parser.UniversalDefinitionsParser", "line_number": 89, "usage_type": "call"}, {"api_name": "typing.Generator", "line_number": 95, "usage_type": "name"}, {"api_name": "lexnlp.extract.common.annotations.definition_annotation.DefinitionAnnotation", "line_number": 95, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 99, "usage_type": "name"}, {"api_name": "lexnlp.extract.common.annotations.definition_annotation.DefinitionAnnotation", "line_number": 99, "usage_type": "name"}, {"api_name": "typing.Generator", "line_number": 103, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 108, "usage_type": "name"}]}
+{"seq_id": "31843693097", "text": "# -*- coding: utf-8 -*-\n\"\"\"upperenvelope\n\nProvides a Numba jit compilled upper envelope for a custom utility function.\n\"\"\"\n\nimport numpy as np\nfrom numba import njit\n\ndef create(ufunc,use_inv_w=False):\n \"\"\" create upperenvelope function from the utility function ufunc\n \n Args:\n\n ufunc (callable): utility function with *args (must be decorated with @njit)\n\n Returns:\n\n upperenvelope (callable): upperenvelope called as (grid_a,m_vec,c_vec,inv_w_vec,use_inv_w,grid_m,c_ast_vec,v_ast_vec,*args)\n use_inv_w (bool,optional): assume that the post decision value-of-choice vector is a negative inverse\n \n \"\"\"\n\n @njit\n def upperenvelope(grid_a,m_vec,c_vec,inv_w_vec,grid_m,c_ast_vec,v_ast_vec,*args):\n \"\"\" upperenvelope function\n \n Args:\n\n grid_a (numpy.ndarray): input, end-of-period asset vector of length Na\n m_vec (numpy.ndarray): input, cash-on-hand vector from egm of length Na\n c_vec (numpy.ndarray): input, consumption vector from egm of length Na\n inv_w_vec (numpy.ndarray): input, post decision value-of-choice vector from egm of length Na\n grid_m (numpy.ndarray): input, common grid for cash-on-hand of length Nm\n c_ast_vec (numpy.ndarray): output, consumption on common grid for cash-on-hand of length Nm\n v_ast_vec (numpy.ndarray): output, value-of-choice on common grid for cash-on-hand of length Nm\n *args: additional arguments to the utility function\n \n \"\"\"\n\n # for given m_vec, c_vec and w_vec (coming from grid_a)\n # find the optimal consumption choices (c_ast_vec) at the common grid (grid_m) \n # using the upper envelope + also value the implied values-of-choice (v_ast_vec)\n\n Na = grid_a.size\n Nm = grid_m.size\n\n c_ast_vec[:] = 0\n v_ast_vec[:] = -np.inf\n\n # constraint\n # the constraint is binding if the common m is smaller\n # than the smallest m implied by EGM step (m_vec[0])\n\n im = 0\n while im < Nm and grid_m[im] <= m_vec[0]:\n \n # a. consume all\n c_ast_vec[im] = grid_m[im] \n\n # b. value of choice\n u = ufunc(c_ast_vec[im],*args)\n if use_inv_w:\n v_ast_vec[im] = u + (-1.0/inv_w_vec[0])\n else:\n v_ast_vec[im] = u + inv_w_vec[0]\n\n im += 1\n\n # upper envellope\n # apply the upper envelope algorithm\n \n for ia in range(Na-1):\n\n # a. a inteval and w slope\n a_low = grid_a[ia]\n a_high = grid_a[ia+1]\n \n inv_w_low = inv_w_vec[ia]\n inv_w_high = inv_w_vec[ia+1]\n\n if a_low > a_high:\n continue\n\n inv_w_slope = (inv_w_high-inv_w_low)/(a_high-a_low)\n \n # b. m inteval and c slope\n m_low = m_vec[ia]\n m_high = m_vec[ia+1]\n\n c_low = c_vec[ia]\n c_high = c_vec[ia+1]\n\n c_slope = (c_high-c_low)/(m_high-m_low)\n\n # c. loop through common grid\n for im in range(Nm):\n\n # i. current m\n m = grid_m[im]\n\n # ii. interpolate?\n interp = (m >= m_low) and (m <= m_high) \n extrap_above = ia == Na-2 and m > m_vec[Na-1]\n\n # iii. interpolation (or extrapolation)\n if interp or extrap_above:\n\n # o. implied guess\n c_guess = c_low + c_slope * (m - m_low)\n a_guess = m - c_guess\n\n # oo. implied post-decision value function\n inv_w = inv_w_low + inv_w_slope * (a_guess - a_low) \n\n # ooo. value-of-choice\n u = ufunc(c_guess,*args)\n if use_inv_w:\n v_guess = u + (-1/inv_w)\n else:\n v_guess = u + inv_w\n\n # oooo. update\n if v_guess > v_ast_vec[im]:\n v_ast_vec[im] = v_guess\n c_ast_vec[im] = c_guess\n \n return upperenvelope", "repo_name": "NumEconCopenhagen/ConsumptionSaving", "sub_path": "consav/upperenvelope.py", "file_name": "upperenvelope.py", "file_ext": "py", "file_size_in_byte": 4256, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 19, "dataset": "github-code", "pt": "24", "api": [{"api_name": "numpy.inf", "line_number": 49, "usage_type": "attribute"}, {"api_name": "numba.njit", "line_number": 24, "usage_type": "name"}]}
+{"seq_id": "3875357075", "text": "from itertools import combinations\n\norders = [\"XYZ\", \"XWY\", \"WXA\"]\ncourse = [2,3,4]\n\nd = dict()\n\nfor o in orders:\n tmp_o = list(o)\n for c_num in course:\n if c_num > len(tmp_o):\n break\n for cand in combinations(tmp_o, c_num):\n cand_str = ''.join(sorted(cand))\n if not d.get(cand_str):\n d[cand_str] = 1\n else:\n d[cand_str] += 1\n\na_dict = dict() \nfor k, v in d.items():\n t_num = len(k)\n if not a_dict.get(t_num):\n if v > 1:\n a_dict[t_num] = ([k], v)\n else:\n if a_dict[t_num][1] < v:\n a_dict[t_num] = ([k], v)\n elif a_dict[t_num][1] == v:\n a_dict[t_num][0].append(k)\n\nanswer = []\nfor c_num in course:\n if a_dict.get(c_num):\n for t_ans in a_dict[c_num][0]:\n answer.append(t_ans)\nanswer.sort()\nprint(answer)", "repo_name": "TTC1018/Algo_Py", "sub_path": "programmers/menu_renewal.py", "file_name": "menu_renewal.py", "file_ext": "py", "file_size_in_byte": 882, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "itertools.combinations", "line_number": 13, "usage_type": "call"}]}
+{"seq_id": "21365989854", "text": "from django import forms\nfrom django.core.exceptions import ValidationError\n\nfrom itou.job_applications.enums import SenderKind\nfrom itou.job_applications.models import JobApplication\n\n\nclass JobApplicationAdminForm(forms.ModelForm):\n class Meta:\n model = JobApplication\n fields = \"__all__\"\n\n def clean(self):\n sender = self.cleaned_data[\"sender\"]\n sender_kind = self.cleaned_data[\"sender_kind\"]\n sender_company = self.cleaned_data.get(\"sender_company\")\n sender_prescriber_organization = self.cleaned_data.get(\"sender_prescriber_organization\")\n\n if sender_kind == SenderKind.JOB_SEEKER:\n if sender is None:\n raise ValidationError(\"Emetteur candidat manquant.\")\n if not sender.is_job_seeker:\n raise ValidationError(\"Emetteur du mauvais type.\")\n\n if sender_kind == SenderKind.EMPLOYER:\n if sender_company is None:\n raise ValidationError(\"SIAE émettrice manquante.\")\n if sender is None:\n raise ValidationError(\"Emetteur SIAE manquant.\")\n else:\n # Sender is optional, but if it exists, check its role.\n if not sender.is_employer:\n raise ValidationError(\"Emetteur du mauvais type.\")\n\n elif sender_company is not None:\n raise ValidationError(\"SIAE émettrice inattendue.\")\n\n if sender_kind == SenderKind.PRESCRIBER:\n if sender:\n # Sender is optional, but if it exists, check its role.\n if not sender.is_prescriber:\n raise ValidationError(\"Emetteur du mauvais type.\")\n # Request organization only if prescriber is actively linked to an organization\n if (\n sender_prescriber_organization is None\n and sender.prescribermembership_set.filter(is_active=True).exists()\n ):\n raise ValidationError(\"Organisation du prescripteur émettrice manquante.\")\n else:\n raise ValidationError(\"Emetteur prescripteur manquant.\")\n elif sender_prescriber_organization is not None:\n raise ValidationError(\"Organisation du prescripteur émettrice inattendue.\")\n\n return\n", "repo_name": "gip-inclusion/les-emplois", "sub_path": "itou/job_applications/admin_forms.py", "file_name": "admin_forms.py", "file_ext": "py", "file_size_in_byte": 2305, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 26, "dataset": "github-code", "pt": "26", "api": [{"api_name": "django.forms.ModelForm", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 8, "usage_type": "name"}, {"api_name": "itou.job_applications.models.JobApplication", "line_number": 10, "usage_type": "name"}, {"api_name": "itou.job_applications.enums.SenderKind.JOB_SEEKER", "line_number": 19, "usage_type": "attribute"}, {"api_name": "itou.job_applications.enums.SenderKind", "line_number": 19, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 21, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 23, "usage_type": "call"}, {"api_name": "itou.job_applications.enums.SenderKind.EMPLOYER", "line_number": 25, "usage_type": "attribute"}, {"api_name": "itou.job_applications.enums.SenderKind", "line_number": 25, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 27, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 29, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 33, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 36, "usage_type": "call"}, {"api_name": "itou.job_applications.enums.SenderKind.PRESCRIBER", "line_number": 38, "usage_type": "attribute"}, {"api_name": "itou.job_applications.enums.SenderKind", "line_number": 38, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 42, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 48, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 50, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 52, "usage_type": "call"}]}
+{"seq_id": "10837055555", "text": "from openerp import models, fields, api, _\nfrom openerp.osv import osv, fields\nimport re\nfrom datetime import datetime, timedelta, date\nfrom openerp import tools, SUPERUSER_ID\nimport base64, csv, StringIO\n\nclass account_invoice(osv.osv):\n\t\n\t_inherit = 'account.invoice'\n\t\n\t@api.multi\n\tdef confirm_paid(self):\n\t\tsuper(account_invoice, self).confirm_paid()\n\t# ketika invoice dibayar, set tanggal bayarnya yaitu tanggal payment pertama\n\t\tpayment_date = date.today()\n\t\tfor payment in self.payment_ids:\n\t\t\tif payment.date:\n\t\t\t\tpayment_date = payment.date\n\t\t\telif payment.date_created:\n\t\t\t\tpayment_date = payment.date_created\n\t\t\tbreak # ambil payment yang pertama aja. asumsi tidak ada backdate payment sedemikian sehingga tanggal payument yang kedua lebih dulu dari yang pertama\n\t\treturn self.write({'payment_date': payment_date})\n\t\t\n\t@api.model\n\tdef calculate_invoice_point(self, member, invoice_line):\n\t# ambil nilai uang dari mo ini\n\t\tamount = invoice_line.price_subtotal\n\t# ambil settingan terakhir. kalau productnya ngga ada setting, ya udah\n\t\tsetting = invoice_line.product_id.member_point_settings\n\t\tif not setting: return 0\n\t# dapatkan setting line berdasarkan level member saat ini \n\t\tused_setting_line = None\n\t\tfor setting_line in setting:\n\t\t\tif setting_line.membership_level_id.id == member.current_level.id:\n\t\t\t\tused_setting_line = setting_line\n\t\t\t\tbreak\n\t\tif not used_setting_line: return 0\n\t# hitung poin beserta pembulatannya\n\t\tpoint = used_setting_line.factor * amount / 1000\t\t\n\t\tpoint = point - (point % used_setting_line.rounding)\n\t\treturn point\n\n\n", "repo_name": "miebakso/ciptadlab2", "sub_path": "account_invoice.py", "file_name": "account_invoice.py", "file_ext": "py", "file_size_in_byte": 1561, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "openerp.osv.osv.osv", "line_number": 8, "usage_type": "attribute"}, {"api_name": "openerp.osv.osv", "line_number": 8, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 16, "usage_type": "name"}, {"api_name": "openerp.api.multi", "line_number": 12, "usage_type": "attribute"}, {"api_name": "openerp.api", "line_number": 12, "usage_type": "name"}, {"api_name": "openerp.api.model", "line_number": 25, "usage_type": "attribute"}, {"api_name": "openerp.api", "line_number": 25, "usage_type": "name"}]}
+{"seq_id": "41064437649", "text": "#!/usr/bin/env python\n# Implementation of algorithm from http://stackoverflow.com/a/22640362/6029703\nimport numpy as np\nimport pylab\nimport matplotlib.pyplot as plt\n\ndef thresholding_algo(y, lag, threshold, influence):\n signals = np.zeros(len(y)) # Initialise signal results\n filteredY = np.array(y) # Initialise filtered series\n avgFilter = [0]*len(y) # Initialise average filter\n stdFilter = [0]*len(y) # Initialise std. filter\n avgFilter[lag - 1] = np.mean(y[0:lag]) # Initialise first value\n stdFilter[lag - 1] = np.std(y[0:lag]) # Initialise first value \n for i in range(lag, len(y) - 1):\n if abs(y[i] - avgFilter[i-1]) > threshold * stdFilter [i-1]:\n if y[i] > avgFilter[i-1]:\n signals[i] = 1 # Positive signal\n else:\n signals[i] = -1 # Negative signal\n\n # Make influence lower\n filteredY[i] = influence * y[i] + (1 - influence) * filteredY[i-1]\n else:\n signals[i] = 0 # No signal\n filteredY[i] = y[i]\n avgFilter[i] = np.mean(filteredY[(i-lag):i])\n stdFilter[i] = np.std(filteredY[(i-lag):i])\n\n return dict(signals = np.asarray(signals),\n avgFilter = np.asarray(avgFilter),\n stdFilter = np.asarray(stdFilter))\n\n\n\n# Data\ny = np.array([1,1,1.1,1,0.9,1,1,1.1,1,0.9,1,1.1,1,1,0.9,1,1,1.1,1,1,1,1,1.1,0.9,1,1.1,1,1,0.9,\n 1,1.1,1,1,1.1,1,0.8,0.9,1,1.2,0.9,1,1,1.1,1.2,1,1.5,1,3,2,5,3,2,1,1,1,0.9,1,1,3,\n 2.6,4,3,3.2,2,1,1,0.8,4,4,2,2.5,1,1,1])\n\n# Settings: lag = 30, threshold = 5, influence = 0\nlag = 30\nthreshold = 5\ninfluence = 0\n\n# Run algo with settings from above\nresult = thresholding_algo(y, lag=lag, threshold=threshold, influence=influence)\n\n# Plot result\nplt.subplot(211)\nplt.plot(np.arange(1, len(y)+1), y)\n\nplt.plot(np.arange(1, len(y)+1),\n result[\"avgFilter\"], color=\"cyan\", lw=2)\n\nplt.plot(np.arange(1, len(y)+1),\n result[\"avgFilter\"] + threshold * result[\"stdFilter\"], color=\"green\", lw=2)\n\nplt.plot(np.arange(1, len(y)+1),\n result[\"avgFilter\"] - threshold * result[\"stdFilter\"], color=\"green\", lw=2)\n\nplt.subplot(212)\nplt.step(np.arange(1, len(y)+1), result[\"signals\"], color=\"red\", lw=2)\nplt.ylim(-1.5, 1.5)\nplt.show()\n\n\n\n", "repo_name": "bmaz/projet_buzz_twitter", "sub_path": "ThresholdingAlgo.py", "file_name": "ThresholdingAlgo.py", "file_ext": "py", "file_size_in_byte": 2260, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "numpy.zeros", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.step", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}]}
+{"seq_id": "40040031736", "text": "#! /usr/bin/env python\n\n\"\"\"Main script to run training and evaluation of models.\n\"\"\"\n\nimport functools\nimport os\nimport tempfile\nimport yaml\n\nfrom seq2seq import models\nfrom seq2seq.data import data_utils, vocab\nfrom seq2seq.training import HParamsParser\nfrom seq2seq.training import utils as training_utils\nfrom seq2seq.training import metrics\n\nimport tensorflow as tf\nfrom tensorflow.contrib.learn.python.learn import learn_runner\nfrom tensorflow.contrib.learn.python.learn.estimators import run_config\nfrom tensorflow.python.platform import gfile\n\n\n# Input Data\ntf.flags.DEFINE_string(\"train_source\", None,\n \"\"\"Path to the training data source sentences. A raw\n text files with tokens separated by spaces.\"\"\")\ntf.flags.DEFINE_string(\"train_target\", None,\n \"\"\"Path to the training data target sentences. A raw\n text files with tokens separated by spaces.\"\"\")\ntf.flags.DEFINE_string(\"dev_source\", None,\n \"\"\"Path to the development data source sentences.\n Same format as training data.\"\"\")\ntf.flags.DEFINE_string(\"dev_target\", None,\n \"\"\"Path to the development data target sentences.\n Same format as training data.\"\"\")\ntf.flags.DEFINE_string(\"vocab_source\", None,\n \"\"\"Path to the source vocabulary.\n A raw text file with one word per line.\"\"\")\ntf.flags.DEFINE_string(\"vocab_target\", None,\n \"\"\"Path to the target vocabulary.\n A raw text file with one word per line.\"\"\")\ntf.flags.DEFINE_string(\"delimiter\", \" \",\n \"\"\"Split input files into tokens on this delimiter.\n Defaults to \" \" (space).\"\"\")\ntf.flags.DEFINE_string(\"config_path\", None,\n \"\"\"Path to a YAML configuration file defining FLAG\n values and hyperparameters. Refer to the documentation\n for more details.\"\"\")\n\n# Model Configuration\ntf.flags.DEFINE_string(\"model\", \"AttentionSeq2Seq\",\n \"\"\"The model class to use. Refer to the documentation\n for all available models.\"\"\")\ntf.flags.DEFINE_string(\"buckets\", None,\n \"\"\"Buckets input sequences according to these length.\n A comma-separated list of sequence length buckets, e.g.\n \"10,20,30\" would result in 4 buckets:\n <10, 10-20, 20-30, >30. None disabled bucketing. \"\"\")\ntf.flags.DEFINE_integer(\"batch_size\", 16,\n \"\"\"Batch size used for training and evaluation.\"\"\")\ntf.flags.DEFINE_string(\"hparams\", None,\n \"\"\"A comma-separated list of hyeperparameter values that\n overwrite the model defaults, e.g.\n \"optimizer.name=Adam,optimization.learning_rate=0.1\".\n Refer to the documentation for a detailed list of\n available hyperparameters.\"\"\")\ntf.flags.DEFINE_string(\"output_dir\", None,\n \"\"\"The directory to write model checkpoints and summaries\n to. If None, a local temporary directory is created.\"\"\")\n\n# Training parameters\ntf.flags.DEFINE_string(\"schedule\", None,\n \"\"\"Estimator function to call, defaults to\n train_and_evaluate for local run\"\"\")\ntf.flags.DEFINE_integer(\"train_steps\", None,\n \"\"\"Maximum number of training steps to run.\n If None, train forever.\"\"\")\ntf.flags.DEFINE_integer(\"train_epochs\", None,\n \"\"\"Maximum number of training epochs over the data.\n If None, train forever.\"\"\")\ntf.flags.DEFINE_integer(\"eval_every_n_steps\", 1000,\n \"Run evaluation on validation data every N steps.\")\ntf.flags.DEFINE_integer(\"sample_every_n_steps\", 500,\n \"\"\"Sample and print sequence predictions every N steps\n during training.\"\"\")\n\n# RunConfig Flags\ntf.flags.DEFINE_integer(\"tf_random_seed\", None,\n \"\"\"Random seed for TensorFlow initializers. Setting\n this value allows consistency between reruns.\"\"\")\ntf.flags.DEFINE_integer(\"save_checkpoints_secs\", 600,\n \"\"\"Save checkpoints every this many seconds.\n Can not be specified with save_checkpoints_steps.\"\"\")\ntf.flags.DEFINE_integer(\"save_checkpoints_steps\", None,\n \"\"\"Save checkpoints every this many steps.\n Can not be specified with save_checkpoints_secs.\"\"\")\ntf.flags.DEFINE_integer(\"keep_checkpoint_max\", 5,\n \"\"\"Maximum number of recent checkpoint files to keep.\n As new files are created, older files are deleted.\n If None or 0, all checkpoint files are kept.\"\"\")\ntf.flags.DEFINE_integer(\"keep_checkpoint_every_n_hours\", 4,\n \"\"\"In addition to keeping the most recent checkpoint\n files, keep one checkpoint file for every N hours of\n training.\"\"\")\n\nFLAGS = tf.flags.FLAGS\n\ndef create_experiment(output_dir):\n \"\"\"\n Creates a new Experiment instance.\n\n Args:\n output_dir: Output directory for model checkpoints and summaries.\n \"\"\"\n\n config = run_config.RunConfig(\n tf_random_seed=FLAGS.tf_random_seed,\n save_checkpoints_secs=FLAGS.save_checkpoints_secs,\n save_checkpoints_steps=FLAGS.save_checkpoints_steps,\n keep_checkpoint_max=FLAGS.keep_checkpoint_max,\n keep_checkpoint_every_n_hours=FLAGS.keep_checkpoint_every_n_hours\n )\n\n # Load vocabulary info\n source_vocab_info = vocab.get_vocab_info(FLAGS.vocab_source)\n target_vocab_info = vocab.get_vocab_info(FLAGS.vocab_target)\n\n # Find model class\n model_class = getattr(models, FLAGS.model)\n\n # Parse parameter and merge with defaults\n hparams = model_class.default_params()\n if FLAGS.hparams is not None and isinstance(FLAGS.hparams, str):\n hparams = HParamsParser(hparams).parse(FLAGS.hparams)\n elif isinstance(FLAGS.hparams, dict):\n hparams.update(FLAGS.hparams)\n\n # Print hparams\n training_utils.print_hparams(hparams)\n\n # One the main worker, save training options and vocabulary\n if config.is_chief:\n # Copy vocabulary to output directory\n gfile.MakeDirs(output_dir)\n source_vocab_path = os.path.join(output_dir, \"vocab_source\")\n gfile.Copy(FLAGS.vocab_source, source_vocab_path, overwrite=True)\n target_vocab_path = os.path.join(output_dir, \"vocab_target\")\n gfile.Copy(FLAGS.vocab_target, target_vocab_path, overwrite=True)\n # Save train options\n train_options = training_utils.TrainOptions(\n hparams=hparams,\n model_class=FLAGS.model,\n source_vocab_path=source_vocab_path,\n target_vocab_path=target_vocab_path)\n train_options.dump(output_dir)\n\n # Create model\n model = model_class(\n source_vocab_info=source_vocab_info,\n target_vocab_info=target_vocab_info,\n params=hparams)\n\n bucket_boundaries = None\n if FLAGS.buckets:\n bucket_boundaries = list(map(int, FLAGS.buckets.split(\",\")))\n\n # Create training input function\n train_input_fn = training_utils.create_input_fn(\n data_provider_fn=functools.partial(\n data_utils.make_parallel_data_provider,\n data_sources_source=FLAGS.train_source,\n data_sources_target=FLAGS.train_target,\n shuffle=True,\n num_epochs=FLAGS.train_epochs,\n delimiter=FLAGS.delimiter),\n batch_size=FLAGS.batch_size,\n bucket_boundaries=bucket_boundaries)\n\n # Create eval input function\n eval_input_fn = training_utils.create_input_fn(\n data_provider_fn=functools.partial(\n data_utils.make_parallel_data_provider,\n data_sources_source=FLAGS.dev_source,\n data_sources_target=FLAGS.dev_target,\n shuffle=False,\n num_epochs=1,\n delimiter=FLAGS.delimiter),\n batch_size=FLAGS.batch_size)\n\n def model_fn(features, labels, params, mode):\n \"\"\"Builds the model graph\"\"\"\n return model(features, labels, params, mode)\n\n estimator = tf.contrib.learn.estimator.Estimator(\n model_fn=model_fn,\n model_dir=output_dir,\n config=config)\n\n train_hooks = training_utils.create_default_training_hooks(\n estimator=estimator,\n sample_frequency=FLAGS.sample_every_n_steps,\n delimiter=FLAGS.delimiter)\n\n eval_metrics = {\n \"log_perplexity\": metrics.streaming_log_perplexity(),\n \"bleu\": metrics.make_bleu_metric_spec(),\n }\n\n experiment = tf.contrib.learn.experiment.Experiment(\n estimator=estimator,\n train_input_fn=train_input_fn,\n eval_input_fn=eval_input_fn,\n min_eval_frequency=FLAGS.eval_every_n_steps,\n train_steps=FLAGS.train_steps,\n eval_steps=None,\n eval_metrics=eval_metrics,\n train_monitors=train_hooks)\n\n return experiment\n\n\ndef main(_argv):\n \"\"\"The entrypoint for the script\"\"\"\n\n # Load flags from config file\n if FLAGS.config_path:\n with gfile.GFile(FLAGS.config_path) as config_file:\n config_flags = yaml.load(config_file)\n for flag_key, flag_value in config_flags.items():\n setattr(FLAGS, flag_key, flag_value)\n\n if not FLAGS.output_dir:\n FLAGS.output_dir = tempfile.mkdtemp()\n\n learn_runner.run(\n experiment_fn=create_experiment,\n output_dir=FLAGS.output_dir,\n schedule=FLAGS.schedule)\n\n\nif __name__ == \"__main__\":\n tf.logging.set_verbosity(tf.logging.INFO)\n tf.app.run()\n", "repo_name": "ForeverZyh/TensorFlow-Program-Bugs", "sub_path": "Github/UT-4/seq2seq-fix/bin/train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 9655, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 29, "dataset": "github-code", "pt": "24", "api": [{"api_name": "tensorflow.flags.DEFINE_string", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 27, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 27, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 36, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 45, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 45, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 51, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 54, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_integer", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 61, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 67, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 72, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_integer", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 75, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_integer", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 78, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_integer", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 81, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_integer", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 83, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_integer", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 88, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_integer", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 91, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_integer", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_integer", "line_number": 97, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 97, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_integer", "line_number": 101, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 101, "usage_type": "attribute"}, {"api_name": "tensorflow.flags", "line_number": 106, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.learn.python.learn.estimators.run_config.RunConfig", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.contrib.learn.python.learn.estimators.run_config", "line_number": 116, "usage_type": "name"}, {"api_name": "seq2seq.data.vocab.get_vocab_info", "line_number": 125, "usage_type": "call"}, {"api_name": "seq2seq.data.vocab", "line_number": 125, "usage_type": "name"}, {"api_name": "seq2seq.data.vocab.get_vocab_info", "line_number": 126, "usage_type": "call"}, {"api_name": "seq2seq.data.vocab", "line_number": 126, "usage_type": "name"}, {"api_name": "seq2seq.models", "line_number": 129, "usage_type": "argument"}, {"api_name": "seq2seq.training.HParamsParser", "line_number": 134, "usage_type": "call"}, {"api_name": "seq2seq.training.utils.print_hparams", "line_number": 139, "usage_type": "call"}, {"api_name": "seq2seq.training.utils", "line_number": 139, "usage_type": "name"}, {"api_name": "tensorflow.python.platform.gfile.MakeDirs", "line_number": 144, "usage_type": "call"}, {"api_name": "tensorflow.python.platform.gfile", "line_number": 144, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path", "line_number": 145, "usage_type": "attribute"}, {"api_name": "tensorflow.python.platform.gfile.Copy", "line_number": 146, "usage_type": "call"}, {"api_name": "tensorflow.python.platform.gfile", "line_number": 146, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path", "line_number": 147, "usage_type": "attribute"}, {"api_name": "tensorflow.python.platform.gfile.Copy", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.python.platform.gfile", "line_number": 148, "usage_type": "name"}, {"api_name": "seq2seq.training.utils.TrainOptions", "line_number": 150, "usage_type": "call"}, {"api_name": "seq2seq.training.utils", "line_number": 150, "usage_type": "name"}, {"api_name": "seq2seq.training.utils.create_input_fn", "line_number": 168, "usage_type": "call"}, {"api_name": "seq2seq.training.utils", "line_number": 168, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 169, "usage_type": "call"}, {"api_name": "seq2seq.data.data_utils.make_parallel_data_provider", "line_number": 170, "usage_type": "attribute"}, {"api_name": "seq2seq.data.data_utils", "line_number": 170, "usage_type": "name"}, {"api_name": "seq2seq.training.utils.create_input_fn", "line_number": 180, "usage_type": "call"}, {"api_name": "seq2seq.training.utils", "line_number": 180, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 181, "usage_type": "call"}, {"api_name": "seq2seq.data.data_utils.make_parallel_data_provider", "line_number": 182, "usage_type": "attribute"}, {"api_name": "seq2seq.data.data_utils", "line_number": 182, "usage_type": "name"}, {"api_name": "tensorflow.contrib.learn.estimator.Estimator", "line_number": 194, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 194, "usage_type": "attribute"}, {"api_name": "seq2seq.training.utils.create_default_training_hooks", "line_number": 199, "usage_type": "call"}, {"api_name": "seq2seq.training.utils", "line_number": 199, "usage_type": "name"}, {"api_name": "seq2seq.training.metrics.streaming_log_perplexity", "line_number": 205, "usage_type": "call"}, {"api_name": "seq2seq.training.metrics", "line_number": 205, "usage_type": "name"}, {"api_name": "seq2seq.training.metrics.make_bleu_metric_spec", "line_number": 206, "usage_type": "call"}, {"api_name": "seq2seq.training.metrics", "line_number": 206, "usage_type": "name"}, {"api_name": "tensorflow.contrib.learn.experiment.Experiment", "line_number": 209, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 209, "usage_type": "attribute"}, {"api_name": "tensorflow.python.platform.gfile.GFile", "line_number": 227, "usage_type": "call"}, {"api_name": "tensorflow.python.platform.gfile", "line_number": 227, "usage_type": "name"}, {"api_name": "yaml.load", "line_number": 228, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 233, "usage_type": "call"}, {"api_name": "tensorflow.contrib.learn.python.learn.learn_runner.run", "line_number": 235, "usage_type": "call"}, {"api_name": "tensorflow.contrib.learn.python.learn.learn_runner", "line_number": 235, "usage_type": "name"}, {"api_name": "tensorflow.logging.set_verbosity", "line_number": 242, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 242, "usage_type": "attribute"}, {"api_name": "tensorflow.app.run", "line_number": 243, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 243, "usage_type": "attribute"}]}
+{"seq_id": "6164584664", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# @Time : 2022/4/20 9:06\n# @Author : jingwen\n# @File : SocketService.py\n# @Desc: : 将预测代码拆开封装成接口供java客户端调用\nimport os\nimport subprocess\nimport time\n\nimport cv2\nimport numpy as np\n\nfrom model.retinaface import Retinaface\nfrom apiResponse.ApiResponse import ApiResponse\nfrom encoding import encoding\n\nfrom socket_config import *\n\n\n# socket中可能要多次调用下列方法,建立一个service类\nclass SocketService(object):\n def __init__(self):\n super(SocketService, self).__init__()\n self.retinaface = Retinaface()\n\n # 内部函数\n def get_file_save_path(self, image_path):\n file_name = os.path.basename(image_path) # jingwen_undetected.jpg\n save_directory = detection_result\n if not os.path.exists(save_directory):\n os.makedirs(save_directory)\n new_file_name = file_name.replace(\"_undetected\", \"_detected\", 1)\n save_image_path = save_directory + new_file_name\n return save_image_path\n\n # 检测图片\n def predictImage(self, image_path=\"\", save_image=False):\n print(\"检测图片,image_path=\", image_path, )\n image = cv2.imread(image_path)\n if image is None:\n # print(\"图片路径有误\")\n return ApiResponse(code=500, message='图片路径有误')\n else:\n image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)\n r_image, recognition_face_list = self.retinaface.detect_image(image)\n print(f'FaceNet识别出{len(recognition_face_list)}张人脸:{recognition_face_list}')\n\n # r_image = cv2.cvtColor(r_image, cv2.COLOR_RGB2BGR)\n # cv2.imshow(\"after\", r_image)\n # cv2.waitKey(0)\n\n data = {}\n data.setdefault('recognition_face_list', recognition_face_list)\n\n # 是否保存检测图片\n if save_image:\n r_image = cv2.cvtColor(r_image, cv2.COLOR_RGB2BGR)\n save_file_path = self.get_file_save_path(image_path)\n cv2.imwrite(save_file_path, r_image)\n data.setdefault('save_file_path', save_file_path)\n print('保存检测结果: ' + save_file_path)\n return ApiResponse(code=200, message=\"识别成功\", data=data)\n\n # 检测视频,完成学生正脸检测,并返回最终得分\n def predictVideo(self, video_path=\"\", save_video=False, video_fps=30):\n print(\"检测视频,video_path=\", video_path)\n\n data = {}\n # 防止传入数据为摄像头\n if video_path == 0:\n return ApiResponse(code=500, message=\"该方法为检测视频,请更换摄像头检测方法\")\n all_recognition_face_list = [] # 为检测到的所有人脸\n\n capture = cv2.VideoCapture(video_path)\n # 视频保存路径\n video_save_path = ''\n if save_video:\n video_save_path = self.get_file_save_path(video_path)\n # 视频的编码\n fourcc = cv2.VideoWriter_fourcc(*'mp4v')\n # 视频的分辨率\n size = (int(capture.get(cv2.CAP_PROP_FRAME_WIDTH)), int(capture.get(cv2.CAP_PROP_FRAME_HEIGHT)))\n # 定义视频输出\n out = cv2.VideoWriter(video_save_path, fourcc, video_fps, size)\n\n ref, frame = capture.read()\n if not ref: # null 0\n raise ApiResponse(code=500, message=\"未能正确读取视频,请检查视频路径是否正确\")\n\n input_video_fps = capture.get(5) # 获取输入视频帧数\n frameRate = int(input_video_fps) * videoTimeRate # 每隔videoTimeRate秒取一帧\n current_ref = 1 # 当前帧\n fps = 0.0\n while (True):\n\n ref, frame = capture.read()\n\n if not ref:\n break\n \"\"\"\n # 镜头水平反转代码\n \"\"\"\n # frame = cv2.flip(frame, 180)\n recognition_face_list = []\n if (current_ref % frameRate == 0):\n t1 = time.time()\n # 读取某一帧\n # 格式转变,BGRtoRGB\n frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)\n # 进行检测\n frame, recognition_face_list = self.retinaface.detect_image(frame)\n\n frame = np.array(frame)\n # RGBtoBGR满足opencv显示格式\n frame = cv2.cvtColor(frame, cv2.COLOR_RGB2BGR)\n\n fps = (fps + (1. / (time.time() - t1))) / 2\n print(f'fps= {fps:.2f}; '\n f'facenet识别出{len(recognition_face_list)}张人脸:{recognition_face_list} ')\n\n # 根据当前帧识别的人脸 更新总识别人脸数\n if len(recognition_face_list) != 0:\n for face_name in recognition_face_list:\n if face_name not in all_recognition_face_list:\n all_recognition_face_list.append(face_name)\n frame = cv2.putText(frame, \"fps= %.2f, recognition_num= %d\" % (fps, len(recognition_face_list)),\n (0, 40), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 255, 0), 2)\n current_ref += 1\n\n # cv2.imshow(\"video\", frame)\n c = cv2.waitKey(1) & 0xff\n if save_video:\n out.write(frame)\n\n if c == 27:\n capture.release()\n break\n print(\"Video Detection Done!\")\n capture.release()\n if save_video == True:\n out.release()\n \"\"\"\n # 为视频添加音频信息\n video_add_voice_config=True 表示添加\n video_add_voice_config=False 表示不添加 \n \"\"\"\n if video_add_voice_config:\n video_save_path = self.video_add_voice(video_save_path, video_path)\n print(\"Save processed video to the path :\" + video_save_path)\n cv2.destroyAllWindows()\n\n data.setdefault('save_file_path', video_save_path)\n data.setdefault('recognition_face_list', all_recognition_face_list)\n return ApiResponse(code=200, message=\"视频检测完成\", data=data)\n\n # 给数据库人脸编码\n def update_face_dataset(self):\n encoding()\n # 更新初始化的retinaface的人脸库信息\n self.retinaface.known_face_encodings = np.load(\n \"model/model_data/{backbone}_face_encoding.npy\".format(backbone=self.retinaface.facenet_backbone))\n self.retinaface.known_face_names = np.load(\n \"model/model_data/{backbone}_names.npy\".format(backbone=self.retinaface.facenet_backbone))\n return ApiResponse(code=200, message='人脸库更新成功')\n\n def video_add_voice(self, video_file, voice_file):\n \"\"\"\n 视频添加音频\n :param video_file: 传入视频文件的路径\n :param voice_file: 传入音频文件的路径\n :return:\n \"\"\"\n outfile_name = video_file.split('.')[0] + '_voice.mp4'\n subprocess.call('ffmpeg -y -i ' + video_file\n + ' -i ' + voice_file + ' -strict -2 -f mp4 '\n + outfile_name, shell=True)\n # 删除原视频文件,若不想删除可以注释掉下行代码\n os.remove(video_file)\n\n return outfile_name\n", "repo_name": "jing3wen/jw_project", "sub_path": "facenet-retinaface-pytorch/SocketService.py", "file_name": "SocketService.py", "file_ext": "py", "file_size_in_byte": 7363, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 4, "dataset": "github-code", "pt": "24", "api": [{"api_name": "model.retinaface.Retinaface", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 40, "usage_type": "call"}, {"api_name": "apiResponse.ApiResponse.ApiResponse", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 45, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 45, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 58, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 58, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 60, "usage_type": "call"}, {"api_name": "apiResponse.ApiResponse.ApiResponse", "line_number": 63, "usage_type": "call"}, {"api_name": "apiResponse.ApiResponse.ApiResponse", "line_number": 72, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 75, "usage_type": "call"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 81, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FRAME_WIDTH", "line_number": 83, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_HEIGHT", "line_number": 83, "usage_type": "attribute"}, {"api_name": "cv2.VideoWriter", "line_number": 85, "usage_type": "call"}, {"api_name": "apiResponse.ApiResponse.ApiResponse", "line_number": 89, "usage_type": "call"}, {"api_name": "time.time", "line_number": 107, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 110, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 110, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 114, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 116, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 116, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 118, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 127, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 128, "usage_type": "attribute"}, {"api_name": "cv2.waitKey", "line_number": 132, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 151, "usage_type": "call"}, {"api_name": "apiResponse.ApiResponse.ApiResponse", "line_number": 155, "usage_type": "call"}, {"api_name": "encoding.encoding", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 163, "usage_type": "call"}, {"api_name": "apiResponse.ApiResponse.ApiResponse", "line_number": 165, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 175, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 179, "usage_type": "call"}]}
+{"seq_id": "28020732345", "text": "import os\nfrom pathlib import Path\n\nfrom flask import Flask\nfrom flask_bootstrap import Bootstrap\nimport logging\n\ndef create_app(test_config=None):\n #create and configure application\n app = Flask(__name__, instance_relative_config=True)\n Bootstrap(app)\n app.config.from_mapping(\n SECRET_KEY = 'devvvv'\n )\n \n\n\n config_file = Path(app.root_path) / \"config.py\"\n\n if test_config is None:\n #Load instance config, if it exists, when not testing\n app.config.from_pyfile(config_file,silent=True)\n else:\n #Load the test config if passed in\n app.config.from_mapping(test_config)\n \n #ensure the instantce folder exist\n try:\n os.makedirs(app.instance_path)\n except OSError:\n pass\n \n #hello page for test\n @app.route('/hello')\n def hello():\n return \"Hello World\"\n \n #load application modules(blueprints)\n from . import flatfile_operations\n from . import backup_operations\n from . import scp_interface\n with app.app_context():\n app.register_blueprint(flatfile_operations.bp)\n app.register_blueprint(backup_operations.bp)\n app.register_blueprint(scp_interface.bp)\n\n #configure logger\n logging.basicConfig(filename='app.log',\n filemode='a',\n format='%(asctime)s - %(name)s - %(levelname)s - %(message)s',\n datefmt='%Y-%m-%d %H:%M:%S')\n app.logger = logging.getLogger('werkzeug')\n #app.logger.disabled = True\n \n return app", "repo_name": "GregZly/Flatfile_ed", "sub_path": "flatfileed/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1538, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "flask_bootstrap.Bootstrap", "line_number": 11, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 18, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 29, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 48, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 52, "usage_type": "call"}]}
+{"seq_id": "24468062966", "text": "import json\nimport requests\nimport datetime\nfrom datetime import date, timedelta\nimport codecs\n\n\ndef daterange(start_date, end_date):\n for n in range(int((end_date - start_date).days)):\n yield start_date + timedelta(n)\n\n\ndef list_to_str(k):\n d = \"\"\n for x in k:\n d = str(d) + \"-\" + str(x)\n return d[1:]\n\n\n# start_date = date(2007, 1, 1)\n# end_date = date(2007, 1, 2)\n# stations_data = {\"6\": [1711,44,46,202,43,42,45], \"152\": [1747],\n# \"153\": [1752], \"16\": [144,149,146,148,242,143,145], \"149\": [1725,1723,1724,1726],\n# \"7\": [49,54,61,57,211,53,50,55]}\n\nif __name__ == \"__main__\":\n start_date = date(2016, 4, 6)\n end_date = date(2016, 4, 7)\n stations_data = {\"6\": [44, 46, 202, 43, 42, 45], \"152\": [1747],\n \"153\": [1752], \"16\": [144, 149, 146, 148, 242, 143, 145], \"149\": [1725, 1723, 1724, 1726],\n \"7\": [49, 54, 61, 57, 211, 53, 50, 55]}\n\n for stat_number, stat_param in stations_data.items():\n for single_date in daterange(start_date, end_date):\n ref = 'http://monitoring.krakow.pios.gov.pl/dane-pomiarowe/automatyczne/stacja/' + str(\n stat_number) + '/parametry/' + str(list_to_str(stat_param))\n response = requests.post(\n url='http://monitoring.krakow.pios.gov.pl/dane-pomiarowe/pobierz',\n data={\n 'query': json.dumps({\n \"measType\": \"Auto\",\n \"viewType\": \"Station\",\n \"dateRange\": \"Day\",\n \"date\": str(single_date.strftime(\"%d.%m.%Y\")),\n \"viewTypeEntityId\": str(stat_number),\n \"channels\": stat_param\n })\n },\n headers={\n 'Referer': ref,\n 'Cookie': 'start_selector_nth=0; start_selector_hide=yes',\n 'Origin': 'http://monitoring.krakow.pios.gov.pl',\n 'Accept': 'application/json, text/javascript, */*; q=0.01',\n 'Accept-Encoding': 'gzip, deflate, lzma',\n 'DNT': '1',\n 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/48.0.2564.116 Safari/537.36 OPR/35.0.2066.92',\n 'Content-Type': 'application/x-www-form-urlencoded; charset=UTF-8',\n 'Host': 'monitoring.krakow.pios.gov.pl',\n 'X-Requested-With': 'XMLHttpRequest',\n 'Connection': 'keep-alive',\n 'Accept-Language': 'pl-PL,pl;q=0.8,en-US;q=0.6,en;q=0.4'\n }\n )\n\n d = response.json()\n print(str(stat_number) + ' <> ' + str(single_date.strftime(\"%Y-%m-%d\")), '-->', d['success'])\n json_file_name = 'smog/' + str(single_date.strftime(\"%Y-%m-%d\")) + \".json\"\n with codecs.open(json_file_name, \"w\", 'utf-8') as json_file:\n json.dump(d, json_file)\n\n file_name = 'smog/' + str(stat_number) + \"-\" + str(single_date.strftime(\"%Y-%m-%d\")) + \".csv\"\n\n\n def date_str(x):\n return datetime.datetime.fromtimestamp(int(x)).strftime('%H:%M')\n\n\n with codecs.open(file_name, \"w\", 'utf-8') as my_file:\n dane = {':'.join([str(x).zfill(2), '00']): {} for x in range(24)}\n print('CZAS', end=\"\", file=my_file)\n comp = []\n for k in d['data']['series']:\n compound_str = str(k['paramLabel'])\n if compound_str == \"Ozon\":\n compound_str = k['label'][:7]\n if compound_str == \"Tlenek węgla\":\n compound_str = k['label'][:15]\n\n if k['data']:\n comp.append(compound_str)\n\n for t, v in k['data']:\n ds = date_str(t)\n if ds not in dane: continue\n dane[ds][compound_str] = v\n\n print(';' + ';'.join(x for x in comp), file=my_file)\n\n for k in [':'.join([str(x).zfill(2), '00']) for x in range(24)]:\n line = k\n for c in comp:\n line += ';'\n line += \"\" if c not in dane[k] else str(dane[k][c])\n\n print(line, file=my_file)\n", "repo_name": "lukasz149/agh-smog-project", "sub_path": "scripts/download-smog.py", "file_name": "download-smog.py", "file_ext": "py", "file_size_in_byte": 4447, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "datetime.timedelta", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 28, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 37, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 40, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 68, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 69, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 75, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 75, "usage_type": "attribute"}, {"api_name": "codecs.open", "line_number": 78, "usage_type": "call"}]}
+{"seq_id": "72355325828", "text": "#!/usr/bin/env python\nfrom argparse import ArgumentParser\nfrom sklearn.utils import shuffle\nfrom sklearn.linear_model import SGDClassifier\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.metrics import roc_curve, auc\nimport os\nimport numpy\nimport scipy\nimport pylab\nimport torch\nimport torch.nn as nn\n \n \ninput_size = 949\nhidden_size = 1000\nnum_classes = 2\nnum_epochs = 10\nbatch_size = 1000\nlearning_rate = 0.01\n\nif __name__ == \"__main__\":\n cadd_dir = '/home/alex'\n ClinVar_ESP_dir = '/home/alex'\n \n print('Load Data')\n X_tr = numpy.load(os.path.join(cadd_dir, 'training.X.npz'))\n X_tr = scipy.sparse.csr_matrix((X_tr['data'], X_tr['indices'], X_tr['indptr']), shape=X_tr['shape'])\n y_tr = numpy.load(os.path.join(cadd_dir, 'training.y.npy'))\n \n X_va = numpy.load(os.path.join(cadd_dir, 'validation.X.npz'))\n X_va = scipy.sparse.csr_matrix((X_va['data'], X_va['indices'], X_va['indptr']), shape=X_va['shape'])\n y_va = numpy.load(os.path.join(cadd_dir, 'validation.y.npy')) \n \n X_te = numpy.load(os.path.join(cadd_dir, 'testing.X.npz'))\n X_te = scipy.sparse.csr_matrix((X_te['data'], X_te['indices'], X_te['indptr']), shape=X_te['shape'])\n y_te = numpy.load(os.path.join(cadd_dir, 'testing.y.npy'))\n \n X_ClinVar_ESP = numpy.load(os.path.join(ClinVar_ESP_dir, 'ClinVar_ESP.X.npz')) \n X_ClinVar_ESP = scipy.sparse.csr_matrix((X_ClinVar_ESP['data'], X_ClinVar_ESP['indices'], X_ClinVar_ESP['indptr']), shape=X_ClinVar_ESP['shape'])\n y_ClinVar_ESP = numpy.load(os.path.join(ClinVar_ESP_dir, 'ClinVar_ESP.y.npy'))\n\n\n\n\nx_dense_va = scipy.sparse.csc_matrix.todense(X_va)\nx_tensor_va = torch.from_numpy(x_dense_va)\n\nx_tensor_va, y_tensor_va\n\n#need to import data \nimport torch.utils.data\n#Define Batch Size \nbatch_size = 200000\n\n# Dataset \ntrain_loader = torch.utils.data.TensorDataset(X_train_tensor,y_train_tensor)\n \ntest_dataset = torch.utils.data.TensorDataset(x_tensor_va,y_tensor_va)\n# Data loader\ntrain_loader = torch.utils.data.DataLoader(dataset=train_dataset, batch_size=batch_size, shuffle=False)\ntest_loader = torch.utils.data.DataLoader(dataset=test_dataset, \n batch_size=batch_size, \n shuffle=False)\n\n#Defining Dataset from dataloader\ntest_dataset = torch.utils.data.TensorDataset(x_tensor_va,y_tensor_va)\ntest_loader = torch.utils.data.DataLoader(dataset=test_dataset, batch_size=batch_size, shuffle=False)\n\n#Constructing the neuralnet from scratch. We added 3 more layers (6 layers in total)\nclass NeuralNet(nn.Module):\n def __init__(self, input_size, hidden_size, num_classes):\n super(NeuralNet, self).__init__()\n self.fc1 = nn.Linear(input_size, hidden_size) \n self.relu = nn.ReLU()\n self.fc2 = nn.Linear(hidden_size, hidden_size) \n self.relu2 = nn.ReLU()\n self.fc3 = nn.Linear(hidden_size, hidden_size) \n self.relu3 = nn.ReLU()\n self.fc4 = nn.Linear(hidden_size, hidden_size) \n self.relu4 = nn.ReLU()\n self.fc5 = nn.Linear(hidden_size, hidden_size) \n self.relu5 = nn.ReLU()\n self.fc6 = nn.Linear(hidden_size, num_classes) \n self.dropout = nn.Dropout(0.1)\n \n \n def forward(self, x):\n out = self.fc1(x)\n out = self.relu(out)\n out = self.dropout(out)\n out = self.fc2(out)\n out = self.relu2(out)\n out = self.dropout(out)\n out = self.fc3(out)\n out = self.relu3(out)\n out = self.dropout(out)\n out = self.fc4(out)\n out = self.relu4(out)\n out = self.dropout(out)\n out = self.fc5(out)\n out = self.relu5(out)\n out = self.dropout(out)\n out = self.fc6(out)\n \n return out\n#Check if cpu is availble, for user's information\ntorch.cuda.is_available()\n\n# Device configuration \ndevice = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n# Loading the model \n\nmodel = NeuralNet(input_size, hidden_size, num_classes).to(device)\n\n# Loss and optimizer\ncriterion = nn.CrossEntropyLoss()\noptimizer = torch.optim.Adam(model.parameters(), lr=learning_rate) \n\n# Define the number of epochs\nnum_epochs = 10\n# Display the number of epoches\nprint ('batch_size = ' + batch_size)\n\n# Train the model\ntotal_step = len(train_loader)\nfor epoch in range(num_epochs):\n for i, (data, labels) in enumerate(train_loader): \n \n data = data.cuda()\n labels = labels.cuda()\n # Forward pass\n outputs = model(data)\n loss = criterion(outputs, labels)\n \n # Backward and optimize\n optimizer.zero_grad()\n loss.backward()\n optimizer.step()\n \n if (i+1) % 10 == 0:\n print ('Epoch [{}/{}], Step [{}/{}], Loss: {:.4f}' \n .format(epoch+1, num_epochs, i+1, total_step, loss.item()))\n# Let us save the model immediately \ntorch.save(model, 'savedodel.ckpt')\n\n\ntorch.load('model2.ckpt')\n\n# Test the model\n# In test phase, we don't need to compute gradients (for memory efficiency)\nwith torch.no_grad():\n correct = 0\n total = 0\n for images, labels in test_loader:\n images = images.to(device)\n labels = labels.to(device)\n outputs = model(images)\n _, predicted = torch.max(outputs.data, 1)\n total += labels.size(0)\n correct += (predicted == labels).sum().item()\n\n print('Accuracy of the network on the test set: {} %'.format(100 * correct / total))\n\n# Save the model checkpoint\ntorch.save(model.state_dict(), 'model2.ckpt')\n\n", "repo_name": "findalexli/DANN", "sub_path": "dann.py", "file_name": "dann.py", "file_ext": "py", "file_size_in_byte": 5555, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "numpy.load", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 28, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 32, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 32, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 36, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 40, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "scipy.sparse.csc_matrix.todense", "line_number": 46, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 46, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.utils.data.TensorDataset", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 57, "usage_type": "attribute"}, {"api_name": "torch.utils.data.TensorDataset", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 59, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 61, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 62, "usage_type": "attribute"}, {"api_name": "torch.utils.data.TensorDataset", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 67, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 68, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 71, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 71, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 74, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 75, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 77, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 80, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 81, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 82, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 83, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 84, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 85, "usage_type": "name"}, {"api_name": "torch.cuda.is_available", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 108, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 111, "usage_type": "attribute"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 117, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 118, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 118, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 145, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 152, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 159, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 166, "usage_type": "call"}]}
+{"seq_id": "34307733271", "text": "import new_crank\r\nimport original_crank\r\nimport time\r\nimport matplotlib.pyplot as plt\r\n\r\ndef test_new_tridiag():\r\n time1 = time.time()\r\n for _ in range(200):\r\n new_crank.price_american_option_with_divs()\r\n time2 = time.time()\r\n print(\"New: %.4f seconds\" % (time2-time1))\r\n\r\ndef test_one_new():\r\n new_crank.price_american_option_with_divs()\r\n start = time.time()\r\n new_crank.price_american_option_with_divs()\r\n end = time.time()\r\n print(\"Time to calc one option price: %.8f seconds\" % (end-start))\r\n\r\ndef test_old_method():\r\n time1 = time.time()\r\n for _ in range(200):\r\n original_crank.price_american_option_with_divs()\r\n time2 = time.time()\r\n print(\"Original: %.4f seconds\" % (time2-time1))\r\n\r\n# Tests 100 calculations of the original and new crank-nicolson method, measuring runtime\r\ndef test_crank():\r\n time1 = time.time()\r\n for _ in range(200):\r\n original_crank.price_american_option_with_divs()\r\n time2 = time.time()\r\n for _ in range(200):\r\n new_crank.price_american_option_with_divs()\r\n time3 = time.time()\r\n print(\"Original: %.4f seconds\" % (time2-time1))\r\n print(\"New: %.4f seconds\" % (time3-time2))\r\n\r\n# Tests 2000 calculations each of the original and new implementations of the crank-nicolson method, measuring runtime every 100 calculations\r\ndef test_crank_plot():\r\n time_orig_start = time.time()\r\n original_times = [0]\r\n for i in range(20):\r\n for _ in range(100):\r\n original_crank.price_american_option_with_divs()\r\n original_times.append(time.time() - time_orig_start)\r\n print(\"loop %d complete, time = %.4f seconds\" % (i+1, original_times[i+1]))\r\n new_crank.price_american_option_with_divs()\r\n time_new_start = time.time()\r\n new_times = [0]\r\n for i in range(20):\r\n for _ in range(100):\r\n new_crank.price_american_option_with_divs()\r\n new_times.append(time.time() - time_new_start)\r\n print(\"loop %d complete, time = %.4f seconds\" % (i+1, new_times[i+1]))\r\n x = range(0, 2001, 100)\r\n plt.plot(x, original_times, \"r^-\", label=\"Original\")\r\n plt.plot(x, new_times, \"bo-\", label=\"New\")\r\n plt.legend()\r\n plt.title(\"Comparison of Crank-Nicolson Method Implementations\\n(Ryzen 3700X, RTX 2080S, 32GB of RAM)\")\r\n plt.xlabel(\"Number of calculations\")\r\n plt.ylabel(\"Time (seconds)\")\r\n plt.annotate(\"%.2f\" % (original_times[-1]), xy=(x[-1], original_times[-1]), xytext=(x[-1], original_times[-1]))\r\n plt.annotate(\"%.2f\" % (new_times[-1]), xy=(x[-1], new_times[-1]), xytext=(x[-1], new_times[-1]))\r\n plt.show() \r\n plt.grid()\r\n plt.savefig(\"crank_plot.png\")\r\n\r\nif __name__ == \"__main__\":\r\n #test_crank_plot()\r\n test_one_new()\r\n #test_new_tridiag()\r\n", "repo_name": "EzePze/AmericanOptionPricing", "sub_path": "tester.py", "file_name": "tester.py", "file_ext": "py", "file_size_in_byte": 2766, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "26", "api": [{"api_name": "time.time", "line_number": 7, "usage_type": "call"}, {"api_name": "new_crank.price_american_option_with_divs", "line_number": 9, "usage_type": "call"}, {"api_name": "time.time", "line_number": 10, "usage_type": "call"}, {"api_name": "new_crank.price_american_option_with_divs", "line_number": 14, "usage_type": "call"}, {"api_name": "time.time", "line_number": 15, "usage_type": "call"}, {"api_name": "new_crank.price_american_option_with_divs", "line_number": 16, "usage_type": "call"}, {"api_name": "time.time", "line_number": 17, "usage_type": "call"}, {"api_name": "time.time", "line_number": 21, "usage_type": "call"}, {"api_name": "original_crank.price_american_option_with_divs", "line_number": 23, "usage_type": "call"}, {"api_name": "time.time", "line_number": 24, "usage_type": "call"}, {"api_name": "time.time", "line_number": 29, "usage_type": "call"}, {"api_name": "original_crank.price_american_option_with_divs", "line_number": 31, "usage_type": "call"}, {"api_name": "time.time", "line_number": 32, "usage_type": "call"}, {"api_name": "new_crank.price_american_option_with_divs", "line_number": 34, "usage_type": "call"}, {"api_name": "time.time", "line_number": 35, "usage_type": "call"}, {"api_name": "time.time", "line_number": 41, "usage_type": "call"}, {"api_name": "original_crank.price_american_option_with_divs", "line_number": 45, "usage_type": "call"}, {"api_name": "time.time", "line_number": 46, "usage_type": "call"}, {"api_name": "new_crank.price_american_option_with_divs", "line_number": 48, "usage_type": "call"}, {"api_name": "time.time", "line_number": 49, "usage_type": "call"}, {"api_name": "new_crank.price_american_option_with_divs", "line_number": 53, "usage_type": "call"}, {"api_name": "time.time", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.annotate", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.annotate", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}]}
+{"seq_id": "41931699031", "text": "from omegaconf import DictConfig\nfrom claficle.models.sandwich import Sandwich\nfrom claficle.models.plain_gpt2 import PlainGPT2\nfrom claficle.models.gewechselt import Gewechselt\nfrom claficle.models.vessel import Vessel\n\nNAME_TO_CLASS = {\n \"sandwich\": Sandwich,\n \"plain_gpt2\": PlainGPT2,\n \"Gewechselt\": Gewechselt,\n \"vessel\": Vessel,\n}\n\n\ndef get_model_preamble_post_init_kwargs(cfg: DictConfig):\n NAME_TO_KWARGS = {\"vessel\": {\"seed\": cfg.seed}}\n try:\n kwargs = NAME_TO_KWARGS[cfg.model.name]\n except KeyError:\n raise ValueError(\n \"This model class either does not have a post_init or its post_init is not \"\n \"expected in preamble\"\n )\n return kwargs\n", "repo_name": "thesofakillers/CLAfICLe", "sub_path": "claficle/models/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 716, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3, "dataset": "github-code", "pt": "26", "api": [{"api_name": "claficle.models.sandwich.Sandwich", "line_number": 8, "usage_type": "name"}, {"api_name": "claficle.models.plain_gpt2.PlainGPT2", "line_number": 9, "usage_type": "name"}, {"api_name": "claficle.models.gewechselt.Gewechselt", "line_number": 10, "usage_type": "name"}, {"api_name": "claficle.models.vessel.Vessel", "line_number": 11, "usage_type": "name"}, {"api_name": "omegaconf.DictConfig", "line_number": 15, "usage_type": "name"}]}
+{"seq_id": "37590665672", "text": "# -*- coding: utf-8 -*-\nfrom ppdai_utils import *\nimport pickle\nfrom keras.models import load_model\nfrom myloss import focal_loss\n\ncdict, emat = embedding('/files/faust/COMPETITION/ppdai/char_embed.txt')\nquestions = pickle.load(open('/files/faust/COMPETITION/ppdai/questions.pkl', 'rb'))\ntestpair = readtest()\n\n\nMAX_SEQUENCE_LENGTH = 60\ntestpair_idx = []\nfor q0, q1 in testpair:\n q0_chars = questions[q0]['cchars']\n q1_chars = questions[q1]['cchars']\n q0_idx = sent2idx(q0_chars, cdict, MAX_SEQUENCE_LENGTH)\n q1_idx = sent2idx(q1_chars, cdict, MAX_SEQUENCE_LENGTH)\n testpair_idx.append((q0_idx, q1_idx))\nprint(len(testpair_idx))\n\ntest_q1 = np.array([p[0] for p in testpair_idx])\ntest_q2 = np.array([p[1] for p in testpair_idx])\n\n\n###############################################\n# test\n###############################################\nmodel = load_model('ppdai_cnn.model')\npred_y = model.predict([test_q1, test_q2], batch_size=64)\nprint(pred_y.shape)\npred_y = np.reshape(pred_y, (len(pred_y),))\nmake_submission(pred_y, 'submission.csv')\n\n", "repo_name": "MingYates/QMATCH", "sub_path": "ppdai/ppdai_cnn_test.py", "file_name": "ppdai_cnn_test.py", "file_ext": "py", "file_size_in_byte": 1051, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "pickle.load", "line_number": 8, "usage_type": "call"}, {"api_name": "keras.models.load_model", "line_number": 29, "usage_type": "call"}]}
+{"seq_id": "26609709198", "text": "from django.http import HttpRequest, HttpResponse, Http404\r\nfrom django.shortcuts import render\r\n\r\ncity = {\r\n 'msk': 'Москвы',\r\n 'spb': 'Питера',\r\n 'nsb': 'Новосибирска',\r\n 'ekb': 'ЕКБ',\r\n 'kaz': 'Казани',\r\n}\r\n\r\ntour_id = {}\r\nfor i in range(1000):\r\n tour_id[i] = i\r\n\r\n\r\ndef main_view(request: HttpRequest):\r\n return render(request, 'tours/index.html')\r\n\r\n\r\ndef departure_view(request, departure):\r\n try:\r\n departure_city = city[departure]\r\n except KeyError:\r\n raise Http404\r\n return render(request, 'tours/departure.html', context={'departure_view': departure_city})\r\n\r\n\r\ndef tour_view(request, id):\r\n try:\r\n tour_page = tour_id[id]\r\n except KeyError:\r\n raise Http404\r\n return render(request, 'tours/tour.html', context={'tour_view': tour_page})\r\n", "repo_name": "art213/stepik_demo", "sub_path": "stepik_tours/tours/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 847, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "django.http.HttpRequest", "line_number": 17, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 18, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 25, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 26, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 33, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 34, "usage_type": "call"}]}
+{"seq_id": "1314294075", "text": "from django.shortcuts import render, redirect\nfrom django.contrib.auth.decorators import login_required\n\nfrom .forms import CreateProfileForm, CreatePromDateForm\nfrom .models import Profile, PromDate\n\n# Create your views here.\n\ndef index(request):\n return render(request, 'promApp/index.html')\n\ndef about(request):\n return render(request, 'promApp/about.html')\n\ndef handler404(request, exception):\n return render(request, 'promApp/err404.html', status=404)\n \ndef handler500(request, *args, **argv): \n return render(request, 'promApp/err500.html', status=500)\n\ndef sorry(request):\n return render(request, \"promApp/sorry.html\")\n\ndef guide(request):\n return render(request, \"promApp/guide.html\")\n\n@login_required(login_url=\"/google/login\")\ndef app(request):\n if request.method == \"POST\":\n if \"profile\" in request.POST:\n profileForm = CreateProfileForm(data=request.POST)\n if profileForm.is_valid():\n profile = profileForm.save(commit=False)\n profile.owner = request.user \n profile.save()\n return redirect(\"promApp:app\")\n elif \"crush\" in request.POST:\n \n crushNameList = PromDate.objects.filter(profile=Profile.objects.filter(owner=request.user).first()).values_list('name', flat=True)\n \n crushName = request.POST['name'].title().strip()\n crushYear = request.POST['year']\n # if inputted name is same as self\n if crushName == request.user.get_full_name():\n return redirect(\"promApp:app\")\n \n # if inputted name already inputted\n if crushName in crushNameList:\n return redirect(\"promApp:app\")\n \n if crushYear == \"FR\" or crushYear == \"SO\":\n return redirect(\"promApp:sorry\")\n \n crushForm = CreatePromDateForm(data=request.POST)\n if crushForm.is_valid():\n crusher = Profile.objects.filter(owner=request.user).first()\n crush = crushForm.save(commit=False)\n crush.profile = crusher\n crush.name = crushName\n crush.save()\n return redirect(\"promApp:app\")\n \n return render(request, \"promApp/app.html\", context={})\n else:\n profileNotCreated = len(Profile.objects.filter(owner=request.user)) == 0\n profileForm = CreateProfileForm()\n crushForm = CreatePromDateForm()\n\n if not profileNotCreated:\n\n crusher = Profile.objects.filter(owner=request.user).first()\n \n if crusher.year == \"FR\" or crusher.year == \"SO\":\n return redirect(\"promApp:sorry\")\n\n remaining = 5-PromDate.objects.filter(profile=crusher).count()\n\n crushList = PromDate.objects.filter(profile=crusher).all()\n crushingOnMe = PromDate.objects.filter(name=request.user.get_full_name(), year=crusher.year)\n\n numCrushing = crushingOnMe.count()\n\n matches = []\n for person in crushingOnMe:\n for crush in crushList:\n\n if (person.profile.owner.get_full_name() == crush.name) and (crush.year == person.profile.year):\n matches.append((crush.name, person.profile.year))\n \n \n context = {\n \"remaining\" : remaining,\n \"matches\" : matches,\n \"profileNotCreated\" : profileNotCreated,\n \"profileForm\" : profileForm,\n \"crushForm\" : crushForm,\n \"crushList\": crushList,\n \"numCrushing\" : numCrushing,\n }\n else:\n \n context = {\n \"profileNotCreated\" : profileNotCreated,\n \"profileForm\" : profileForm,\n \"crushForm\" : crushForm,\n }\n\n return render(request, \"promApp/app.html\", context=context)\n\n\n\n\n\n", "repo_name": "EdwardX29/nwprom", "sub_path": "promApp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4003, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "django.shortcuts.render", "line_number": 10, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 13, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 16, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 19, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 22, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 25, "usage_type": "call"}, {"api_name": "forms.CreateProfileForm", "line_number": 31, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 36, "usage_type": "call"}, {"api_name": "models.PromDate.objects.filter", "line_number": 39, "usage_type": "call"}, {"api_name": "models.PromDate.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "models.PromDate", "line_number": 39, "usage_type": "name"}, {"api_name": "models.Profile.objects.filter", "line_number": 39, "usage_type": "call"}, {"api_name": "models.Profile.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "models.Profile", "line_number": 39, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 45, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 49, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 52, "usage_type": "call"}, {"api_name": "forms.CreatePromDateForm", "line_number": 54, "usage_type": "call"}, {"api_name": "models.Profile.objects.filter", "line_number": 56, "usage_type": "call"}, {"api_name": "models.Profile.objects", "line_number": 56, "usage_type": "attribute"}, {"api_name": "models.Profile", "line_number": 56, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 61, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Profile.objects.filter", "line_number": 65, "usage_type": "call"}, {"api_name": "models.Profile.objects", "line_number": 65, "usage_type": "attribute"}, {"api_name": "models.Profile", "line_number": 65, "usage_type": "name"}, {"api_name": "forms.CreateProfileForm", "line_number": 66, "usage_type": "call"}, {"api_name": "forms.CreatePromDateForm", "line_number": 67, "usage_type": "call"}, {"api_name": "models.Profile.objects.filter", "line_number": 71, "usage_type": "call"}, {"api_name": "models.Profile.objects", "line_number": 71, "usage_type": "attribute"}, {"api_name": "models.Profile", "line_number": 71, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 74, "usage_type": "call"}, {"api_name": "models.PromDate.objects.filter", "line_number": 76, "usage_type": "call"}, {"api_name": "models.PromDate.objects", "line_number": 76, "usage_type": "attribute"}, {"api_name": "models.PromDate", "line_number": 76, "usage_type": "name"}, {"api_name": "models.PromDate.objects.filter", "line_number": 78, "usage_type": "call"}, {"api_name": "models.PromDate.objects", "line_number": 78, "usage_type": "attribute"}, {"api_name": "models.PromDate", "line_number": 78, "usage_type": "name"}, {"api_name": "models.PromDate.objects.filter", "line_number": 79, "usage_type": "call"}, {"api_name": "models.PromDate.objects", "line_number": 79, "usage_type": "attribute"}, {"api_name": "models.PromDate", "line_number": 79, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 108, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 27, "usage_type": "call"}]}
+{"seq_id": "13897086625", "text": "\"\"\"\nbase_trainer.py has one class BaseTrainer.\n\nBaseTrainer provides base functionality for any trainer object.\nIt provides functionality to:\n - train for one epoch\n - validation \n - testing\n - freezing model\n - unfreezing model\n - saving checkpoint\n - loading checkpoint\n\"\"\"\nimport torch\nfrom torchvision import models\nimport math\nimport matplotlib.pyplot as plt\nimport copy\nimport os\nimport datetime\nfrom base import base_trainer\n\n\nclass BaseTrainer:\n \"\"\"Base Class for trainer/train.py.\"\"\"\n\n def __init__(self, model, train_dl, valid_dl, test_dl, criterion):\n \"\"\"Initialize the BaseTrainer object.\"\"\"\n self.model = model\n self.train_dl = train_dl\n self.valid_dl = valid_dl\n self.test_dl = test_dl\n self.criterion = criterion\n self.opt_lrs = []\n\n def train_epoch(self, optimizer, scheduler):\n \"\"\"\n Perform one training epoch.\n\n Parameters:\n optimizer - optimizer to use while training\n scheduler - scheduler to use while training\n Returns: training loss after epoch\n\n \"\"\"\n self.model.train()\n\n final_loss = None\n\n for inputs, labels in self.train_dl:\n\n inputs = inputs.to(self.device)\n labels = labels.to(self.device)\n loss = None\n\n optimizer.zero_grad()\n\n outputs = self.model(inputs)\n loss = self.criterion(outputs, labels)\n loss.backward()\n\n optimizer.step()\n self.opt_lrs.append(optimizer.state_dict()['param_groups'][0]['lr'])\n scheduler.step()\n\n final_loss = float(loss)\n\n del(loss)\n\n torch.cuda.empty_cache()\n\n del(inputs)\n del(labels)\n del(outputs)\n\n return final_loss\n\n def valid_epoch(self):\n \"\"\"\n Perform one validation epoch.\n\n Returns:\n val_loss - Validation loss \n val_acc - Validation accuracy\n\n \"\"\"\n self.model.eval()\n\n running_loss = 0.0\n running_corrects = 0\n\n with torch.no_grad():\n\n for val_inputs, val_labels in self.valid_dl:\n val_inputs = val_inputs.to(self.device)\n val_labels = val_labels.to(self.device)\n\n val_loss = None\n\n val_outputs = self.model(val_inputs)\n val_preds, val_indices = torch.max(val_outputs, dim=1)\n\n val_loss = self.criterion(val_outputs, val_labels)\n\n # float(val_loss) => the float() is the most important thing\n # to avoid cuda out of memory error\n # https://pytorch.org/docs/stable/notes/faq.html#my-model-reports-cuda-runtime-error-2-out-of-memory\n running_loss += float(val_loss)\n running_corrects += torch.sum(val_indices == val_labels)\n\n del(val_loss)\n\n torch.cuda.empty_cache()\n\n del(val_inputs)\n del(val_labels)\n del(val_outputs)\n\n epoch_loss = running_loss / len(self.valid_dl.dataset)\n epoch_acc = running_corrects.double() / len(self.valid_dl.dataset)\n\n return [epoch_loss, epoch_acc]\n\n def test_epoch(self):\n \"\"\"\n Perform a test epoch using current model.\n\n Returns accuracy on test set\n \"\"\"\n self.model.eval()\n\n corrects = 0\n print('Performing Test Epoch')\n for test_inputs, test_labels in self.test_dl:\n test_inputs = test_inputs.to(self.device)\n test_labels = test_labels.to(self.device)\n\n test_output = self.model(test_inputs)\n test_output_labels = torch.max(test_output, dim=1)[1]\n corrects += int(torch.sum(test_output_labels == test_labels))\n\n del(test_inputs)\n del(test_labels)\n del(test_output)\n\n return corrects/len(self.test_dl.dataset)\n\n def unfreeze(self):\n \"\"\"Unfreeze all layers of model.\"\"\"\n for param in self.model.parameters():\n param.requires_grad = True\n\n def freeze(self):\n \"\"\"Freeze all layers of model.\"\"\"\n for param in self.model.parameters():\n param.requires_grad = False\n\n def save_checkpoint(self, path, optimizer, scheduler, cycle, train_loss,\n valid_loss, valid_acc):\n \"\"\"\n Save checkpoint into given path.\n\n Parameters:\n path - path where checkpoint would be saved\n optimizer - optimizer's current state_dict to save\n scheduler - scheduler's current state_dict to save\n cycle - current cycle number (in SGDR) used for filename\n train_loss - training loss at the end of cycle \n valid_loss - validation loss at the end of cycle\n valid_acc - validation accuracy at the end of cycle\n\n \"\"\"\n state = {\n 'model': self.model.state_dict(),\n 'optimizer': optimizer.state_dict(),\n 'scheduler': scheduler.state_dict(),\n 'cycle': cycle,\n 'train_loss': train_loss,\n 'valid_loss': valid_loss,\n 'valid_acc': valid_acc\n }\n\n filename = f'cycle_{cycle}'\n torch.save(state, f'{path}/{filename}')\n\n def load_checkpoint(self, path):\n \"\"\"\n Load checkpoint from given path.\n\n Returns state dictionary with keys:\n model - model's state dictionary\n optimizer - optimizer's state dictionary\n scheduler - scheduler's state dictionary\n cycle - cycle number when checkpoint was saved\n train_loss - training loss when checkpoint was saved\n valid_loss - validation loss when checkpoint was saved\n valid_acc - validation accuracy when checkpoint was saved\n\n \"\"\"\n state = torch.load(path)\n return state\n", "repo_name": "jayeshsaita/Speech-Commands-Recognition", "sub_path": "base/base_trainer.py", "file_name": "base_trainer.py", "file_ext": "py", "file_size_in_byte": 5873, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 35, "dataset": "github-code", "pt": "26", "api": [{"api_name": "torch.cuda.empty_cache", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 70, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.cuda.empty_cache", "line_number": 113, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 113, "usage_type": "attribute"}, {"api_name": "torch.max", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 184, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 200, "usage_type": "call"}]}
+{"seq_id": "70448737030", "text": "import itertools\nimport json\nimport math\nimport pprint\nimport time\n\nimport nltk\nfrom nltk import PorterStemmer\nfrom sklearn.linear_model import Perceptron\nfrom sklearn.metrics import precision_score, f1_score, recall_score, average_precision_score\nfrom truecase import get_true_case\n\nfrom project_ex2 import getDataFromDir\nfrom project_ex3 import getAllCandidates, getTFIDFScore, listOfTaggedToString, getBM25Score\nimport sklearn.metrics\n\n# nltk.download('maxent_ne_chunker')\n# nltk.download('words')\nfrom project_p2_ex1 import Metrics, Helper\n\n\ndef createTargetList(reference, term_list):\n target = {}\n with open(reference) as f:\n reference_results = json.load(f)\n for i, (name, doc_values) in enumerate(reference_results.items()):\n classes = []\n\n for term in term_list[i]:\n\n values = list(itertools.chain.from_iterable(doc_values))\n\n porter = PorterStemmer()\n\n stemmed = \"\"\n\n for w in term.split():\n stemmed += porter.stem(w) + \" \"\n stemmed = stemmed[:-1]\n\n # s_term = porter.stem(term)\n\n if stemmed in values:\n classes.append(1)\n else:\n classes.append(0)\n\n target.update({name: classes})\n return target\n\n\ndef calculateParameters(all_cands, doc, scores):\n params = []\n\n max_cand_score = max(scores.values())\n\n for cand in all_cands:\n\n freq = doc.count(cand)\n\n if cand not in scores:\n cand_score = 0.\n else:\n cand_score = scores[cand] # / max_cand_score\n\n cand_len = len(cand)\n cand_term_count = len(cand.split())\n ne_cand = get_true_case(cand)\n words = nltk.pos_tag(nltk.word_tokenize(ne_cand))\n ne = nltk.tree2conlltags(nltk.ne_chunk(words))\n ne = [' '.join(word for word, pos, chunk in group).lower()\n for key, group in itertools.groupby(ne, lambda tpl: tpl[2] != 'O') if key]\n\n ne_cnt = len(ne[0].split()) if ne else 0\n\n first_match = doc.find(cand) / len(doc)\n last_match = doc.rfind(cand) / len(doc)\n\n # if cand_term_count == 1:\n # cohesion = 0.\n # else:\n # cohesion = cand_term_count * (1 + math.log(freq, 10)) * freq /\n\n if first_match == last_match:\n spread = 0.\n else:\n spread = last_match - first_match\n\n # print([cand_score, freq, cand_len, cand_term_count, first_match, last_match, spread, ne_cnt])\n\n params.append([cand_score, cand_len, cand_term_count, first_match, last_match, spread, ne_cnt]) #cand_score,\n return params\n\n\ndef calcResults(predicted, true):\n # , average_precision_score(true, predicted)\n return precision_score(true, predicted), recall_score(true, predicted), f1_score(true, predicted)\n\np_classifier = Perceptron(alpha=0.1)\n\ntrain = getDataFromDir('ake-datasets-master/datasets/500N-KPCrowd/train', mode='list')\ntrainStr = listOfTaggedToString(train)\n\ntest = getDataFromDir('ake-datasets-master/datasets/500N-KPCrowd/test', mode='list')\ntestStr = listOfTaggedToString(test)\n\nallCandidatesTrain = getAllCandidates(train)\nallCandidatesTest = getAllCandidates(test)\n\n\n# bm25\n# 0.3558736870896098\n# 0.7640337163696295\n# 0.4607649659785287\n\n# TF IDF\n# 0.37863851957992073\n# 0.31571002226187983\n# 0.3159382700815522\n\nbm25train = getBM25Score(train, mergetype='dict', min_df=1)\nbm25test = getBM25Score(test, mergetype='dict', min_df=1)\n\ntargets = createTargetList('ake-datasets-master/datasets/500N-KPCrowd/references/train.reader.stem.json',\n allCandidatesTrain)\n\ntestTargets = createTargetList('ake-datasets-master/datasets/500N-KPCrowd/references/test.reader.stem.json',\n allCandidatesTest)\n\nfor doc_index, doc_name in enumerate(train.keys()):\n allParams = calculateParameters(allCandidatesTrain[doc_index], trainStr[doc_index], bm25train[doc_name])\n if not targets[doc_name].count(0) == len(targets[doc_name]):\n p_classifier.fit(allParams, targets[doc_name])\n\nprint('predict')\n\nprecision = []\nrecall = []\nf1 = []\nap = []\ntr = Helper.getTrueKeyphrases('ake-datasets-master/datasets/500N-KPCrowd/references/test.reader.stem.json')\nkfs = {}\n\nfor doc_index, doc_name in enumerate(test.keys()):\n params = calculateParameters(allCandidatesTest[doc_index], testStr[doc_index], bm25test[doc_name])\n\n predicted = p_classifier.predict(params)\n plane = p_classifier.decision_function(params)\n true = testTargets[doc_name]\n\n print('PERCEPTRON')\n print(predicted)\n print('[P2]', plane)\n print('REALITY')\n print(true)\n\n rnk = {list(bm25test[doc_name].keys())[i]: v for i, v in enumerate(plane) if v > 0}\n rnk = list(dict(Helper.dictToOrderedList(rnk, rev=True)).keys())\n p, r, f = calcResults(predicted, true)\n kfs[doc_name] = rnk\n\n #precision.append(p)\n #recall.append(r)\n #f1.append(f)\n #ap.append(aps)\n\nmeanAPre, meanPre, meanRe, meanF1 = Helper.results(kfs, 'ake-datasets-master/datasets/500N-KPCrowd/references/test'\n '.reader.stem.json')\nprint('--RESULTS--')\nprint('Precision = ', meanPre)\nprint('Recall = ', meanRe)\nprint('F1 = ', meanF1)\nprint('Mean AVG Precision = ', meanAPre)\n\n# for dos documentos\n# para cada doc_name extrair candidatos\n# para cada candidato calcular os parâmetros\n\n# for dos documentos\n# passamos a lista que contem os parametros de todos os candidatos\n# calculamos a lista de resultados [ 0 0 0 1 0 0 1 ]\n# fit\n", "repo_name": "davidfrickert/PRI-19-20", "sub_path": "project_ex4.py", "file_name": "project_ex4.py", "file_ext": "py", "file_size_in_byte": 5597, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "json.load", "line_number": 25, "usage_type": "call"}, {"api_name": "itertools.chain.from_iterable", "line_number": 31, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 31, "usage_type": "attribute"}, {"api_name": "nltk.PorterStemmer", "line_number": 33, "usage_type": "call"}, {"api_name": "truecase.get_true_case", "line_number": 68, "usage_type": "call"}, {"api_name": "nltk.pos_tag", "line_number": 69, "usage_type": "call"}, {"api_name": "nltk.word_tokenize", "line_number": 69, "usage_type": "call"}, {"api_name": "nltk.tree2conlltags", "line_number": 70, "usage_type": "call"}, {"api_name": "nltk.ne_chunk", "line_number": 70, "usage_type": "call"}, {"api_name": "itertools.groupby", "line_number": 72, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 97, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 97, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 97, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Perceptron", "line_number": 99, "usage_type": "call"}, {"api_name": "project_ex2.getDataFromDir", "line_number": 101, "usage_type": "call"}, {"api_name": "project_ex3.listOfTaggedToString", "line_number": 102, "usage_type": "call"}, {"api_name": "project_ex2.getDataFromDir", "line_number": 104, "usage_type": "call"}, {"api_name": "project_ex3.listOfTaggedToString", "line_number": 105, "usage_type": "call"}, {"api_name": "project_ex3.getAllCandidates", "line_number": 107, "usage_type": "call"}, {"api_name": "project_ex3.getAllCandidates", "line_number": 108, "usage_type": "call"}, {"api_name": "project_ex3.getBM25Score", "line_number": 121, "usage_type": "call"}, {"api_name": "project_ex3.getBM25Score", "line_number": 122, "usage_type": "call"}, {"api_name": "project_p2_ex1.Helper.getTrueKeyphrases", "line_number": 141, "usage_type": "call"}, {"api_name": "project_p2_ex1.Helper", "line_number": 141, "usage_type": "name"}, {"api_name": "project_p2_ex1.Helper.dictToOrderedList", "line_number": 158, "usage_type": "call"}, {"api_name": "project_p2_ex1.Helper", "line_number": 158, "usage_type": "name"}, {"api_name": "project_p2_ex1.Helper.results", "line_number": 167, "usage_type": "call"}, {"api_name": "project_p2_ex1.Helper", "line_number": 167, "usage_type": "name"}]}
+{"seq_id": "34087072610", "text": "# %% [markdown]\n# # Credit-Score-Classification\n# \n# - Projeto de classificação de clientes de acordo com seus dados pessoais e financeiros. Dataset disponível em https://www.kaggle.com/laotse/credit-risk-dataset.\n# - GitHub : https://github.com/Ewertonv90/Credit-Score-Classification\n# \n# \n# \n# # EN\n# \n# Problem Statement\n# You are working as a data scientist in a global finance company. Over the years, the company has collected basic bank details and gathered a lot of credit-related information. The management wants to build an intelligent system to segregate the people into credit score brackets to reduce the manual efforts.\n# \n# Task\n# Given a person’s credit-related information, build a machine learning model that classifies the customer if credit can be released or not.\n# \n# # PT-BR\n# \n# Declaração do problema\n# Você está trabalhando como cientista de dados em uma empresa financeira global. Ao longo dos anos, a empresa coletou dados bancários básicos e reuniu muitas informações relacionadas a crédito. A gerência quer construir um sistema inteligente para segregar as pessoas em faixas de pontuação de crédito para reduzir os esforços manuais.\n# \n# Tarefa\n# Dadas as informações relacionadas ao crédito de uma pessoa, construa um modelo de aprendizado de máquina que possa classificar o cliente e se o crédito deve ser liberado ou não.\n# \n# Bussiness Sucess Criteria : more or equal to 85%\n# \n# \n# # Data dictonary\n# \n# \n# \n# - person_age = Age\n# - person_income\t = Annual Income\n# - personhomeownership\t = Home ownership\n# - personemplength\t = Employment length (in years)\n# - loan_intent\t = Loan intent\n# - loan_grade\t = Loan grade\n# - loan_amnt\t = Loan amount\n# - loanintrate\t = Interest rate\n# - loan_status\t = Loan status (0 is non default 1 is default)\n# - loanpercentincome\t = Percent income\n# - cbpersondefaultonfile\t = Historical default\n# - cbpresoncredhistlength\t= Credit history length\n\nimport pandas as pd\nimport numpy as np\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.preprocessing import Normalizer\nfrom sklearn.compose import ColumnTransformer\nfrom sklearn.impute import SimpleImputer\nfrom sklearn.preprocessing import OneHotEncoder\nfrom sklearn.model_selection import GridSearchCV\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.ensemble import GradientBoostingClassifier\nfrom sklearn.metrics import f1_score,accuracy_score, precision_score, recall_score\nimport joblib\n# Data Engeneering\n\ndf = pd.read_csv('data/credit_risk_dataset.csv')\n\nindex = df[ df['person_age'] >= 100 ].index\ndf.drop(index, inplace=True)\n\ndf = df.dropna(how=\"all\")\n\ndf = df.drop_duplicates(keep='first')\n\nX = df.drop(columns=['loan_status'])\ny = df['loan_status']\n\n\n\n# Train and Test split\nX_train, X_test, y_train, y_test = train_test_split(X, y , test_size=0.2, random_state=42)\n\n# PIPELINE\n\npipe = Pipeline(steps=[\n \n ('encoder', ColumnTransformer(\n [\n ('encoder_type', OneHotEncoder(drop='first'),['loan_intent','loan_grade','person_home_ownership', 'cb_person_default_on_file'])\n ]\n )\n ),\n ('imputer', SimpleImputer(strategy='mean',fill_value=np.nan)),\n ('normalizer', Normalizer()),\n ('classifier', GradientBoostingClassifier(max_depth=5, min_samples_split=3,\n n_estimators=300))\n ]\n)\n\n\npipe.fit(X_train, y_train)\n\nprint(round(pipe.score(X_train, y_train),4)*100, \"%\")\nprint(round(pipe.score(X_test, y_test),4)*100 , \"%\")\n\ny_pred = pipe.predict(X_test)\n\nparam_grid = {\n 'classifier__n_estimators' : [300],\n 'classifier__max_depth' : [5],\n 'classifier__min_samples_split': [3],\n}\n\n# param_grid = {\n# 'classifier__n_estimators' : [100, 200, 300],\n# 'classifier__learning_rate' : [0.1, 0.5, 1.0],\n# 'classifier__subsample': [0.1, 0.3, 0.5, 0.8, 1.0],\n# 'classifier__max_depth' : [1, 3, 5, 10],\n# 'classifier__min_samples_leaf' : [1,3, 5, 10],\n# 'classifier__min_samples_split': [2, 3, 5, 10],\n# }\n\ngrid = GridSearchCV(estimator=pipe, param_grid=param_grid,cv=3 , n_jobs= -1, verbose=0, )\ngrid.fit(X_train, y_train)\nprint(grid.score(X_train, y_train))\nprint(grid.score(X_test, y_test))\n\nf1 = round(f1_score(y_test, y_pred, average=\"micro\")*100, 2)\naccuracy = round(accuracy_score(y_test, y_pred)*100, 2)\nprecision = round(precision_score(y_test, y_pred)*100 , 2)\nrecall = round(recall_score(y_test, y_pred, average='micro')*100 , 2)\n\nprint(f\"F1 Score: {f1}%\")\nprint(f\"Accuracy Score: {accuracy}%\")\nprint(f\"Precision Score: {precision}%\")\nprint(f\"Recall Score: {recall}%\")\n\nmetrics = {}\n\nmetrics = {\n\n \"accuracy\" : accuracy,\n \"precision\" : precision,\n \"recall\" : recall,\n \"f1\" : f1\n }\n\nbest_pipeline = grid.best_estimator_\n\njoblib.dump(metrics, filename=\"log/model_score_train.pkl\")\njoblib.dump(best_pipeline, filename=\"src/deployment/models/model_pipeline.pkl\")\n\n\n\n\n", "repo_name": "Ewertonv90/Credit-Score-Classification_Risk", "sub_path": "src/deployment/train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 5160, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "26", "api": [{"api_name": "pandas.read_csv", "line_number": 59, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 74, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 78, "usage_type": "call"}, {"api_name": "sklearn.compose.ColumnTransformer", "line_number": 80, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.OneHotEncoder", "line_number": 82, "usage_type": "call"}, {"api_name": "sklearn.impute.SimpleImputer", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 86, "usage_type": "attribute"}, {"api_name": "sklearn.preprocessing.Normalizer", "line_number": 87, "usage_type": "call"}, {"api_name": "sklearn.ensemble.GradientBoostingClassifier", "line_number": 88, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 116, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 121, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 122, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 123, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 124, "usage_type": "call"}, {"api_name": "joblib.dump", "line_number": 143, "usage_type": "call"}, {"api_name": "joblib.dump", "line_number": 144, "usage_type": "call"}]}
+{"seq_id": "73372784383", "text": "\"\"\"\nA module for aggregating the data from the off-chain sources. For the\nperp futures data there is an option of adding funding rates data.\n#####################################################################\nData format\nSPOT MARKET: 'ETHUSDT' or 'ethusdt'\nPERP MARKET: 'ALTUSDT-PERP' or 'altusdt-perp'\nFUTURES MARKET: 'BTCUSD-CURRENT' or 'btcusd-next'\nDVOL DATA: 'ETH-DVOL' or 'eth-dvol'\n\nOPTION DATA: 'ETH-OPT' or 'eth-opt'\n\"\"\"\n\n# pylint: disable=anomalous-backslash-in-string, too-many-instance-attributes, too-many-public-methods, too-many-arguments, too-many-locals, invalid-name\nimport re\nimport datetime\nimport itertools\nfrom typing import Union, List, Tuple, Dict, Any\nimport requests\nimport numpy as np\nimport pandas as pd\n\n\nclass DeFibulizer():\n \"\"\"\n Class for data aggregation and processing of the on-chain and off-chain data.\n \"\"\"\n\n def __init__(self):\n \"\"\"\n Class constructor.\n \"\"\"\n\n self.asset_names = None\n self.asset_types = None\n self.start_date = None\n self.end_date = None\n self.frequency = None\n self.get_funding_rate = None\n self.joint_dataframe = None\n self.allowed_dvol = ['ETH', 'BTC']\n\n @staticmethod\n def set_asset_name(asset_names: Union[str, List[str]]) -> np.ndarray:\n \"\"\"\n Checks whither the provided asset_names parameter type is correct and\n transforms the data into an uppercase list of strings if needed.\n :param asset_names: (str, List[str]) Provided list of asset names to\n retrieve data for.\n :return: (List[str]) List of the asset names to retrieve data for.\n \"\"\"\n # Performing the typing check and transformation.\n\n if isinstance(asset_names, str):\n\n # If the singular asset is provided transforming it into the list.\n asset_names = np.array([asset_names.upper()])\n\n elif isinstance(asset_names, list) and all(isinstance(asset, str) for asset in asset_names):\n\n # If multiple assets are provided in a form of the list keeping it intact.\n asset_names = np.array([name.upper() for name in asset_names])\n\n else:\n\n # Else raise type error.\n raise TypeError('Incorrect asset type, please provide either str or List[str].')\n\n return asset_names\n\n @staticmethod\n def set_date(date: Union[datetime.datetime, str]) -> int:\n \"\"\"\n Checks whither the provided dates parameter types are correct and\n transforms the data into an timestamps.\n :param date: (datetime.datetime/str)\n :return: (datetime.timestamp) Timestamp to retrieve data for.\n \"\"\"\n # Performing the type check and transformation.\n\n if isinstance(date, str):\n # Splitting the string into year month and day component.\n temp = date.split('-')\n\n # Transform the date into timestamp.\n date = datetime.datetime(int(temp[0]), int(temp[1]), int(temp[2])).timestamp()\n\n elif isinstance(date, datetime.datetime):\n\n # If multiple assets are provided in a form of the list keeping it intact.\n date = date.timestamp()\n else:\n\n # Else raise type error.\n raise TypeError('Incorrect date type, please provide either YYYY-MM-DD '\n 'or datetime.datetime.')\n\n return int(date)\n\n @staticmethod\n def set_frequency(frequency: str = 'h') -> Tuple[int, str]:\n \"\"\"\n Checks whether the provided frequency belongs to the range of available options\n and convert it into seconds.\n Options:\n m - minutely\n h - hourly\n d - daily\n :param frequency: (str) Desired data frequency in str format.\n :return: (int) Data frequency in seconds.\n \"\"\"\n # Initializing\n freq_str = ['m', 'h', 'd']\n freq_seconds = [60, 3600, 86400]\n\n adjust_for_binance = lambda el: '1' + el\n\n if frequency not in freq_str:\n # Raise type error if the frequency provided does not belong to the\n # list of possible options.\n raise TypeError('Incorrect frequency type, please provide either m, h or d.')\n\n # Finding the corresponding index in the freq_str\n index = freq_str.index(frequency)\n\n # Taking the equivalent value in seconds from the frec_seconds\n frequency = (freq_seconds[index], adjust_for_binance(freq_str[index]))\n\n return frequency\n\n def is_spot(self) -> list:\n \"\"\"\n Creates a boolean mask that returns true values if the asset belongs to spot market\n based on the provided asset symbol formatting.\n :return: (list) List of boolean indexes that showcase whether the asset\n belongs to the spot market.\n \"\"\"\n # Creating a regular expression pattern to recognize the assets from spot market.\n spot_pattern = re.compile('^([A-Z]+)$')\n\n # Applying the pattern to the asset names to create the boolean mask\n spot_mask = [bool(spot_pattern.match(name)) for name in self.asset_names]\n\n return spot_mask\n\n def is_perp(self) -> list:\n \"\"\"\n Creates a boolean mask that returNs true values if the asset belongs to future market\n and is a perpetual based on the provided asset symbol formatting.\n :return: (list) List of boolean indexes that showcase whether the asset belongs to\n the spot market.\n \"\"\"\n # Creating a regular expression pattern to recognize the assets from spot market.\n perp_pattern = re.compile('^([A-Z]+)[\\-](PERP)$')\n\n # Applying the pattern to the asset names to create the boolean mask\n perp_mask = [bool(perp_pattern.match(name)) for name in self.asset_names]\n\n return perp_mask\n\n def is_futures(self) -> list:\n \"\"\"\n Creates a boolean mask that returns true values if the asset belongs to future market\n and has an expiry date based on the provided asset symbol formatting.\n :return: (list) List of boolean indexes that showcase whether the asset belongs\n to the spot market.\n \"\"\"\n # Creating a regular expression pattern to recognize the assets from spot market.\n futures_pattern = re.compile('^([A-Z]+)[\\-](CURRENT|NEXT)$')\n\n # Applying the pattern to the asset names to create the boolean mask\n futures_mask = [bool(futures_pattern.match(name)) for name in self.asset_names]\n\n return futures_mask\n\n def is_dvol(self) -> list:\n \"\"\"\n Creates a boolean mask that returns true values if the asset asset_name reffers to DVOL\n data based on the provided asset symbol formatting.\n :return: (list) List of boolean indexes that showcase whether the asset belongs\n to the spot market.\n \"\"\"\n # Creating a regular expression pattern to recognize the assets from spot market.\n dvol_pattern = re.compile('^([A-Z]+)[\\-](DVOL)$')\n\n # Applying the pattern to the asset names to create the boolean mask\n dvol_mask = [bool(dvol_pattern.match(name)) for name in self.asset_names]\n\n return dvol_mask\n\n def is_option(self) -> list:\n \"\"\"\n Creates a boolean mask that returns true values if the asset asset_name reffers to option\n data based on the provided asset symbol formatting.\n\n :return: (list) List of boolean indexes that showcase whether the asset belongs\n to the spot market.\n \"\"\"\n # Creating a regular expression pattern to recognize the assets from spot market.\n dvol_pattern = re.compile('^([A-Z]+)[\\-](OPT)$')\n\n # Applying the pattern to the asset names to create the boolean mask\n dvol_mask = [bool(dvol_pattern.match(name)) for name in self.asset_names]\n\n return dvol_mask\n\n def asset_mapping(self):\n \"\"\"\n Maps the provided asset/assets to the class of asset it belongs to, returning the\n array of mapped values.\n Mappings:\n s - spot market assets\n p - perpetual futures\n f - regular futures\n v - volatility index\n o - options\n\n :return: (list) Asset mapping to the asset types.\n \"\"\"\n # Initializing the array for mapped values.\n asset_types = self.asset_names.copy()\n\n # Assigning the list of possible labels for the assets.\n asset_labels = ['s', 'p', 'f', 'v', 'o']\n\n # Creating the boolean masks for each type of assets.\n asset_masks = [self.is_spot(), self.is_perp(), self.is_futures(), self.is_dvol(), self.is_option()]\n\n # Mapping the labels.\n for label, mask in zip(asset_labels, asset_masks):\n asset_types[mask] = label\n\n for index, element in enumerate(asset_types):\n\n if element not in asset_labels:\n raise ValueError(f'The {index} element asset_name is provided incorrectly')\n\n return asset_types\n\n def get_binance_spot_query(self, asset_name: str) -> Dict[str, Union[Union[str, int], Any]]:\n \"\"\"\n Prepares the provided general data and the asset_name to transform into\n the required format for the Binance spot data request.\n :param asset_name: (str) The asset_name of the asset.\n :return: (list) Query details for the Binance data.\n \"\"\"\n\n start_date = int(self.start_date * 1e3)\n end_date = int(self.end_date * 1e3)\n interval = self.frequency[1]\n limit = 1000\n\n query = {'symbol': asset_name,\n 'interval': interval,\n 'startTime': start_date,\n 'endTime': end_date,\n 'limit': limit\n }\n\n return query\n\n def get_binance_futures_query(self, asset_name: str, asset_type: str):\n \"\"\"\n Prepares the provided general data and the asset_name to transform into\n the required format for the Binance futures data request.\n :param asset_name: (str) The asset_name of the asset.\n :param asset_type: (str) The type of the futures - finite or perpetual.\n :return: (list) Query details for the Binance data.\n \"\"\"\n # Setup the initial data parameters\n start_date = int(self.start_date * 1e3)\n end_date = int(self.end_date * 1e3)\n limit = 1000\n interval = self.frequency[1]\n name, contract_type = asset_name.split('-')\n\n # Complete the string for the contract type\n if asset_type == 'p':\n string_completion = 'ETUAL'\n elif asset_type == 'f':\n string_completion = '_QUARTER'\n else:\n raise ValueError('Incorrect asset type literal.')\n\n contract_type += string_completion\n\n query = {'pair': name,\n 'contractType': contract_type,\n 'interval': interval,\n 'startTime': start_date,\n 'endTime': end_date,\n 'limit': limit\n }\n\n return query\n\n def get_binance_funding_query(self, asset_name: str):\n \"\"\"\n Prepares the provided general data and the asset_name to transform into\n the required format for the Binance funding rate data request.\n :param asset_name: (str) The asset_name of the asset.\n :return: (list) Query details for the Binance data.\n \"\"\"\n # Setup the initial data parameters\n start_date = int(self.start_date * 1e3)\n end_date = int(self.end_date * 1e3)\n limit = 1000\n name, _ = asset_name.split('-')\n\n query = {'symbol': name,\n 'startTime': start_date,\n 'endTime': end_date,\n 'limit': limit\n }\n\n return query\n\n @staticmethod\n def get_url(query: dict, data_type: str) -> List[Union[str, dict]]:\n \"\"\"\n Creates url string to be used in API call\n :param query: (dict) Dict of the key query parameters.\n :param data_type: (str) The type of the URL request to construct.\n Input options: 'binance_spot'; 'binance_funding'; 'deribit_dvol'\n :return: (list) URL for the data request and a dict of request parameters.\n \"\"\"\n params = query\n\n if data_type == 'binance_spot':\n # Constructing an URL for the spot/futures/perp price data request\n url = 'https://api.binance.com/api/v3/klines'\n\n if data_type == 'binance_futures':\n url = 'https://fapi.binance.com/fapi/v1/continuousKlines'\n\n elif data_type == 'binance_funding':\n\n # Constructing an URL for the funding rate request\n url = 'https://fapi.binance.com/fapi/v1/fundingRate'\n\n elif data_type == 'deribit_dvol':\n\n # Constructing an URL for the DVOL price data request\n url = 'https://www.deribit.com/api/v2/public/get_volatility_index_data?'\n\n elif data_type == 'deribit_option':\n\n # Constructing an URL for the Derobit option trade data request\n url = 'https://history.deribit.com/api/v2/public/get_last_trades_by_currency_and_time?'\n\n else:\n\n # Else raise type error.\n raise TypeError('Incorrect URL type')\n\n output = [url, params]\n\n return output\n\n def get_deribit_query(self, name: str, asset_type: str) -> Dict[str, Union[Union[str, int, float], Any]]:\n \"\"\"\n Prepares the provided general data and the asset_name to transform into\n the required format for the Deribit data.\n :param name: (str) The asset_name of the asset.\n :param asset_type: (str) The type of the deribit data to request - dvol or options.\n :return: (list) Query details for the Deribit data.\n \"\"\"\n # Select only the part relevant to the query\n name = name.split('-')[0]\n\n # Adjust the timestamp to be in ms\n start_date = int(self.start_date * 1e3)\n end_date = int(self.end_date * 1e3)\n\n frequency = self.frequency[0]\n\n # Adjust the symbol for daily data\n if frequency == 86400:\n frequency = '1D'\n\n # Set up the base deribit query parameters\n query = {'currency': name,\n 'start_timestamp': start_date,\n 'end_timestamp': end_date\n }\n\n # Add asset-specific query parameters\n if asset_type == 'v':\n query['resolution'] = frequency\n\n elif asset_type == 'o':\n query['kind'] = 'option'\n query['count'] = 1e4\n query['include_old'] = 'true'\n\n else:\n raise ValueError('Incorrect asset type literal.')\n\n # query = [name, start_date, end_date, frequency]\n\n return query\n\n def get_fractional_binance_request(self, query: dict, data_type: str) -> list:\n \"\"\"\n Generator function that generates Binance request results for the fractions of\n the data range until the whole data range is processed.\n :param query: (list) The list of API query parameters:\n [asset_names, start_date, end_date, frequency].\n :param data_type: (str) The type of data that needs to be downloaded 'binance_spot'or\n 'binance_funding'\n \"\"\"\n # Getting the start and end date from the initial query\n start_date = query['startTime']\n end_date = query['endTime']\n\n # Establishing an appropriate step for the request, -1 to avoid jumping over the 9th hours\n step = int(query['limit'] * self.frequency[0] * 1e3 - 1)\n\n # Starting a cycle of generating the outcomes of fractional requests until\n # the whole data range is processed\n\n\n for start_time in range(start_date, end_date, step):\n\n # Calculating an end_time of a fraction\n end_time = min(start_time + step, end_date)\n\n # Set the new start/end time for the fractional query\n query['startTime'] = start_time\n query['endTime'] = end_time\n # Creating an URL for the request\n url, params = self.get_url(query, data_type)\n\n # Trying getting the results of the query\n resp = requests.get(url, params=params).json()\n\n if isinstance(resp, dict):\n msg = 'msg'\n raise ValueError(f'{resp[msg]}')\n\n # Yielding the request results\n yield resp\n\n def get_fractional_deribit_request(self, query: dict) -> list:\n \"\"\"\n Generator function that generates deribit request results for the fractions\n of the data range until the whole data range is processed.\n :param query: (list) The list of API query parameters:\n [asset_names, start_date, end_date, frequency].\n \"\"\"\n # Setting the continuation as the end date\n continuation = query['end_timestamp']\n currency = 'currency'\n\n while continuation is not None:\n\n query['end_timestamp'] = continuation\n\n # Creating an URL for the request\n continuation_url, params = self.get_url(query, 'deribit_dvol')\n\n # Trying to get the request and return an exception otherwise\n try:\n resp = requests.get(continuation_url, params=params).json()['result']\n except TypeError as type_err: # pragma: no cover\n raise TypeError(f'The provided {query[currency]} deribit data is incorrect.') from type_err\n except KeyError as key_err: # pragma: no cover\n raise KeyError(requests.get(continuation_url, params=params).json()['error']) from key_err\n\n data = resp['data']\n\n yield data\n\n continuation = resp['continuation'] # pragma: no cover\n\n def get_fractional_deribit_option_request(self, query: dict) -> list:\n \"\"\"\n Generator function that generates deribit request option trades results for the fractions\n of the data range until the whole data range is processed.\n\n :param query: (list) The list of API query parameters:\n [asset_names, start_date, end_date, frequency].\n \"\"\"\n # Setting the continuation as the end date\n continuation = query['start_timestamp']\n\n currency = 'currency'\n\n has_more = True\n\n while has_more:\n\n query['start_timestamp'] = continuation\n\n # Creating an URL for the request\n continuation_url, params = self.get_url(query, 'deribit_option')\n\n # Trying to get the request and return an exception otherwise\n try:\n resp = requests.get(continuation_url, params=params).json()['result']\n except TypeError as type_err: # pragma: no cover\n raise TypeError(f'The provided {query[currency]} option data is incorrect.') from type_err\n except KeyError as key_err: # pragma: no cover\n raise KeyError(requests.get(continuation_url, params=params).json()['error']) from key_err\n\n data = resp['trades']\n\n yield data\n\n continuation = resp['trades'][-1]['timestamp'] # pragma: no cover\n\n has_more = resp['has_more']\n\n def get_binance_data(self, query: dict, data_type: str) -> list:\n \"\"\"\n Retrieves the raw json data from the Binance API according to passed query\n parameters. The query provided has to include the asset asset_name, start and end\n dates for the required data and the frequency. The retrieved data is then\n transformed into a single list.\n :param query: (list) The list of API query parameters:\n [asset_names, start_date, end_date, frequency].\n :return: (list) List of data for the provided query date range, data format:\n [startTime, time, open, high, low, close, volume].\n \"\"\"\n\n # Get the Binance request generator\n binance_generator = self.get_fractional_binance_request(query=query,\n data_type=data_type)\n\n # Create a list from the generator for the Binance requests\n output = list(itertools.chain.from_iterable(binance_generator))\n\n return output\n\n def get_deribit_data(self, query: dict, is_option: bool = True) -> dict:\n \"\"\"\n Retrieves the raw vol data from the Deribit API according to passed query\n parameters. The query provided has to include the asset asset_name, start and end dates for\n the required data and the frequency. The retrieved data is then transformed into\n a single list.\n :param query: (list) The list of API query parameters:\n [asset_names, start_date, end_date, frequency]\n :return: (list) List of data for the provided query date range, data format:\n [time, open, high, low, close].\n \"\"\"\n # Check whether the token has data available\n if not query['currency'] in self.allowed_dvol:\n raise ValueError('The dvol data for this asset is not available')\n\n # Getting the deribit request generator\n if is_option:\n deribit_generator = self.get_fractional_deribit_option_request(query=query)\n else:\n deribit_generator = self.get_fractional_deribit_request(query=query)\n\n # Getting the full list of data for the whole range of dates from generator\n unchained_output = list(itertools.chain.from_iterable(deribit_generator))\n output = list(itertools.chain(unchained_output))\n\n return output\n\n @staticmethod\n def dvol_data_processing(raw_data: dict, asset_names: str) -> pd.DataFrame:\n \"\"\"\n Creates an organized pandas DataFrame from the raw data collected from the\n Deribit API calls.\n :param raw_data: (pd.DataFrame) Raw list of dictionaries containing the API request output.\n :param asset_names: (str) The asset_name of the asset.\n :return: (pd.DataFrame) Query details for the Deribit data.\n \"\"\"\n\n # Creating a pd.DataFrame from the raw data\n dataframe = pd.DataFrame(raw_data)\n\n # Establishing the column names\n dataframe.columns = ['time', 'open', 'high', 'low', 'close']\n\n dataframe = dataframe.sort_values(by=['time'])\n\n # Add the datetime index\n timestamp_temp = pd.to_datetime(dataframe.time, unit='ms', utc=True)\n\n dataframe.set_index(timestamp_temp, inplace=True)\n\n # Add a MultiIndex\n dataframe.columns = pd.MultiIndex.from_product([[asset_names], dataframe.columns])\n\n return dataframe\n\n @staticmethod\n def option_data_processing(raw_data: dict) -> pd.DataFrame:\n \"\"\"\n Creates an organized pandas DataFrame from the raw options data collected from the\n Deribit API calls.\n\n :param raw_data: (pd.DataFrame) Raw list of dictionaries containing the API request output.\n\n :return: (pd.DataFrame) Query details for the Deribit data.\n \"\"\"\n\n # Creating a pd.DataFrame from the raw data\n dataframe = pd.DataFrame.from_dict(raw_data)\n\n # Establishing the column names\n temp_timestamp = pd.to_datetime(dataframe.timestamp, unit='ms')\n\n dataframe = dataframe.sort_values(by=['timestamp'])\n\n temp = dataframe.instrument_name.str.split('-').tolist()\n\n dataframe[['asset', 'expiry_date', 'strike', 'option_type']] = pd.DataFrame(temp, index=dataframe.index)\n\n convert_expiry = lambda date: datetime.datetime.strptime(date, '%d%b%y')\n\n dataframe.expiry_date = dataframe.expiry_date.apply(convert_expiry)\n\n dataframe.strike = pd.to_numeric(dataframe.strike)\n\n # Add the datetime index\n\n dataframe.set_index(temp_timestamp, inplace=True)\n\n return dataframe\n\n @staticmethod\n def binance_data_processing(raw_data: list,\n asset_name: str,\n raw_funding_rates=None) -> pd.DataFrame:\n \"\"\"\n Creates an organized pandas DataFrame from the raw data collected from the\n Binance API calls.\n :param raw_data: (pd.DataFrame) Raw list of dictionaries containing the API request output\n :param asset_name: (str) The asset_name of the asset.\n :param raw_funding_rates: (pd.DataFrame) Raw list of dictionaries containing the API\n request output for the funding rates.\n :return: (pd.DataFrame) Query details for the Deribit data.\n \"\"\"\n\n # Creating a pd.DataFrame from the raw data\n dataframe = pd.DataFrame(raw_data)\n\n columns = ['open_time', 'open', 'high',\n 'low', 'close', 'volume',\n 'close_time', 'quote_asset_volume',\n 'number_of_trades', 'taker_buy_volume',\n 'taker_buy_quote_asset_volume', 'drop']\n\n dataframe.columns = columns\n\n convert_to_date = lambda col: pd.to_datetime(col, unit='ms', utc=True)\n\n if raw_funding_rates is not None:\n # Creating the funding rates dataframe\n funding_dataframe = pd.DataFrame(raw_funding_rates)\n # Merging the funding data with the original perp dataframe\n dataframe = dataframe.merge(funding_dataframe, how='left',\n left_on='open_time', right_on='fundingTime')\n\n dataframe.drop(['fundingTime', 'symbol'], axis=1, inplace=True)\n\n dataframe.drop(['drop'], axis=1, inplace=True)\n\n time_related_subset = dataframe.filter(regex='time')\n not_time_subset = dataframe.filter(regex='^((?!time).)*$')\n\n dataframe[time_related_subset.columns] = time_related_subset.apply(convert_to_date)\n\n # Add the datetime index\n timestamp_temp = pd.to_datetime(dataframe.open_time, unit='ms', utc=True)\n\n dataframe.set_index(timestamp_temp, inplace=True)\n\n dataframe[not_time_subset.columns] = dataframe[not_time_subset.columns].astype(float)\n\n # Adding a MultiIndex\n dataframe.columns = pd.MultiIndex.from_product([[asset_name], dataframe.columns])\n\n return dataframe\n\n def get_spot_data(self, data: list, asset_name: str):\n \"\"\"\n Runs the procedure to form a request, get and postprocess the spot\n data. Then the dataframe is appended to the list of the dataframes.\n\n :param data: (list) The list of dataframes for all the retrieved data.\n :param asset_name: (str) The name of the asset to retrieve.\n \"\"\"\n # Creating a query variables list\n query = self.get_binance_spot_query(asset_name)\n # Getting the raw data from the Binance API\n raw_data = self.get_binance_data(query, 'binance_spot')\n # Appending the processed dataframe\n data.append(self.binance_data_processing(raw_data, asset_name))\n\n def get_perp_data(self, data: list, asset_name: str):\n \"\"\"\n Runs the procedure to form a request, get and postprocess the perpetual futures\n data. Then the dataframe is appended to the list of the dataframes.\n\n :param data: (list) The list of dataframes for all the retrieved data.\n :param asset_name: (str) The name of the asset to retrieve.\n \"\"\"\n asset_type = 'p'\n\n # Creating a query variables list for the perp\n query = self.get_binance_futures_query(asset_name, asset_type)\n # Getting the raw data from the Binance API\n raw_data = self.get_binance_data(query, 'binance_futures')\n\n # If there is a flag for getting the funding rate start the procedure\n if self.get_funding_rate:\n # Creating a query variables list for the perp funding rates\n query = self.get_binance_funding_query(asset_name)\n # Getting the raw funding rates data from the Binance API\n raw_funding_rates = self.get_binance_data(query, 'binance_funding')\n # Appending the processed dataframe\n data.append(self.binance_data_processing(raw_data, asset_name, raw_funding_rates))\n\n else:\n # Appending the processed dataframe\n data.append(self.binance_data_processing(raw_data, asset_name))\n\n def get_futures_data(self, data: list, asset_name: str):\n \"\"\"\n Runs the procedure to form a request, get and postprocess the finite futures\n data. Then the dataframe is appended to the list of the dataframes.\n\n :param data: (list) The list of dataframes for all the retrieved data.\n :param asset_name: (str) The name of the asset to retrieve.\n \"\"\"\n asset_type = 'f'\n\n # Creating a query variables list for the future\n query = self.get_binance_futures_query(asset_name, asset_type)\n # Getting the raw data from the Binance API\n raw_data = self.get_binance_data(query, 'binance_futures')\n # Appending the processed dataframe\n data.append(self.binance_data_processing(raw_data, asset_name))\n\n def get_options_data(self, data: list, asset_name: str):\n \"\"\"\n Runs the procedure to form a request, get and postprocess the Deribit options\n data. Then the dataframe is appended to the list of the dataframes.\n\n :param data: (list) The list of dataframes for all the retrieved data.\n :param asset_name: (str) The name of the asset to retrieve.\n \"\"\"\n asset_type = 'o'\n # Creating a query variables list for the future\n query = self.get_deribit_query(asset_name, asset_type)\n # Getting the raw data from the Deribit API\n raw_data = self.get_deribit_data(query, is_option=True)\n # Appending the processed dataframe\n data.append(self.option_data_processing(raw_data))\n\n def get_dvol_data(self, data: list, asset_name: str):\n \"\"\"\n Runs the procedure to form a request, get and postprocess the volatility index\n data. Then the dataframe is appended to the list of the dataframes.\n\n :param data: (list) The list of dataframes for all the retrieved data.\n :param asset_name: (str) The name of the asset to retrieve.\n \"\"\"\n asset_type = 'v'\n\n # Creating a query variables list for the future\n query = self.get_deribit_query(asset_name, asset_type)\n # Getting the raw data from the Deribit API\n raw_data = self.get_deribit_data(query, is_option=False)\n # Appending the processed dataframe\n data.append(self.dvol_data_processing(raw_data, asset_name))\n\n def init_asset_getter_dict(self):\n \"\"\"\n Creates a dictionary of data retrieving procedures corresponding to\n respective types of assets.\n\n :return: (dict) Dictionary with the asset types as key and processing\n functions as values.\n \"\"\"\n asset_types = ['s', 'p', 'f', 'v', 'o']\n\n data_getters = [self.get_spot_data,\n self.get_perp_data,\n self.get_futures_data,\n self.get_dvol_data,\n self.get_options_data]\n\n output = dict(map(lambda i, j: (i, j), asset_types, data_getters))\n\n return output\n\n def get_data(self,\n asset_names: Union[str, List[str]] = None,\n start_date: Union[datetime.datetime, str] = None,\n end_date: Union[datetime.datetime, str] = None,\n frequency: str = 'm',\n get_funding_rate: bool = False,\n joint_dataframe=True,\n save_csv: bool = False) -> Union[pd.DataFrame, List[pd.DataFrame]]:\n \"\"\"\n Gathers the available instrument data in a requested data range for every\n instance in the provided asset_names list. The data is gathered using the\n respective API - Binance or Deribit. Obtained raw data is then processed into\n a single concatenated dataframe for all assets/list of data frames for\n each individual asset data.\n :param asset_names: (str, List[str]) List of asset names to retrieve data for.\n :param start_date: (datetime.datetime, str) Beginning of data range. Str format\n should be YYYY-MM-DD.\n :param end_date: (datetime.datetime, str) End of the data range. tr format\n should be YYYY-MM-DD.\n :param frequency: (str) The wanted data frequency. Available formats:\n 'm' - minutely; 'h' - hourly; 'd' - daily.\n :param get_funding_rate: (bool) Flag whether to retrieve the funding rate data\n for the perpetual futures.\n :param joint_dataframe: (bool) Flag whether to return a joint dataframe or a list\n of separate dataframes.\n :param save_csv: (bool) Flag whether to save the dataframe on users system.\n :return: (pd.DataFrame, List[pd.DataFrame]) Processed data.\n \"\"\"\n # Setting all the attributes from the provided arguments\n self.asset_names = self.set_asset_name(asset_names)\n # Mapping the asset names to the asset type\n self.asset_types = self.asset_mapping()\n # Transforming the start and end date into int(timestamps)\n self.start_date = self.set_date(start_date)\n self.end_date = self.set_date(end_date)\n # Setting the data frequency\n self.frequency = self.set_frequency(frequency)\n # Setting the flags rom the provided\n self.get_funding_rate = get_funding_rate\n self.joint_dataframe = joint_dataframe\n\n # Checking whether the start and end dates are valid\n if self.start_date > self.end_date:\n raise ValueError('End date can not precede the start date.')\n\n # Initializing the data list\n data = []\n\n fill_in_parameters = lambda func: func(data=data, asset_name=asset_name)\n\n # Initialize the dictionary of a getter functions for respective asset types\n data_getter = self.init_asset_getter_dict()\n\n if 'o' in self.asset_types:\n joint_dataframe = False\n\n # For every asset in provided list get the data from the respective API\n for asset_type, asset_name in zip(self.asset_types, self.asset_names):\n\n # If the asset belongs to spot market, perpetual futures, finite futures\n # options or Deribit volatility indexes - start the respective procedure\n fill_in_parameters(data_getter[asset_type])\n\n # Concatenating the dataframes if the flag is true\n if joint_dataframe:\n for el in data:\n ind = el[el.index.duplicated()].index\n el.drop(ind, axis=0, inplace=True)\n output_data = pd.concat(data, axis=1)\n\n output_data.index.name = 'time'\n\n # Saving the data if the flag is true\n self.save_csv(data, save_csv)\n else:\n output_data = data\n\n return output_data\n\n def save_csv(self, data: Union[list, pd.DataFrame] = None, flag: bool = True, test: bool = False) -> None:\n \"\"\"\n Saves data in a *.csv data format.\n :param data: (list/pd.DataFrame) Data to be saved.\n :param flag: (bool) A flag signifying whether to save the files\n :param test: (bool) A flag signifying whether it is in a unit test or not.\n \"\"\"\n if flag:\n if isinstance(data, pd.DataFrame):\n if not test: # pragma: no cover\n dataframe_name = '_'.join(self.asset_names)\n data.to_csv(f'{dataframe_name}.csv')\n\n else:\n # If data is a list - save as separate entities\n if not test: # pragma: no cover\n for name, dataframe in zip(self.asset_names, data):\n dataframe.to_csv(f'{name}.csv')\n\n @staticmethod\n def load_csv(file_name: str) -> pd.DataFrame:\n \"\"\"\n Assists the users with reading the MultiIndex csv DataFrame.\n :param file_name: (str) Name of the saved DeFibulizer dataframe or the path to the file.\n :return: (pd.DataFrame) The correctly formatted dataframe from the saved csv file.\n \"\"\"\n # Reading the csv specifying the levels and indexes\n output = pd.read_csv(file_name, header=[0, 1], index_col=0)\n\n return output\n", "repo_name": "Brahma-fi/brahma-vaults-backtests", "sub_path": "modules/defibulizer/offchain_defibulizer.py", "file_name": "offchain_defibulizer.py", "file_ext": "py", "file_size_in_byte": 36717, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "24", "api": [{"api_name": "typing.Union", "line_number": 44, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 44, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 44, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 72, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 72, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 86, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 88, "usage_type": "attribute"}, {"api_name": "typing.Tuple", "line_number": 101, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 139, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 154, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 169, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 184, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 200, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 240, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 240, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 240, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 319, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 319, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 360, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 360, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 360, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 435, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 464, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 468, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 500, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 504, "usage_type": "call"}, {"api_name": "itertools.chain.from_iterable", "line_number": 531, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 531, "usage_type": "attribute"}, {"api_name": "itertools.chain.from_iterable", "line_number": 557, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 557, "usage_type": "attribute"}, {"api_name": "itertools.chain", "line_number": 558, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 573, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 581, "usage_type": "call"}, {"api_name": "pandas.MultiIndex.from_product", "line_number": 586, "usage_type": "call"}, {"api_name": "pandas.MultiIndex", "line_number": 586, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 563, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 602, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 602, "usage_type": "attribute"}, {"api_name": "pandas.to_datetime", "line_number": 605, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 611, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 613, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 613, "usage_type": "attribute"}, {"api_name": "pandas.to_numeric", "line_number": 617, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 591, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 640, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 650, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 654, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 669, "usage_type": "call"}, {"api_name": "pandas.MultiIndex.from_product", "line_number": 676, "usage_type": "call"}, {"api_name": "pandas.MultiIndex", "line_number": 676, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 628, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 794, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 794, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 795, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 795, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 796, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 796, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 861, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 800, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 800, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 800, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 872, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 872, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 880, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 899, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 892, "usage_type": "attribute"}]}
+{"seq_id": "4045120204", "text": "import json\nimport random\n\nfrom data.dimo_loader import DimoLoader\nfrom pathlib import Path\nfrom typing import List, Tuple\nimport numpy as np\nfrom bop_toolkit_lib import misc, visibility, inout, renderer\nimport os\nimport shutil\nimport cv2\n\n\ndimo_data = {\n 'im_width': 2560,\n 'im_height': 2048\n}\n\n\ndef create_or_ignore_folder(path: str):\n if not os.path.exists(path):\n os.mkdir(path)\n\n\ndef create_or_empty_folder(path: str):\n if os.path.exists(path):\n shutil.rmtree(path)\n os.mkdir(path)\n\n\ndef get_file_count(path: str) -> int:\n if not os.path.exists(path):\n return 0\n else:\n return len(os.listdir(path))\n\n\ndef create_dimo_masks(path: str, subsets: List[str], override: bool = False) -> None:\n \"\"\"\n Generates the visible mask for each object in each image.\n Masks are saved separately as a binary image for each object under {scene_id}/masks/{image_id}/{object_no}.png\n :param override: masks that are already generated are ignored if set to false, otherwise new masks are generated\n :param path: path to DIMO dataset\n :param subsets: subsets of dimo dataset to generate masks for (eg. sim_jaigo)\n \"\"\"\n dimo_loader = DimoLoader()\n dimo_ds = dimo_loader.load(Path(path), cameras=subsets)\n\n for subset_name in subsets:\n subset = dimo_ds[subset_name]\n models = dimo_ds['models']\n\n ren = renderer.create_renderer(dimo_data['im_width'], dimo_data['im_height'], renderer_type='vispy', mode='depth')\n for model in models:\n ren.add_object(model['id'], model['cad'])\n\n for scene in subset:\n masks_path = os.path.join(scene['path'], 'mask_visib/')\n print(f\"Processing {scene['path']}\")\n\n if get_file_count(masks_path) == sum([len(image['objects']) for image in scene['images']]):\n continue\n\n if override:\n create_or_empty_folder(masks_path)\n else:\n create_or_ignore_folder(masks_path)\n\n render_scene_masks(scene, ren)\n\n\ndef render_scene_masks(scene: dict, ren: renderer):\n masks_path = os.path.join(scene['path'], 'mask_visib/')\n\n for image in scene['images']:\n camera = image['camera']\n\n K = camera['K']\n fx, fy, cx, cy = K[0, 0], K[1, 1], K[0, 2], K[1, 2]\n\n dist_image = np.array(np.ones((dimo_data['im_height'], dimo_data['im_width'])) * np.inf)\n distance_maps = []\n for object in image['objects']:\n depth_gt = \\\n ren.render_object(object['id'], np.array(object['cam_R_m2c']).reshape(3, 3), np.array(object['cam_t_m2c']),\n fx, fy, cx, cy)['depth']\n dist_gt = misc.depth_im_to_dist_im_fast(depth_gt, K)\n\n # set zero depths to infinity to compute closest object for total depth map\n dist_gt[dist_gt == 0] = np.inf\n dist_image = np.minimum(dist_image, dist_gt)\n\n dist_gt[dist_gt == np.inf] = 0\n distance_maps.append(dist_gt)\n\n dist_image[dist_image == np.inf] = 0\n\n for object_no, dist_map in enumerate(distance_maps):\n object_mask_path = os.path.join(masks_path, f\"{str(image['id']).zfill(6)}_{str(object_no).zfill(6)}.png\")\n mask_visib = visibility.estimate_visib_mask_gt(dist_image, dist_map, 1, visib_mode='bop19') * 255\n inout.save_im(object_mask_path, np.array(mask_visib, dtype=np.uint8))\n\n\ndef create_dimo_scene_masks(path: str, subset: str, scene_id: int) -> None:\n dimo_loader = DimoLoader()\n dimo_ds = dimo_loader.load(Path(path), cameras=[subset])\n\n subset = dimo_ds[subset]\n models = dimo_ds['models']\n\n ren = renderer.create_renderer(dimo_data['im_width'], dimo_data['im_height'], renderer_type='vispy', mode='depth')\n for model in models:\n ren.add_object(model['id'], model['cad'])\n\n for scene in subset:\n if scene['id'] == scene_id:\n masks_path = os.path.join(scene['path'], 'mask_visib/')\n\n print(f\"Processing {scene['path']}\")\n create_or_empty_folder(masks_path)\n render_scene_masks(scene, ren)\n\n\ndef create_dimo_train_split(path: str, subsets: List[str], train: float = 0.9, val: float = 0.05, test: float = 0.05, seed: int = None, split_scenes: bool = False) -> None:\n \"\"\"\n Given the path to a dimo dataset, this function will split the scenes of the given subsets in training, validation\n and testing dataset. The function creates files of name {split}.txt\n :param path: path to the dimo dataset\n :param subsets: subsets to generate split for\n :param train: portion of scenes to be training dataset\n :param val: portion of scenes to be validation dataset\n :param test: portion of scenes to be test dataset\n :param seed: optionally set random seed\n :param split_scenes:if set to true the split are based on the scenes, otherwise on the images\n :return:\n \"\"\"\n def get_scenes_split(subset: list) -> Tuple[List[str], List[str], List[str]]:\n scenes = subset\n random.shuffle(scenes)\n\n train_ids = [f\"{scene['id']}_{image['id']}\" for scene in scenes[:int(train * len(scenes))] for image in scene['images']]\n val_ids = [f\"{scene['id']}_{image['id']}\" for scene in scenes[int(train * len(scenes)):int((val + train) * len(scenes))] for image in scene['images']]\n test_ids = [f\"{scene['id']}_{image['id']}\" for scene in scenes[int((val + train) * len(scenes)):] for image in scene['images']]\n\n return train_ids, val_ids, test_ids\n\n def get_images_split(subset: list) -> Tuple[List[str], List[str], List[str]]:\n image_ids = [f\"{scene['id']}_{image['id']}\" for scene in subset for image in scene['images']]\n random.shuffle(image_ids)\n\n train_ids = image_ids[:int(train * len(image_ids))]\n val_ids = image_ids[int(train * len(image_ids)):int((val + train) * len(image_ids))]\n test_ids = image_ids[int((val + train) * len(image_ids)):]\n\n return train_ids, val_ids, test_ids\n\n def write_to_file(file: str, data: list):\n with open(file, 'w') as f:\n for element in data:\n f.write(f\"{element}\\n\")\n\n if seed:\n random.seed(seed)\n\n train /= sum([train, val, test])\n test /= sum([train, val, test])\n val /= sum([train, val, test])\n\n dimo_loader = DimoLoader()\n dimo_ds = dimo_loader.load(Path(path), cameras=subsets, models_dir=None)\n\n for subset_name in subsets:\n subset = dimo_ds[subset_name]\n\n train_ids, val_ids, test_ids = get_scenes_split(subset) if split_scenes else get_images_split(subset)\n\n subset_path = os.path.join(path, f\"{subset_name}/\")\n write_to_file(os.path.join(subset_path, \"train.txt\"), train_ids)\n write_to_file(os.path.join(subset_path, \"val.txt\"), val_ids)\n write_to_file(os.path.join(subset_path, \"test.txt\"), test_ids)\n\n\nif __name__ == \"__main__\":\n create_dimo_scene_masks(\"D:/Datasets/DIMO/dimo\", \"sim_jaigo_rand_light_rand_pose\", 3649)", "repo_name": "EDM-Research/DIMO_ObjectDetection", "sub_path": "data/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 7051, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 7, "dataset": "github-code", "pt": "26", "api": [{"api_name": "os.path.exists", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 27, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 35, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 38, "usage_type": "name"}, {"api_name": "data.dimo_loader.DimoLoader", "line_number": 46, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 47, "usage_type": "call"}, {"api_name": "bop_toolkit_lib.renderer.create_renderer", "line_number": 53, "usage_type": "call"}, {"api_name": "bop_toolkit_lib.renderer", "line_number": 53, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "bop_toolkit_lib.renderer", "line_number": 72, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 85, "usage_type": "call"}, {"api_name": "bop_toolkit_lib.misc.depth_im_to_dist_im_fast", "line_number": 87, "usage_type": "call"}, {"api_name": "bop_toolkit_lib.misc", "line_number": 87, "usage_type": "name"}, {"api_name": "numpy.inf", "line_number": 90, "usage_type": "attribute"}, {"api_name": "numpy.minimum", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 93, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 96, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "bop_toolkit_lib.visibility.estimate_visib_mask_gt", "line_number": 100, "usage_type": "call"}, {"api_name": "bop_toolkit_lib.visibility", "line_number": 100, "usage_type": "name"}, {"api_name": "bop_toolkit_lib.inout.save_im", "line_number": 101, "usage_type": "call"}, {"api_name": "bop_toolkit_lib.inout", "line_number": 101, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 101, "usage_type": "attribute"}, {"api_name": "data.dimo_loader.DimoLoader", "line_number": 105, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 106, "usage_type": "call"}, {"api_name": "bop_toolkit_lib.renderer.create_renderer", "line_number": 111, "usage_type": "call"}, {"api_name": "bop_toolkit_lib.renderer", "line_number": 111, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path", "line_number": 117, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 124, "usage_type": "name"}, {"api_name": "random.shuffle", "line_number": 139, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 137, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 137, "usage_type": "name"}, {"api_name": "random.shuffle", "line_number": 149, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 147, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 147, "usage_type": "name"}, {"api_name": "data.dimo_loader", "line_number": 159, "usage_type": "name"}, {"api_name": "random.seed", "line_number": 163, "usage_type": "call"}, {"api_name": "data.dimo_loader.DimoLoader", "line_number": 169, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 177, "usage_type": "call"}, {"api_name": "os.path", "line_number": 177, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 178, "usage_type": "call"}, {"api_name": "os.path", "line_number": 178, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path", "line_number": 179, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path", "line_number": 180, "usage_type": "attribute"}]}
+{"seq_id": "27746688628", "text": "import datetime\nimport logging\nimport os\nimport re\nfrom functools import wraps\nfrom random import randint\n\nimport sqlalchemy\nfrom telegram import Update, ReplyKeyboardMarkup, InlineKeyboardButton, InlineKeyboardMarkup\nfrom telegram.ext import ApplicationBuilder, CallbackContext, CommandHandler, MessageHandler, filters, \\\n ConversationHandler, CallbackQueryHandler\nfrom transmission_rpc import TransmissionError\n\nfrom bot_csv import csv_upload_handler, csv_download_handler\nfrom bot_get_progress import get_progress\nfrom bot_utils import make_movie_reply, update_torr_db, \\\n exclude_torrents_from_watchlist, get_movie_from_all_databases, search_imdb_title, add_to_watchlist, \\\n get_telegram_users, invite_friend\nfrom command_regex_handler import RegexpCommandHandler\nfrom bot_watchlist import get_torrents_for_imdb_id, update_watchlist_item_status\nfrom utils import deconvert_imdb_id, send_torrent, compose_link, get_user_by_tgram_id, get_my_movie_by_imdb, \\\n update_many, Movie, convert_imdb_id, check_database, User, get_onetimepasswords, remove_onetimepassword, \\\n insert_onetimepasswords, get_movies_for_bulk_rating, make_client, update_torrent_status, update_torrent_grace_days, \\\n get_torrent_by_torr_id_user\n\n\"\"\"\nIMPORTANT for SSL: add verify=False in\nAnaconda3\\envs\\mSquaredPlex\\Lib\\site-packages\\telegram\\request\\_httpxrequest.py\nin self._client_kwargs = dict()\n\"\"\"\n\n# Enable logging\nlogging.basicConfig(\n format='[%(asctime)s] {%(filename)s:%(lineno)d} [%(name)s] [%(levelname)s] --> %(message)s', level=logging.INFO\n)\n\nlogger = logging.getLogger('MovieTimeBot')\n\nSUPERADMIN_PASSWORD = os.getenv('SUPERADMIN_PASSWORD')\n\nTELEGRAM_TOKEN = os.getenv('TELEGRAM_TOKEN')\nTELEGRAM_AUTH_TEST_PATH = os.getenv('TELEGRAM_AUTH_TEST_PATH')\nTELEGRAM_AUTH_APPROVE = os.getenv('TELEGRAM_AUTH_APPROVE')\nTELEGRAM_IMDB_RATINGS = os.getenv('TELEGRAM_IMDB_RATINGS')\nTELEGRAM_NETFLIX_PNG = os.getenv('TELEGRAM_NETFLIX_PNG')\nTELEGRAM_RESET_PNG = os.getenv('TELEGRAM_RESET_PNG')\n\n# State definitions for top level conversation\nCHOOSE_TASK, REGISTER_USER, DOWNLOAD_MOVIE, CHECK_PROGRESS, UPLOAD_ACTIVITY, RATE_TITLE = range(6)\n\n# State definitions DOWNLOAD_MOVIE\nCHOOSE_MULTIPLE, CHOOSE_ONE, CONFIRM_REDOWNLOAD_ACTION, SEARCH_FOR_TORRENTS, \\\nDOWNLOAD_TORRENT, WATCHLIST_NO_TORR = range(6, 12)\n\n# State definitions for CHECK_PROGRESS\nDOWNLOAD_PROGRESS = 12\n\n# State definitions for UPLOAD_ACTIVITY\nNETFLIX_CSV, UPLOAD_CSV = 13, 14\n\n# State definitions for DOWNLOAD_ACTIVITY\nDOWNLOAD_CSV = 15\n\n# State definitions for RATE_TITLE\nCHOOSE_WHAT_TO_RATE = 16\nSUBMIT_RATING = 17\n\n# State definitions for REGISTER_USER\nCHECK_EMAIL, GIVE_IMDB, CHECK_IMDB = range(18, 21)\n\nUSERS = None\n\nmenu_keyboard = [\n [\"📥 Download a movie\"],\n [\"📈 Check download progress\", \"🌡️ Rate a title\"],\n [\"❤☠🤖 Upload Netflix activity\", '💾 Download my movies']\n]\nbool_keyboard = [\n ['Yes'],\n ['No']\n]\nmovie_selection_keyboard = [\n ['Yes'],\n ['No'],\n ['Exit'],\n]\nrate_keyboard = [\n ['1', '2'],\n ['3', '4'],\n ['5', '6'],\n ['7', '8'],\n ['9', '10'],\n [\"I've changed my mind\"]\n]\nrate_keyboard_bulk = [\n ['1', '2'],\n ['3', '4'],\n ['5', '6'],\n ['7', '8'],\n ['9', '10'],\n [\"Skip this movie.\"],\n [\"Exit rating process.\"]\n]\n\n\ndef auth_wrap(f):\n @wraps(f)\n async def wrap(update: Update, context: CallbackContext, *optional_args):\n # print(f\"Wrapped {f.__name__}\", )\n user = update.effective_user['id']\n if user in USERS.keys():\n # User is registered\n result = f(update, context, *optional_args)\n # print(result)\n return await result\n else:\n if USERS:\n # New user ask for password\n await update.message.reply_text(\"Please input the password provided by the admin\")\n context.user_data['user_type'] = 'user'\n else:\n # No users registered, probably admin but check.\n await update.effective_message.reply_photo(\n photo=open(TELEGRAM_AUTH_TEST_PATH, 'rb'),\n caption=\"Looks like you're new here. Answer correctly and you may enter.\",\n )\n context.user_data['user_type'] = 'admin'\n return REGISTER_USER\n\n return wrap\n\n\n\"\"\"<< DOWNLOAD A MOVIE >>\"\"\"\n\n\n@auth_wrap\nasync def start(update: Update, context: CallbackContext, message: str = '') -> int:\n \"\"\"Send a message when the command /start is issued.\"\"\"\n\n user = update.effective_user\n if not message:\n message = f\"Hi {user.mention_markdown_v2()}\\!\\n\" \\\n fr\"Please select one of the options or type /help for more options\\.\"\n await update.message.reply_markdown_v2(\n message,\n reply_markup=ReplyKeyboardMarkup(menu_keyboard, one_time_keyboard=True, resize_keyboard=True),\n )\n return CHOOSE_TASK\n\n\n@auth_wrap\nasync def reset(update: Update, context: CallbackContext.DEFAULT_TYPE) -> int:\n \"\"\"End Conversation by command.\"\"\"\n\n await update.message.reply_text(\"See you next time. Type anything to get started.\")\n\n return ConversationHandler.END\n\n\n@auth_wrap\nasync def choose_task(update: Update, context: CallbackContext) -> int:\n \"\"\"Choose between download a movie, get download progress, upload activity\n or download my viewing activity\"\"\"\n\n if update.message.text == menu_keyboard[0][0]:\n message = 'Great, give me an IMDB id, a title or an IMDB link.'\n context.user_data['action'] = 'download'\n await update.message.reply_text(message)\n return DOWNLOAD_MOVIE\n\n elif update.message.text == menu_keyboard[1][0]:\n await update.effective_message.reply_text('Ok, retrieving data... (might be slow sometimes)')\n context.user_data['download_progress'] = 0\n return await get_download_progress(update, context)\n\n elif update.message.text == menu_keyboard[1][1]:\n await update.effective_message.reply_text('Do you wish to rate a new title or a seen but unrated movie?',\n reply_markup=ReplyKeyboardMarkup([['New title', 'Rate seen movies']],\n one_time_keyboard=True,\n resize_keyboard=True,\n ))\n return RATE_TITLE\n\n elif update.message.text == menu_keyboard[2][0]:\n await update.message.reply_photo(photo=open(TELEGRAM_NETFLIX_PNG, 'rb'),\n caption=\"Ok, follow the instructions, \"\n \"hit `Download all` and upload here the resulting .csv.\\n\"\n \"You may add any other records to \"\n \"the CSV, given that you don't change the column names.\\n\"\n \"If you want to also add ratings, create an extra column \"\n \"named 'Ratings' and it will be picked up.\\n\"\n \"Any overlapping ratings/seen dates will be overwritten. However, \"\n \"IMDB ratings, seen dates and PLEX seen dates will have prevalence.\\n\",\n )\n await update.message.reply_text(\"Now please update the .csv file.\")\n return NETFLIX_CSV\n elif update.message.text == menu_keyboard[2][1]:\n await update.message.reply_text(\"Ok, we've started the process, we'll let you know once it's done\\.\")\n\n return await download_csv(update, context)\n message = 'Please choose one of the options\\.'\n return await start(update, context, message)\n\n\n@auth_wrap\nasync def parse_imdb_id(update: Update, context: CallbackContext) -> int:\n \"\"\"Get data about movie, using the IMDB id.\n Queries the internal database and calls the APIs if necessary.\"\"\"\n\n await update.message.reply_text(\"Just a sec until we get data about this title...\")\n # We need number so filter out the number from the user input:\n imdb_id = ''.join([x for x in update.message.text if x.isdigit()]).lstrip('0')\n\n # Get IMDB data\n pkg = get_movie_from_all_databases(imdb_id, update.effective_user['id'])\n context.user_data['pkg'] = pkg\n context.user_data['more_options'] = False\n\n message, image = make_movie_reply(pkg)\n await update.effective_message.reply_photo(\n photo=image,\n caption=message,\n reply_markup=ReplyKeyboardMarkup(movie_selection_keyboard,\n one_time_keyboard=True,\n resize_keyboard=True,\n ),\n )\n return CHOOSE_ONE\n\n\n@auth_wrap\nasync def parse_imdb_text(update: Update, context: CallbackContext) -> int:\n \"\"\"Route the query if the string is a link or a title.\"\"\"\n\n await update.message.reply_text(\"Just a sec until we get data about this title...\")\n # Is it a link?\n try:\n imdb_id = re.search(r\"[-a-zA-Z0-9@:%._\\+~#=]{1,256}\\.[a-zA-Z0-9()]{1,6}\\b([-a-zA-Z0-9()@:%_\\+.~#?&//=]*)\",\n update.message.text).group(0)\n\n # We need number so filter out the number from the user input:\n imdb_id = ''.join([x for x in imdb_id if x.isdigit()]).lstrip('0')\n\n # Get IMDB data\n pkg = get_movie_from_all_databases(imdb_id, update.effective_user['id'])\n context.user_data['pkg'] = pkg\n\n message, image = make_movie_reply(pkg)\n await update.effective_message.reply_photo(\n photo=image,\n caption=message,\n reply_markup=ReplyKeyboardMarkup(movie_selection_keyboard,\n one_time_keyboard=True,\n resize_keyboard=True,\n ),\n )\n return CHOOSE_ONE\n except AttributeError:\n # Yes but wrong link or no match\n if 'https://' in update.message.text:\n await update.effective_message.reply_text(\"Couldn't find the specified ID in the link, are you\"\n \"sure it's an IMDB link? Try pasting only the ID, as in\"\n \"`tt0903624`.\")\n return DOWNLOAD_MOVIE\n else:\n # Test for title\n movies = search_imdb_title(update.message.text)\n context.user_data['potential_titles'] = movies\n\n return await choose_multiple(update, context)\n\n\n@auth_wrap\nasync def choose_multiple(update: Update, context: CallbackContext) -> int:\n \"\"\"Loop through potential movie matches\"\"\"\n\n movies = context.user_data['potential_titles']\n if movies:\n if type(movies) == str:\n await update.effective_message.reply_text(\"We're having trouble with our IMDB API, please\"\n \"insert an IMDB ID or paste a link.\")\n return DOWNLOAD_MOVIE\n movie = movies.pop(0)\n # Check again if we can find it\n pkg = get_movie_from_all_databases(movie['id'], update.effective_user['id'])\n # context.user_data['pkg'] = pkg\n if pkg:\n context.user_data['pkg'] = pkg\n message, image = make_movie_reply(pkg)\n await update.effective_message.reply_photo(\n photo=image,\n caption=message,\n reply_markup=ReplyKeyboardMarkup(movie_selection_keyboard,\n one_time_keyboard=True,\n resize_keyboard=True,\n ),\n )\n return CHOOSE_ONE\n else:\n return await choose_multiple(update, context)\n else:\n await update.effective_message.reply_text(\"Couldn't find the specified movie.\"\n \" Check your spelling or try pasting the IMDB id or a link\"\n \"`tt0903624`.\")\n return DOWNLOAD_MOVIE\n\n\n@auth_wrap\nasync def accept_reject_title(update: Update, context: CallbackContext) -> int:\n \"\"\"Accept, reject the match or exit\"\"\"\n if update.message.text == 'Yes':\n if context.user_data['action'] == 'download':\n return await check_movie_status(update, context)\n else:\n return await rating_movie_info(update, context)\n elif update.message.text == 'No':\n if context.user_data['potential_titles']:\n await update.effective_message.reply_text(\"Ok, trying next hit...\")\n return await choose_multiple(update, context)\n else:\n await update.effective_message.reply_text(\"These were all the hits. Sorry. Feel free to type anything \"\n \"to start again\")\n return await reset(update, context)\n elif update.message.text == 'Exit':\n return await reset(update, context)\n else:\n await update.effective_message.reply_text(\"Please choose one of the options.\")\n\n\n@auth_wrap\nasync def check_movie_status(update: Update, context: CallbackContext) -> int:\n \"\"\"Customise the message depending on the status of the movie\n (seen, already downloaded)\"\"\"\n\n movie = context.user_data['pkg']\n # Check if you've already seen it and send info\n if movie['already_in_my_movies']:\n message = f\"Looks like you've already seen this movie.\"\n if 'my_score' in movie.keys():\n message += f\"\\nYour score: {movie['my_score']}\"\n if 'seen_date' in movie.keys():\n message += f\"\\nAnd you've seen it on {movie['seen_date']}\"\n await update.message.reply_text(message)\n if movie['torr_result']:\n message = f\"Looks like the movie is also downloaded in {movie['resolution']}p\\n\" \\\n f\"Torrent status: {movie['torr_status']}\\n\" \\\n f\"Would you still like to proceed to download?\"\n else:\n message = f\"\\nWould you still like to proceed to download?\"\n\n await update.effective_message.reply_html(message, reply_markup=ReplyKeyboardMarkup(bool_keyboard,\n one_time_keyboard=True,\n resize_keyboard=True,\n ))\n return CONFIRM_REDOWNLOAD_ACTION\n else:\n return await search_for_torrents(update, context)\n\n\n@auth_wrap\nasync def search_for_torrents(update: Update, context: CallbackContext) -> int:\n \"\"\"Search for torrents for the given IMDB id\n Build keyboard to show them\"\"\"\n\n await update.message.reply_text('Searching for available torrents...')\n torrents = get_torrents_for_imdb_id(context.user_data['pkg']['imdb'])\n torrents = sorted(torrents, key=lambda k: k['size'])\n if torrents:\n context.user_data['pkg']['torrents'] = sorted(torrents, key=lambda k: k['size'])\n keyboard = [[]]\n \"\"\"\n # Check if we need to filter excluded resolutions\n # Removed functionality, left here for reference or other further uses\n if 'from_watchlist' in context.user_data.keys():\n movie_id = deconvert_imdb_id(torrents[0]['imdb'])\n excluded_resolutions = get_excluded_resolutions(movie_id, update.effective_user['id'])\n torrents = [x for x in torrents if x['id'] not in excluded_resolutions]\n \"\"\"\n for pos, item in enumerate(torrents):\n pos += 1 # exclude 0\n btn_text = (\n f\"🖥 Q: {str(item['resolution'])}\"\n f\"🗳 S: {str(round(item['size'] / 1000000000, 2))} GB\"\n f\"🌱 S/P: {str(item['seeders'])}/{str(item['leechers'])}\"\n )\n btn = InlineKeyboardButton(btn_text, callback_data=item['id'])\n keyboard.append([btn])\n # Add button for None\n keyboard.append([InlineKeyboardButton('None, thanks', callback_data=0)])\n await update.message.reply_text(\n f\"Please select one of the torrents\",\n reply_markup=InlineKeyboardMarkup(keyboard, one_time_keyboard=True),\n )\n return DOWNLOAD_TORRENT\n else:\n await update.message.reply_text(\n \"We couldn't find any torrents for this title.\\n\"\n \"Would you like to add it to your watchlist?\",\n reply_markup=ReplyKeyboardMarkup(bool_keyboard, one_time_keyboard=True, resize_keyboard=True),\n )\n return WATCHLIST_NO_TORR\n\n\n@auth_wrap\nasync def confirm_redownload_action(update: Update, context: CallbackContext) -> int:\n \"\"\"Ask user if he really wants to download the movie given that\n he's aready seen it or it's already in transmission\"\"\"\n\n if update.message.text == 'Yes':\n if context.user_data['action'] == 'download':\n return await search_for_torrents(update, context)\n else:\n return await rate_title(update, context)\n else:\n return await reset(update, context)\n\n\n@auth_wrap\nasync def download_torrent(update: Update, context: CallbackContext) -> int:\n \"\"\"Send download request to transmission\"\"\"\n\n query = update.callback_query\n await query.answer()\n if query.data != '0':\n await query.edit_message_text(text=f\"Thanks, sending download request...\")\n # Send download request\n try:\n torr = [x for x in context.user_data['pkg']['torrents'] if query.data == str(x['id'])][0]\n # Send torrent to daemon\n torr_client_resp = send_torrent(compose_link(query.data))\n # Update torrent DB\n update_torr_db(torr, torr_client_resp, update.effective_user['id'])\n message = f\"Download started, have a great day!\"\n except TransmissionError as e:\n logger.error(f\"Error on torrent send: {e}\")\n message = f\"Download failed, please check logs and try again.\"\n await query.edit_message_text(text=message)\n if 'from_watchlist' in context.user_data.keys():\n return ConversationHandler.END\n return CHOOSE_TASK\n else:\n if 'from_watchlist' in context.user_data.keys():\n await query.edit_message_text(text=\"Ok, i'll remove these torrent options from future watchlist alerts \"\n \"regarding this movie.\")\n return await exclude_res_from_watchlist(update, context)\n else:\n await update.effective_message.reply_text('Would you like to add it to your watchlist?',\n reply_markup=ReplyKeyboardMarkup(bool_keyboard,\n one_time_keyboard=True,\n resize_keyboard=True))\n return WATCHLIST_NO_TORR\n\n\n@auth_wrap\nasync def exclude_res_from_watchlist(update: Update, context: CallbackContext) -> int:\n \"\"\"Function gets called only when a watchlist allert appears.\n If the user refuses to download, the current movie resolution will be\n excluded from further alerts.\"\"\"\n\n torrents = context.user_data['pkg']['torrents']\n movie_id = deconvert_imdb_id(torrents[0]['imdb'])\n exclude_torrents_from_watchlist(movie_id, update.effective_user['id'], [x['id'] for x in torrents])\n await update.callback_query.edit_message_text(text=\"Removed these torrent quality for future recommendations \"\n \"on this title. Type anything to get started\")\n return ConversationHandler.END\n\n\n@auth_wrap\nasync def add_to_watchlist_no_torrent(update: Update, context: CallbackContext) -> int:\n \"\"\"Add movie to watchlist if no torrent is currently available.\n If torrents for the movie ARE available but the user still wants to add to\n watchlist, the movie is added but user will be notified only when another\n resolution for the movie becomes available.\"\"\"\n\n if update.message.text == 'Yes':\n # Try to get user's IMDB id if he has any\n try:\n user = get_user_by_tgram_id(update.effective_user['id'])\n except Exception as e:\n logger.warning(f\"Error upon retrieving IMDB id for user with tgram_id {update.effective_user['id']} \"\n f\"might not have any. Error: {e}\")\n user = None\n if 'torrents' not in context.user_data['pkg'].keys():\n add_to_watchlist(deconvert_imdb_id(context.user_data['pkg']['imdb']), user, 'new')\n else:\n add_to_watchlist(deconvert_imdb_id(context.user_data['pkg']['imdb']), user, 'new',\n [x['id'] for x in context.user_data['pkg']['torrents']])\n message = \"Added to watchlist!\"\n await update.message.reply_text(message)\n\n return await reset(update, context)\n\n\n\"\"\"<< CHECK DOWNLOAD PROGRESS >>\"\"\"\n\n\n@auth_wrap\nasync def get_download_progress(update: Update, context: CallbackContext) -> int:\n \"\"\"Return the status for the last 10 torrents for this user\"\"\"\n\n user = update.effective_user['id']\n torrents = get_progress(user, logger=logger)\n if torrents:\n for torrent in torrents[:5]:\n await update.message.reply_text(f\"{torrent['TorrentName']}\\n\"\n f\"Resolution: {torrent['Resolution']}\\n\"\n f\"Status: {torrent['Status']}\\n\"\n f\"Progress: {torrent['Progress']}\\n\"\n f\"ETA: {torrent['ETA']}\")\n else:\n await update.message.reply_text(\"No torrents to show :(.\")\n return await reset(update, context)\n\n\n\"\"\"<< UPLOAD ACTIVITY >>\"\"\"\n\n\n@auth_wrap\nasync def netflix_rate_or_not(update: Update, context: CallbackContext) -> int:\n \"\"\"User chooses to receive or not notifications for unrated titles.\"\"\"\n\n if update.message.text == 'No':\n context.user_data['send_notifications'] = False\n else:\n context.user_data['send_notifications'] = True\n await update.message.reply_text(\"K, now upload the .csv file please.\")\n return NETFLIX_CSV\n\n\n@auth_wrap\nasync def netflix_csv(update: Update, context: CallbackContext) -> int:\n \"\"\"Sends CSV file in order to be processed\"\"\"\n\n csv_context = {\n 'user': update.effective_user.id,\n 'file': update.message.document.file_id,\n }\n await update.message.reply_text(\"Thanks!We started the upload process, we'll let you know \"\n \"when it's done or if there's any trouble.\")\n context.job_queue.run_once(\n callback=csv_upload_handler,\n context=csv_context,\n when=10\n )\n return await reset(update, context)\n\n\n@auth_wrap\nasync def netflix_no_csv(update: Update, context: CallbackContext) -> int:\n await update.message.reply_text(\"Upload the .csv or hit /reset.\")\n return NETFLIX_CSV\n\n\n\"\"\"<< RATE A TITLE >>\"\"\"\n\n\n@auth_wrap\nasync def choose_what_to_rate(update: Update, context: CallbackContext) -> int:\n if update.message.text == 'New title':\n message = 'Great, give me an IMDB id, a title or an IMDB link.'\n context.user_data['action'] = 'rate'\n await update.message.reply_text(message)\n return DOWNLOAD_MOVIE\n\n elif update.message.text == 'Rate seen movies':\n await update.message.reply_text(\"Preparing movies...\")\n context.user_data['unrated_movies'] = get_movies_for_bulk_rating(update.effective_user['id'])\n if context.user_data['unrated_movies']:\n return await rate_multiple(update, context)\n else:\n await update.message.reply_text(\"You have no unrated movies!\")\n return await start(update, context, \"Can i help you with anything else?\")\n\n\n@auth_wrap\nasync def rate_multiple(update: Update, context: CallbackContext) -> int:\n movies = context.user_data['unrated_movies']\n if movies:\n movie = movies.pop(0)\n # Check again if we can find it\n pkg = get_movie_from_all_databases(movie['imdb_id'], update.effective_user['id'])\n if pkg:\n context.user_data['pkg'] = pkg\n message, image = make_movie_reply(pkg)\n message += \"\\nPlease choose a rating\"\n await update.effective_message.reply_photo(\n photo=image,\n caption=message,\n reply_markup=ReplyKeyboardMarkup(rate_keyboard_bulk,\n one_time_keyboard=True,\n resize_keyboard=True,\n ),\n )\n context.user_data['rate_origin'] = 'multiple'\n return SUBMIT_RATING\n else:\n return await rate_multiple(update, context)\n else:\n await update.effective_message.reply_text(\"No more movies left, good job!\")\n return await start(update, context, \"Can i help you with anything else?\")\n\n\n@auth_wrap\nasync def rating_movie_info(update: Update, context: CallbackContext) -> int:\n movie = context.user_data['pkg']\n # Check if you've already seen it and send info\n if movie['already_in_my_movies']:\n message = f\"Movie seen\"\n if 'seen_date' in movie.keys():\n message = message + f\" on {movie['seen_date']}\"\n if 'my_score' in movie.keys():\n message += f\"\\nYour score: {movie['my_score']}\"\n await update.message.reply_text(message)\n message = f\"\\nWould you like to rate it again?\"\n await update.effective_message.reply_html(message, reply_markup=ReplyKeyboardMarkup(bool_keyboard,\n one_time_keyboard=True,\n resize_keyboard=True,\n ))\n return CONFIRM_REDOWNLOAD_ACTION\n else:\n return await rate_title(update, context)\n\n\n@auth_wrap\nasync def rate_title(update: Update, context: CallbackContext) -> int:\n context.user_data['rate_origin'] = 'simple'\n message = f\"Great, please choose a rating:\"\n await update.effective_message.reply_html(message, reply_markup=ReplyKeyboardMarkup(rate_keyboard,\n one_time_keyboard=True,\n resize_keyboard=True,\n ))\n return SUBMIT_RATING\n\n\n@auth_wrap\nasync def rate_title_plex_triggered(update: Update, context: CallbackContext, passed_args) -> int:\n context.user_data['pkg'] = {\n 'imdb': int(passed_args)\n }\n return await rate_title(update, context)\n\n\n@auth_wrap\nasync def submit_rating(update: Update, context: CallbackContext) -> int:\n \"\"\"User receives a message from an external routine\n If he clicks on it this function gets triggered.\"\"\"\n if context.user_data['rate_origin'] == 'simple':\n return1_func = reset\n message1 = \"Ok, no worries! I won't bother you about this title anymore.\\n\" \\\n \"Have a great day!\"\n return2_func = reset\n else:\n return1_func = rate_multiple\n message1 = \"Ok, no worries! I won't bother you about this title anymore.\\n\"\n return2_func = rate_multiple\n\n if update.message.text in [str(x) for x in list(range(1, 11))]:\n # Got rating\n item = get_my_movie_by_imdb(context.user_data['pkg']['imdb'], update.effective_user['id'])\n if item:\n item['rating_status'] = 'rated in telegram'\n item['my_score'] = int(update.message.text)\n item['seen_date'] = datetime.datetime.now()\n else:\n item = {\n 'imdb_id': context.user_data['pkg']['imdb'],\n 'my_score': int(update.message.text),\n 'rating_status': 'rated in telegram',\n 'user_id': update.effective_user['id'],\n 'seen_date': datetime.datetime.now(),\n }\n update_many([item], Movie, Movie.id)\n await update.effective_message.reply_text(f\"Ok, great! Here's a link if you also want to rate it on IMDB:\\n\"\n f\"https://www.imdb.com/title/\"\n f\"{convert_imdb_id(context.user_data['pkg']['imdb'])}/\",\n disable_web_page_preview=True)\n return await return1_func(update, context)\n\n elif update.message.text in [rate_keyboard[-1][0], rate_keyboard_bulk[-2][0]]:\n item = get_my_movie_by_imdb(context.user_data['pkg']['imdb'], update.effective_user['id'])\n if item:\n item['rating_status'] = 'refused to rate'\n else:\n item = {\n 'imdb_id': context.user_data['pkg']['imdb'],\n 'rating_status': 'refused to rate',\n 'user_id': update.effective_user['id'],\n }\n update_many([item], Movie, Movie.id)\n await update.effective_message.reply_text(message1)\n return await return2_func(update, context)\n elif update.message.text == rate_keyboard_bulk[-1][0]:\n await update.effective_message.reply_text(\"Ok, your progress is saved, come back anytime.\")\n return await start(update, context)\n else:\n await update.effective_message.reply_text(\"Please choose an option from the keyboard.\")\n return SUBMIT_RATING\n\n\n\"\"\"<< AUTHENTICATION >>\"\"\"\n\n\nasync def check_user(update: Update, context: CallbackContext):\n if context.user_data['user_type'] == 'user':\n onetime_passwords = get_onetimepasswords()\n passwords = [x['password'] for x in onetime_passwords]\n expiry_dates = [x['expiry'] for x in onetime_passwords]\n user_types = [x['user_type'] for x in onetime_passwords]\n try:\n pwd = int(update.message.text)\n if pwd in passwords and datetime.datetime.now() < expiry_dates[passwords.index(pwd)]:\n remove_onetimepassword(pwd)\n context.user_data['user_type'] = user_types[passwords.index(pwd)]\n return await password_ok(update, context)\n else:\n await update.effective_message.reply_text(\"Password expired, please contact \"\n \"the admin and try again.\")\n return REGISTER_USER\n except ValueError:\n pass\n else:\n if update.message.text.lower() == SUPERADMIN_PASSWORD:\n return await password_ok(update, context)\n await update.effective_message.reply_text(\"Incorrect password\")\n return REGISTER_USER\n\n\nasync def password_ok(update: Update, context: CallbackContext):\n await update.effective_message.reply_photo(\n photo=open(TELEGRAM_AUTH_APPROVE, 'rb'),\n caption=\"Welcome! Just a few more steps to configure your preferences. \"\n \"First, please type in your email so that we can add you \"\n \"to our PLEX users.\",\n )\n return CHECK_EMAIL\n\n\nasync def check_email(update: Update, context: CallbackContext):\n # Invite to PLEX server\n email_invite = invite_friend(update.message.text)\n if email_invite:\n message = \"Great! An invitation for PLEX has been sent to your email.\\n\"\n else:\n message = \"Looks like either this email is already in our PLEX users database \" \\\n \"OR you're not planning to use PLEX.\\n\" \\\n \"If this is not the case, please contact the admin.\\n\\n\"\n\n # Continue to IMDB stuff\n context.user_data['new_user'] = {\n 'telegram_chat_id': update.message.chat_id,\n 'telegram_name': update.effective_user.first_name,\n 'email': update.message.text,\n 'email_newsletters': True,\n 'scan_watchlist': False,\n 'user_type': context.user_data['user_type']\n\n }\n message += \"Would you like to connect your IMDB account? \" \\\n \"In this way we'll be able to pull your movie \" \\\n \"ratings and warn you when you'll search for a movie \" \\\n \"you've already seen.\\n\" \\\n \"We'll also scan your watchlist periodically and notify you \" \\\n \"when we'll be able to download any of the titles there.\\n\" \\\n \"In the future we're planning to be able to \" \\\n \"give ratings here and transfer them to IMDB.\"\n await update.effective_message.reply_text(message, reply_markup=ReplyKeyboardMarkup(bool_keyboard,\n one_time_keyboard=True,\n resize_keyboard=True,\n ))\n return GIVE_IMDB\n\n\nasync def give_imdb(update: Update, context: CallbackContext):\n if update.message.text == 'Yes':\n await update.effective_message.reply_photo(\n photo=open(TELEGRAM_IMDB_RATINGS, 'rb'),\n caption=\"I'll need you to go to your IMDB account and copy here your user ID, like the one in the photo, \"\n \"ur77571297. Also make sure that your Ratings are PUBLIC and so is your Watchlist (10 pages max).\\n\"\n \"If this is too much, just type 'fuck it' and skip this step.\\n\"\n \"https://www.imdb.com/\",\n )\n return CHECK_IMDB\n else:\n return await register_user(update, context)\n\n\nasync def check_imdb(update: Update, context: CallbackContext):\n if update.message.text.lower() != 'fuck it':\n context.user_data['new_user']['scan_watchlist'] = True\n context.user_data['new_user']['imdb_id'] = ''.join([x for x in update.message.text if x.isdigit()])\n return await register_user(update, context)\n\n\nasync def register_user(update: Update, context: CallbackContext):\n global USERS\n # Update user to database\n update_many([context.user_data['new_user']], User, User.telegram_chat_id)\n USERS = get_telegram_users()\n message = \"Ok, that's it\\. I'll take care of the rest, from now on \" \\\n \"anytime you type something i'll be here to help you out\\. Enjoy\\!\\n\" \\\n \"Type /help to find out more\\.\"\n return await start(update, context, message)\n\n\ndef wrong_input(update: Update, context: CallbackContext):\n update.effective_message.reply_text(\"Wrong input, please try again.\")\n return CHECK_EMAIL\n\n\ndef wrong_input_imdb(update: Update, context: CallbackContext):\n update.effective_message.reply_text(\"Wrong input, please try again.\")\n return CHECK_IMDB\n\n\n\"\"\"<< DOWNLOAD MOVIE DATABASE (CSV) >>\"\"\"\n\n\n@auth_wrap\nasync def download_csv(update: Update, context: CallbackContext) -> None:\n csv_context = {\n 'user': update.effective_user.id,\n }\n context.job_queue.run_once(\n callback=csv_download_handler,\n context=csv_context,\n when=0\n )\n return None\n\n\n\"\"\"<< OTHER FUNCTIONS >>\"\"\"\n\n\n@auth_wrap\nasync def help_command(update: Update, context: CallbackContext) -> None:\n \"\"\"Displays info on how to use the bot.\"\"\"\n\n watchlist_status = 'MONITORING' if USERS[update.effective_user.id]['scan_watchlist'] == 1 else 'NOT MONITORING'\n email_status = 'RECEIVING' if USERS[update.effective_user.id]['email_newsletters'] else 'NOT RECEIVING'\n generate_password = '\\n\\nGENERATE CODE FOR NEW USER: run command /generate_pass. If you want the user to have ' \\\n 'admin privileges, use -admin flag (/generate_pass -admin)' if \\\n USERS[update.effective_user.id]['user_type'] == 'admin' else ' '\n await update.message.reply_text(\"Type anything for the bot to start.\\n\\n\"\n \"If i lose my shit just type /reset anytime.\\n\\n\"\n f\"Right now we are {watchlist_status} your watchlist. \"\n f\"Type /change_watchlist \"\n \"to reverse the status.\\n\\n\"\n f\"Right now you are {email_status} the email newsletters. Type /change_newsletter \"\n \"to reverse the status.\\n\\n\"\n \"If you want to change your email address or your imdb ID type /update_user \"\n \"and we'll ask you to retake the login process. Once started, you must complete \"\n \"the entire process.\"\n f\"{generate_password}\")\n # don't change state\n return None\n\n\n@auth_wrap\nasync def generate_password(update: Update, context: CallbackContext) -> None:\n def insert_pwd(pwd):\n try:\n insert_onetimepasswords(pwd)\n except sqlalchemy.exc.IntegrityError:\n pwd['password'] = randint(10000, 99999)\n insert_pwd(pwd)\n return pwd['password']\n\n if not USERS[update.effective_user.id]['user_type'] == 'admin':\n return None\n else:\n arguments = (' '.join(context.args)).strip()\n pwd = {\n 'password': randint(10000, 99999),\n 'expiry': datetime.datetime.now() + datetime.timedelta(days=1)\n }\n if arguments == '-admin':\n pwd['user_type'] = 'admin'\n else:\n pwd['user_type'] = 'user'\n pwd = insert_pwd(pwd)\n await update.message.reply_text(f\"Token {pwd} available for 24 hours\")\n return None\n\n\n@auth_wrap\nasync def update_user(update: Update, context: CallbackContext) -> None:\n await update.message.reply_text('Type anything to get started.')\n del USERS[update.effective_user.id]\n\n\n@auth_wrap\nasync def watchlist_entry(update: Update, context: CallbackContext, *passed_args) -> int:\n context.user_data['pkg'] = {\n 'imdb': int(passed_args[0]),\n }\n context.user_data['from_watchlist'] = True\n await update.message.reply_text('Watchlist entry')\n return await search_for_torrents(update, context)\n\n\n@auth_wrap\nasync def remove_watchlist_entry(update: Update, context: CallbackContext, *passed_args) -> int:\n movie_id = int(passed_args[0])\n update_watchlist_item_status(movie_id, update.effective_user['id'], 'closed')\n await update.message.reply_text(\"Done, no more watchlist updates for this movie.\")\n return ConversationHandler.END\n\n\n@auth_wrap\nasync def keep_torrent(update: Update, context: CallbackContext, *passed_args) -> int:\n torr_id = int(passed_args[0])\n update_torrent_grace_days(torr_id, update.effective_user['id'])\n await update.message.reply_text(\"Ok, done.\")\n return ConversationHandler.END\n\n\n@auth_wrap\nasync def remove_torrent(update: Update, context: CallbackContext, *passed_args) -> int:\n db_torr_id = int(passed_args[0])\n # remove torrent and data\n client = make_client()\n torrents = client.get_torrents()\n db_torr = get_torrent_by_torr_id_user(db_torr_id, update.effective_user['id'])\n try:\n torrent = [x for x in torrents if x.hashString == db_torr['torr_hash']][0]\n except IndexError:\n await update.message.reply_text(\"Error while removing torrent,\\n\"\n \"please contact admin:).\")\n return ConversationHandler.END\n client.remove_torrent(torrent.id, delete_data=True)\n # change status\n update_torrent_status(db_torr_id, 'removed')\n await update.message.reply_text(\"Torrent and files removed.\")\n return ConversationHandler.END\n\n\n@auth_wrap\nasync def seed_forever_torrent(update: Update, context: CallbackContext, *passed_args) -> int:\n torr_id = int(passed_args[0])\n update_torrent_grace_days(torr_id, update.effective_user['id'], 99999)\n await update.message.reply_text(\"Done, SeedMaster.\")\n return ConversationHandler.END\n\n\n@auth_wrap\nasync def change_watchlist_command(update: Update, context: CallbackContext) -> None:\n pkg = USERS[update.effective_user.id]\n pkg['telegram_chat_id'] = update.effective_user.id\n if pkg['scan_watchlist'] == 0:\n pkg['scan_watchlist'] = 1\n else:\n pkg['scan_watchlist'] = 0\n update_many([pkg], User, User.telegram_chat_id)\n await update.message.reply_text(\"Updated your watchlist preferences.\")\n\n\n@auth_wrap\nasync def change_newsletter_command(update: Update, context: CallbackContext) -> None:\n pkg = USERS[update.effective_user.id]\n pkg['telegram_chat_id'] = update.effective_user.id\n if pkg['email_newsletters'] == 0:\n pkg['email_newsletters'] = 1\n else:\n pkg['email_newsletters'] = 0\n update_many([pkg], User, User.telegram_chat_id)\n await update.message.reply_text(\"Updated your newsletter preferences.\")\n\n\ndef main() -> None:\n \"\"\"\n Main function, runs bot and all other services\n \"\"\"\n\n global USERS\n check_database()\n USERS = get_telegram_users()\n\n application = ApplicationBuilder().token(TELEGRAM_TOKEN).build()\n\n download_movie_conversation_handler = ConversationHandler(\n entry_points=[\n RegexpCommandHandler(r'WatchMatch_[\\d]+', watchlist_entry),\n MessageHandler(filters.Regex('^([tT]{2})?\\d+$') & (~filters.COMMAND), parse_imdb_id),\n MessageHandler(filters.TEXT & (~filters.COMMAND), parse_imdb_text),\n ],\n states={\n CHOOSE_MULTIPLE: [\n MessageHandler(filters.TEXT & (~filters.COMMAND), choose_multiple)],\n CHOOSE_ONE: [\n MessageHandler(filters.TEXT & (~filters.COMMAND), accept_reject_title)],\n CONFIRM_REDOWNLOAD_ACTION: [\n MessageHandler(filters.TEXT & (~filters.COMMAND), confirm_redownload_action)],\n SEARCH_FOR_TORRENTS: [\n MessageHandler(filters.TEXT & (~filters.COMMAND), search_for_torrents)],\n DOWNLOAD_TORRENT: [\n CallbackQueryHandler(download_torrent)],\n WATCHLIST_NO_TORR: [\n MessageHandler(filters.TEXT & (~filters.COMMAND), add_to_watchlist_no_torrent), ],\n },\n fallbacks=[CommandHandler(\"reset\", reset)],\n map_to_parent={\n CHOOSE_TASK: CHOOSE_TASK,\n DOWNLOAD_MOVIE: DOWNLOAD_MOVIE,\n SUBMIT_RATING: SUBMIT_RATING,\n ConversationHandler.END: ConversationHandler.END\n }\n )\n\n check_progress_conversation_handler = ConversationHandler(\n entry_points=[\n MessageHandler(filters.TEXT, get_download_progress),\n CallbackQueryHandler(get_download_progress)],\n states={},\n fallbacks=[],\n map_to_parent={}\n )\n\n rate_title_conversation_handler = ConversationHandler(\n entry_points=[\n RegexpCommandHandler(r'RateTitle_[\\d]+', rate_title_plex_triggered),\n MessageHandler(filters.TEXT & (~filters.COMMAND), choose_what_to_rate), ],\n states={\n DOWNLOAD_MOVIE: [\n download_movie_conversation_handler\n ],\n SUBMIT_RATING: [\n MessageHandler(filters.TEXT & (~filters.COMMAND), submit_rating, )],\n },\n fallbacks=[],\n map_to_parent={\n CHOOSE_TASK: CHOOSE_TASK,\n RATE_TITLE: RATE_TITLE,\n ConversationHandler.END: ConversationHandler.END\n }\n )\n\n register_user_conversation_handler = ConversationHandler(\n entry_points=[\n MessageHandler(filters.TEXT & (~filters.COMMAND), check_user)],\n states={\n CHECK_EMAIL: [\n MessageHandler(filters.Regex('[^@]+@[^@]+\\.[^@]+') & (~filters.COMMAND), check_email),\n MessageHandler(filters.TEXT & (~filters.COMMAND), wrong_input, )],\n GIVE_IMDB: [\n MessageHandler(filters.TEXT & (~filters.COMMAND), give_imdb, )],\n CHECK_IMDB: [\n MessageHandler(filters.Regex('^[u]?[r]?\\d+$') & (~filters.COMMAND), check_imdb),\n MessageHandler(filters.TEXT & (~filters.COMMAND), wrong_input_imdb),\n ],\n },\n fallbacks=[CommandHandler('reset', reset)],\n map_to_parent={\n REGISTER_USER: REGISTER_USER,\n CHOOSE_TASK: CHOOSE_TASK,\n ConversationHandler.END: ConversationHandler.END\n }\n )\n\n conversation_handler = ConversationHandler(\n entry_points=[\n RegexpCommandHandler(r'WatchMatch_[\\d]+', watchlist_entry),\n RegexpCommandHandler(r'RateTitle_[\\d]+', rate_title_plex_triggered),\n MessageHandler(filters.TEXT & (~filters.COMMAND), start),\n ],\n states={\n CHOOSE_TASK: [MessageHandler(filters.TEXT & (~filters.COMMAND), choose_task)],\n DOWNLOAD_MOVIE: [download_movie_conversation_handler],\n CHECK_PROGRESS: [check_progress_conversation_handler],\n NETFLIX_CSV: [\n MessageHandler(filters.Document.FileExtension('csv') & (~filters.COMMAND), netflix_csv),\n MessageHandler(filters.TEXT & (~filters.COMMAND), netflix_no_csv), ],\n RATE_TITLE: [rate_title_conversation_handler],\n REGISTER_USER: [register_user_conversation_handler],\n DOWNLOAD_TORRENT: [CallbackQueryHandler(download_torrent)],\n SUBMIT_RATING: [MessageHandler(filters.TEXT & (~filters.COMMAND), submit_rating)],\n },\n fallbacks=[\n CommandHandler(\"reset\", reset),\n CommandHandler('help', help_command),\n CommandHandler('generate_pass', generate_password),\n CommandHandler('update_user', update_user),\n CommandHandler('change_watchlist', change_watchlist_command),\n CommandHandler('change_newsletter', change_newsletter_command),\n (RegexpCommandHandler(r'RateTitle_[\\d]+', rate_title_plex_triggered)),\n (RegexpCommandHandler(r'WatchMatch_[\\d]+', watchlist_entry)),\n (RegexpCommandHandler(r'UnWatchMatch_[\\d]+', remove_watchlist_entry)),\n (RegexpCommandHandler(r'Keep_[\\d]+', keep_torrent)),\n (RegexpCommandHandler(r'Remove_[\\d]+', remove_torrent)),\n (RegexpCommandHandler(r'SeedForever_[\\d]+', seed_forever_torrent)),\n\n ]\n )\n\n application.add_handler(conversation_handler)\n application.add_handler(CommandHandler('help', help_command))\n application.add_handler(CommandHandler('generate_pass', generate_password))\n application.add_handler(CommandHandler('update_user', update_user))\n application.add_handler(RegexpCommandHandler(r'RateTitle_[\\d]+', rate_title_plex_triggered))\n application.add_handler(RegexpCommandHandler(r'WatchMatch_[\\d]+', watchlist_entry))\n application.add_handler(RegexpCommandHandler(r'UnWatchMatch_[\\d]+', remove_watchlist_entry))\n application.add_handler(RegexpCommandHandler(r'Keep_[\\d]+', keep_torrent))\n application.add_handler(RegexpCommandHandler(r'Remove_[\\d]+', remove_torrent))\n application.add_handler(RegexpCommandHandler(r'SeedForever_[\\d]+', seed_forever_torrent))\n application.add_handler(CommandHandler('change_watchlist', change_watchlist_command))\n application.add_handler(CommandHandler('change_newsletter', change_newsletter_command))\n application.run_polling(stop_signals=None)\n\n\nif __name__ == '__main__':\n main()\n", "repo_name": "ironsilk/mSquaredPlex", "sub_path": "telegram_service/bot.py", "file_name": "bot.py", "file_ext": "py", "file_size_in_byte": 47995, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 5, "dataset": "github-code", "pt": "25", "api": [{"api_name": "logging.basicConfig", "line_number": 33, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 34, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 37, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 39, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 41, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 42, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 43, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 44, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 45, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 46, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 108, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 108, "usage_type": "name"}, {"api_name": "functools.wraps", "line_number": 107, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 137, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 137, "usage_type": "name"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 146, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 152, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext.DEFAULT_TYPE", "line_number": 152, "usage_type": "attribute"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 152, "usage_type": "name"}, {"api_name": "telegram.ext.ConversationHandler.END", "line_number": 157, "usage_type": "attribute"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 157, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 161, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 161, "usage_type": "name"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 178, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 206, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 206, "usage_type": "name"}, {"api_name": "bot_utils.get_movie_from_all_databases", "line_number": 215, "usage_type": "call"}, {"api_name": "bot_utils.make_movie_reply", "line_number": 219, "usage_type": "call"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 223, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 232, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 232, "usage_type": "name"}, {"api_name": "re.search", "line_number": 238, "usage_type": "call"}, {"api_name": "bot_utils.get_movie_from_all_databases", "line_number": 245, "usage_type": "call"}, {"api_name": "bot_utils.make_movie_reply", "line_number": 248, "usage_type": "call"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 252, "usage_type": "call"}, {"api_name": "bot_utils.search_imdb_title", "line_number": 267, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 274, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 274, "usage_type": "name"}, {"api_name": "bot_utils.get_movie_from_all_databases", "line_number": 285, "usage_type": "call"}, {"api_name": "bot_utils.make_movie_reply", "line_number": 289, "usage_type": "call"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 293, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 309, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 309, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 331, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 331, "usage_type": "name"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 351, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 361, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 361, "usage_type": "name"}, {"api_name": "bot_watchlist.get_torrents_for_imdb_id", "line_number": 366, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 386, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 389, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardMarkup", "line_number": 392, "usage_type": "call"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 399, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 405, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 405, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 419, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 419, "usage_type": "name"}, {"api_name": "utils.send_torrent", "line_number": 430, "usage_type": "call"}, {"api_name": "utils.compose_link", "line_number": 430, "usage_type": "call"}, {"api_name": "bot_utils.update_torr_db", "line_number": 432, "usage_type": "call"}, {"api_name": "transmission_rpc.TransmissionError", "line_number": 434, "usage_type": "name"}, {"api_name": "telegram.ext.ConversationHandler.END", "line_number": 439, "usage_type": "attribute"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 439, "usage_type": "name"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 448, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 455, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 455, "usage_type": "name"}, {"api_name": "utils.deconvert_imdb_id", "line_number": 461, "usage_type": "call"}, {"api_name": "bot_utils.exclude_torrents_from_watchlist", "line_number": 462, "usage_type": "call"}, {"api_name": "telegram.ext.ConversationHandler.END", "line_number": 465, "usage_type": "attribute"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 465, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 469, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 469, "usage_type": "name"}, {"api_name": "utils.get_user_by_tgram_id", "line_number": 478, "usage_type": "call"}, {"api_name": "bot_utils.add_to_watchlist", "line_number": 484, "usage_type": "call"}, {"api_name": "utils.deconvert_imdb_id", "line_number": 484, "usage_type": "call"}, {"api_name": "bot_utils.add_to_watchlist", "line_number": 486, "usage_type": "call"}, {"api_name": "utils.deconvert_imdb_id", "line_number": 486, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 498, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 498, "usage_type": "name"}, {"api_name": "bot_get_progress.get_progress", "line_number": 502, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 519, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 519, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 531, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 531, "usage_type": "name"}, {"api_name": "bot_csv.csv_upload_handler", "line_number": 541, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 549, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 549, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 558, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 558, "usage_type": "name"}, {"api_name": "utils.get_movies_for_bulk_rating", "line_number": 567, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 576, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 576, "usage_type": "name"}, {"api_name": "bot_utils.get_movie_from_all_databases", "line_number": 581, "usage_type": "call"}, {"api_name": "bot_utils.make_movie_reply", "line_number": 584, "usage_type": "call"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 589, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 604, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 604, "usage_type": "name"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 615, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 625, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 625, "usage_type": "name"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 628, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 636, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 636, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 644, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 644, "usage_type": "name"}, {"api_name": "utils.get_my_movie_by_imdb", "line_number": 659, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 663, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 663, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 670, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 670, "usage_type": "attribute"}, {"api_name": "utils.update_many", "line_number": 672, "usage_type": "call"}, {"api_name": "utils.Movie", "line_number": 672, "usage_type": "argument"}, {"api_name": "utils.Movie.id", "line_number": 672, "usage_type": "attribute"}, {"api_name": "utils.convert_imdb_id", "line_number": 675, "usage_type": "call"}, {"api_name": "utils.get_my_movie_by_imdb", "line_number": 680, "usage_type": "call"}, {"api_name": "utils.update_many", "line_number": 689, "usage_type": "call"}, {"api_name": "utils.Movie", "line_number": 689, "usage_type": "argument"}, {"api_name": "utils.Movie.id", "line_number": 689, "usage_type": "attribute"}, {"api_name": "telegram.Update", "line_number": 703, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 703, "usage_type": "name"}, {"api_name": "utils.get_onetimepasswords", "line_number": 705, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 711, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 711, "usage_type": "attribute"}, {"api_name": "utils.remove_onetimepassword", "line_number": 712, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 728, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 728, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 738, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 738, "usage_type": "name"}, {"api_name": "bot_utils.invite_friend", "line_number": 740, "usage_type": "call"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 766, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 773, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 773, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 787, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 787, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 794, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 794, "usage_type": "name"}, {"api_name": "utils.update_many", "line_number": 797, "usage_type": "call"}, {"api_name": "utils.User", "line_number": 797, "usage_type": "argument"}, {"api_name": "utils.User.telegram_chat_id", "line_number": 797, "usage_type": "attribute"}, {"api_name": "bot_utils.get_telegram_users", "line_number": 798, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 805, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 805, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 810, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 810, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 819, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 819, "usage_type": "name"}, {"api_name": "bot_csv.csv_download_handler", "line_number": 824, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 835, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 835, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 859, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 859, "usage_type": "name"}, {"api_name": "utils.insert_onetimepasswords", "line_number": 862, "usage_type": "call"}, {"api_name": "sqlalchemy.exc", "line_number": 863, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 864, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 873, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 874, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 874, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 874, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 886, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 886, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 892, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 892, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 902, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 902, "usage_type": "name"}, {"api_name": "bot_watchlist.update_watchlist_item_status", "line_number": 904, "usage_type": "call"}, {"api_name": "telegram.ext.ConversationHandler.END", "line_number": 906, "usage_type": "attribute"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 906, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 910, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 910, "usage_type": "name"}, {"api_name": "utils.update_torrent_grace_days", "line_number": 912, "usage_type": "call"}, {"api_name": "telegram.ext.ConversationHandler.END", "line_number": 914, "usage_type": "attribute"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 914, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 918, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 918, "usage_type": "name"}, {"api_name": "utils.make_client", "line_number": 921, "usage_type": "call"}, {"api_name": "utils.get_torrent_by_torr_id_user", "line_number": 923, "usage_type": "call"}, {"api_name": "telegram.ext.ConversationHandler.END", "line_number": 929, "usage_type": "attribute"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 929, "usage_type": "name"}, {"api_name": "utils.update_torrent_status", "line_number": 932, "usage_type": "call"}, {"api_name": "telegram.ext.ConversationHandler.END", "line_number": 934, "usage_type": "attribute"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 934, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 938, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 938, "usage_type": "name"}, {"api_name": "utils.update_torrent_grace_days", "line_number": 940, "usage_type": "call"}, {"api_name": "telegram.ext.ConversationHandler.END", "line_number": 942, "usage_type": "attribute"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 942, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 946, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 946, "usage_type": "name"}, {"api_name": "utils.update_many", "line_number": 953, "usage_type": "call"}, {"api_name": "utils.User", "line_number": 953, "usage_type": "argument"}, {"api_name": "utils.User.telegram_chat_id", "line_number": 953, "usage_type": "attribute"}, {"api_name": "telegram.Update", "line_number": 958, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 958, "usage_type": "name"}, {"api_name": "utils.update_many", "line_number": 965, "usage_type": "call"}, {"api_name": "utils.User", "line_number": 965, "usage_type": "argument"}, {"api_name": "utils.User.telegram_chat_id", "line_number": 965, "usage_type": "attribute"}, {"api_name": "utils.check_database", "line_number": 975, "usage_type": "call"}, {"api_name": "bot_utils.get_telegram_users", "line_number": 976, "usage_type": "call"}, {"api_name": "telegram.ext.ApplicationBuilder", "line_number": 978, "usage_type": "call"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 980, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 982, "usage_type": "call"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 983, "usage_type": "call"}, {"api_name": "telegram.ext.filters.Regex", "line_number": 983, "usage_type": "call"}, {"api_name": "telegram.ext.filters", "line_number": 983, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 983, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 984, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 984, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 984, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 984, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 988, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 988, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 988, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 988, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 990, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 990, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 990, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 990, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 992, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 992, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 992, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 992, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 994, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 994, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 994, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 994, "usage_type": "attribute"}, {"api_name": "telegram.ext.CallbackQueryHandler", "line_number": 996, "usage_type": "call"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 998, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 998, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 998, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 998, "usage_type": "attribute"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1000, "usage_type": "call"}, {"api_name": "telegram.ext.ConversationHandler.END", "line_number": 1005, "usage_type": "attribute"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 1005, "usage_type": "name"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 1009, "usage_type": "call"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1011, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 1011, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 1011, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackQueryHandler", "line_number": 1012, "usage_type": "call"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 1018, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1020, "usage_type": "call"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1021, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 1021, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 1021, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1021, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1027, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 1027, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 1027, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1027, "usage_type": "attribute"}, {"api_name": "telegram.ext.ConversationHandler.END", "line_number": 1033, "usage_type": "attribute"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 1033, "usage_type": "name"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 1037, "usage_type": "call"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1039, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 1039, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 1039, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1039, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1042, "usage_type": "call"}, {"api_name": "telegram.ext.filters.Regex", "line_number": 1042, "usage_type": "call"}, {"api_name": "telegram.ext.filters", "line_number": 1042, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1042, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1043, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 1043, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 1043, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1043, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1045, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 1045, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 1045, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1045, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1047, "usage_type": "call"}, {"api_name": "telegram.ext.filters.Regex", "line_number": 1047, "usage_type": "call"}, {"api_name": "telegram.ext.filters", "line_number": 1047, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1047, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1048, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 1048, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 1048, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1048, "usage_type": "attribute"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1051, "usage_type": "call"}, {"api_name": "telegram.ext.ConversationHandler.END", "line_number": 1055, "usage_type": "attribute"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 1055, "usage_type": "name"}, {"api_name": "telegram.ext.ConversationHandler", "line_number": 1059, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1061, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1062, "usage_type": "call"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1063, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 1063, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 1063, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1063, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1066, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 1066, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 1066, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1066, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1070, "usage_type": "call"}, {"api_name": "telegram.ext.filters.Document.FileExtension", "line_number": 1070, "usage_type": "call"}, {"api_name": "telegram.ext.filters.Document", "line_number": 1070, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 1070, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1070, "usage_type": "attribute"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1071, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 1071, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 1071, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1071, "usage_type": "attribute"}, {"api_name": "telegram.ext.CallbackQueryHandler", "line_number": 1074, "usage_type": "call"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 1075, "usage_type": "call"}, {"api_name": "telegram.ext.filters.TEXT", "line_number": 1075, "usage_type": "attribute"}, {"api_name": "telegram.ext.filters", "line_number": 1075, "usage_type": "name"}, {"api_name": "telegram.ext.filters.COMMAND", "line_number": 1075, "usage_type": "attribute"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1078, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1079, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1080, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1081, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1082, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1083, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1084, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1085, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1086, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1087, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1088, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1089, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1095, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1096, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1097, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1098, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1099, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1100, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1101, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1102, "usage_type": "call"}, {"api_name": "command_regex_handler.RegexpCommandHandler", "line_number": 1103, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1104, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 1105, "usage_type": "call"}]}
+{"seq_id": "31212702663", "text": "'''\nYou are given a stream of points on the X-Y plane. Design an algorithm that:\n\n- Adds new points from the stream into a data structure. Duplicate points are allowed and should be treated as different points.\n- Given a query point, counts the number of ways to choose three points from the data structure such that the three points and the query point form an axis-aligned square with positive area.\n\nAn axis-aligned square is a square whose edges are all the same length and are either parallel or perpendicular to the x-axis and y-axis.\n\nImplement the DetectSquares class:\n\n- DetectSquares() Initializes the object with an empty data structure.\n- void add(int[] point) Adds a new point point = [x, y] to the data structure.\n- int count(int[] point) Counts the number of ways to form axis-aligned squares with point point = [x, y] as described above.\n'''\nfrom time import time\nfrom typing import List\n\n\nclass DetectSquares:\n def __init__(self, debug=False):\n # Need at least 2x 2 different points for each\n self.p_count = {}\n self.p_arr = []\n self.debug = debug\n\n def add(self, point: List[int]) -> None:\n x, y = point\n if (x, y) not in self.p_count:\n self.p_count[(x, y)] = 1\n else:\n self.p_count[(x, y)] += 1\n self.p_arr.append(point)\n\n def count(self, point: List[int]) -> int:\n res = 0\n px, py = point\n for x, y in self.p_arr:\n if (abs(px - x) != abs(py - y)) or x == px or y == py:\n continue\n a = self.p_count[(x, py)]\n b = self.p_count[(px, y)]\n res += a * b\n\n if self.debug:\n print(res)\n\n return res\n\n\nif __name__ == '__main__':\n test = DetectSquares(debug=True)\n sol_start = time()\n test.add([3, 10])\n test.add([11, 2])\n test.add([3, 2])\n test.count([11, 10]) # Return 1\n test.count([14, 8]) # Return 0\n test.add([11, 2]) # Duplicate allowed\n test.add([11, 1]) # Testing, should not add another square\n test.add([3, 1]) # Testing, should not add another square\n test.count([11, 10]) # Return 2\n print(f'Runtime for our solution: {time() - sol_start}\\n')\n", "repo_name": "stevenxchung/l33t-code-problems", "sub_path": "NeetCode +75/17 - Math and Geometry/2013-detect-squares.py", "file_name": "2013-detect-squares.py", "file_ext": "py", "file_size_in_byte": 2199, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "typing.List", "line_number": 26, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 34, "usage_type": "name"}, {"api_name": "time.time", "line_number": 52, "usage_type": "call"}, {"api_name": "time.time", "line_number": 62, "usage_type": "call"}]}
+{"seq_id": "41061890223", "text": "import os\nimport pandas as pd\nimport numpy as np\n#import nltk\nimport cytoolz\n#%%\nfrom suggestion import clustering\nfrom scipy.misc import logsumexp\n#%%\nfrom suggestion.paths import paths\nclizer = pd.read_pickle(os.path.join(paths.parent, 'models', 'goal_oriented_suggestion_data.pkl'))\n#%%\ndata = pd.read_pickle(os.path.join(paths.parent, 'yelp_preproc/all_data.pkl'))\nreviews = data['data'].reset_index(drop=True)\n\n#%%\ndef get_topic_distribution(clizer, target_dist, sent_cluster_distribs, new_dists_opts):\n from scipy.special import kl_div\n with_new_dist = np.array([np.concatenate((sent_cluster_distribs, new_dist_opt[None]), axis=0) for new_dist_opt in new_dists_opts])\n dist_with_new_dist = clustering.normalize_dists(np.mean(with_new_dist, axis=1))\n return kl_div(dist_with_new_dist, target_dist).sum(axis=1)\n\n#%%\nif False:\n scores_by_cluster = clizer.scores_by_cluster.copy()\n likelihood_bias = logsumexp(scores_by_cluster, axis=1, keepdims=True)\n scores_by_cluster -= .85 * likelihood_bias\n scores_by_cluster[suggested_already] = -np.inf\n scores_by_cluster[clizer.omit] = -np.inf\n most_distinctive = np.argmax(scores_by_cluster, axis=0)\n#%%\ntopic_tags = [f'' for i in range(10)]\n\n#%%\n\n# FIXME: this approach is super-biased for predicting topic tags because the topic tags repeat for every sentence.\n# Better would be to separately train topics.\ndef review_to_tagged_sents(sents):\n cluster_distances = cytoolz.thread_first(\n sents,\n clizer.vectorize_sents,\n clustering.normalize_vecs,\n clizer.clusterer.transform)\n clusters_for_sents = np.argmin(cluster_distances, axis=1)\n\n res = []\n for i, sent in enumerate(sents):\n res.append([topic_tags[c] for c in clusters_for_sents[:i+1][-4:]] + sent.lower().split())\n return res\n\nimport tqdm\nfrom suggestion import util\nutil.dump_kenlm('yelp_topic_tagged', [\n ' '.join(s)\n for tokenized in tqdm.tqdm(reviews.tokenized)\n for s in review_to_tagged_sents(tokenized.split('\\n'))])\n#%%\nfrom suggestion import lang_model\ntopic2sentence_lm = lang_model.Model.from_basename(paths.model_basename('yelp_topic_tagged'))\n#%%\nimport itertools\ntopic_transitions_indices = list(itertools.product(range(10), range(10)))\nrev_topic_transitions_indices = [10*i+i for i in range(10)]\n#%%\ntransition_log_likelihoods = np.array([[topic2sentence_lm.score_seq(topic2sentence_lm.get_state([topic_tags[c1], topic_tags[c2]], bos=True)[0], k)[0] for c1, c2 in itertools.product(range(10), range(10))] for k in tqdm.tqdm(clizer.unique_starts, desc=\"Score starts\")])\n#%%\n#scores_by_cluster = scores_by_cluster_raw.copy()\n#likelihood_bias = logsumexp(scores_by_cluster, axis=1, keepdims=True)\n#%%\n#unconditional_likelihood_bias = np.array([[topic2sentence_lm.score_seq(topic2sentence_lm.get_state([topic_tags[c]], bos=True)[0], k)[0] for c in range(10)] for k in tqdm.tqdm(clizer.unique_starts, desc=\"Score starts\")])\nunconditional_likelihood_bias_2 = np.array([\n logsumexp(scores_by_cluster_raw[:,10*i:10*(i+1)], axis=1) for i in range(10)]).T\n#%%\nscores_by_cluster = transition_log_likelihoods - .9*logsumexp(transition_log_likelihoods, axis=1, keepdims=True)#[:,rev_topic_transitions_indices] - 1. * unconditional_likelihood_bias_2\nscores_by_cluster = scores_by_cluster[:,rev_topic_transitions_indices]\nfor cluster_idx in range(clizer.n_clusters):\n i = cluster_idx# + cluster_idx*10\n# print(topic_transitions_indices[i])\n print(i)\n for idx in np.argsort(scores_by_cluster[:,i])[-5:][::-1]:\n print(' '.join(clizer.unique_starts[idx]))\n print('\\n\\n')\n#%%\nfor i in np.argsort(scores_by_cluster[:,8])[-10:]: print(' '.join(clizer.unique_starts[i]))\n#%%\nnp.save('topic_continuation_scores.npy', scores_by_cluster)\n\n#%%\n#%%\ndef get_topic_seq(sents):\n cluster_distances = cytoolz.thread_first(\n sents,\n clizer.vectorize_sents,\n clustering.normalize_vecs,\n clizer.clusterer.transform)\n return np.argmin(cluster_distances, axis=1)\ntopic_seqs = [get_topic_seq(tokenized.split('\\n')) for tokenized in tqdm.tqdm(reviews.tokenized)]\n#%%\n# TODO: This actually needs to pass --discount_fallback to lmplz, and may want to use a high order like 12-gram\nutil.dump_kenlm('yelp_topic_seqs', [' '.join(topic_tags[c] for c in seq) for seq in topic_seqs])\n#%%\ndef review_to_tagged_sents(topic_seq, sents):\n res = []\n for i, sent in enumerate(sents):\n res.append([topic_tags[c] for c in topic_seq[:i+1][-4:]] + sent.lower().split())\n return res\n#%%\n\ntopic_seq_model = suggestion_generator.get_model('yelp_topic_seqs')\ntopic_word_indices = [topic_seq_model.model.vocab_index(tag) for tag in suggestion_generator.topic_tags]", "repo_name": "kcarnold/suggestion", "sub_path": "tmp/tagged_sents.py", "file_name": "tagged_sents.py", "file_ext": "py", "file_size_in_byte": 4707, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "26", "api": [{"api_name": "pandas.read_pickle", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "suggestion.paths.paths.parent", "line_number": 11, "usage_type": "attribute"}, {"api_name": "suggestion.paths.paths", "line_number": 11, "usage_type": "name"}, {"api_name": "pandas.read_pickle", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "suggestion.paths.paths.parent", "line_number": 13, "usage_type": "attribute"}, {"api_name": "suggestion.paths.paths", "line_number": 13, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 19, "usage_type": "call"}, {"api_name": "suggestion.clustering.normalize_dists", "line_number": 20, "usage_type": "call"}, {"api_name": "suggestion.clustering", "line_number": 20, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 20, "usage_type": "call"}, {"api_name": "scipy.special.kl_div", "line_number": 21, "usage_type": "call"}, {"api_name": "scipy.misc.logsumexp", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 30, "usage_type": "call"}, {"api_name": "cytoolz.thread_first", "line_number": 39, "usage_type": "call"}, {"api_name": "suggestion.clustering.normalize_vecs", "line_number": 42, "usage_type": "attribute"}, {"api_name": "suggestion.clustering", "line_number": 42, "usage_type": "name"}, {"api_name": "numpy.argmin", "line_number": 44, "usage_type": "call"}, {"api_name": "suggestion.util.dump_kenlm", "line_number": 53, "usage_type": "call"}, {"api_name": "suggestion.util", "line_number": 53, "usage_type": "name"}, {"api_name": "tqdm.tqdm", "line_number": 55, "usage_type": "call"}, {"api_name": "suggestion.lang_model.Model.from_basename", "line_number": 59, "usage_type": "call"}, {"api_name": "suggestion.lang_model.Model", "line_number": 59, "usage_type": "attribute"}, {"api_name": "suggestion.lang_model", "line_number": 59, "usage_type": "name"}, {"api_name": "suggestion.paths.paths.model_basename", "line_number": 59, "usage_type": "call"}, {"api_name": "suggestion.paths.paths", "line_number": 59, "usage_type": "name"}, {"api_name": "itertools.product", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 65, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 65, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "scipy.misc.logsumexp", "line_number": 72, "usage_type": "call"}, {"api_name": "scipy.misc.logsumexp", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 86, "usage_type": "call"}, {"api_name": "cytoolz.thread_first", "line_number": 91, "usage_type": "call"}, {"api_name": "suggestion.clustering.normalize_vecs", "line_number": 94, "usage_type": "attribute"}, {"api_name": "suggestion.clustering", "line_number": 94, "usage_type": "name"}, {"api_name": "numpy.argmin", "line_number": 96, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 97, "usage_type": "call"}, {"api_name": "suggestion.util.dump_kenlm", "line_number": 100, "usage_type": "call"}, {"api_name": "suggestion.util", "line_number": 100, "usage_type": "name"}]}
+{"seq_id": "14244890128", "text": "import os\nimport requests\nimport time\nimport sched, time\nimport get_temp\n\nserver = \"https://nikitech.eu/\"\ncontroller = \"nikihome/\"\nfunction = \"temperature_post.php\"\nurl = server + controller + function\n\nhour = 60 * 60\n\nscheduler = sched.scheduler(time.time, time.sleep)\n\ndef post_temperature(scheduler_param):\n \n inside_temp = get_temp.read_temp(get_temp.inside_device)\n outside_temp = get_temp.read_temp(get_temp.outside_device)\n date = int(round(time.time() * 1000))\n \n inside_temp = str(inside_temp)\n outside_temp = str(outside_temp)\n date = str(date)\n \n print(\"Posting temperature: \")\n print(\"inside: \" + inside_temp)\n print(\"outside: \" + outside_temp)\n print(\"______________________\")\n print(\"\\n\")\n \n data = {\n 'inside': inside_temp,\n 'outside': outside_temp,\n 'date': date\n }\n\n\n request = requests.post(url, data = data)\n print(\"Response: \" + request.text)\n \n if scheduler_param is not None:\n scheduler.enter(hour, 1, post_temperature, (scheduler_param,))\n\nprint(\"Uploading current temperature\")\npost_temperature(None)\n\nprint(\"Next upload will be in one hour\")\nscheduler.enter(hour, 1, post_temperature, (scheduler,))\nscheduler.run()\n\n\n\n\n", "repo_name": "Nikituh/scripts", "sub_path": "nikiberry/post_temp.py", "file_name": "post_temp.py", "file_ext": "py", "file_size_in_byte": 1239, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "sched.scheduler", "line_number": 14, "usage_type": "call"}, {"api_name": "time.time", "line_number": 14, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 14, "usage_type": "attribute"}, {"api_name": "get_temp.read_temp", "line_number": 18, "usage_type": "call"}, {"api_name": "get_temp.inside_device", "line_number": 18, "usage_type": "attribute"}, {"api_name": "get_temp.read_temp", "line_number": 19, "usage_type": "call"}, {"api_name": "get_temp.outside_device", "line_number": 19, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 20, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 39, "usage_type": "call"}]}
+{"seq_id": "40564563018", "text": "# coding=utf-8\n\nimport requests\nimport grequests\nfrom lxml import html\n\nfrom metacrawler.base import Element\nfrom metacrawler.fields import Field\n\n\nclass Crawler(Element):\n\n \"\"\"Crawler parse page use fields as rules. May be nested.\"\"\"\n\n def __init__(self):\n super().__init__()\n\n self.data = []\n\n if not self.fields:\n raise ValueError('Cannot use `Crawler` without fields.')\n\n if len(self.fields) > 1 and self.collapse:\n raise ValueError(\n 'Must not use `collapse` with few fields or/and crawlers.'\n )\n\n @property\n def fields(self):\n fields = {}\n\n candidates = {}\n candidates.update(self.__dict__)\n candidates.update(self.__class__.__dict__)\n\n for name, attribute in candidates.items():\n if isinstance(attribute, (Crawler, Field)):\n fields[name] = attribute\n\n return fields\n\n def get_url(self):\n return getattr(self.__class__, 'url', None)\n\n def get_pagination(self):\n return getattr(self.__class__, 'pagination', None)\n\n def get_collapse(self):\n return getattr(self.__class__, 'collapse', False)\n\n def get_limit(self):\n return getattr(self.__class__, 'limit', None)\n\n def get_session(self):\n return getattr(self.__class__, 'session', requests.Session())\n\n def get_timeout(self):\n return getattr(self.__class__, 'timeout', 3.0)\n\n def get_authentication(self):\n return getattr(self.__class__, 'authentication', None)\n\n def crawl(self, *args, **kwargs):\n \"\"\"Crawl page.\n\n :returns: `dict` data.\n \"\"\"\n self.before()\n data = []\n\n if self.authentication is not None:\n self.session = self.authentication.authentication(self.session)\n\n def parse(response):\n page = html.fromstring(response.content)\n\n if self.limit is not None and int(self.limit) <= len(data):\n return page\n\n if self.collapse:\n field = list(self.fields.items())[0][1]\n if field.to is list:\n data.extend(field.crawl(page))\n else:\n data.append(field.crawl(page))\n else:\n data.append({n: f.crawl(page) for n, f in self.fields.items()})\n\n return page\n\n try:\n iterator = iter(self.pagination)\n except TypeError:\n iterator = None\n\n if iterator:\n requests_list = []\n\n for url in iterator:\n requests_list.append(grequests.request(\n 'GET', url, session=self.session, timeout=self.timeout\n ))\n\n for response in grequests.map(requests_list):\n if response:\n parse(response)\n else:\n while self.url:\n page = parse(self.session.get(self.url, verify=False))\n self.paginate(page)\n\n if self.pagination:\n self.data.extend(data)\n else:\n self.data = data[0]\n\n return self.clean(self.data)\n\n def paginate(self, page):\n \"\"\"Paginate.\n\n :param page: `lxml.Element` instance.\n \"\"\"\n if self.pagination:\n self.url = self.pagination.next(page)\n else:\n self.url = None\n", "repo_name": "dvemnt/metacrawler", "sub_path": "metacrawler/crawlers.py", "file_name": "crawlers.py", "file_ext": "py", "file_size_in_byte": 3365, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "metacrawler.base.Element", "line_number": 11, "usage_type": "name"}, {"api_name": "metacrawler.fields.Field", "line_number": 37, "usage_type": "name"}, {"api_name": "requests.Session", "line_number": 55, "usage_type": "call"}, {"api_name": "lxml.html.fromstring", "line_number": 75, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 75, "usage_type": "name"}, {"api_name": "grequests.request", "line_number": 100, "usage_type": "call"}, {"api_name": "grequests.map", "line_number": 104, "usage_type": "call"}]}
+{"seq_id": "37642706514", "text": "from flask import Flask, render_template, redirect, request, url_for\nfrom functools import wraps\nfrom functions import get_necessary_number_of_card_names\napp = Flask(__name__)\n\n\n@app.route('/')\ndef index_page():\n return render_template('index.html')\n\n\n@app.route('/game', methods=['GET'])\ndef game_page():\n rows = int(request.args.get('row_num'))\n cols = int(request.args.get('column_num'))\n if rows == 0 or cols == 0:\n return redirect(url_for('index_page'))\n cards = get_necessary_number_of_card_names(cols, rows)\n return render_template('game.html', rows=rows, cols=cols, cards=cards)\n\n\n@app.errorhandler(404)\ndef handle_404(e):\n return render_template('error.html', code=404), 404\n\n\nif __name__ == '__main__':\n app.secret_key = 'magic'\n app.run(debug=True, port=5000)\n", "repo_name": "CodecoolBP20172/wswp-memory-game-DanielKnoll", "sub_path": "server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 805, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "flask.Flask", "line_number": 4, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 9, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 14, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 15, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 17, "usage_type": "call"}, {"api_name": "functions.get_necessary_number_of_card_names", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 24, "usage_type": "call"}]}
+{"seq_id": "35641561695", "text": "from tqdm import tqdm\nimport torch\nimport csv\nimport numpy as np\nfrom loss import IOU\n\ndef Train(model, epochs, optimizer, scheduler, loss_fn, train_iter, val_iter, device, early_stopping, DeepLab=False):\n # Create dictionary to store history\n Loss = {\"val_iou\":[]}\n\n with open('./models/readme.txt', 'w') as f:\n f.write('Model: {} \\n Max Epochs: {} \\n Optimizer: {} \\n Scheduler: {} \\n Loss Function: {} \\n Device: {} \\n Early Stopping: {}'.format(model, epochs, optimizer, scheduler, loss_fn, device, early_stopping))\n\n with open('./models/log.csv', 'w') as f:\n writer = csv.writer(f)\n writer.writerow([\"epoch\",\"train_loss\",\"train_iou\",\"val_loss\",\"val_iou\"])\n \n # Set patience to zero.\n patience = 0\n\n for epoch in range(epochs):\n # Set model in training mode\n model.train()\n\n # Initialise cumulative loss\n train_loss, train_iou, val_loss, val_iou = 0, 0, 0, 0\n \n # Print LR if it has decreased.\n if epoch != 0:\n if optimizer.param_groups[0]['lr'] < LR:\n print('Learning rate decreased to ', optimizer.param_groups[0]['lr'])\n else:\n print('Initial learning rate set to ', optimizer.param_groups[0]['lr'])\n LR = optimizer.param_groups[0]['lr']\n\n # Loop over the training set\n for i, data in enumerate(tqdm(train_iter)):\n inputs, labels = data[0].to(device), data[1].to(device)\n\n # Zero previous gradients\n optimizer.zero_grad()\n \n \n # Generate predictions and loss with current model parameters\n if DeepLab == True:\n outputs = model(inputs)[\"out\"]\n else:\n outputs = model(inputs)\n loss = loss_fn(outputs, labels)\n\n # Initiate backpropagation to adjust loss weights\n loss.backward()\n optimizer.step()\n\n # Update total training loss\n train_loss += loss\n train_iou += IOU(outputs, labels, device)\n train_steps = i+1\n\n with torch.no_grad():\n # Set the model to evaluation mode\n model.eval()\n\n # Loop over the validation set\n for i, data in enumerate(tqdm(val_iter)):\n inputs, labels = data[0].to(device), data[1].to(device)\n\n # Calculate validation loss\n if DeepLab == True:\n outputs = model(inputs)[\"out\"]\n else:\n outputs = model(inputs)\n val_loss += loss_fn(outputs, labels)\n val_iou += IOU(outputs, labels, device)\n val_steps = i+1\n \n # Calculate the average training and validation loss\n avg_train_loss = float(train_loss / train_steps)\n avg_train_iou = float(train_iou / train_steps)\n avg_val_loss = float(val_loss / val_steps)\n avg_val_iou = float(val_iou / val_steps)\n \n if scheduler is not None:\n scheduler.step(avg_val_loss)\n\n # Save the best model if appropriate, else continue.\n if epoch == 0:\n torch.save(model.state_dict(), \"./models/model.pth\")\n print(\"Saved best model!\")\n elif avg_val_iou > np.max(Loss[\"val_iou\"]):\n torch.save(model.state_dict(), './models/model.pth')\n print(\"Saved best model!\")\n patience = 0\n else:\n patience += 1\n\n # Update train and val loss history\n Loss[\"val_iou\"].append(avg_val_iou)\n\n with open('./models/log.csv', 'a') as csv_file:\n dict_object = csv.DictWriter(csv_file, fieldnames=[\"epoch\",\"train_loss\",\"train_iou\",\"val_loss\",\"val_iou\"])\n dict_object.writerow({\"epoch\":epoch,\"train_loss\":avg_train_loss,\"train_iou\":avg_train_iou,\"val_loss\":avg_val_loss,\"val_iou\":avg_val_iou})\n\n print(\"Epoch {}, Train Loss {:3f}, Train IOU {:3f}, Val Loss {:3f}, Val IOU {:3f}\".format(\n epoch, avg_train_loss, avg_train_iou, avg_val_loss, avg_val_iou))\n \n if patience > early_stopping:\n print(\"Early stopping triggered, best val IOU: {}\".format(np.max(Loss[\"val_iou\"])))\n break", "repo_name": "christianmcb/Deep-Learning-Project", "sub_path": "code/training.py", "file_name": "training.py", "file_ext": "py", "file_size_in_byte": 4212, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "csv.writer", "line_number": 15, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 37, "usage_type": "call"}, {"api_name": "loss.backward", "line_number": 52, "usage_type": "call"}, {"api_name": "loss.IOU", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 60, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 65, "usage_type": "call"}, {"api_name": "loss.IOU", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 91, "usage_type": "call"}, {"api_name": "csv.DictWriter", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 108, "usage_type": "call"}]}
+{"seq_id": "17126482545", "text": "from selenium import webdriver\nfrom selenium.webdriver.common.keys import Keys\n\n# Specify the path to your browser driver executable\n# Example: For Chrome, you may use chromedriver.exe\ndriver_path = '/path/to/driver'\n\n# Create a new instance of the browser driver\ndriver = webdriver.Chrome(executable_path=driver_path)\n\n# Open the website in the browser\nwebsite_url = 'https://www.example.com'\ndriver.get(website_url)\n\n# Perform an action on the website\nelement = driver.find_element_by_id('button-id') # Replace 'button-id' with the actual ID of the button\nelement.click()\n\n# Additional actions can be performed here, such as filling forms or navigating through the website\n\n# Extract information from the web page\nelement = driver.find_element_by_xpath('//div[@class=\"info\"]') # Replace with the appropriate XPath to locate the desired element\ntext = element.text\nprint(f\"Extracted information: {text}\")\n\n# Close the browser\ndriver.quit()\n", "repo_name": "drissraki/test1", "sub_path": "arkx/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 943, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 9, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 9, "usage_type": "name"}]}
+{"seq_id": "1800374324", "text": "from flask import Flask, request, render_template\nimport Bot\nimport wikipedia as wk\nimport wolframalpha\n\nAPIkey = \"R62U64-JYYVGAVHHU\"\nwolf = wolframalpha.Client(APIkey)\n\n\napp = Flask(__name__)\n\n@app.route(\"/\")\ndef index():\n return render_template(\"index.html\")\n@app.route(\"/get\")\ndef chat():\n flag = True\n while (flag == True):\n user_response = request.args.get('msg')\n\n user_response = user_response.lower()\n if (user_response != 'bye'):\n if (user_response == 'thanks' or user_response == 'thank you'):\n flag = False\n response = \"You are welcome..\"\n return response\n else:\n if (Bot.greeting(user_response) != None):\n response = Bot.greeting(user_response)\n return response\n else:\n try:\n res = wolf.query(user_response)\n response = next(res.results).text\n except:\n try:\n response = wk.summary(user_response, sentences = 3)\n except:\n response = \"May you try to refine your query, I am still learning please be patient with me\"\n\n return response\n else:\n flag = False\n response = \"Good Bye! \"\n return response\n\nif __name__ == \"__main__\":\n app.run()\n", "repo_name": "rapha18th/phonicwolf", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1453, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "wolframalpha.Client", "line_number": 7, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 19, "usage_type": "name"}, {"api_name": "Bot.greeting", "line_number": 28, "usage_type": "call"}, {"api_name": "Bot.greeting", "line_number": 29, "usage_type": "call"}, {"api_name": "wikipedia.summary", "line_number": 37, "usage_type": "call"}]}
+{"seq_id": "45085745510", "text": "\"\"\"\nCreated on Fri Feb 12 12:54:33 2021\n\n@author: Ben Horowitz (and friends)\n\nFrom lenspack implementation\n\"\"\"\nimport tensorflow as tf\nimport numpy as np\nimport tensorflow_probability as tfp\nfrom DifferentiableHOS.transforms import starlet2d\n\n\ndef _kernel(bw, X, x):\n \"\"\"Gaussian kernel for KDE\"\"\"\n return (1.0 / np.sqrt(2 * np.pi) / bw) * tf.math.exp(-((X - x)**2) /\n (bw**2 * 2.0))\n\n\ndef _get_wavelet_normalization(image, nscales):\n \"\"\" Computes normalizing constant for starlet, for given image.\n \"\"\"\n _, nx, ny = image.get_shape()\n knorm = tf.ones((1, 1, 1, 1), dtype=tf.float32)\n knorm = tf.image.resize_with_crop_or_pad(knorm, nx, ny)\n wt = starlet2d(knorm[..., 0], nscales=nscales)\n return [tf.math.sqrt(tf.reduce_sum(c**2)) for c in wt]\n\n\n@tf.function\ndef find_peaks2d_tf(image, mask=None, ordered=True, threshold=None):\n if mask is not None:\n # mask = np.atleast_2d(mask)\n if mask.shape != image.shape:\n print(\"Warning: mask not compatible with image -> ignoring.\")\n mask = tf.ones(image.shape)\n else:\n # Make sure mask is binary, i.e. turn nonzero values into ones\n mask = tf.cast(tf.cast(mask, bool), float)\n else:\n mask = tf.ones(image.shape)\n\n if threshold is None:\n threshold = tf.math.reduce_min(image)\n else:\n threshold = tf.math.reduce_max((threshold, tf.math.reduce_min(image)))\n\n offset = tf.math.reduce_min(image)\n threshold = threshold - offset\n image = image - offset\n\n map0 = image[1:-1, 1:-1]\n\n # Extract shifted maps\n map1 = image[0:-2, 0:-2]\n map2 = image[1:-1, 0:-2]\n map3 = image[2:, 0:-2]\n map4 = image[0:-2, 1:-1]\n map5 = image[2:, 1:-1]\n map6 = image[0:-2, 2:]\n map7 = image[1:-1, 2:]\n map8 = image[2:, 2:]\n\n merge = ((map0 > map1) & (map0 > map2) & (map0 > map3) & (map0 > map4) &\n (map0 > map5) & (map0 > map6) & (map0 > map7) & (map0 > map8))\n\n bordered = tf.pad(merge,\n tf.constant(((1, 1), (1, 1))),\n constant_values=0.0)\n peaksmap = tf.cast(bordered, float) * image * mask\n XY = tf.where(peaksmap > threshold)\n heights = tf.gather_nd(image, XY) + offset\n\n if ordered:\n ind = tf.argsort(heights)[::-1]\n return tf.gather(XY[:, 0],\n ind), tf.gather(XY[:, 1],\n ind), tf.gather(heights, ind)\n return XY[:, 0], XY[:, 1], heights\n\n\n@tf.function\ndef peaks_histogram_tf_mulscale(image,\n nscales=3,\n bins=None,\n mask=None,\n name='peakscount',\n bw_factor=2.):\n \"\"\"Compute a histogram of peaks in a 2d Starlet transform of the input image.\n\n Parameters\n ----------\n image : tensor (2D)\n Two-dimensional input tensor\n nscales: int\n Number of wavelet scales to include\n in the decomposition.\n value_range: Shape [2] Tensor of same dtype as image\n Range of values in the Histogram.\n bins : int or tensor (1D), optional\n Specification of centers or the number of bins to use for the\n histogram. If not provided, a default of 10 bins linearly spaced\n between the image minimum and maximum (inclusive) is used.\n mask : tensor (same shape as `image`), optional\n Tensor identifying which pixels of `image` to consider/exclude\n in finding peaks. Can either either be numeric (0 or 1) or boolean \n (false or true)\n bw_factor: float\n Factor by which to divide the bin width to define the bandwidth of the\n smoothing kernel.\n Returns\n -------\n results, bins : list of 1D tensors\n Histogram and bin boundary values. \n \"\"\"\n with tf.name_scope(name):\n image = tf.cast(image, dtype=tf.float32)\n\n # Compute the wavelet normalization factor\n norm_factors = _get_wavelet_normalization(image, nscales)\n\n # Compute wavelet transform\n wt = starlet2d(image, nscales)\n results = []\n # Loop over all wavelet scales\n for coeffs, factor in zip(wt, norm_factors):\n # Normalizing coefficients to preserve standard deviations\n # across scales\n coeffs = coeffs / factor\n\n # Histogram the coefficient values\n image = tf.reshape(coeffs, [coeffs.shape[1], coeffs.shape[2]])\n if bins is None:\n bins = tf.linspace(tf.math.reduce_min(image),\n tf.math.reduce_max(image), 10)\n elif isinstance(bins, int):\n bins = tf.linspace(tf.math.reduce_min(image),\n tf.math.reduce_max(image), bins)\n else:\n bins = bins\n\n x, y, heights = find_peaks2d_tf(image, threshold=None, mask=mask)\n\n # To avoid issues, we clip the image to within the peaks\n heights = tf.clip_by_value(heights, bins[0], bins[-1])\n w = tf.reshape(tf.ones_like(heights), [-1])\n k = _kernel(\n tf.reduce_mean((bins[1:] - bins[:-1])) / bw_factor,\n tf.reshape(heights, [-1, 1]), bins)\n k = k / tf.reduce_sum(k, axis=1, keepdims=True)\n counts = tf.tensordot(k, w, axes=[[0], [0]])\n results.append(counts)\n return results, bins\n\n\n@tf.function\ndef peaks_histogram_tf(image, bins=None, mask=None, bw_factor=2.):\n \"\"\"Compute a histogram of peaks in a 2d image.\n Parameters\n ----------\n image : tensor (2D)\n Two-dimensional input tensor\n bins : int or tensor (1D), optional\n Specification of centers or the number of bins to use for the\n histogram. If not provided, a default of 10 bins linearly spaced\n between the image minimum and maximum (inclusive) is used.\n mask : tensor (same shape as `image`), optional\n Tensor identifying which pixels of `image` to consider/exclude\n in finding peaks. Can either either be numeric (0 or 1) or boolean \n (false or true)\n bw_factor: float\n Factor by which to divide the bin width to define the bandwidth of the\n smoothing kernel.\n Returns\n -------\n counts, bins : tuple of 1D tensors\n Histogram and bin boundary values. If the returned `counts` has \n N values, `bin_edges` will have N + 1 values.\n \"\"\"\n image = tf.cast(image, dtype=tf.float32)\n if bins is None:\n bins = tf.linspace(tf.math.reduce_min(image),\n tf.math.reduce_max(image), 10)\n elif isinstance(bins, int):\n bins = tf.linspace(tf.math.reduce_min(image),\n tf.math.reduce_max(image), bins)\n else:\n bins = bins\n\n x, y, heights = find_peaks2d_tf(image, threshold=None, mask=mask)\n\n # To avoid issues, we clip the image to within the peaks\n heights = tf.clip_by_value(heights, bins[0], bins[-1])\n\n w = tf.reshape(tf.ones_like(heights), [-1])\n k = _kernel(\n tf.reduce_mean((bins[1:] - bins[:-1])) / bw_factor,\n tf.reshape(heights, [-1, 1]), bins)\n k = k / tf.reduce_sum(k, axis=1, keepdims=True)\n counts = tf.tensordot(k, w, axes=[[0], [0]])\n\n return counts, bins\n\n\n@tf.function\ndef non_diffable_peaks_histogram_tf(image, bins=None, mask=None):\n \"\"\"Compute a histogram of peaks in a 2d image.\n\n CAREFULL: This implementation is not differentiable\n\n Parameters\n ----------\n image : tensor (2D)\n Two-dimensional input tensor\n bins : int or tensor (1D), optional\n Specification of bin edges or the number of bins to use for the\n histogram. If not provided, a default of 10 bins linearly spaced\n between the image minimum and maximum (inclusive) is used.\n mask : tensor (same shape as `image`), optional\n Tensor identifying which pixels of `image` to consider/exclude\n in finding peaks. Can either either be numeric (0 or 1) or boolean \n (false or true)\n Returns\n -------\n counts, bins : tuple of 1D tensors\n Histogram and bin boundary values. If the returned `counts` has \n N values, `bin_edges` will have N + 1 values.\n \"\"\"\n if bins is None:\n bins = tf.linspace(tf.math.reduce_min(image),\n tf.math.reduce_max(image), 10)\n elif isinstance(bins, int):\n bins = tf.linspace(tf.math.reduce_min(image),\n tf.math.reduce_max(image), bins)\n else:\n bins = bins\n\n x, y, heights = find_peaks2d_tf(image, threshold=None, mask=mask)\n # To avoid issues, we clip the image to within the peaks\n heights = tf.clip_by_value(heights, bins[0], bins[-1])\n\n counts = tfp.stats.histogram(heights, bins)\n return counts, bins\n", "repo_name": "LSSTDESC/DifferentiableHOS", "sub_path": "DifferentiableHOS/statistics/peak_counts_tf.py", "file_name": "peak_counts_tf.py", "file_ext": "py", "file_size_in_byte": 8896, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3, "dataset": "github-code", "pt": "26", "api": [{"api_name": "numpy.sqrt", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 16, "usage_type": "attribute"}, {"api_name": "tensorflow.math.exp", "line_number": 16, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 16, "usage_type": "attribute"}, {"api_name": "tensorflow.ones", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tensorflow.image.resize_with_crop_or_pad", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 25, "usage_type": "attribute"}, {"api_name": "DifferentiableHOS.transforms.starlet2d", "line_number": 26, "usage_type": "call"}, {"api_name": "tensorflow.math.sqrt", "line_number": 27, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 27, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 27, "usage_type": "call"}, {"api_name": "tensorflow.ones", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.ones", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.math.reduce_min", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tensorflow.math.reduce_max", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 46, "usage_type": "attribute"}, {"api_name": "tensorflow.math.reduce_min", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.math.reduce_min", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 48, "usage_type": "attribute"}, {"api_name": "tensorflow.pad", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.where", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.gather_nd", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.argsort", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.gather", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.gather", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.gather", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.function", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 117, "usage_type": "attribute"}, {"api_name": "DifferentiableHOS.transforms.starlet2d", "line_number": 123, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 132, "usage_type": "call"}, {"api_name": "tensorflow.linspace", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.math.reduce_min", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 134, "usage_type": "attribute"}, {"api_name": "tensorflow.math.reduce_max", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 135, "usage_type": "attribute"}, {"api_name": "tensorflow.linspace", "line_number": 137, "usage_type": "call"}, {"api_name": "tensorflow.math.reduce_min", "line_number": 137, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 137, "usage_type": "attribute"}, {"api_name": "tensorflow.math.reduce_max", "line_number": 138, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 138, "usage_type": "attribute"}, {"api_name": "tensorflow.clip_by_value", "line_number": 145, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 146, "usage_type": "call"}, {"api_name": "tensorflow.ones_like", "line_number": 146, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 149, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.tensordot", "line_number": 151, "usage_type": "call"}, {"api_name": "tensorflow.function", "line_number": 82, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 180, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 180, "usage_type": "attribute"}, {"api_name": "tensorflow.linspace", "line_number": 182, "usage_type": "call"}, {"api_name": "tensorflow.math.reduce_min", "line_number": 182, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 182, "usage_type": "attribute"}, {"api_name": "tensorflow.math.reduce_max", "line_number": 183, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 183, "usage_type": "attribute"}, {"api_name": "tensorflow.linspace", "line_number": 185, "usage_type": "call"}, {"api_name": "tensorflow.math.reduce_min", "line_number": 185, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 185, "usage_type": "attribute"}, {"api_name": "tensorflow.math.reduce_max", "line_number": 186, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 186, "usage_type": "attribute"}, {"api_name": "tensorflow.clip_by_value", "line_number": 193, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 195, "usage_type": "call"}, {"api_name": "tensorflow.ones_like", "line_number": 195, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 197, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 198, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 199, "usage_type": "call"}, {"api_name": "tensorflow.tensordot", "line_number": 200, "usage_type": "call"}, {"api_name": "tensorflow.function", "line_number": 156, "usage_type": "attribute"}, {"api_name": "tensorflow.linspace", "line_number": 230, "usage_type": "call"}, {"api_name": "tensorflow.math.reduce_min", "line_number": 230, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 230, "usage_type": "attribute"}, {"api_name": "tensorflow.math.reduce_max", "line_number": 231, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 231, "usage_type": "attribute"}, {"api_name": "tensorflow.linspace", "line_number": 233, "usage_type": "call"}, {"api_name": "tensorflow.math.reduce_min", "line_number": 233, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 233, "usage_type": "attribute"}, {"api_name": "tensorflow.math.reduce_max", "line_number": 234, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 234, "usage_type": "attribute"}, {"api_name": "tensorflow.clip_by_value", "line_number": 240, "usage_type": "call"}, {"api_name": "tensorflow_probability.stats.histogram", "line_number": 242, "usage_type": "call"}, {"api_name": "tensorflow_probability.stats", "line_number": 242, "usage_type": "attribute"}, {"api_name": "tensorflow.function", "line_number": 205, "usage_type": "attribute"}]}
+{"seq_id": "38595173578", "text": "from django.shortcuts import render\nfrom .models import Book, Person\nfrom datetime import date\n\n# Create your views here.\ndef home(request):\n # pass Books and people from database to template\n books = Book.objects.all()\n people = Person.objects.all().count()\n\n # keeps count of how many books are checked out, late, and due today\n late = 0\n due_today = 0\n\n # counts how many books are late and due today\n today = date.today()\n today = int(str(today.year)+str(today.month).zfill(2)+str(today.day).zfill(2))\n for book in books.filter(available=False):\n if today > int(book.due_date): # count how many books are late\n late = late + 1\n if today == int(book.due_date): # count how many books are due today\n due_today = due_today + 1\n checked_out = books.filter(available=False).count() - due_today - late\n context ={\n 'books': books,\n 'person': people,\n 'out': checked_out,\n 'late': late,\n 'today': due_today,\n 'Home': \"active\",\n }\n return render(request, 'active_template/index.html', context)\n\ndef checkedout_books(request):\n today = date.today()\n todays_date = int(str(today.year)+str(today.month).zfill(2)+str(today.day).zfill(2))\n books_out = Book.objects.filter(available=False).order_by('due_date')\n\n context = {\n 'books_out': books_out,\n 'today': todays_date\n }\n return render(request, 'active_template/checkedout.html', context)\n\ndef due_today(request):\n today = date.today()\n todays_date = int(str(today.year)+str(today.month).zfill(2)+str(today.day).zfill(2))\n due = Book.objects.filter(due_date=todays_date)\n return render(request, 'active_template/duetoday.html', {'due': due})\n\ndef overdue(request):\n today = date.today()\n todays_date = int(str(today.year) + str(today.month).zfill(2) + str(today.day).zfill(2))\n overdue = Book.objects.filter(available=False).order_by('-due_date')\n\n context = {\n 'overdue': overdue,\n 'today': todays_date\n }\n return render(request, 'active_template/overdue.html', context)\n\ndef help(request):\n return render(request, 'active_template/help.html', {'Help': 'active'})", "repo_name": "luismoralesXD/FBLA_Coding_and_Programming", "sub_path": "BookManager/Database/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2210, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "models.Book.objects.all", "line_number": 8, "usage_type": "call"}, {"api_name": "models.Book.objects", "line_number": 8, "usage_type": "attribute"}, {"api_name": "models.Book", "line_number": 8, "usage_type": "name"}, {"api_name": "models.Person.objects.all", "line_number": 9, "usage_type": "call"}, {"api_name": "models.Person.objects", "line_number": 9, "usage_type": "attribute"}, {"api_name": "models.Person", "line_number": 9, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 16, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 32, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 35, "usage_type": "name"}, {"api_name": "models.Book.objects.filter", "line_number": 37, "usage_type": "call"}, {"api_name": "models.Book.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.Book", "line_number": 37, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 46, "usage_type": "name"}, {"api_name": "models.Book.objects.filter", "line_number": 48, "usage_type": "call"}, {"api_name": "models.Book.objects", "line_number": 48, "usage_type": "attribute"}, {"api_name": "models.Book", "line_number": 48, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 52, "usage_type": "name"}, {"api_name": "models.Book.objects.filter", "line_number": 54, "usage_type": "call"}, {"api_name": "models.Book.objects", "line_number": 54, "usage_type": "attribute"}, {"api_name": "models.Book", "line_number": 54, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 60, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 63, "usage_type": "call"}]}
+{"seq_id": "660704872", "text": "import os\nimport re\nfrom utils import getLines, mkFileIfNot\n\n\nINIT_ENVS = {\n \"DJA_PATH\": os.path.dirname(__file__)\n}\n\n\nclass Env:\n def __init__(self, v=False):\n self.v = v\n mkFileIfNot('.djaenv')\n\n \n def getEnvs(self):\n envs = {}\n lines = getLines('.djaenv')\n\n for i in range(len(lines)):\n match = re.match(r'([A-Z_]+)=(.+)', lines[i])\n \n if re.match:\n envs[match.group(1)] = match.group(2)\n \n return envs\n\n\n def getEnv(self, name):\n envs = self.getEnvs()\n if name in envs:\n return envs[name]\n else:\n if self.v == True:\n print(f'env \"{name}\" not found /_/')\n return -1\n\n\n def setEnv(self, name, value):\n \"\"\"\n If env found set value of env and return 0\n If env not found create env and return 1 \n \"\"\"\n rv = 0\n envs = self.getEnvs()\n if not name in envs:\n rv = 1\n\n envs[name] = value\n with open(\".djaenv\", \"w\") as f:\n f.write(\"\\n\".join([f'{k}={envs[k]}' for k in envs]))\n \n return rv\n\nENV = Env()\n\nfor k, v in INIT_ENVS.items():\n ENV.setEnv(k, v)", "repo_name": "naraSokami/djaVite", "sub_path": "env.py", "file_name": "env.py", "file_ext": "py", "file_size_in_byte": 1242, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "utils.mkFileIfNot", "line_number": 14, "usage_type": "call"}, {"api_name": "utils.getLines", "line_number": 19, "usage_type": "call"}, {"api_name": "re.match", "line_number": 22, "usage_type": "call"}, {"api_name": "re.match", "line_number": 24, "usage_type": "attribute"}]}
+{"seq_id": "39481577064", "text": "import os\nfrom typing import Optional\n\nimport yaml\n\nfrom agent.commons.collector_exceptions import CollectorException\n\n\nclass CollectorDefinitions:\n COLLECTOR_DEFINITIONS_FILENAME = \"collector_definitions.yaml\"\n\n _collector_globals: Optional[dict] = None\n _collector_inputs: Optional[dict] = None\n\n @staticmethod\n def get_collector_globals() -> dict:\n if CollectorDefinitions._collector_globals is None:\n CollectorDefinitions._load_collector_definitions()\n\n return CollectorDefinitions._collector_globals\n\n @staticmethod\n def get_input_definitions(input_name: str) -> dict:\n if CollectorDefinitions._collector_inputs is None:\n CollectorDefinitions._load_collector_definitions()\n\n return CollectorDefinitions._collector_inputs.get(input_name)\n\n @staticmethod\n def _load_collector_definitions() -> None:\n loaded_collector_definitions = False\n with open(\n os.path.join(\n os.getcwd(),\n \"config_internal\",\n CollectorDefinitions.COLLECTOR_DEFINITIONS_FILENAME\n )\n ) as service_definitions_file:\n service_definitions_content = yaml.safe_load(service_definitions_file)\n if service_definitions_content:\n CollectorDefinitions._collector_globals = service_definitions_content.get(\"collector_globals\")\n CollectorDefinitions._collector_inputs = service_definitions_content.get(\"collector_inputs\")\n loaded_collector_definitions = True\n if loaded_collector_definitions is False:\n raise CollectorException(0, \"Collector definitions were not loaded\")\n", "repo_name": "sahil-metron/All-neccessary-code-check", "sub_path": "agent/collectordefinitions/collector_definitions.py", "file_name": "collector_definitions.py", "file_ext": "py", "file_size_in_byte": 1704, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "typing.Optional", "line_number": 12, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 13, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 34, "usage_type": "call"}, {"api_name": "yaml.safe_load", "line_number": 39, "usage_type": "call"}, {"api_name": "agent.commons.collector_exceptions.CollectorException", "line_number": 45, "usage_type": "call"}]}
+{"seq_id": "74408059907", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# @Refer from: https://github.com/MaybeShewill-CV/CRNN_Tensorflow/blob/master/crnn_model/crnn_model.py\n# @Site : http://github.com/TJCVRS\n\n\"\"\"\nImplement the crnn model mentioned in An End-to-End Trainable Neural Network for Image-based Sequence\nRecognition and Its Application to Scene Text Recognition paper\n\"\"\"\nimport os\nimport os.path as ops\nimport time\n\nimport numpy as np\nfrom collections import namedtuple\nfrom utils import tf_utils,log_utils\n\nimport tensorflow as tf\n#from utils import tf_extended\n\nfrom tensorflow.contrib import layers as tflayers\nfrom tensorflow.contrib import rnn\n\n\nfrom nets.crnn import cnn_basenet\nlogger = log_utils.init_logger()\n\nHIDDEN_NUMS = 256\nNUM_CLASSES = 10\n\n# Test setting\nis_recursive = False # Recursively test the dataset \nloops_nums = 100\n\nclass CRNNnet(cnn_basenet.CNNBaseModel):\n \"\"\"\n Implement the crnn model for squence recognition\n \"\"\"\n #def __init__(self, phase, hidden_nums, layers_nums, seq_length, num_classes):\n def __init__(self, phase, num_classes=NUM_CLASSES+1):\n \"\"\"\n\n :param phase:\n \"\"\"\n super(CRNNnet, self).__init__()\n self.__phase = phase\n #self.__hidden_nums = hidden_nums\n #self.__layers_nums = layers_nums\n #self.__seq_length = seq_length\n self.__num_classes = num_classes\n return\n\n @property\n def phase(self):\n \"\"\"\n\n :return:\n \"\"\"\n return self.__phase\n\n @phase.setter\n def phase(self, value):\n \"\"\"\n\n :param value:\n :return:\n \"\"\"\n if not isinstance(value, str):\n raise TypeError('value should be a str \\'Test\\' or \\'Train\\'')\n if value.lower() not in ['test', 'train']:\n raise ValueError('value should be a str \\'Test\\' or \\'Train\\'')\n self.__phase = value.lower()\n return\n\n def __conv_stage(self, inputdata, out_dims, name=None):\n \"\"\"\n Traditional conv stage in VGG format\n :param inputdata:\n :param out_dims:\n :return:\n \"\"\"\n conv = self.conv2d(inputdata=inputdata, out_channel=out_dims, kernel_size=3, stride=1, use_bias=False, name=name)\n relu = self.relu(inputdata=conv)\n max_pool = self.maxpooling(inputdata=relu, kernel_size=2, stride=2)\n return max_pool\n\n def __feature_sequence_extraction(self, inputdata):\n \"\"\"\n Implement the 2.1 Part Feature Sequence Extraction\n :param inputdata: eg. batch*32*128*3 NHWC format\n :return:\n end_points: a set of activations for external use, for example summaries or\n losses.\n \"\"\"\n # end_points collect relevant activations for external use.\n end_points = {}\n \n conv1 = self.__conv_stage(inputdata=inputdata, out_dims=64, name='conv1') # batch*16*64*64\n end_points['conv1'] = conv1\n conv2 = self.__conv_stage(inputdata=conv1, out_dims=128, name='conv2') # batch*8*32*128\n end_points['conv2'] = conv2\n conv3 = self.conv2d(inputdata=conv2, out_channel=256, kernel_size=3, stride=1, use_bias=False, name='conv3') # batch*8*32*256 \n relu3 = self.relu(conv3) # batch*8*32*256 \n end_points['conv3'] = relu3\n conv4 = self.conv2d(inputdata=relu3, out_channel=256, kernel_size=3, stride=1, use_bias=False, name='conv4') # batch*8*32*256\n relu4 = self.relu(conv4) # batch*8*32*256\n max_pool4 = self.maxpooling(inputdata=relu4, kernel_size=[2, 1], stride=[2, 1], padding='VALID') # batch*4*32*256\n end_points['conv4'] = max_pool4\n conv5 = self.conv2d(inputdata=max_pool4, out_channel=512, kernel_size=3, stride=1, use_bias=False, name='conv5') # batch*4*32*512\n relu5 = self.relu(conv5) # batch*4*32*512\n if self.phase.lower() == 'train':\n bn5 = self.layerbn(inputdata=relu5, is_training=True)\n else:\n bn5 = self.layerbn(inputdata=relu5, is_training=False) # batch*4*32*512\n max_pool5 = self.maxpooling(inputdata=bn5, kernel_size=2, stride=2) #maxpool H and W\n end_points['conv5'] = max_pool5\n conv6 = self.conv2d(inputdata=max_pool5, out_channel=512, kernel_size=3, stride=1, use_bias=False, name='conv6') # batch*4*32*512\n relu6 = self.relu(conv6) # batch*4*32*512\n if self.phase.lower() == 'train':\n bn6 = self.layerbn(inputdata=relu6, is_training=True)\n else:\n bn6 = self.layerbn(inputdata=relu6, is_training=False) # batch*4*32*512\n max_pool6 = self.maxpooling(inputdata=bn6, kernel_size=[2, 1], stride=[2, 1]) # batch*2*32*512\n end_points['conv6'] = max_pool6\n conv7 = self.conv2d(inputdata=max_pool6, out_channel=512, kernel_size=2, stride=[2, 1], use_bias=False, name='conv7') # batch*1*32*512\n relu7 = self.relu(conv7) # batch*1*32*512\n end_points['conv7'] = relu7\n return relu7, end_points\n\n def __map_to_sequence(self, inputdata):\n \"\"\"\n Implement the map to sequence part of the network mainly used to convert the cnn feature map to sequence used in\n later stacked lstm layers\n :param inputdata:\n :return:\n \"\"\"\n shape = inputdata.get_shape().as_list()\n assert shape[1] == 1 # H of the feature map must equal to 1\n return self.squeeze(inputdata=inputdata, axis=1)\n\n def __sequence_label(self, inputdata):\n \"\"\"\n Implement the sequence label part of the network\n :param inputdata:\n :return:\n \"\"\"\n list_n_hidden = [HIDDEN_NUMS, HIDDEN_NUMS]\n \n with tf.variable_scope('LSTMLayers'):\n # construct stack lstm rcnn layer\n # forward lstm cell\n fw_cell_list = [rnn.BasicLSTMCell(nh, forget_bias=1.0) for nh in list_n_hidden]\n # Backward direction cells\n bw_cell_list = [rnn.BasicLSTMCell(nh, forget_bias=1.0) for nh in list_n_hidden]\n\n stack_lstm_layer, _, _ = rnn.stack_bidirectional_dynamic_rnn(fw_cell_list, bw_cell_list, inputdata,\n dtype=tf.float32)\n\n if self.phase.lower() == 'train':\n stack_lstm_layer = self.dropout(inputdata=stack_lstm_layer, keep_prob=0.5)\n\n [batch_s, _, hidden_nums] = inputdata.get_shape().as_list() # [batch, width, 2*n_hidden]\n rnn_reshaped = tf.reshape(stack_lstm_layer, [-1, hidden_nums]) # [batch x width, 2*n_hidden]\n\n w = tf.Variable(tf.truncated_normal([hidden_nums, self.__num_classes], stddev=0.1), name=\"w\")\n # Doing the affine projection\n\n logits = tf.matmul(rnn_reshaped, w)\n\n logits = tf.reshape(logits, [batch_s, -1, self.__num_classes])\n\n raw_pred = tf.argmax(tf.nn.softmax(logits), axis=2, name='raw_prediction')\n\n # Swap batch and batch axis\n rnn_out = tf.transpose(logits, (1, 0, 2), name='transpose_time_major') # [width, batch, n_classes]\n\n return rnn_out, raw_pred\n\n def build_CRNNnet(self, inputdata):\n \"\"\"\n\n :param inputdata:\n :return:\n net_out:output tensor corresponding to the final_endpoint.\n end_points: a set of activations for external use, for example summaries or\n losses.\n \"\"\"\n # first apply the cnn feature extraction stage\n cnn_out,end_points = self.__feature_sequence_extraction(inputdata=inputdata)\n print(\"====##===cnn_out:%s\",cnn_out.get_shape().as_list() )\n # second apply the map to sequence stage\n sequence = self.__map_to_sequence(inputdata=cnn_out)\n print(\"====##===sequence:%s\",sequence.get_shape().as_list() )\n # third apply the sequence label stage\n net_out, raw_pred = self.__sequence_label(inputdata=sequence)\n print(\"====net_out===:\",net_out.get_shape().as_list() )\n print(\"====Predictions===:\",raw_pred.get_shape().as_list() )\n end_points['Logits'] = net_out\n end_points['Predictions'] = raw_pred\n\n return net_out, end_points\n\n\n def train_crnn(self, FLAGS,global_step,cost,sequence_dist,input_labels,pred_labels):\n \"\"\"\n Train crnn model, collect summaries, save model.\n :param inputdata:\n FLAGS: config parameters\n global_step: global step of optimizer\n cost: loss cost\n sequence_dist:\n input_labels: ground true labels\n pred_labels: predict labels\n :return:\n \"\"\"\n learning_rate = tf.train.exponential_decay(FLAGS.LEARNING_RATE, global_step,\n FLAGS.LR_DECAY_STEPS, FLAGS.LR_DECAY_RATE,\n staircase=True)\n update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)\n \n with tf.control_dependencies(update_ops):\n optimizer = tf.train.AdadeltaOptimizer(learning_rate=learning_rate).minimize(loss=cost, global_step=global_step)\n \n # Set tf summary\n tboard_save_path = 'tboard/crnn'\n if not ops.exists(tboard_save_path):\n os.makedirs(tboard_save_path)\n tf.summary.scalar(name='Cost', tensor=cost)\n tf.summary.scalar(name='Learning_Rate', tensor=learning_rate)\n tf.summary.scalar(name='Seq_Dist', tensor=sequence_dist)\n merge_summary_op = tf.summary.merge_all()\n \n # Set the training parameters\n train_epochs = FLAGS.EPOCHS\n checkpoint_path = FLAGS.checkpoint_path\n \n # Set saver configuration\n #saver = tf.train.Saver(write_version = saver_pb2.SaverDef.V1)\n saver = tf.train.Saver()\n model_save_dir = 'checkpoints/crnn'\n #model_save_dir = FLAGS.checkpoint_path\n if not ops.exists(model_save_dir):\n os.makedirs(model_save_dir)\n train_start_time = time.strftime('%Y-%m-%d-%H-%M-%S', time.localtime(time.time()))\n model_name = 'crnn_{:s}.ckpt'.format(str(train_start_time))\n model_save_path = ops.join(model_save_dir, model_name)\n \n # Set sess configuration\n sess_config = tf.ConfigProto()\n sess_config.gpu_options.per_process_gpu_memory_fraction = FLAGS.gpu_memory_fraction\n sess_config.gpu_options.allow_growth = FLAGS.TF_ALLOW_GROWTH\n \n sess = tf.Session(config=sess_config)\n \n summary_writer = tf.summary.FileWriter(tboard_save_path)\n summary_writer.add_graph(sess.graph)\n \n with sess.as_default():\n if checkpoint_path is None:\n logger.info('Training from scratch')\n init = tf.global_variables_initializer()\n sess.run(init)\n else:\n logger.info('Restore model from {:s}'.format(checkpoint_path))\n saver.restore(sess=sess, save_path=checkpoint_path)\n \n coord = tf.train.Coordinator()\n threads = tf.train.start_queue_runners(sess=sess, coord=coord)\n \n for epoch in range(train_epochs):\n _, c, seq_distance, preds, gt_labels, summary = sess.run(\n [optimizer, cost, sequence_dist, pred_labels, input_labels, merge_summary_op])\n \n # calculate the precision \n preds = tf_utils.sparse_tensor_to_str(preds[0])\n gt_labels = tf_utils.sparse_tensor_to_str(gt_labels)\n \n accuracy = [] \n \n for index, gt_label in enumerate(gt_labels):\n pred = preds[index]\n totol_count = len(gt_label)\n correct_count = 0\n try:\n for i, tmp in enumerate(gt_label):\n #import ipdb; ipdb.set_trace()\n #print(\"tmp,pred:\",tmp, pred[i])\n if tmp == pred[i]:\n correct_count += 1\n except IndexError:\n continue\n finally:\n try:\n accuracy.append(correct_count / totol_count)\n except ZeroDivisionError:\n if len(pred) == 0:\n accuracy.append(1)\n else:\n accuracy.append(0)\n accuracy = np.mean(np.array(accuracy).astype(np.float32), axis=0)\n #\n if epoch % FLAGS.DISPLAY_STEP == 0:\n logger.info('Epoch: {:d} cost= {:9f} seq distance= {:9f} train accuracy= {:9f}'.format(\n epoch + 1, c, seq_distance, accuracy))\n \n summary_writer.add_summary(summary=summary, global_step=epoch)\n #logger.info('Save model to {:s}'.format(model_save_path))\n saver.save(sess=sess, save_path=model_save_path, global_step=epoch)\n #saver.save(sess,\"/tmp/crnn.ckpt\")\n \n coord.request_stop()\n coord.join(threads=threads)\n \n sess.close()\n \n return\n\n\n def eval_crnn(self, FLAGS, decoded, images_sh, labels_sh):\n \"\"\"\n Evaluation crnn model, collect summaries, save model.\n :param inputdata:\n FLAGS: config parameters\n decoded: predict labels\n images_sh: image data\n labels_sh: ground true labels\n :return:\n \"\"\" \n # config tf session\n sess_config = tf.ConfigProto()\n sess_config.gpu_options.per_process_gpu_memory_fraction = FLAGS.gpu_memory_fraction\n sess_config.gpu_options.allow_growth = FLAGS.TF_ALLOW_GROWTH\n \n # config tf saver\n #saver = tf.train.Saver(variables_to_restore)\n saver = tf.train.Saver()\n \n if tf.gfile.IsDirectory(FLAGS.checkpoint_path):\n checkpoint_path = tf.train.latest_checkpoint(FLAGS.checkpoint_path)\n else:\n checkpoint_path = FLAGS.checkpoint_path\n \n sess = tf.Session(config=sess_config)\n #checkpoint_path = \"checkpoints/crnn/crnn_2018-07-20-12-09-01.ckpt-99\"\n #module_file = tf.train.latest_checkpoint('/Users/simon/Desktop/OCR/Handwritting_recognition/checkpoints/crnn')\n \"\"\"\n test_sample_count = 0\n for record in tf.python_io.tf_record_iterator(ops.join(dataset_dir, 'test_feature.tfrecords')):\n test_sample_count += 1\n loops_nums = int(math.ceil(test_sample_count / 32))\n # loops_nums = 100\n \"\"\"\n \n with sess.as_default():\n \n # restore the model weights\n saver.restore(sess=sess, save_path=checkpoint_path)\n #saver.restore(sess,\"/tmp/crnn.ckpt\")\n \n coord = tf.train.Coordinator()\n threads = tf.train.start_queue_runners(sess=sess, coord=coord)\n \n print('Start predicting ......')\n #if not is_recursive:\n if not is_recursive:\n predictions, images, labels = sess.run([decoded, images_sh, labels_sh])\n \n preds_res = tf_utils.sparse_tensor_to_str(predictions[0])\n gt_res = tf_utils.sparse_tensor_to_str(labels)\n #preds_res = tf_utils.decode_sparse_tensor(predictions[0])\n #gt_res = tf_utils.decode_sparse_tensor(labels)\n print(\"$$$$$preds:\",preds_res)\n print(\"$$$$$gt_labels:\",gt_res)\n accuracy = []\n #true_numer = 0\n \n for index, gt_label in enumerate(gt_res):\n pred = preds_res[index]\n total_count = len(gt_label)\n\n correct_count = 0\n try:\n for i, tmp in enumerate(gt_label):\n if tmp == pred[i]:\n correct_count += 1\n except IndexError:\n continue\n finally:\n try:\n accuracy.append(correct_count / total_count)\n except ZeroDivisionError:\n if len(pred) == 0:\n accuracy.append(1)\n else:\n accuracy.append(0)\n\n accuracy = np.mean(np.array(accuracy).astype(np.float32), axis=0)\n logger.info('Mean test accuracy is {:5f}'.format(accuracy))\n print('Mean test accuracy is {:5f}'.format(accuracy))\n \n #for index, image in enumerate(images):\n #print('Predict image with gt label: {:s} **** predict label: {:s}'.format(\n # gt_res[index], preds_res[index]))\n #if is_vis:\n # plt.imshow(image[:, :, (2, 1, 0)])\n # plt.show()\n else:\n accuracy = []\n for epoch in range(loops_nums):\n predictions, images, labels = sess.run([decoded, images_sh, labels_sh])\n\n preds_res = tf_utils.sparse_tensor_to_str(predictions[0])\n gt_res = tf_utils.sparse_tensor_to_str(labels)\n \n for index, gt_label in enumerate(gt_res):\n pred = preds_res[index]\n totol_count = len(gt_label)\n correct_count = 0\n try:\n for i, tmp in enumerate(gt_label):\n if tmp == pred[i]:\n correct_count += 1\n except IndexError:\n continue\n finally:\n try:\n accuracy.append(correct_count / totol_count)\n except ZeroDivisionError:\n if len(pred) == 0:\n accuracy.append(1)\n else:\n accuracy.append(0)\n \n for index, image in enumerate(images):\n print('Predict image with gt label: {:s} **** predict label: {:s}'.format(\n gt_res[index], preds_res[index]))\n \n # if is_vis:\n # plt.imshow(image[:, :, (2, 1, 0)])\n # plt.show()\n \n accuracy = np.mean(np.array(accuracy).astype(np.float32), axis=0)\n print('Test accuracy is {:5f}'.format(accuracy))\n \n \n coord.request_stop()\n coord.join(threads=threads)\n \n sess.close()\n \n return\n\n\n# Temporary setting for OCR Telnumber \nCRNNnet.default_image_size = (32,256)\n#CRNNnet.default_image_size = (64,512)\n\n \ndef crnn_arg_scope(weight_decay=0.0005, data_format='NHWC'):\n \"\"\"Defines the arg scope.\n Args:\n weight_decay: The l2 regularization coefficient.\n Returns:\n An arg_scope.\n \"\"\"\n \"\"\"\n with slim.arg_scope([slim.conv2d, slim.fully_connected],\n activation_fn=tf.nn.relu,\n weights_regularizer=slim.l2_regularizer(weight_decay),\n weights_initializer=tf.contrib.layers.xavier_initializer(),\n biases_initializer=tf.zeros_initializer()):\n with slim.arg_scope([slim.conv2d, slim.max_pool2d],\n padding='SAME',\n data_format=data_format):\n with slim.arg_scope([custom_layers.pad2d,\n custom_layers.l2_normalization,\n custom_layers.channel_to_last],\n data_format=data_format) as sc:\n return sc\n \"\"\" ", "repo_name": "funhere/Handwritting_recognition", "sub_path": "nets/crnn/crnn.py", "file_name": "crnn.py", "file_ext": "py", "file_size_in_byte": 20009, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "26", "api": [{"api_name": "utils.log_utils.init_logger", "line_number": 26, "usage_type": "call"}, {"api_name": "utils.log_utils", "line_number": 26, "usage_type": "name"}, {"api_name": "nets.crnn.cnn_basenet.CNNBaseModel", "line_number": 35, "usage_type": "attribute"}, {"api_name": "nets.crnn.cnn_basenet", "line_number": 35, "usage_type": "name"}, {"api_name": "tensorflow.variable_scope", "line_number": 149, "usage_type": "call"}, {"api_name": "tensorflow.contrib.rnn.BasicLSTMCell", "line_number": 152, "usage_type": "call"}, {"api_name": "tensorflow.contrib.rnn", "line_number": 152, "usage_type": "name"}, {"api_name": "tensorflow.contrib.rnn.BasicLSTMCell", "line_number": 154, "usage_type": "call"}, {"api_name": "tensorflow.contrib.rnn", "line_number": 154, "usage_type": "name"}, {"api_name": "tensorflow.contrib.rnn.stack_bidirectional_dynamic_rnn", "line_number": 156, "usage_type": "call"}, {"api_name": "tensorflow.contrib.rnn", "line_number": 156, "usage_type": "name"}, {"api_name": "tensorflow.float32", "line_number": 157, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 163, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 165, "usage_type": "call"}, {"api_name": "tensorflow.truncated_normal", "line_number": 165, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 168, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 170, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 172, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax", "line_number": 172, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 172, "usage_type": "attribute"}, {"api_name": "tensorflow.transpose", "line_number": 175, "usage_type": "call"}, {"api_name": "tensorflow.train.exponential_decay", "line_number": 216, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 216, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 219, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 219, "usage_type": "attribute"}, {"api_name": "tensorflow.control_dependencies", "line_number": 221, "usage_type": "call"}, {"api_name": "tensorflow.train.AdadeltaOptimizer", "line_number": 222, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 222, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 226, "usage_type": "call"}, {"api_name": "os.path", "line_number": 226, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 227, "usage_type": "call"}, {"api_name": "tensorflow.summary.scalar", "line_number": 228, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 228, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 229, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 229, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 230, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 230, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.merge_all", "line_number": 231, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 231, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Saver", "line_number": 239, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 239, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 242, "usage_type": "call"}, {"api_name": "os.path", "line_number": 242, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 243, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 244, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 244, "usage_type": "call"}, {"api_name": "time.time", "line_number": 244, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 246, "usage_type": "call"}, {"api_name": "os.path", "line_number": 246, "usage_type": "name"}, {"api_name": "tensorflow.ConfigProto", "line_number": 249, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 253, "usage_type": "call"}, {"api_name": "tensorflow.summary.FileWriter", "line_number": 255, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 255, "usage_type": "attribute"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 261, "usage_type": "call"}, {"api_name": "tensorflow.train.Coordinator", "line_number": 267, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 267, "usage_type": "attribute"}, {"api_name": "tensorflow.train.start_queue_runners", "line_number": 268, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 268, "usage_type": "attribute"}, {"api_name": "utils.tf_utils.sparse_tensor_to_str", "line_number": 275, "usage_type": "call"}, {"api_name": "utils.tf_utils", "line_number": 275, "usage_type": "name"}, {"api_name": "utils.tf_utils.sparse_tensor_to_str", "line_number": 276, "usage_type": "call"}, {"api_name": "utils.tf_utils", "line_number": 276, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 300, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 300, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 300, "usage_type": "attribute"}, {"api_name": "tensorflow.ConfigProto", "line_number": 330, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 336, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 336, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.IsDirectory", "line_number": 338, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 338, "usage_type": "attribute"}, {"api_name": "tensorflow.train.latest_checkpoint", "line_number": 339, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 339, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 343, "usage_type": "call"}, {"api_name": "tensorflow.train.Coordinator", "line_number": 360, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 360, "usage_type": "attribute"}, {"api_name": "tensorflow.train.start_queue_runners", "line_number": 361, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 361, "usage_type": "attribute"}, {"api_name": "utils.tf_utils.sparse_tensor_to_str", "line_number": 368, "usage_type": "call"}, {"api_name": "utils.tf_utils", "line_number": 368, "usage_type": "name"}, {"api_name": "utils.tf_utils.sparse_tensor_to_str", "line_number": 369, "usage_type": "call"}, {"api_name": "utils.tf_utils", "line_number": 369, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 397, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 397, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 397, "usage_type": "attribute"}, {"api_name": "utils.tf_utils.sparse_tensor_to_str", "line_number": 412, "usage_type": "call"}, {"api_name": "utils.tf_utils", "line_number": 412, "usage_type": "name"}, {"api_name": "utils.tf_utils.sparse_tensor_to_str", "line_number": 413, "usage_type": "call"}, {"api_name": "utils.tf_utils", "line_number": 413, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 442, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 442, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 442, "usage_type": "attribute"}]}
+{"seq_id": "16069374510", "text": "import xml.etree.ElementTree as ET\nimport re\nimport chardet\nimport argparse\nimport os.path\n\n# Create an argument parser\nparser = argparse.ArgumentParser(description='ADMX Policy Parser')\n\n# Add a command-line argument for the input filename\nparser.add_argument('filename', help='Path to the ADMX file')\n\n# Parse the command-line arguments\nargs = parser.parse_args()\n\n# Validate the input filename\nif not os.path.isfile(args.filename):\n print(f\"Error: The provided file '{args.filename}' does not exist.\")\n exit(1)\n\n# Detect the encoding of the XML file\nwith open(args.filename, 'rb') as file:\n detector = chardet.universaldetector.UniversalDetector()\n for line in file.readlines():\n detector.feed(line)\n if detector.done:\n break\n encoding = detector.result['encoding']\n\n# Read the XML file\nwith open(args.filename, 'r', encoding=encoding) as file:\n xml_data = file.read()\n\n# Remove the entire xmlns attribute using regular expressions\nmodified_xml_data = re.sub(r'\\s?xmlns=\"[^\"]+\"', '', xml_data)\n\n# Parse the modified XML\nroot = ET.fromstring(modified_xml_data)\n\n# Initialiser un dictionnaire pour stocker les policies par catégorie\npolicies_by_category = {}\n\n# Parcourir toutes les catégories\nfor category in root.iter('category'):\n category_name = category.attrib['name']\n \n # Créer une liste pour chaque catégorie dans le dictionnaire\n if category_name not in policies_by_category:\n policies_by_category[category_name] = []\n \n # Parcourir toutes les policies qui ont cette catégorie comme parentCategory\n for policy in root.iter('policy'):\n parent_category = policy.find('parentCategory')\n \n # Si la policy appartient à la catégorie courante, ajouter son nom à la liste\n if parent_category is not None and parent_category.attrib['ref'] == category_name:\n policies_by_category[category_name].append(policy.attrib['name'])\n\n# Afficher les policies triées par catégorie\nfor category, policies in policies_by_category.items():\n print(f'Category: {category}')\n for policy in policies:\n print(f' - {policy}')\n", "repo_name": "SeiyaGame/admx-policy-parser", "sub_path": "admx-policy-parser.py", "file_name": "admx-policy-parser.py", "file_ext": "py", "file_size_in_byte": 2144, "program_lang": "python", "lang": "fr", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.path.isfile", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 17, "usage_type": "name"}, {"api_name": "chardet.universaldetector.UniversalDetector", "line_number": 23, "usage_type": "call"}, {"api_name": "chardet.universaldetector", "line_number": 23, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 35, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 38, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 38, "usage_type": "name"}]}
+{"seq_id": "26286285095", "text": "import re\nimport json\nfrom nltk.corpus import stopwords\nfrom rake_nltk import Rake\nimport nltk\n\n\n# this line for download of nltk files here we dowmload stopwords and keyword etractons \n# nltk.download()\n\n## read json file\ndata_json = json.load(open('data.json'))\n\n\n## this will only work when you download stopwords from nltk to your local machine alternatively you can run this ode on colab or kaggle. process is automatic there\nstop_words =stopwords.words('english')\n\n## remove html tages and get pure text\ndef get_pure_text(temp):\n pure_text= ''\n for item in temp:\n# item = re.sub(r'[^\\w\\s]','',item) \n pure_text+=re.sub(r'<.*?>', \"\", item)\n return pure_text\n\n\n## stop words remove\ndef remove_stop_words(temp):\n pure_text = get_pure_text(temp)\n final=''\n for word in pure_text.split():\n if word.lower() not in stop_words:\n# text = re.sub(r'[^a-zA-Z0-9. ]','',word) # remove punctuations\n final += word+' '\n return final\n\n## this will extract keywords from given text\ndef get_keywords(json_text):\n stop_words_removed_text = remove_stop_words(json_text)\n r= Rake()\n r.extract_keywords_from_text(stop_words_removed_text)\n keywords = r.get_ranked_phrases()\n # print(f'html text : {json_text} \\n\\n stopword removed text : {stop_words_removed_text}\\n\\n keywords : {set(keywords)}')\n# print(keywords)\n# print(len(keywords))\n return list(set(keywords))\n\n\n## this will update json data by adding keywords to the file\ndef update_json_with_keywords(jsonData):\n for i in range(len(jsonData)):\n jsonData[i].update({\n 'keywords' : get_keywords(jsonData[i]['data']),\n 'data' : get_pure_text(jsonData[i]['data'])\n })\n return jsonData\n\n## following is the driver code\nnew_json = json.dumps(update_json_with_keywords(data_json))\nwith open(\"datawithkeywords.json\", \"w\") as outfile:\n outfile.write(new_json)\n ", "repo_name": "vattevaii/scraping", "sub_path": "keyword extraction.py", "file_name": "keyword extraction.py", "file_ext": "py", "file_size_in_byte": 1933, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "json.load", "line_number": 12, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords.words", "line_number": 16, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords", "line_number": 16, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 23, "usage_type": "call"}, {"api_name": "rake_nltk.Rake", "line_number": 40, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 59, "usage_type": "call"}]}
+{"seq_id": "12736564103", "text": "from newspaper import Article, ArticleException\nfrom multiprocessing import Pool, cpu_count\nfrom functools import wraps, partial\nfrom urllib.parse import urlparse\nimport traceback\nimport datetime\nimport calendar\nimport psycopg2\nimport requests\nimport tempfile\nimport zipfile\nimport shutil\nimport json\nimport sys\nimport csv\nimport re\nimport os\n\n\nclass Extractor(object):\n\n def __init__(self, config):\n\n self.config = self.read_config(config)\n\n self.db_name = self.config['db_name']\n self.db_user = self.config['db_user']\n self.db_pass = self.config['db_pass']\n self.db_host = self.config['db_host']\n\n @staticmethod\n def read_config(config):\n\n try:\n return config if isinstance(config, dict) else json.load(open(config))\n\n except ValueError as val_err:\n print(f'Configuration Input \"{config}\" is Not Valid: {val_err}')\n sys.exit(1)\n\n @staticmethod\n def get_date_range(y, m):\n\n return [\n datetime.date(y, m, day).strftime('%Y%m%d') for day in range(1, calendar.monthrange(y, m)[1] + 1)\n ]\n\n @staticmethod\n def extract_daily_csv(target_date):\n\n # Pull CSV from GDELT Repository\n date_zip = '{}.export.CSV.zip'.format(target_date)\n event_url = 'http://data.gdeltproject.org/events/{}'.format(date_zip)\n response = requests.get(event_url, stream=True)\n\n if response.status_code != 200:\n return None\n\n # Dumpt to Local CSV\n temp_dir = tempfile.mkdtemp(dir=r'C:\\Temp', prefix='{}_'.format(target_date))\n zip_file = '{}/{}.zip'.format(temp_dir, target_date)\n with open(zip_file, 'wb') as f: f.write(response.content)\n with zipfile.ZipFile(zip_file, 'r') as the_zip: the_zip.extractall(temp_dir)\n\n return '{}/{}.export.CSV'.format(temp_dir, target_date)\n\n @staticmethod\n def text_filter(text):\n\n return re.sub('[^a-zA-Z0-9 \\n]', '', text)\n\n def get_connection(self):\n\n return psycopg2.connect(dbname=self.db_name, user=self.db_user, password=self.db_pass, host=self.db_host)\n\n def process_article(self, source_url):\n\n # Parse GDELT Source\n article = Article(source_url)\n article.download()\n article.parse()\n article.nlp()\n\n # Unpack Article Properties & Replace Special Characters\n title = self.text_filter(article.title)\n summary = '{} . . . '.format(self.text_filter(article.summary)[:500])\n keywords = ', '.join(sorted([self.text_filter(key) for key in article.keywords]))\n meta_keys = ', '.join(sorted([self.text_filter(key) for key in article.meta_keywords]))\n site = urlparse(article.source_url).netloc\n\n return [title, site, summary, keywords, meta_keys]\n\n def process_events(self, year, target_csv):\n\n # Tracking\n seen_urls = []\n proc_urls = 0\n\n # Extract Records\n with open(target_csv, newline='', encoding='utf8') as the_csv:\n\n the_reader = csv.reader(the_csv, delimiter='\\t')\n\n for idx, row in enumerate(the_reader, start=1):\n\n # Pull Filter Attributes\n avg_tone = float(row[34]) # Average Tone\n src_url = row[57] # Source URL\n a1_geo_lat = row[39] # Latitude Check\n a1_gc = row[37] # Actor1Geo_Country\n a2_geo_lat = row[39] # Longitude Check\n a2_gc = row[44] # Actor1Geo_Country\n\n try:\n # TODO - Actor1Geo_Type in ('2', '3')\n if all([v == 'US' for v in [a1_gc, a2_gc]]) \\\n and avg_tone < 0 \\\n and src_url not in seen_urls \\\n and all([a1_geo_lat, a2_geo_lat]):\n\n # Extract NLP Values with Article\n derived_attributes = self.process_article(src_url)\n\n # Push Values into Master Table\n with self.get_connection() as conn:\n with conn.cursor() as cur:\n cur.execute(\n '''\n insert into gdelt_{}\n values {}\n '''.format(year, tuple(row + derived_attributes))\n )\n\n proc_urls += 1\n\n except ArticleException:\n pass\n\n except:\n print(f'{traceback.format_exc()}')\n\n finally:\n seen_urls.append(src_url)\n\n def process_day(self, year, the_day):\n\n print(f'Processing Day: {the_day}')\n\n # Download GDELT Records Locally for Processing\n daily_csv = self.extract_daily_csv(the_day)\n\n # Ignore Bad CSV Requests\n if not daily_csv: return\n\n # Collect Enriched Values & Push Into Table\n self.process_events(year, daily_csv)\n\n # Remove Temporary Directory\n shutil.rmtree(os.path.dirname(daily_csv))\n\n def run_month(self, month, year):\n\n date_range = self.get_date_range(year, month)\n\n # Create Pool & Run Records\n pool = Pool(processes=cpu_count() - 1)\n pool.map(partial(self.process_day, year), date_range)\n pool.close()\n pool.join()\n", "repo_name": "Jwmazzi/media_research", "sub_path": "py/extractor.py", "file_name": "extractor.py", "file_ext": "py", "file_size_in_byte": 5398, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "json.load", "line_number": 35, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 45, "usage_type": "call"}, {"api_name": "calendar.monthrange", "line_number": 45, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 54, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 60, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 63, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 70, "usage_type": "call"}, {"api_name": "psycopg2.connect", "line_number": 74, "usage_type": "call"}, {"api_name": "newspaper.Article", "line_number": 79, "usage_type": "call"}, {"api_name": "urllib.parse.urlparse", "line_number": 89, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 102, "usage_type": "call"}, {"api_name": "newspaper.ArticleException", "line_number": 136, "usage_type": "name"}, {"api_name": "traceback.format_exc", "line_number": 140, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path", "line_number": 159, "usage_type": "attribute"}, {"api_name": "multiprocessing.Pool", "line_number": 166, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 166, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 167, "usage_type": "call"}]}
+{"seq_id": "71825724547", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\n\"\"\"\nInput: \nconfig - filename of json file containing the configuration parameters.\n\nIf no input is passed, run with hard-coded config that does not download \nviews and scheduled queries.\n\nThis function downloads all table names and types from BigQuery, \noptionally downloads view and scheduled queries or loads the previously\ndownloaded (and locally stored) views and scheduled queries, cleanes queries of comments,\nclassify tables into types EXTERNAL, TABLE, VIEW and SCHEDULED. It then creates a \nflowchart that is saved to a file.\n\nCreated by: Henrique S. Xavier, hsxavier@if.usp.br, 12/sep/2019.\n\"\"\"\n\nimport sys\nimport create_flowchart_functions as cf\nimport download_bigquery_info as di\nimport json\n\n# Docstring output:\nif len(sys.argv) > 1 + 1: \n print(__doc__)\n sys.exit(1)\n\n# Get input config:\nelif len(sys.argv) == 1 + 1:\n config = sys.argv[1]\n \n# Set default config:\nelse:\n config = {\n \"credentials\": \"/home/skems/gabinete/projetos/keys-configs/gabinete-compartilhado.json\",\n \"printout\": False,\n \"get_views\": False,\n \"views_path\": \"../views/\",\n \"get_scheduled\": False,\n \"scheduled_path\": \"../scheduled_queries/\",\n \"flowchart\": True,\n \"flowchart_file\": \"this_file.pdf\"\n }\n\n \n### MAIN CODE ###\n\n# Load config from file:\nif type(config)==str:\n with open(config, 'r') as f:\n config = json.load(f) \n\nif config['get_scheduled']:\n di.get_scheduled_queries(config)\n \nif config['flowchart']:\n all_table_query_dict, all_table_type_dict = cf.structure_bigquery_data(config)\n all_tables_list = list(all_table_type_dict.keys())\n\n cf.create_flowchart(all_table_query_dict, all_table_type_dict, all_tables_list, config['flowchart_file'])\n", "repo_name": "hsxavier/bigQuery_mapper", "sub_path": "create_flowchart.py", "file_name": "create_flowchart.py", "file_ext": "py", "file_size_in_byte": 1788, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "25", "api": [{"api_name": "sys.argv", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 32, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 53, "usage_type": "call"}, {"api_name": "download_bigquery_info.get_scheduled_queries", "line_number": 56, "usage_type": "call"}, {"api_name": "create_flowchart_functions.structure_bigquery_data", "line_number": 59, "usage_type": "call"}, {"api_name": "create_flowchart_functions.create_flowchart", "line_number": 62, "usage_type": "call"}]}
+{"seq_id": "41847149548", "text": "\"\"\"\nHOW-TO\n\nfrom server import BaseExceptionContainer,declare_exc\n\nclass my_exceptions(BaseExceptionContainer):\n exc_code_prefix = '11'\n ExceptionA = declare_exc(code='01',default_message='Error MSG',base=True)\n ExceptionB = declare_exc('02','Error MSG')\n\nraise my_exceptions.ExceptionA\nraise my_exceptions.ExceptionA(data=None)\n\n所有用BaseExceptionContainer实现的异常均为NLPException的子类\n同一个BaseExceptionContainer下均以base=True的异常作为父类\n\"\"\"\n\nfrom dataclasses import dataclass\nfrom typing import Dict, Tuple\n\n\nclass NLPException(Exception):\n default_message = 'unknown error'\n code = '999999'\n\n def __init__(self, message: str = '', data: dict = None) -> None:\n self.message = message or self.default_message\n self.data = data\n\n def __str__(self) -> str:\n return f'{self.code}-{self.message}-{self.data}'\n\n\n@dataclass\nclass declared_exc:\n code: str\n default_message: str = '未知错误'\n base: bool = False\n\n\ndeclared_exceptions: Dict[str, NLPException] = {}\n\n\nclass BaseExceptionContainerMeta(type):\n\n def __new__(cls, name: str, bases: tuple, attrs: dict):\n new_cls = super().__new__(cls, name, bases, attrs)\n if not bases:\n return new_cls\n\n assert isinstance(\n getattr(new_cls, 'exc_code_prefix', None),\n str\n ), (f'{new_cls.__module__}-{new_cls.__name__} expected \"exc_code_prefix\" is a string.')\n\n def new_exc(\n exc_name: str,\n declared_exc: declared_exc,\n bases: Tuple = (NLPException,),\n ) -> NLPException:\n exc_attrs = declared_exc.__dict__.copy()\n exc_attrs['code'] = f'{new_cls.exc_code_prefix}{declared_exc.code}'\n exc_attrs['_container'] = new_cls\n\n return type(exc_name, bases, exc_attrs)\n\n cls_declared_exceptions: Dict[str, Tuple[str, declared_exc]] = {}\n base_exc_code = None\n for k, v in attrs.items():\n if isinstance(v, declared_exc):\n code = f'{new_cls.exc_code_prefix}{v.code}'\n if code in declared_exceptions or code in cls_declared_exceptions:\n e = declared_exceptions.get(code)\n if not e:\n _, e = cls_declared_exceptions[code]\n\n raise RuntimeError(\n 'Multiple exceptions were defined with code'\n f\"{v.code}:{new_cls.__name__}.{k},\"\n f'{e._container.__name__}.{e.__name__}'\n )\n\n if v.base:\n if base_exc_code:\n raise RuntimeError(\n 'Expected only one declared_exc(base=True) in'\n f'{new_cls.__name__}'\n )\n base_exc_code = code\n\n cls_declared_exceptions[code] = (k, v)\n\n if not base_exc_code:\n base_exc_name, base_exc_declared = (\n 'BaseException', declared_exc('0000', base=True))\n else:\n base_exc_name, base_exc_declared = (\n cls_declared_exceptions[base_exc_code])\n\n base_exc = new_exc(base_exc_name, base_exc_declared)\n\n declared_exceptions[base_exc.code] = base_exc\n setattr(new_cls, base_exc_name, base_exc)\n\n for code, (exc_name, exc_declared) in cls_declared_exceptions.items():\n if exc_name != base_exc_name:\n exc = new_exc(exc_name, exc_declared, bases=(base_exc,))\n setattr(new_cls, exc_name, exc)\n declared_exceptions[code] = exc\n\n return new_cls\n\n\nclass BaseExceptionContainer(metaclass=BaseExceptionContainerMeta):\n exc_code_prefix = None\n", "repo_name": "niubiqigai/address_ner", "sub_path": "server/exceptions.py", "file_name": "exceptions.py", "file_ext": "py", "file_size_in_byte": 3787, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 16, "dataset": "github-code", "pt": "25", "api": [{"api_name": "dataclasses.dataclass", "line_number": 34, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 41, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 59, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 67, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 67, "usage_type": "name"}]}
+{"seq_id": "20556829729", "text": "from django.urls import path\nfrom . import views\n\napp_name = 'member'\nurlpatterns = [\n path('registration/', views.RegisterUserView.as_view(), name='registration'),\n path('login', views.BlogLoginView.as_view(), name='login'),\n path('logout', views.LogoutUserView.as_view(), name='logout'),\n\n]", "repo_name": "Cyber0x/News_site", "sub_path": "mysite/member/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 301, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}]}
+{"seq_id": "5990130185", "text": "\nimport csv\nimport logging\nimport os\n\nICGC_SAMPLE_ID_KEY=\"ICGC Submitted Sample ID\"\nSEQUENCING_STRATEGY_KEY='ICGC Submitted Sequencing Strategy'\nEGAF_ACCESSION_KEY='EGA File Accession'\nDELIMITER='\\t'\n\nclass EGAAudit:\n\n def __init__(self, csv_file):\n self._csv_file = open(csv_file,'r')\n self._rows = []\n self._load_csv()\n return\n\n def _load_csv(self):\n csv_reader = csv.DictReader(self._csv_file, delimiter=DELIMITER)\n for row in csv_reader:\n self._rows.append(row)\n\n def get_ids(self, seq_strategy=None):\n ids = []\n for row in self._rows:\n if seq_strategy != None:\n if row['ICGC Submitted Sequencing Strategy'] == seq_strategy:\n ids.append({'ICGC_SAMPLE_ID_KEY':row[ICGC_SAMPLE_ID_KEY],'SEQUENCING_STRATEGY_KEY':row[SEQUENCING_STRATEGY_KEY],'EGAF_ACCESSION_KEY':row[EGAF_ACCESSION_KEY]})\n else:\n ids.append({'ICGC_SAMPLE_ID_KEY':row[ICGC_SAMPLE_ID_KEY],'SEQUENCING_STRATEGY_KEY':row[SEQUENCING_STRATEGY_KEY],'EGAF_ACCESSION_KEY':row[EGAF_ACCESSION_KEY]})\n return ids\n\n def get_row(self,icgc_sample_id, egaf, seq_strategy):\n for row in self._rows:\n if row[ICGC_SAMPLE_ID_KEY] == icgc_sample_id and row[SEQUENCING_STRATEGY_KEY] == seq_strategy and row[EGAF_ACCESSION_KEY] == egaf:\n return row\n raise Exception(\"There is no row found with the following informations: icgc_sample_id=%s, egafid=%s, seq_strategy=%s\" % (icgc_sample_id, egaf, seq_strategy))\n\n def _get_key(self, icgc_sample_id, egaf, seq_strategy, key):\n self._csv_file.seek(0)\n row = self.get_row(icgc_sample_id, egaf, seq_strategy)\n return row[key]\n\n def get_egaf_id(self, icgc_sample_id, egaf, seq_strategy):\n return self._get_key(icgc_sample_id, egaf, seq_strategy,EGAF_ACCESSION_KEY)\n\n def get_egaf_ids(self, ids=[]):\n if len(ids) == 0:\n ids = self.get_ids()\n fids = []\n for id in ids:\n logging.debug(\"EGAFID added: %s\" % (id))\n fids.append(self.get_egaf_id(id['ICGC_SAMPLE_ID_KEY'],id['EGAF_ACCESSION_KEY'],id['SEQUENCING_STRATEGY_KEY']))\n return fids\n\n def find_rows(self, egaf_id):\n self._csv_file.seek(0)\n rows = []\n csv_reader = csv.DictReader(self._csv_file, delimiter=DELIMITER)\n for row in csv_reader:\n if row[EGAF_ACCESSION_KEY] == egaf_id:\n rows.append(row)\n return rows\n\n def find_rows_with_bundle_id(self, bundle_id):\n rows = []\n for row in self._rows:\n if row['EGA Analysis Accession'] == bundle_id or row['EGA Run Accession'] == bundle_id:\n rows.append(row)\n return rows\n\n def get_job(self, egaf_id, metadata_repo):\n job = {}\n rows = self.find_rows(egaf_id)\n\n job['bundle_id'] = self._get_bundle_id(rows[0]['EGA Analysis Accession'],rows[0]['EGA Run Accession'])\n job['name'] = job['bundle_id']\n job['bundle_type'] = self._get_bundle_type(rows[0]['EGA Analysis Accession'],rows[0]['EGA Run Accession'])\n job['donor_gender'] = rows[0][\"Donor Gender\"] if rows[0][\"Donor Gender\"] in ['male','female'] else 'unspecified'\n job['ega_analysis_id'] = rows[0]['EGA Analysis Accession']\n job['ega_dataset_id'] = rows[0][\"EGA Dataset Accession\"]\n job['ega_experiment_id'] = rows[0][\"EGA Experiment Accession\"]\n job['ega_metadata_file_name'] = 'bundle.'+job['bundle_id']+'.xml'\n job['ega_metadata_repo'] = metadata_repo\n job['ega_run_id'] = rows[0][\"EGA Run Accession\"]\n job['ega_sample_id'] = rows[0][\"EGA Sample Accession\"]\n job['ega_study_id'] = rows[0][\"EGA Study Accession\"]\n job['insert_size'] = rows[0][\"Insert Size\"]\n job['library_strategy'] = rows[0][SEQUENCING_STRATEGY_KEY]\n job['paired_end'] = rows[0][\"Paired-End\"]\n job['project_code'] = rows[0][\"ICGC DCC Project Code\"]\n job['reference_genome'] = rows[0][\"Reference Genome\"]\n job['submitter'] = rows[0][\"ICGC DCC Project Code\"]\n job['submitter_donor_id'] = rows[0][\"ICGC Submitted Donor ID\"]\n job['submitter_sample_id'] = rows[0][\"ICGC Submitted Sample ID\"]\n job['submitter_specimen_id'] = rows[0][\"ICGC Submitted Specimen ID\"]\n job['submitter_specimen_type'] = rows[0][\"ICGC Submitted Specimen Type\"]\n\n job['files'] = []\n for row in self.find_rows_with_bundle_id(job['bundle_id']):\n file = {}\n file['ega_file_id'] = row[EGAF_ACCESSION_KEY]\n file['file_md5sum'] = row['Unencrypted Checksum']\n file['file_name'] = row['Unencrypted Checksum']+'.'+os.path.basename(row['EGA Raw Sequence Filename'][:-4] if str(row['EGA Raw Sequence Filename']).endswith('.gpg') else row['EGA Raw Sequence Filename'])\n file['size'] = row['File Size']\n job['files'].append(file)\n\n return \".\".join(['job',job['bundle_id'],job['project_code'],job['submitter_sample_id'],job['ega_sample_id']]),job\n\n def _get_bundle_type(self, analysis_accession, run_accession):\n if analysis_accession != None and analysis_accession != \"\":\n return \"analysis\"\n if run_accession != None and run_accession != \"\":\n return \"run\"\n return None\n\n def _get_bundle_id(self, analysis_accession, run_accession):\n if analysis_accession != None and analysis_accession != \"\":\n return analysis_accession\n if run_accession != None and run_accession != \"\":\n return run_accession\n return None\n\n def get_info_from_egafid(self, egafid, *args):\n print(egafid)\n row = self.find_rows(egafid)[0]\n return [row[x] for x in args]\n\n", "repo_name": "icgc-dcc/JTrackerTransferOperations", "sub_path": "operations/ega/utils/ega_audit.py", "file_name": "ega_audit.py", "file_ext": "py", "file_size_in_byte": 5778, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "csv.DictReader", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 53, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}]}
+{"seq_id": "39839761163", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Dec 26 13:58:02 2017\nTesting suite for landspy Grid class\n@author: J. Vicente Perez\n@email: geolovic@hotmail.com\n@last_modified: 04 october, 2022\n\"\"\"\n\nimport unittest\nimport sys\nimport numpy as np\nfrom osgeo import gdal\n# Add to the path code folder and data folder\nsys.path.append(\"../src/\")\nfrom landspy import PRaster\ninfolder = \"data/in\"\n\nclass TestPRaster00(unittest.TestCase):\n \n def test_properties(self):\n # Test PRaster getters\n files = [\"small25.tif\", \"tunez.tif\", \"jebja30.tif\"]\n for file in files:\n raster = PRaster(infolder + \"/{0}\".format(file))\n size = raster.getSize()\n dims = raster.getDims()\n ncells = raster.getNCells()\n cellsize = raster.getCellSize()\n geot = raster.getGeot()\n computed = (size, dims, ncells, cellsize, geot)\n graster = gdal.Open(infolder + \"/{0}\".format(file))\n band = graster.GetRasterBand(1)\n arr = band.ReadAsArray()\n ggeot = graster.GetGeoTransform()\n expected = ((band.XSize, band.YSize), arr.shape, arr.size, (ggeot[1], ggeot[5]), ggeot)\n self.assertEqual(computed, expected)\n\n def test_projections(self):\n # Test PRaster getters\n files = [\"small25.tif\", \"tunez.tif\", \"jebja30.tif\"]\n \n for file in files:\n # Test PRaster getters for a raster\n raster = PRaster(infolder + \"/{0}\".format(file))\n graster = gdal.Open(infolder + \"/{0}\".format(file))\n self.assertEqual(raster.getCRS(), graster.GetProjection())\n \n def test_empty(self):\n # Test PRaster getters in an empty PRaster\n raster = PRaster()\n size = raster.getSize()\n dims = raster.getDims()\n ncells = raster.getNCells()\n cellsize = raster.getCellSize()\n geot = raster.getGeot()\n proj = raster.getCRS()\n computed = (size, dims, ncells, cellsize, geot, proj)\n expected = ((1, 1), (1, 1), 1, (1.0, -1.0), (0., 1., 0., 1., 0., -1.), \"\")\n self.assertEqual(computed, expected)\n \n def test_copy_layout(self):\n # Test copy_layout for some rasters\n files = [\"small25.tif\", \"tunez.tif\", \"jebja30.tif\"]\n for file in files:\n b_raster = PRaster()\n c_raster = PRaster(infolder + \"/{0}\".format(file))\n b_raster.copyLayout(c_raster)\n \n size = c_raster.getSize()\n dims = c_raster.getDims()\n ncells = c_raster.getNCells()\n cellsize = c_raster.getCellSize()\n geot = c_raster.getGeot()\n proj = c_raster.getCRS()\n computed = (size, dims, ncells, cellsize, geot, proj)\n \n size = b_raster.getSize()\n dims = b_raster.getDims()\n ncells = b_raster.getNCells()\n cellsize = b_raster.getCellSize()\n geot = b_raster.getGeot()\n proj = b_raster.getCRS()\n expected = (size, dims, ncells, cellsize, geot, proj)\n \n self.assertEqual(computed, expected)\n\nclass TestPRaster01(unittest.TestCase):\n \n def setUp(self): \n # Load test data\n self.ids = np.load(infolder + \"/np_files/small25_100rnd_id.npy\")\n self.rows = np.load(infolder + \"/np_files/small25_100rnd_row.npy\")\n self.cols = np.load(infolder + \"/np_files/small25_100rnd_col.npy\")\n self.xi = np.load(infolder + \"/np_files/small25_100rnd_X.npy\")\n self.yi = np.load(infolder + \"/np_files/small25_100rnd_Y.npy\")\n \n def test_xy_2_cell_01(self): \n raster = PRaster(infolder + \"/small25.tif\")\n xi = self.xi\n yi = self.yi\n rows = self.rows\n cols = self.cols\n c_rows, c_cols = raster.xyToCell(xi, yi)\n res = (np.array_equal(rows, c_rows), np.array_equal(cols, c_cols))\n self.assertEqual(res, (True, True))\n \n def test_xy_2_cell_02(self):\n raster = PRaster(infolder + \"/small25.tif\")\n x = 471927\n y = 4116048\n row, col = raster.xyToCell(x, y)\n self.assertEqual((43, 71), (row, col))\n \n def test_xy_2_cell_03(self):\n raster = PRaster(infolder + \"/small25.tif\")\n xi = self.xi.tolist()\n yi = self.yi.tolist()\n rows = self.rows\n cols = self.cols\n c_rows, c_cols = raster.xyToCell(xi, yi)\n res = (np.array_equal(rows, c_rows), np.array_equal(cols, c_cols))\n self.assertEqual(res, (True, True))\n \n \n def test_cell_2_xy_01(self): \n raster = PRaster(infolder + \"/small25.tif\")\n xi = self.xi\n yi = self.yi\n rows = self.rows\n cols = self.cols\n cxi, cyi = raster.cellToXY(rows, cols)\n res = (np.array_equal(cxi, xi), np.array_equal(cyi, yi))\n self.assertEqual(res, (True, True))\n \n def test_cell_2_xy_02(self):\n raster = PRaster(infolder + \"/small25.tif\")\n x = 471927.1\n y = 4116048.5\n row = 43\n col = 71\n cx, cy = raster.cellToXY(row, col)\n self.assertEqual((x, y), (cx, cy))\n \n def test_cell_2_xy_03(self):\n raster = PRaster(infolder + \"/small25.tif\")\n xi = self.xi\n yi = self.yi\n rows = self.rows.tolist()\n cols = self.cols.tolist()\n cxi, cyi = raster.cellToXY(rows, cols)\n res = (np.array_equal(xi, cxi), np.array_equal(yi, cyi))\n self.assertEqual(res, (True, True)) \n \n def test_cell_2_ind_01(self): \n raster = PRaster(infolder + \"/small25.tif\")\n rows = self.rows\n cols = self.cols\n ids = self.ids\n cids = raster.cellToInd(rows, cols)\n self.assertEqual(np.array_equal(ids, cids), True)\n \n def test_cell_2_ind_02(self):\n raster = PRaster(infolder + \"/small25.tif\")\n row = 25\n col = 11\n ids = 4986\n cids = raster.cellToInd(row, col)\n self.assertEqual(ids, cids)\n \n def test_cell_2_ind_03(self): \n raster = PRaster(infolder + \"/small25.tif\")\n rows = self.rows.tolist()\n cols = self.cols.tolist()\n ids = self.ids\n cids = raster.cellToInd(rows, cols)\n self.assertEqual(np.array_equal(ids, cids), True)\n \n def test_is_inside(self):\n # points 1, 2, 4, 8 --> -Fuera de ráster\n # points 3, 5 --> Dentro de ráster, pero en NoData\n # points 6, 7, 9 --> Dentro de ráster\n \n puntos = np.array([[476154., 4115084.],\n [472289., 4112838.],\n [471317., 4114050.],\n [472874., 4117717.],\n [472205., 4114091.],\n [470795., 4116411.],\n [472257., 4115565.],\n [469572., 4115376.],\n [473877., 4114844.]])\n x = puntos[:,0]\n y = puntos[:,1]\n raster = PRaster(infolder + \"/small25.tif\")\n computed = raster.isInside(x, y)\n expected = np.array([False, False, True, False, True, True, True, False, True])\n self.assertEqual(np.array_equal(computed, expected), True)\n \nif __name__ == \"__main__\":\n unittest.main()", "repo_name": "geolovic/landspy", "sub_path": "tests/test_00_PRaster.py", "file_name": "test_00_PRaster.py", "file_ext": "py", "file_size_in_byte": 7308, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 7, "dataset": "github-code", "pt": "25", "api": [{"api_name": "sys.path.append", "line_number": 15, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 19, "usage_type": "attribute"}, {"api_name": "landspy.PRaster", "line_number": 25, "usage_type": "call"}, {"api_name": "osgeo.gdal.Open", "line_number": 32, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 32, "usage_type": "name"}, {"api_name": "landspy.PRaster", "line_number": 45, "usage_type": "call"}, {"api_name": "osgeo.gdal.Open", "line_number": 46, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 46, "usage_type": "name"}, {"api_name": "landspy.PRaster", "line_number": 51, "usage_type": "call"}, {"api_name": "landspy.PRaster", "line_number": 66, "usage_type": "call"}, {"api_name": "landspy.PRaster", "line_number": 67, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 96, "usage_type": "call"}, {"api_name": "landspy.PRaster", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 105, "usage_type": "call"}, {"api_name": "landspy.PRaster", "line_number": 109, "usage_type": "call"}, {"api_name": "landspy.PRaster", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 122, "usage_type": "call"}, {"api_name": "landspy.PRaster", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 133, "usage_type": "call"}, {"api_name": "landspy.PRaster", "line_number": 137, "usage_type": "call"}, {"api_name": "landspy.PRaster", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 152, "usage_type": "call"}, {"api_name": "landspy.PRaster", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 161, "usage_type": "call"}, {"api_name": "landspy.PRaster", "line_number": 164, "usage_type": "call"}, {"api_name": "landspy.PRaster", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 184, "usage_type": "call"}, {"api_name": "landspy.PRaster", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 198, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 201, "usage_type": "call"}]}
+{"seq_id": "19200079903", "text": "\"\"\"\nDefines:\n - Cart3D(log=None, debug=False)\n - read_cart3d(self, infilename, result_names=None)\n - write_cart3d(self, outfilename, is_binary=False, float_fmt='%6.7f')\n\n - flip_model()\n - make_mirror_model(self, nodes, elements, regions, loads, axis='y', tol=0.000001)\n - make_half_model(self, axis='y', remap_nodes=True)\n - get_free_edges(self, elements)\n - get_area(self)\n - get_normals(self)\n - get_normals_at_nodes(self, cnormals)\n\n - comp2tri(in_filenames, out_filename,\n is_binary=False, float_fmt='%6.7f')\n\"\"\"\nfrom __future__ import print_function, unicode_literals\nimport sys\nfrom struct import pack, unpack\nfrom math import ceil\nfrom collections import defaultdict\nfrom codecs import open as codec_open\n\nfrom six import iteritems, PY2\nfrom six.moves import zip, range\n\nimport numpy as np\n\nfrom pyNastran.utils import is_binary_file, _filename, b\nfrom pyNastran.utils.log import get_logger2\n\nif PY2:\n string_type = unicode\n bytes_type = str\n write_ascii = 'wb'\nelse:\n string_type = str\n bytes_type = bytes\n write_ascii = 'wb'\n\n\ndef read_cart3d(cart3d_filename, log=None, debug=False, result_names=None):\n \"\"\"loads a Cart3D file\"\"\"\n model = Cart3D(log=log, debug=debug)\n model.read_cart3d(cart3d_filename, result_names)\n return model\n\n\ndef comp2tri(in_filenames, out_filename,\n is_binary=False, float_fmt='%6.7f'):\n \"\"\"\n Combines multiple Cart3d files (binary or ascii) into a single file.\n\n Parameters\n ----------\n in_filenames : List[str]\n list of filenames\n out_filename : str\n output filename\n is_binary : bool; default=False\n is the output filename binary\n float_fmt : str; default='%6.7f'\n the format string to use for ascii writing\n\n .. note:: assumes loads is None\n \"\"\"\n points = []\n elements = []\n regions = []\n\n #ne = 0\n npoints = 0\n nregions = 0\n model = Cart3D()\n for infilename in in_filenames:\n model.read_cart3d(infilename)\n npointsi = model.nodes.shape[0]\n nregionsi = len(np.unique(model.regions))\n #element += npoints - 1\n #region += nregions\n\n points.append(model.nodes)\n elements.append(model.elements)\n regions.append(model.regions)\n npoints += npointsi\n nregions += nregionsi\n\n points = np.vstack(points)\n elements = np.vstack(elements)\n regions = np.vstack(regions)\n model.points = points\n model.elements = elements\n model.regions = regions\n\n model.write_cart3d(out_filename, is_binary=False, float_fmt=float_fmt)\n\n\nclass Cart3dIO(object):\n \"\"\"\n Cart3d IO class\n \"\"\"\n def __init__(self, log=None, debug=False):\n self.log = get_logger2(log, debug=debug)\n self._endian = b''\n self._encoding = 'latin1'\n self.n = 0\n self.infile = None\n # self.readHalf = False\n # self.nPoints = None\n # self.nElements = None\n self.infilename = None\n self.points = None\n self.elements = None\n self.regions = None\n self.loads = {}\n\n def _write_header(self, outfile, points, elements, is_loads, is_binary=False):\n \"\"\"\n writes the cart3d header\n\n Without results\n ---------------\n npoints nelements\n\n With results\n ------------\n npoints nelements nresults\n\n \"\"\"\n npoints = points.shape[0]\n nelements = elements.shape[0]\n\n if is_binary:\n if is_loads:\n fmt = self._endian + b('iiiii')\n msg = pack(fmt, 3*4, npoints, nelements, 6, 4)\n else:\n fmt = self._endian + b('iiii')\n msg = pack(fmt, 2*4, npoints, nelements, 4)\n\n int_fmt = None\n else:\n # this is ASCII data\n if is_loads:\n msg = b(\"%i %i 6\\n\" % (npoints, nelements))\n else:\n msg = b(\"%i %i\\n\" % (npoints, nelements))\n\n # take the max value, string it, and length it\n # so 123,456 is length 6\n int_fmt = b('%%%si' % len(str(nelements)))\n outfile.write(msg)\n return int_fmt\n\n def _write_points(self, outfile, points, is_binary, float_fmt='%6.6f'):\n \"\"\"writes the points\"\"\"\n if is_binary:\n four = pack(self._endian + b('i'), 4)\n outfile.write(four)\n\n npoints = points.shape[0]\n fmt = self._endian + b('%if' % (npoints * 3))\n floats = pack(fmt, *np.ravel(points))\n\n outfile.write(floats)\n outfile.write(four)\n else:\n if isinstance(float_fmt, bytes_type):\n fmt_ascii = float_fmt\n else:\n fmt_ascii = float_fmt.encode('latin1')\n np.savetxt(outfile, points, fmt_ascii)\n\n def _write_elements(self, outfile, elements, is_binary, int_fmt='%6i'):\n \"\"\"writes the triangles\"\"\"\n min_e = elements.min()\n assert min_e == 0, 'min(elements)=%s' % min_e\n if is_binary:\n fmt = self._endian + b('i')\n four = pack(fmt, 4)\n outfile.write(four)\n nelements = elements.shape[0]\n fmt = self._endian + b('%ii' % (nelements * 3))\n ints = pack(fmt, *np.ravel(elements+1))\n\n outfile.write(ints)\n outfile.write(four)\n else:\n if isinstance(int_fmt, bytes_type):\n fmt_ascii = int_fmt\n else:\n fmt_ascii = int_fmt.encode('latin1')\n np.savetxt(outfile, elements+1, fmt_ascii)\n\n def _write_regions(self, outfile, regions, is_binary):\n \"\"\"writes the regions\"\"\"\n if is_binary:\n fmt = self._endian + b('i')\n four = pack(fmt, 4)\n outfile.write(four)\n\n nregions = len(regions)\n fmt = self._endian + b('%ii' % nregions)\n ints = pack(fmt, *regions)\n outfile.write(ints)\n\n outfile.write(four)\n else:\n fmt = b'%i'\n np.savetxt(outfile, regions, fmt)\n\n def _write_loads(self, outfile, loads, is_binary, float_fmt='%6.6f'):\n \"\"\"writes the *.triq loads\"\"\"\n if is_binary:\n raise NotImplementedError('is_binary=%s' % is_binary)\n else:\n Cp = loads['Cp']\n rho = loads['rho']\n rhoU = loads['rhoU']\n rhoV = loads['rhoV']\n rhoW = loads['rhoW']\n E = loads['E']\n npoints = self.points.shape[0]\n assert len(Cp) == npoints, 'len(Cp)=%s npoints=%s' % (len(Cp), npoints)\n #nrows = len(Cp)\n fmt = '%s\\n%s %s %s %s %s\\n' % (float_fmt, float_fmt, float_fmt,\n float_fmt, float_fmt, float_fmt)\n for (cpi, rhoi, rhou, rhov, rhoe, e) in zip(Cp, rho, rhoU, rhoV, rhoW, E):\n outfile.write(fmt % (cpi, rhoi, rhou, rhov, rhoe, e))\n\n def _read_header_ascii(self):\n line = self.infile.readline()\n sline = line.strip().split()\n if len(sline) == 2:\n npoints, nelements = int(sline[0]), int(sline[1])\n nresults = 0\n elif len(sline) == 3:\n npoints = int(sline[0])\n nelements = int(sline[1])\n nresults = int(sline[2])\n else:\n raise ValueError('invalid result type')\n return npoints, nelements, nresults\n\n @property\n def nresults(self):\n \"\"\"get the number of results\"\"\"\n if isinstance(self.loads, dict):\n return len(self.loads)\n return 0\n\n @property\n def nnodes(self):\n \"\"\"alternate way to access number of points\"\"\"\n return self.npoints\n\n @property\n def npoints(self):\n \"\"\"get the number of points\"\"\"\n return self.points.shape[0]\n\n @property\n def nodes(self):\n \"\"\"alternate way to access the points\"\"\"\n return self.points\n\n @nodes.setter\n def nodes(self, points):\n \"\"\"alternate way to access the points\"\"\"\n self.points = points\n\n @property\n def nelements(self):\n \"\"\"get the number of elements\"\"\"\n return self.elements.shape[0]\n\n def _read_points_ascii(self, npoints):\n \"\"\"\n A point is defined by x,y,z and the ID is the location in points.\n \"\"\"\n p = 0\n data = []\n assert npoints > 0, 'npoints=%s' % npoints\n points = np.zeros((npoints, 3), dtype='float32')\n while p < npoints:\n data += self.infile.readline().strip().split()\n while len(data) > 2:\n x = data.pop(0)\n y = data.pop(0)\n z = data.pop(0)\n points[p] = [x, y, z]\n p += 1\n\n #maxX = self.get_max(points, 0)\n #maxY = self.get_max(points, 1)\n #maxZ = self.get_max(points, 2)\n\n #minX = self.get_min(points, 0)\n #minY = self.get_min(points, 1)\n #minZ = self.get_min(points, 2)\n\n #self.log.debug(\"X max=%g min=%g\" % (maxX, minX))\n #self.log.debug(\"Y max=%g min=%g\" % (maxY, minY))\n #self.log.debug(\"Z max=%g min=%g\" % (maxZ, minZ))\n return points\n\n def _read_elements_ascii(self, nelements):\n \"\"\"\n An element is defined by n1,n2,n3 and the ID is the location in elements.\n \"\"\"\n assert nelements > 0, 'npoints=%s nelements=%s' % (self.npoints, nelements)\n elements = np.zeros((nelements, 3), dtype='int32')\n\n e = 0\n data = []\n while e < nelements:\n data += self.infile.readline().strip().split()\n while len(data) > 2:\n n1 = int(data.pop(0))\n n2 = int(data.pop(0))\n n3 = int(data.pop(0))\n elements[e] = [n1, n2, n3]\n e += 1\n assert elements.min() == 1, elements.min()\n return elements - 1\n\n def _read_regions_ascii(self, nelements):\n regions = np.zeros(nelements, dtype='int32')\n r = 0\n data = []\n while r < nelements:\n data = self.infile.readline().strip().split()\n ndata = len(data)\n regions[r : r + ndata] = data\n r += ndata\n return regions\n\n def _read_header_binary(self):\n data = self.infile.read(4)\n size_little, = unpack(b'i', data)\n if size_big in [12, 8]:\n self._endian = b'>'\n size = size_big\n elif size_little in [8, 12]:\n self._endian = b'<'\n size = size_little\n else:\n self._rewind()\n self.show(100)\n raise RuntimeError('unknown endian')\n\n self.n += 4\n data = self.infile.read(size)\n self.n += size\n\n so4 = size // 4 # size over 4\n if so4 == 3:\n (npoints, nelements, nresults) = unpack(self._endian + b('iii'), data)\n self.log.info(\"npoints=%s nelements=%s nresults=%s\" % (npoints, nelements, nresults))\n elif so4 == 2:\n (npoints, nelements) = unpack(self._endian + b('ii'), data)\n nresults = 0\n self.log.info(\"npoints=%s nelements=%s\" % (npoints, nelements))\n else:\n self._rewind()\n self.show(100)\n raise RuntimeError('in the wrong spot...endian...size/4=%s' % so4)\n self.infile.read(8) # end of first block, start of second block\n return (npoints, nelements, nresults)\n\n def _read_points_binary(self, npoints):\n \"\"\"reads the xyz points\"\"\"\n size = npoints * 12 # 12=3*4 all the points\n data = self.infile.read(size)\n\n dtype = np.dtype(self._endian + b('f4'))\n points = np.fromstring(data, dtype=dtype).reshape((npoints, 3))\n\n self.infile.read(8) # end of second block, start of third block\n return points\n\n def _read_elements_binary(self, nelements):\n \"\"\"reads the triangles\"\"\"\n size = nelements * 12 # 12=3*4 all the elements\n data = self.infile.read(size)\n\n dtype = np.dtype(self._endian + b('i4'))\n elements = np.fromstring(data, dtype=dtype).reshape((nelements, 3))\n\n self.infile.read(8) # end of third (element) block, start of regions (fourth) block\n assert elements.min() == 1, elements.min()\n return elements - 1\n\n def _read_regions_binary(self, nelements):\n \"\"\"reads the regions\"\"\"\n size = nelements * 4 # 12=3*4 all the elements\n data = self.infile.read(size)\n\n regions = np.zeros(nelements, dtype='int32')\n dtype = self._endian + b'i'\n regions = np.fromstring(data, dtype=dtype)\n\n self.infile.read(4) # end of regions (fourth) block\n return regions\n\n def _read_results_binary(self, i, infile, result_names=None):\n \"\"\"binary results are not supported\"\"\"\n pass\n\n def _rewind(self):\n \"\"\"go back to the beginning of the file\"\"\"\n self.n = 0\n self.infile.seek(self.n)\n\n def show(self, n, types='ifs', endian=None):\n assert self.n == self.infile.tell(), 'n=%s tell=%s' % (self.n, self.infile.tell())\n nints = n // 4\n data = self.infile.read(4 * n)\n strings, ints, floats = self.show_data(data, types=types, endian=endian)\n self.infile.seek(self.n)\n return strings, ints, floats\n\n def show_data(self, data, types='ifs', endian=None):\n return self._write_data(sys.stdout, data, types=types, endian=endian)\n\n def _write_data(self, outfile, data, types='ifs', endian=None):\n \"\"\"\n Useful function for seeing what's going on locally when debugging.\n \"\"\"\n n = len(data)\n nints = n // 4\n ndoubles = n // 8\n strings = None\n ints = None\n floats = None\n longs = None\n\n if endian is None:\n endian = self._endian\n\n if 's' in types:\n strings = unpack(b'%s%is' % (endian, n), data)\n outfile.write(\"strings = %s\\n\" % str(strings))\n if 'i' in types:\n ints = unpack(b'%s%ii' % (endian, nints), data)\n outfile.write(\"ints = %s\\n\" % str(ints))\n if 'f' in types:\n floats = unpack(b'%s%if' % (endian, nints), data)\n outfile.write(\"floats = %s\\n\" % str(floats))\n\n if 'l' in types:\n longs = unpack(b'%s%il' % (endian, nints), data)\n outfile.write(\"long = %s\\n\" % str(longs))\n if 'I' in types:\n ints2 = unpack(b'%s%iI' % (endian, nints), data)\n outfile.write(\"unsigned int = %s\\n\" % str(ints2))\n if 'L' in types:\n longs2 = unpack(b'%s%iL' % (endian, nints), data)\n outfile.write(\"unsigned long = %s\\n\" % str(longs2))\n if 'q' in types:\n longs = unpack(b'%s%iq' % (endian, ndoubles), data[:ndoubles*8])\n outfile.write(\"long long = %s\\n\" % str(longs))\n return strings, ints, floats\n\n def show_ndata(self, n, types='ifs'):\n return self._write_ndata(sys.stdout, n, types=types)\n\n def _write_ndata(self, outfile, n, types='ifs'):\n \"\"\"\n Useful function for seeing what's going on locally when debugging.\n \"\"\"\n nold = self.n\n data = self.infile.read(n)\n self.n = nold\n self.infile.seek(self.n)\n return self._write_data(outfile, data, types=types)\n\n\nclass Cart3D(Cart3dIO):\n \"\"\"\n Cart3d interface class\n \"\"\"\n model_type = 'cart3d'\n isStructured = False\n isOutwardNormals = True\n\n def __init__(self, log=None, debug=False):\n Cart3dIO.__init__(self, log=log, debug=debug)\n self.loads = {}\n self.points = None\n self.elements = None\n\n def flip_model(self):\n \"\"\"flip the model about the y-axis\"\"\"\n self.points[:, 1] *= -1.\n self.elements = np.hstack([\n self.elements[:, 0:1],\n self.elements[:, 2:3],\n self.elements[:, 1:2],\n ])\n print(self.elements.shape)\n\n def make_mirror_model(self, nodes, elements, regions, loads, axis='y', tol=0.000001):\n \"\"\"\n Makes a full cart3d model about the given axis.\n\n Parameters\n ----------\n nodes : (nnodes, 3) ndarray\n the nodes\n elements : (nelements, 3) ndarray\n the elmements\n regions : (nelements) ndarray\n the regions\n loads : dict[str] = (nnodes) ndarray\n not supported\n axis : str; {\"x\", \"y\", \"z\", \"-x\", \"-y\", \"-z\"}\n a string of the axis\n tol : float; default=0.000001\n the tolerance for the centerline points\n \"\"\"\n raise NotImplementedError()\n self.log.info('---starting make_mirror_model---')\n assert tol >= 0, 'tol=%r' % tol # prevents hacks to the axis\n\n nnodes = nodes.shape[0]\n assert nnodes > 0, 'nnodes=%s' % nnodes\n\n nelements = elements.shape[0]\n assert nelements > 0, 'nelements=%s' % nelements\n\n ax = self._get_ax(axis)\n if ax in [0, 1, 2]: # positive x, y, z values; mirror to -side\n iy0 = np.where(nodes[:, ax] > tol)[0]\n ax2 = ax\n elif ax in [3, 4, 5]: # negative x, y, z values; mirror to +side\n iy0 = np.where(nodes[:, ax-3] < -tol)[0]\n ax2 = ax - 3 # we create ax2 in order to generalize the data updating\n else:\n raise NotImplementedError(axis)\n\n # the nodes to be duplicated are the nodes that aren't below the tolerance\n nodes_upper = nodes[iy0]\n nodes_upper[:, ax2] *= -1.0 # flip the nodes about the axis\n\n nodes2 = np.vstack([nodes, nodes_upper])\n nnodes2 = nodes2.shape[0]\n assert nnodes2 > nnodes, 'nnodes2=%s nnodes=%s' % (nnodes2, nnodes)\n\n nnodes_upper = nodes_upper.shape[0]\n elements_upper = elements.copy()\n nelements = elements.shape[0]\n\n # remap the mirrored nodes with the new node ids\n for eid in range(nelements):\n element = elements_upper[eid, :]\n for i, eidi in enumerate(element):\n if eidi in iy0:\n elements_upper[eid][i] = nnodes_upper + eidi\n\n # we need to reverse the element in order to get\n # the proper normal vector\n elements_upper[eid] = elements_upper[eid, ::-1]\n\n elements2 = np.vstack([elements, elements_upper])\n nelements2 = elements2.shape[0]\n assert nelements2 > nelements, 'nelements2=%s nelements=%s' % (nelements2, nelements)\n\n nregions = len(np.unique(regions))\n regions_upper = regions.copy() + nregions\n regions2 = np.hstack([regions, regions_upper])\n\n loads2 = {}\n for key, data in iteritems(loads):\n\n # flip the sign on the flipping axis terms\n if((key in ['U', 'rhoU'] and ax2 == 0) or\n (key in ['V', 'rhoV'] and ax2 == 1) or\n (key in ['W', 'rhoW'] and ax2 == 2)):\n data_upper = -data[iy0]\n else:\n data_upper = data[iy0]\n loads2[key] = np.hstack([data, data_upper])\n\n self.log.info('---finished make_mirror_model---')\n return (nodes2, elements2, regions2, loads2)\n\n def _get_ax(self, axis):\n \"\"\"helper method to convert an axis_string into an integer\"\"\"\n axis = axis.lower().strip()\n if axis in ['+x', 'x', 0]:\n ax = 0\n elif axis in ['+y', 'y', 1]:\n ax = 1\n elif axis in ['+z', 'z', 2]:\n ax = 2\n\n elif axis in ['-x', 3]:\n ax = 3\n elif axis == ['-y', 4]:\n ax = 4\n elif axis == ['-z', 5]:\n ax = 5\n else:\n raise NotImplementedError('axis=%r' % axis)\n self.log.info(\"axis=%r ax=%s\" % (axis, ax))\n return ax\n\n def make_half_model(self, axis='y', remap_nodes=True):\n \"\"\"\n Makes a half model from a full model\n\n ... note:: Cp is really loads['Cp'] and was meant for loads analysis only\n \"\"\"\n nodes = self.nodes\n elements = self.elements\n regions = self.regions\n loads = self.loads\n if loads is None:\n loads = {}\n\n nnodes = nodes.shape[0]\n assert nnodes > 0, 'nnodes=%s' % nnodes\n\n nelements = elements.shape[0]\n assert nelements > 0, 'nelements=%s' % nelements\n\n inodes_remove = set([])\n self.log.info('---starting make_half_model---')\n ax = self._get_ax(axis)\n\n if ax in [0, 1, 2]: # remove values > 0\n inodes_save = np.where(nodes[:, ax] >= 0.0)[0]\n elif ax in [3, 4, 5]: # remove values < 0\n inodes_save = np.where(nodes[:, ax-3] <= 0.0)[0]\n else:\n raise NotImplementedError('axis=%r ax=%s' % (axis, ax))\n inodes_save.sort()\n\n inodes_map = np.arange(len(inodes_save))\n if not(0 < len(inodes_save) < nnodes):\n msg = 'len(inodes_save)=%s nnodes=%s' % (len(inodes_save), nnodes)\n raise RuntimeError(msg)\n\n nodes2 = nodes[inodes_save, :]\n nnodes2 = nodes2.shape[0]\n assert 0 < nnodes2 < nnodes, 'nnodes=%s nnodes2=%s' % (nnodes, nnodes2)\n\n inodes_save += 1 # +1 is so we don't have to shift inode\n # .. todo:: still need to handle element's node id renumbering\n ielements_save = set([])\n for ielement in range(nelements):\n save_element = True\n element = elements[ielement, :]\n\n # could be faster...\n for inode in element:\n if inode not in inodes_save:\n save_element = False\n break\n\n if save_element:\n ielements_save.add(ielement)\n\n ielements_save_lst = list(ielements_save)\n ielements_save_lst.sort()\n\n elements2 = elements[ielements_save_lst]\n regions2 = regions[ielements_save_lst]\n\n # renumbers mesh\n nelements2 = elements2.shape[0]\n assert 0 < nelements2 < nelements, 'nelements=%s nelements2=%s' % (nelements, nelements2)\n\n remap_nodes = False\n if np.amax(elements2) > len(inodes_save):\n # build a dictionary of old node ids to new node ids\n nodes_map = {}\n for i in range(1, len(inodes_save) + 1):\n nid = inodes_save[i - 1]\n nodes_map[nid] = i\n\n # update the node ids\n for ielement in range(nelements2):\n element = elements2[ielement, :]\n elements[ielement, :] = [nodes_map[nid] for nid in element]\n\n loads2 = {} # 'Cp', 'Mach', 'U', etc.\n for key, load in iteritems(loads):\n loads2[key] = load[inodes_save]\n\n self.log.info('---finished make_half_model---')\n return (nodes2, elements2, regions2, loads2)\n\n def get_free_edges(self, elements):\n \"\"\"\n Cart3d must be a closed model with each edge shared by 2 elements\n The free edges indicate the problematic areas.\n\n Returns\n -------\n free edges : (nedges, 2) int ndarray\n the free edge node ids\n \"\"\"\n edge_to_eid_map = defaultdict(list)\n for i, element in enumerate(elements):\n edge1 = tuple(sorted([element[0], element[1]]))\n edge2 = tuple(sorted([element[1], element[2]]))\n edge3 = tuple(sorted([element[2], element[0]]))\n edge_to_eid_map[edge1].append(i)\n edge_to_eid_map[edge2].append(i)\n edge_to_eid_map[edge3].append(i)\n\n free_edges = []\n for edge, eids in sorted(iteritems(edge_to_eid_map)):\n if len(eids) != 2:\n free_edges.append(edge)\n return np.array(free_edges, dtype='int32')\n\n def read_cart3d(self, infilename, result_names=None):\n \"\"\"extracts the points, elements, and Cp\"\"\"\n self.infilename = infilename\n self.log.info(\"---reading cart3d...%r---\" % self.infilename)\n\n self.infilename = infilename\n if is_binary_file(infilename):\n with open(infilename, 'rb') as self.infile:\n try:\n npoints, nelements, nresults = self._read_header_binary()\n self.points = self._read_points_binary(npoints)\n self.elements = self._read_elements_binary(nelements)\n self.regions = self._read_regions_binary(nelements)\n # TODO: loads\n except:\n msg = 'failed reading %r' % infilename\n self.log.error(msg)\n raise\n\n else:\n with codec_open(_filename(infilename), 'r', encoding=self._encoding) as self.infile:\n try:\n npoints, nelements, nresults = self._read_header_ascii()\n self.points = self._read_points_ascii(npoints)\n self.elements = self._read_elements_ascii(nelements)\n self.regions = self._read_regions_ascii(nelements)\n self._read_results_ascii(0, self.infile, nresults, result_names=result_names)\n except:\n msg = 'failed reading %r' % infilename\n self.log.error(msg)\n raise\n\n self.log.debug(\"npoints=%s nelements=%s\" % (self.npoints, self.nelements))\n assert self.npoints > 0, 'npoints=%s' % self.npoints\n assert self.nelements > 0, 'nelements=%s' % self.nelements\n\n def write_cart3d(self, outfilename, is_binary=False, float_fmt='%6.7f'):\n \"\"\"\n writes a cart3d file\n \"\"\"\n assert len(self.points) > 0, 'len(self.points)=%s' % len(self.points)\n\n if self.loads is None or self.loads == {}:\n loads = {}\n is_loads = False\n else:\n is_loads = True\n\n self.log.info(\"---writing cart3d...%r---\" % outfilename)\n if is_binary:\n form = 'wb'\n else:\n form = write_ascii\n\n with codec_open(outfilename, form) as outfile:\n int_fmt = self._write_header(outfile, self.points, self.elements, is_loads, is_binary)\n self._write_points(outfile, self.points, is_binary, float_fmt)\n self._write_elements(outfile, self.elements, is_binary, int_fmt)\n self._write_regions(outfile, self.regions, is_binary)\n\n if is_loads:\n assert is_binary is False, 'is_binary=%r is not supported for loads' % is_binary\n self._write_loads(outfile, self.loads, is_binary, float_fmt)\n\n\n def get_min(self, points, i):\n return np.amin(points[:, i])\n\n def get_max(self, points, i):\n return np.amax(points[:, i])\n\n def _read_results_ascii(self, i, infile, nresults, result_names=None):\n \"\"\"\n Reads the Cp results.\n Results are read on a nodal basis from the following table:\n Cp\n rho,rhoU,rhoV,rhoW,rhoE\n\n With the following definitions:\n Cp = (p - 1/gamma) / (0.5*M_inf*M_inf)\n rhoVel^2 = rhoU^2+rhoV^2+rhoW^2\n M^2 = rhoVel^2/rho^2\n\n Thus:\n p = (gamma-1)*(e- (rhoU**2+rhoV**2+rhoW**2)/(2.*rho))\n p_dimensional = qInf * Cp + pInf\n\n # ???\n rho,rhoU,rhoV,rhoW,rhoE\n\n Parameters\n ----------\n result_names : List[str]; default=None (All)\n result_names = ['Cp', 'rho', 'rhoU', 'rhoV', 'rhoW', 'rhoE',\n 'Mach', 'U', 'V', 'W', 'E']\n \"\"\"\n if nresults == 0:\n return\n loads = {}\n if result_names is None:\n result_names = ['Cp', 'rho', 'rhoU', 'rhoV', 'rhoW', 'rhoE',\n 'Mach', 'U', 'V', 'W', 'E', 'a', 'T', 'Pressure', 'q']\n self.log.debug('---starting read_results---')\n\n results = np.zeros((self.npoints, 6), dtype='float32')\n\n nresult_lines = int(ceil(nresults / 5.)) - 1\n for ipoint in range(self.npoints):\n # rho rhoU,rhoV,rhoW,pressure/rhoE/E\n sline = infile.readline().strip().split()\n i += 1\n for n in range(nresult_lines):\n sline += infile.readline().strip().split() # Cp\n i += 1\n #gamma = 1.4\n #else:\n # p=0.\n sline = _get_list(sline)\n\n # Cp\n # rho rhoU rhoV rhoW E\n # 0.416594\n # 1.095611 0.435676 0.003920 0.011579 0.856058\n results[ipoint, :] = sline\n\n #p=0\n #cp = sline[0]\n #rho = float(sline[1])\n #if(rho > abs(0.000001)):\n #rhoU = float(sline[2])\n #rhoV = float(sline[3])\n #rhoW = float(sline[4])\n #rhoE = float(sline[5])\n #mach2 = (rhoU) ** 2 + (rhoV) ** 2 + (rhoW) ** 2 / rho ** 2\n #mach = sqrt(mach2)\n #if mach > 10:\n #print(\"nid=%s Cp=%s mach=%s rho=%s rhoU=%s rhoV=%s rhoW=%s\" % (\n #pointNum, cp, mach, rho, rhoU, rhoV, rhoW))\n #print(\"pt=%s i=%s Cp=%s p=%s\" %(pointNum,i,sline[0],p))\n del sline\n self.loads = self._calculate_results(result_names, results)\n\n def _calculate_results(self, result_names, results, loads=None):\n \"\"\"\n Takes the Cart3d variables and calculates additional variables\n\n Parameters\n ----------\n result_names : List[str]\n the variables to calculate\n results : (n,6) ndarray\n the non-dimensional prmitive flow variables\n loads : dict; default=None -> {}\n key : ???\n value : ???\n \"\"\"\n if loads is None:\n loads = {}\n Cp = results[:, 0]\n rho = results[:, 1]\n rhoU = results[:, 2]\n rhoV = results[:, 3]\n rhoW = results[:, 4]\n E = results[:, 5]\n\n ibad = np.where(rho <= 0.000001)[0]\n if len(ibad) > 0:\n\n if 'Mach' in result_names:\n Mach = np.sqrt(rhoU**2 + rhoV**2 + rhoW**2)# / rho\n Mach[ibad] = 0.0\n if 'U' in result_names:\n U = rhoU / rho\n U[ibad] = 0.0\n if 'U' in result_names:\n V = rhoV / rho\n V[ibad] = 0.0\n if 'W' in result_names:\n W = rhoW / rho\n W[ibad] = 0.0\n #if 'rhoE' in result_names:\n #rhoE = rhoE / rho\n #e[ibad] = 0.0\n\n is_bad = True\n n = 0\n #for i in ibad:\n #print(\"nid=%s Cp=%s mach=%s rho=%s rhoU=%s rhoV=%s rhoW=%s\" % (\n #i, Cp[i], Mach[i], rho[i], rhoU[i], rhoV[i], rhoW[i]))\n #Mach[i] = 0.0\n #n += 1\n #if n > 10:\n # break\n else:\n is_bad = False\n\n\n #loc = locals()\n if 'Cp' in result_names:\n loads['Cp'] = Cp\n if 'rhoU' in result_names:\n loads['rhoU'] = rhoU\n if 'rhoV' in result_names:\n loads['rhoV'] = rhoV\n if 'rhoW' in result_names:\n loads['rhoW'] = rhoW\n #if 'rhoE' in result_names:\n #loads['rhoE'] = rhoE\n\n if 'rho' in result_names:\n loads['rho'] = rho\n\n if 'Mach' in result_names:\n if not is_bad:\n #Mach = np.sqrt(rhoU**2 + rhoV**2 + rhoW**2) / rho\n Mach = np.sqrt(rhoU**2 + rhoV**2 + rhoW**2)\n loads['Mach'] = Mach\n\n if 'U' in result_names:\n if not is_bad:\n U = rhoU / rho\n loads['U'] = U\n if 'V' in result_names:\n if not is_bad:\n V = rhoV / rho\n loads['V'] = V\n if 'W' in result_names:\n if not is_bad:\n W = rhoW / rho\n loads['W'] = W\n if 'E' in result_names:\n #if not is_bad:\n #E = rhoE / rho\n loads['E'] = E\n\n gamma = 1.4\n qinf = 1.0\n pinf = 1. / gamma\n Tinf = 1.0\n #Cp = (p - pinf) / qinf\n p = Cp * qinf + pinf\n\n T = (Tinf * gamma) * p / rho\n q = 0.5 * rho * Mach ** 2\n\n if 'a' in result_names:\n #print('T: min=%s max=%s' % (T.min(), T.max()))\n loads['a'] = np.sqrt(T)\n if 'T' in result_names:\n loads['T'] = T\n\n if 'Pressure' in result_names:\n loads['Pressure'] = p\n if 'q' in result_names:\n loads['q'] = q\n # dynamic pressure\n # speed of sound\n # total pressure = p0/rhoi*ainf**2\n # total density\n # entropy\n # kinetic energy\n # enthalpy\n # energy, E\n # total energy\n # total enthalpy\n\n #i = where(Mach == max(Mach))[0][0]\n #self.log.info(\"i=%s Cp=%s rho=%s rhoU=%s rhoV=%s rhoW=%s Mach=%s\" % (\n #i, Cp[i], rho[i], rhoU[i], rhoV[i], rhoW[i], Mach[i]))\n self.log.debug('---finished read_results---')\n return loads\n\n def _get_area_vector(self):\n \"\"\"\n Gets the area vector (unnormalized normal vector)\n Returns\n -------\n normals : (n, 3) ndarray\n unnormalized centroidal normal vectors\n \"\"\"\n elements = self.elements\n nodes = self.nodes\n p1 = nodes[elements[:, 0], :]\n p2 = nodes[elements[:, 1], :]\n p3 = nodes[elements[:, 2], :]\n\n ne = elements.shape[0]\n avec = p2 - p1\n bvec = p3 - p1\n n = np.cross(avec, bvec)\n assert len(n) == ne, 'len(n)=%s ne=%s' % (len(n), ne)\n\n return n\n\n def get_area(self):\n \"\"\"\n Gets the element area\n\n Returns\n -------\n area : (n, 3) ndarray\n the element areas\n \"\"\"\n ne = self.elements.shape[0]\n n = self._get_area_vector()\n ni = np.linalg.norm(n, axis=1)\n assert len(ni) == ne, 'len(ni)=%s ne=%s' % (len(ni), ne)\n return 0.5 * ni\n\n def get_normals(self):\n \"\"\"\n Gets the centroidal normals\n\n Returns\n -------\n cnormals : (n, 3) ndarray\n normalized centroidal normal vectors\n \"\"\"\n ne = self.elements.shape[0]\n n = self._get_area_vector()\n ni = np.linalg.norm(n, axis=1)\n assert len(ni) == ne, 'len(ni)=%s ne=%s' % (len(ni), ne)\n\n assert ni.min() > 0.0, ni[np.where(ni <= 0.0)[0]]\n n /= ni[:, None] # normal vector\n return n\n\n def get_normals_at_nodes(self, cnormals):\n \"\"\"\n Gets the nodal normals\n\n Parameters\n ----------\n cnormals : (n, 3) ndarray\n normalized centroidal normal vectors\n\n Returns\n -------\n nnormals : (n, 3) ndarray\n normalized nodal normal vectors\n \"\"\"\n elements = self.elements\n nodes = self.nodes\n nnodes = self.nnodes\n nid_to_eids = defaultdict(list)\n\n # find the elements to consider for each node\n for eid, element in enumerate(elements):\n n1, n2, n3 = element\n nid_to_eids[n1].append(eid)\n nid_to_eids[n2].append(eid)\n nid_to_eids[n3].append(eid)\n\n nnormals = np.zeros((nnodes, 3), dtype='float64')\n for nid in range(nnodes):\n eids = nid_to_eids[nid]\n if len(eids) == 0:\n raise RuntimeError('nid=%s is not used' % nid)\n ni_avg = cnormals[eids, :]\n nnormals[nid] = cnormals[eids, :].sum(axis=0)\n ni = np.linalg.norm(nnormals, axis=1)\n assert ni.min() > 0, ni\n nnormals /= ni[:, None] # normal vector\n return nnormals\n\ndef convert_to_float(svalues):\n \"\"\"Takes a list of strings and converts them to floats.\"\"\"\n values = []\n for value in svalues:\n values.append(float(value))\n return values\n\ndef _get_list(sline):\n \"\"\"Takes a list of strings and converts them to floats.\"\"\"\n try:\n sline2 = convert_to_float(sline)\n except ValueError:\n print(\"sline = %s\" % sline)\n raise SyntaxError('cannot parse %s' % sline)\n return sline2\n", "repo_name": "EmanueleCannizzaro/pynastran2", "sub_path": "pynastran2/cart3d.py", "file_name": "cart3d.py", "file_ext": "py", "file_size_in_byte": 36284, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 3, "dataset": "github-code", "pt": "25", "api": [{"api_name": "six.PY2", "line_number": 33, "usage_type": "name"}, {"api_name": "numpy.unique", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 91, "usage_type": "call"}, {"api_name": "pyNastran.utils.log.get_logger2", "line_number": 104, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 136, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 137, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 139, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 140, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 146, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 148, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 152, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 159, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 159, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 163, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 173, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 180, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 181, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 184, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 194, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 199, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 200, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 204, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 211, "usage_type": "call"}, {"api_name": "six.moves.zip", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 313, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 329, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 341, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 342, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 360, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 360, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 363, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 363, "usage_type": "call"}, {"api_name": "numpy.dtype", "line_number": 378, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 378, "usage_type": "call"}, {"api_name": "numpy.fromstring", "line_number": 379, "usage_type": "call"}, {"api_name": "numpy.dtype", "line_number": 389, "usage_type": "call"}, {"api_name": "pyNastran.utils.b", "line_number": 389, "usage_type": "call"}, {"api_name": "numpy.fromstring", "line_number": 390, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.fromstring", "line_number": 403, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 426, "usage_type": "attribute"}, {"api_name": "struct.unpack", "line_number": 444, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 447, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 450, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 454, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 457, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 460, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 463, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 468, "usage_type": "attribute"}, {"api_name": "numpy.hstack", "line_number": 498, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 536, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 539, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 548, "usage_type": "call"}, {"api_name": "six.moves.range", "line_number": 557, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 567, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 571, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 573, "usage_type": "call"}, {"api_name": "six.iteritems", "line_number": 576, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 585, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 635, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 637, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 642, "usage_type": "call"}, {"api_name": "six.moves.range", "line_number": 654, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 678, "usage_type": "call"}, {"api_name": "six.moves.range", "line_number": 681, "usage_type": "call"}, {"api_name": "six.moves.range", "line_number": 686, "usage_type": "call"}, {"api_name": "six.iteritems", "line_number": 691, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 707, "usage_type": "call"}, {"api_name": "six.iteritems", "line_number": 717, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 720, "usage_type": "call"}, {"api_name": "pyNastran.utils.is_binary_file", "line_number": 728, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 742, "usage_type": "call"}, {"api_name": "pyNastran.utils._filename", "line_number": 742, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 776, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 788, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 791, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 826, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 828, "usage_type": "call"}, {"api_name": "six.moves.range", "line_number": 829, "usage_type": "call"}, {"api_name": "six.moves.range", "line_number": 833, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 887, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 891, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 937, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 969, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 1011, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 1027, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1027, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 1042, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1042, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 1045, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 1066, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1075, "usage_type": "call"}, {"api_name": "six.moves.range", "line_number": 1076, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 1082, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1082, "usage_type": "attribute"}]}
+{"seq_id": "17702451898", "text": "from .. import PloneMessageFactory as _\nfrom ..browser import formhelper\nfrom ..portlets import base\nfrom Acquisition import aq_base\nfrom Acquisition import aq_inner\nfrom plone.base.interfaces.controlpanel import ISiteSchema\nfrom plone.i18n.normalizer.interfaces import IIDNormalizer\nfrom plone.memoize.instance import memoize\nfrom plone.portlets.interfaces import IPortletDataProvider\nfrom plone.registry.interfaces import IRegistry\nfrom Products.CMFCore.utils import getToolByName\nfrom Products.Five.browser.pagetemplatefile import ViewPageTemplateFile\nfrom zope import schema\nfrom zope.component import getMultiAdapter\nfrom zope.component import getUtility\nfrom zope.component import queryUtility\nfrom zope.interface import implementer\n\n\nclass IReviewPortlet(IPortletDataProvider):\n no_icons = schema.Bool(\n title=_(\"Suppress Icons\"),\n description=_(\"If enabled, the portlet will not show document type icons\"),\n required=False,\n default=False,\n )\n\n thumb_scale = schema.TextLine(\n title=_(\"Override thumb scale\"),\n description=_(\n \"Enter a valid scale name\"\n \" (see 'Image Handling' control panel) to override\"\n \" (e.g. icon, tile, thumb, mini, preview, ... ).\"\n \" Leave empty to use default (see 'Site' control panel).\"\n ),\n required=False,\n default=\"\",\n )\n\n no_thumbs = schema.Bool(\n title=_(\"Suppress thumbs\"),\n description=_(\"If enabled, the portlet will not show thumbs.\"),\n required=False,\n default=False,\n )\n\n\n@implementer(IReviewPortlet)\nclass Assignment(base.Assignment):\n no_icons = False\n thumb_scale = None\n no_thumbs = False\n\n def __init__(self, no_icons=False, thumb_scale=None, no_thumbs=False):\n self.no_icons = no_icons\n self.thumb_scale = thumb_scale\n self.no_thumbs = no_thumbs\n\n @property\n def title(self):\n return _(\"Review list\")\n\n\nclass Renderer(base.Renderer):\n render = ViewPageTemplateFile(\"review.pt\")\n\n title = _(\"box_review_list\", default=\"Review List\")\n\n def __init__(self, *args):\n base.Renderer.__init__(self, *args)\n\n @property\n def anonymous(self):\n context = aq_inner(self.context)\n portal_state = getMultiAdapter(\n (context, self.request), name=\"plone_portal_state\"\n )\n return portal_state.anonymous()\n\n @property\n def available(self):\n return not self.anonymous and len(self._data())\n\n def review_items(self):\n return self._data()\n\n def full_review_link(self):\n context = aq_inner(self.context)\n mtool = getToolByName(context, \"portal_membership\")\n # check if user is allowed to Review Portal Content here\n if mtool.checkPermission(\"Review portal content\", context):\n return \"%s/full_review_list\" % context.absolute_url()\n else:\n return None\n\n @memoize\n def _data(self):\n if self.anonymous:\n return []\n context = aq_inner(self.context)\n workflow = getToolByName(context, \"portal_workflow\")\n\n plone_view = getMultiAdapter((context, self.request), name=\"plone\")\n getMember = getToolByName(context, \"portal_membership\").getMemberById\n toLocalizedTime = plone_view.toLocalizedTime\n\n idnormalizer = queryUtility(IIDNormalizer)\n norm = idnormalizer.normalize\n objects = workflow.getWorklistsResults()\n items = []\n for obj in objects:\n review_state = workflow.getInfoFor(obj, \"review_state\")\n creator_id = obj.Creator()\n creator = getMember(creator_id)\n if creator:\n creator_name = creator.getProperty(\"fullname\", \"\") or creator_id\n else:\n creator_name = creator_id\n hasImage = True if getattr(aq_base(obj), \"image\", None) else False\n images = obj.restrictedTraverse(\"@@images\") if hasImage else None\n items.append(\n dict(\n path=obj.absolute_url(),\n title=obj.pretty_title_or_id(),\n item_class=\"contenttype-\" + norm(obj.portal_type),\n description=obj.Description(),\n creator=creator_name,\n review_state=review_state,\n review_state_class=\"state-%s \" % norm(review_state),\n mod_date=toLocalizedTime(obj.ModificationDate()),\n hasImage=hasImage,\n images=images,\n )\n )\n return items\n\n @memoize\n def thumb_scale(self):\n \"\"\"Use override value or read thumb_scale from registry.\n Image sizes must fit to value in allowed image sizes.\n None will suppress thumb.\n \"\"\"\n if getattr(self.data, \"no_thumbs\", False):\n # Individual setting overrides ...\n return None\n thsize = getattr(self.data, \"thumb_scale\", \"\")\n if thsize:\n return thsize\n registry = getUtility(IRegistry)\n settings = registry.forInterface(ISiteSchema, prefix=\"plone\", check=False)\n thumb_scale_portlet = settings.thumb_scale_portlet\n return thumb_scale_portlet\n\n\nclass AddForm(formhelper.AddForm):\n schema = IReviewPortlet\n label = _(\"Add Review Portlet\")\n description = _(\"This portlet displays a queue of documents awaiting \" \"review.\")\n\n def create(self, data):\n return Assignment(**data)\n\n\nclass EditForm(formhelper.EditForm):\n schema = IReviewPortlet\n label = _(\"Edit Review Portlet\")\n description = _(\"displays a queue of documents awaiting \" \"review.\")\n", "repo_name": "plone/plone.app.portlets", "sub_path": "plone/app/portlets/portlets/review.py", "file_name": "review.py", "file_ext": "py", "file_size_in_byte": 5686, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 12, "dataset": "github-code", "pt": "25", "api": [{"api_name": "plone.portlets.interfaces.IPortletDataProvider", "line_number": 20, "usage_type": "name"}, {"api_name": "zope.schema.Bool", "line_number": 21, "usage_type": "call"}, {"api_name": "zope.schema", "line_number": 21, "usage_type": "name"}, {"api_name": "zope.schema.TextLine", "line_number": 28, "usage_type": "call"}, {"api_name": "zope.schema", "line_number": 28, "usage_type": "name"}, {"api_name": "zope.schema.Bool", "line_number": 40, "usage_type": "call"}, {"api_name": "zope.schema", "line_number": 40, "usage_type": "name"}, {"api_name": "portlets.base.Assignment", "line_number": 49, "usage_type": "attribute"}, {"api_name": "portlets.base", "line_number": 49, "usage_type": "name"}, {"api_name": "zope.interface.implementer", "line_number": 48, "usage_type": "call"}, {"api_name": "portlets.base.Renderer", "line_number": 64, "usage_type": "attribute"}, {"api_name": "portlets.base", "line_number": 64, "usage_type": "name"}, {"api_name": "Products.Five.browser.pagetemplatefile.ViewPageTemplateFile", "line_number": 65, "usage_type": "call"}, {"api_name": "portlets.base.Renderer.__init__", "line_number": 70, "usage_type": "call"}, {"api_name": "portlets.base.Renderer", "line_number": 70, "usage_type": "attribute"}, {"api_name": "portlets.base", "line_number": 70, "usage_type": "name"}, {"api_name": "Acquisition.aq_inner", "line_number": 74, "usage_type": "call"}, {"api_name": "zope.component.getMultiAdapter", "line_number": 75, "usage_type": "call"}, {"api_name": "Acquisition.aq_inner", "line_number": 88, "usage_type": "call"}, {"api_name": "Products.CMFCore.utils.getToolByName", "line_number": 89, "usage_type": "call"}, {"api_name": "Acquisition.aq_inner", "line_number": 100, "usage_type": "call"}, {"api_name": "Products.CMFCore.utils.getToolByName", "line_number": 101, "usage_type": "call"}, {"api_name": "zope.component.getMultiAdapter", "line_number": 103, "usage_type": "call"}, {"api_name": "Products.CMFCore.utils.getToolByName", "line_number": 104, "usage_type": "call"}, {"api_name": "zope.component.queryUtility", "line_number": 107, "usage_type": "call"}, {"api_name": "plone.i18n.normalizer.interfaces.IIDNormalizer", "line_number": 107, "usage_type": "argument"}, {"api_name": "Acquisition.aq_base", "line_number": 119, "usage_type": "call"}, {"api_name": "plone.memoize.instance.memoize", "line_number": 96, "usage_type": "name"}, {"api_name": "zope.component.getUtility", "line_number": 149, "usage_type": "call"}, {"api_name": "plone.registry.interfaces.IRegistry", "line_number": 149, "usage_type": "argument"}, {"api_name": "plone.base.interfaces.controlpanel.ISiteSchema", "line_number": 150, "usage_type": "argument"}, {"api_name": "plone.memoize.instance.memoize", "line_number": 137, "usage_type": "name"}, {"api_name": "browser.formhelper.AddForm", "line_number": 155, "usage_type": "attribute"}, {"api_name": "browser.formhelper", "line_number": 155, "usage_type": "name"}, {"api_name": "zope.schema", "line_number": 156, "usage_type": "name"}, {"api_name": "browser.formhelper.EditForm", "line_number": 164, "usage_type": "attribute"}, {"api_name": "browser.formhelper", "line_number": 164, "usage_type": "name"}, {"api_name": "zope.schema", "line_number": 165, "usage_type": "name"}]}
+{"seq_id": "33855853629", "text": "import common\nimport f_main\nfrom f_main import ( bail, diag, DIAG_COLUMN, DIAG_ERR, DIAG_FILE, DIAG_LINE )\nimport f_stmt\nimport f_token as tk\nimport f_expr\nimport b_opcode\n\nAREA_TOP = 0\nAREA_LOCAL = 1\nAREA_FOR = 2\nAREA_PARAM = 3\n\nMAX_MAP_LOCATIONS = 128\nMAX_WORLD_LOCATIONS = 256\nMAX_GLOBAL_LOCATIONS = 64\n\nSCRIPT_MIN_NUM = 0\nSCRIPT_MAX_NUM = 999\nSCRIPT_MAX_PARAMS = 3\n\ndef is_dec( front ):\n return front.tk in [ tk.T_INT, tk.T_STR, tk.T_BOOL, tk.T_VOID,\n tk.T_FUNCTION, tk.T_WORLD, tk.T_GLOBAL, tk.T_STATIC ]\n\ndef read( front, area ):\n dec = {\n 'area' : area,\n 'pos' : front.tk_pos,\n 'storage' : common.STORAGE_LOCAL,\n 'storage_pos' : None,\n 'storage_name' : 'local',\n 'value' : False,\n 'name' : '',\n 'name_pos' : None,\n 'dim' : [],\n 'dim_implicit' : None,\n 'initials' : []\n }\n func = False\n if front.tk == tk.T_FUNCTION and area == AREA_TOP:\n tk.read( front )\n func = True\n\n # Storage.\n if not func:\n if front.tk == tk.T_GLOBAL:\n dec[ 'storage' ] = common.STORAGE_GLOBAL\n dec[ 'storage_pos' ] = front.tk_pos\n dec[ 'storage_name' ] = front.tk_text\n tk.read( front )\n elif front.tk == tk.T_WORLD:\n dec[ 'storage' ] = common.STORAGE_WORLD\n dec[ 'storage_pos' ] = front.tk_pos\n dec[ 'storage_name' ] = front.tk_text\n tk.read( front )\n elif front.tk == tk.T_STATIC:\n dec[ 'storage' ] = common.STORAGE_MAP\n dec[ 'storage_name' ] = 'map'\n if area == AREA_FOR:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, 'static variable in for loop initialization' )\n elif area == AREA_PARAM:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, '\\'static\\' used in parameter' )\n elif area == AREA_TOP:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, '\\'static\\' used in top scope' )\n tk.read( front )\n elif area == AREA_TOP:\n dec[ 'storage' ] = common.STORAGE_MAP\n dec[ 'storage_name' ] = 'map'\n\n # Type.\n if front.tk in [ tk.T_INT, tk.T_STR, tk.T_BOOL ]:\n # Scripts can only have integer parameters.\n if area == AREA_PARAM and front.dec_params[ 'is_script' ] and \\\n front.tk != tk.T_INT:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, 'script parameter not of \\'int\\' type' )\n dec[ 'value' ] = True\n tk.read( front )\n elif front.tk == tk.T_VOID:\n tk.read( front )\n else:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, 'expecting type in declaration' )\n raise Exception\n bail( front )\n\n while True:\n # Storage index.\n if not func:\n if front.tk == tk.T_LIT_DECIMAL:\n pos = front.tk_pos\n dec[ 'storage_index' ] = f_expr.read_literal( front )\n tk.test( front, tk.T_COLON )\n tk.read( front )\n max_loc = MAX_WORLD_LOCATIONS\n if dec[ 'storage' ] != common.STORAGE_WORLD:\n if dec[ 'storage' ] == common.STORAGE_GLOBAL:\n max_loc = MAX_GLOBAL_LOCATIONS\n else:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n pos, 'index specified for %s storage',\n dec[ 'storage_name' ] )\n bail( front )\n if dec[ 'storage_index' ] >= max_loc:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n pos, 'index specified for %s storage',\n dec[ 'storage_name' ] )\n bail( front )\n else:\n # Index must be explicitly specified for these storages.\n if dec[ 'storage' ] == common.STORAGE_WORLD or \\\n dec[ 'storage' ] == common.STORAGE_GLOBAL:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, 'missing index for %s storage',\n dec[ 'storage_name' ] )\n bail( front )\n\n # Name.\n if front.tk == tk.T_ID:\n dec[ 'name_pos' ] = front.tk_pos\n dec[ 'name' ] = read_unique_name( front )\n else:\n # Parameters don't require a name.\n if area != AREA_PARAM:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, 'missing name in declaration' )\n bail( front )\n\n if func:\n tk.test( front, tk.T_PAREN_L )\n tk.read( front )\n f_main.add_scope( front )\n num_param = read_param_list( front, False )\n tk.test( front, tk.T_PAREN_R )\n tk.read( front )\n func = common.func_t()\n func.pos = dec[ 'name_pos' ]\n func.value = dec[ 'value' ]\n func.name = dec[ 'name' ]\n func.min_param = num_param\n func.max_param = num_param\n front.scopes[ 0 ].names[ func.name ] = func\n block = common.block_t()\n front.block = block\n front.func = func\n f_stmt.read_block( front )\n f_main.pop_scope( front )\n front.block = None\n front.func = None\n func.impl = {\n 'def' : True,\n 'def_params' : True,\n 'body' : block,\n 'index' : 0,\n 'size' : 0\n }\n front.module.funcs.append( func )\n break\n else:\n # Array dimension.\n if front.tk == tk.T_BRACKET_L:\n read_dim( front, dec )\n else:\n dec[ 'dim' ] = []\n dec[ 'dim_implicit' ] = None\n read_init( front, dec )\n if area == AREA_PARAM:\n if dec[ 'name' ]:\n pass\n break\n else:\n finish_var( front, dec )\n if front.tk == tk.T_COMMA:\n tk.read( front )\n else:\n tk.test( front, tk.T_SEMICOLON )\n tk.read( front )\n break\n\ndef read_unique_name( front ):\n tk.test( front, tk.T_ID )\n if front.tk_text in front.scope.names:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, 'name \\'%s\\' already used', front.tk_text )\n object = front.scope.names[ front.tk_text ]\n pos = None\n if object.node == common.NODE_VAR or \\\n object.node == common.NODE_CONSTANT:\n pos = object.pos\n elif object.node == common.NODE_FUNC:\n # These builtin functions are loaded internally by the compiler and\n # have no position of their definition given.\n if object.type != common.FUNC_DED and \\\n object.type != common.FUNC_FORMAT:\n pos = object.pos\n if pos:\n diag( front, DIAG_FILE | DIAG_LINE | DIAG_COLUMN, pos,\n 'name previously used here' )\n bail( front )\n else:\n name = front.tk_text\n tk.read( front )\n return name\n\ndef read_dim( front, dec ):\n # At this time, a local array is not allowed.\n if dec[ 'storage' ] == common.STORAGE_LOCAL:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, 'array in local scope' )\n bail( front )\n while front.tk == tk.T_BRACKET_L:\n pos = front.tk_pos\n tk.read( front )\n expr = None\n # Implicit size.\n if front.tk == tk.T_BRACKET_R:\n # Only the first dimension can have an implicit size.\n if len( dec[ 'dim' ] ):\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN, pos,\n 'implicit size in subsequent dimension' )\n bail( front )\n tk.read( front )\n else:\n expr = f_expr.read( front, True )\n tk.test( front, tk.T_BRACKET_R )\n tk.read( front )\n if not expr.folded:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n expr.pos, 'array size not a constant expression' )\n bail( front )\n elif expr.value <= 0:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n expr.pos, 'invalid array size' )\n bail( front )\n dim = common.dim_t()\n dec[ 'dim' ].append( dim )\n if not expr:\n dec[ 'dim_implicit' ] = dim\n else:\n dim.size = expr.value\n i = len( dec[ 'dim' ] ) - 1\n # For now, each element of the last dimension is 1 integer in size.\n dec[ 'dim' ][ i ].element_size = 1\n while i > 0:\n dec[ 'dim' ][ i - 1 ].element_size = (\n dec[ 'dim' ][ i ].element_size *\n dec[ 'dim' ][ i ].size )\n i -= 1\n\ndef read_init( front, dec ):\n if front.tk != tk.T_ASSIGN:\n if dec[ 'dim_implicit' ] and ( (\n dec[ 'storage' ] != common.STORAGE_WORLD and\n dec[ 'storage' ] != common.STORAGE_GLOBAL ) or\n len( dec[ 'dim' ] ) > 1 ):\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, 'missing initialization' )\n bail( front )\n return\n # At this time, there is no way to initialize at the top scope an array with\n # world or global storage.\n if ( dec[ 'storage' ] == common.STORAGE_WORLD or\n dec[ 'storage' ] == common.STORAGE_GLOBAL ) and \\\n len( front.scopes ) == 1:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, 'initialization of variable with %s storage '\n 'at top scope', dec[ 'storage_name' ] )\n bail( front )\n tk.read( front )\n if front.tk == tk.T_BRACE_L:\n read_initz( front, dec )\n else:\n if len( dec[ 'dim' ] ):\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, 'array initialization missing initializer' )\n bail( front )\n expr = f_expr.read( front, True )\n if dec[ 'storage' ] == common.STORAGE_MAP and not expr.folded:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN, expr.pos,\n 'initial value not constant' )\n bail( front )\n initial = common.initial_t()\n initial.value = expr.value\n dec[ 'initials' ].append( initial )\n\ndef finish_var( front, dec ):\n var = common.var_t()\n var.pos = dec[ 'name_pos' ]\n var.name = dec[ 'name' ]\n var.storage = dec[ 'storage' ]\n var.dim = dec[ 'dim' ]\n if var.dim:\n var.size = var.dim[ 0 ].size * var.dim[ 0 ].element_size\n var.initial = dec[ 'initials' ]\n else:\n var.size = 1\n if dec[ 'initials' ]:\n var.initial = dec[ 'initials' ].pop()\n front.scope.names[ var.name ] = var\n if var.dim:\n front.module.arrays.append( var )\n elif dec[ 'area' ] == AREA_TOP:\n # Variables with initials appear first.\n if var.initial:\n front.module.vars.insert( 0, var )\n else:\n front.module.vars.append( var )\n elif dec[ 'area' ] == AREA_LOCAL:\n var.index = alloc_index( front )\n front.block.stmts.append( var )\n else:\n var.index = alloc_index( front )\n front.dec_for_init.append( var )\n\ndef read_param_list( front, is_script ):\n return 0\n\ndef read_script( front ):\n tk.test( front, tk.T_SCRIPT )\n script = common.script_t()\n script.pos = front.tk_pos\n tk.read( front )\n # Script number.\n number_pos = None\n if front.tk == tk.T_SHIFT_L:\n tk.read( front )\n # The token between the << and >> tokens must be the digit zero.\n if front.tk == tk.T_LIT_DECIMAL and front.tk_text == '0':\n number_pos = front.tk_pos\n tk.read( front )\n tk.test( front, tk.T_SHIFT_R )\n tk.read( front )\n else:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, 'missing the digit 0' )\n bail( front )\n else:\n front.reading_script_number = True\n expr = f_expr.read( front, True )\n front.reading_script_number = False\n if not expr.folded:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n expr.pos, 'script number not a constant expression' )\n bail( front )\n elif expr.value < SCRIPT_MIN_NUM or expr.value > SCRIPT_MAX_NUM:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n expr.pos, 'script number not between %d and %d', SCRIPT_MIN_NUM,\n SCRIPT_MAX_NUM )\n bail( front )\n elif expr.value == 0:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n expr.pos, 'script number 0 not between << and >>' )\n bail( front )\n else:\n script.number = expr.value\n number_pos = expr.pos\n # There should be no duplicate scripts in the same module.\n for older_script in front.module.scripts:\n if script.number == older_script.number:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n number_pos, 'script number %d already used', script.number )\n diag( front, DIAG_FILE | DIAG_LINE | DIAG_COLUMN, older_script.pos,\n 'first script to use number found here' )\n break\n f_main.add_scope( front )\n # Parameter list.\n param_pos = None\n if front.tk == tk.T_PAREN_L:\n param_pos = front.tk_pos\n tk.read( front )\n script.num_param = read_param_list( front, True )\n tk.test( front, tk.T_PAREN_R )\n tk.read( front )\n # Script type.\n types = {\n tk.T_OPEN : common.SCRIPT_TYPE_OPEN,\n tk.T_RESPAWN : common.SCRIPT_TYPE_RESPAWN,\n tk.T_DEATH : common.SCRIPT_TYPE_DEATH,\n tk.T_ENTER : common.SCRIPT_TYPE_ENTER,\n tk.T_PICKUP : common.SCRIPT_TYPE_PICKUP,\n tk.T_BLUE_RETURN : common.SCRIPT_TYPE_BLUE_RETURN,\n tk.T_RED_RETURN : common.SCRIPT_TYPE_RED_RETURN,\n tk.T_WHITE_RETURN : common.SCRIPT_TYPE_WHITE_RETURN,\n tk.T_LIGHTNING : common.SCRIPT_TYPE_LIGHTNING,\n tk.T_DISCONNECT : common.SCRIPT_TYPE_DISCONNECT,\n tk.T_UNLOADING : common.SCRIPT_TYPE_UNLOADING,\n tk.T_RETURN : common.SCRIPT_TYPE_RETURN }\n name = ''\n if front.tk in types:\n script.type = types[ front.tk ]\n name = front.tk_text\n tk.read( front )\n if script.type == common.SCRIPT_TYPE_CLOSED:\n if script.num_param > SCRIPT_MAX_PARAMS:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n param_pos, 'script has over maximum %d parameters',\n SCRIPT_MAX_PARAMS )\n elif script.type == common.SCRIPT_TYPE_DISCONNECT:\n # A disconnect script must have a single parameter. It is the number of\n # the player who disconnected from the server.\n if script.num_param != 1:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n param_pos,\n 'disconnect script missing one player-number parameter' )\n else:\n if script.num_param != 0:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n param_pos, 'parameter list specified for %s script', name )\n # Script flags.\n while True:\n flag = common.SCRIPT_FLAG_NET\n if front.tk != tk.T_NET:\n if front.tk == tk.T_CLIENTSIDE:\n flag = common.SCRIPT_FLAG_CLIENTSIDE\n else:\n break\n if not ( script.flags & flag ):\n script.flags |= flag\n tk.read( front )\n else:\n diag( front, DIAG_ERR | DIAG_FILE | DIAG_LINE | DIAG_COLUMN,\n front.tk_pos, '%s flag already set', front.tk_text )\n tk.read( front )\n # Body.\n block = common.block_t()\n block.in_script = True\n front.block = block\n f_stmt.read_block( front )\n front.block = None\n script.body = block\n f_main.pop_scope( front )\n front.module.scripts.append( script )\n\ndef read_bfunc_list( front ):\n tk.test( front, tk.T_SPECIAL )\n tk.read( front )\n while True:\n ext = False\n if front.tk == tk.T_MINUS:\n tk.read( front )\n ext = True\n code = f_expr.read_literal( front )\n tk.test( front, tk.T_COLON )\n tk.read( front )\n name_pos = front.tk_pos\n name = read_unique_name( front )\n tk.test( front, tk.T_PAREN_L )\n tk.read( front )\n min_param = 0\n max_param = f_expr.read_literal( front )\n if front.tk == tk.T_COMMA:\n min_param = max_param\n tk.read( front )\n max_param = f_expr.read_literal( front )\n tk.test( front, tk.T_PAREN_R )\n tk.read( front )\n func = common.func_t()\n func.pos = name_pos\n func.value = True\n func.min_param = min_param\n func.max_param = max_param\n front.scopes[ 0 ].names[ name ] = func\n if ext:\n func.type = common.FUNC_EXT\n func.impl[ 'id' ] = code\n else:\n func.type = common.FUNC_ASPEC\n func.impl[ 'id' ] = code\n if front.tk == tk.T_COMMA:\n tk.read( front )\n else:\n tk.test( front, tk.T_SEMICOLON )\n tk.read( front )\n break\n\ndef load_ded_format_funcs( front ):\n ded = [\n # Format: name / returns-value / parameter-count / parameter-minimum /\n # opcode / is-latent.\n [ 'delay', False, 1, 1, b_opcode.k_delay, True ],\n [ 'random', True, 2, 2, b_opcode.k_random, False ],\n [ 'thingcount', True, 2, 2, b_opcode.k_thing_count, False ],\n [ 'tagwait', False, 1, 1, b_opcode.k_tag_wait, True ],\n [ 'polywait', False, 1, 1, b_opcode.k_poly_wait, True ],\n [ 'changefloor', False, 2, 2, b_opcode.k_change_floor, False ],\n [ 'changeceiling', False, 2, 2, b_opcode.k_change_ceiling, False ],\n [ 'lineside', True, 0, 0, b_opcode.k_line_side, False ],\n [ 'scriptwait', False, 1, 1, b_opcode.k_script_wait, True ],\n [ 'clearlinespecial', False, 0, 0, b_opcode.k_clear_line_special, False ],\n [ 'playercount', True, 0, 0, b_opcode.k_player_count, False ],\n [ 'gametype', True, 0, 0, b_opcode.k_game_type, False ],\n [ 'gameskill', True, 0, 0, b_opcode.k_game_skill, False ],\n [ 'timer', True, 0, 0, b_opcode.k_timer, False ],\n [ 'sectorsound', False, 2, 2, b_opcode.k_sector_sound, False ],\n [ 'ambientsound', False, 2, 2, b_opcode.k_ambient_sound, False ],\n [ 'soundsequence', False, 1, 1, b_opcode.k_sound_sequence, False ],\n [ 'setlinetexture', False, 4, 4, b_opcode.k_set_line_texture, False ],\n [ 'setlineblocking', False, 2, 2, b_opcode.k_set_line_blocking, False ],\n [ 'setlinespecial', False, 7, 2, b_opcode.k_set_line_special, False ],\n [ 'thingsound', False, 3, 3, b_opcode.k_thing_sound, False ],\n [ 'activatorsound', False, 2, 2, b_opcode.k_activator_sound, False ],\n [ 'localambientsound', False, 2, 2, b_opcode.k_local_ambient_sound,\n False ],\n [ 'setlinemonsterblocking', False, 2, 2,\n b_opcode.k_set_line_monster_blocking, False ],\n [ 'ismultiplayer', True, 0, 0, b_opcode.k_is_multiplayer, False ],\n [ 'playerteam', True, 0, 0, b_opcode.k_player_team, False ],\n [ 'playerhealth', True, 0, 0, b_opcode.k_player_health, False ],\n [ 'playerarmorpoints', True, 0, 0, b_opcode.k_player_armor_points,\n False ],\n [ 'playerfrags', True, 0, 0, b_opcode.k_player_frags, False ],\n [ 'bluecount', True, 0, 0, b_opcode.k_blue_team_count, False ],\n [ 'blueteamcount', True, 0, 0, b_opcode.k_blue_team_count, False ],\n [ 'redcount', True, 0, 0, b_opcode.k_red_team_count, False ],\n [ 'redteamcount', True, 0, 0, b_opcode.k_red_team_count, False ],\n [ 'bluescore', True, 0, 0, b_opcode.k_blue_team_score, False ],\n [ 'blueteamscore', True, 0, 0, b_opcode.k_blue_team_score, False ],\n [ 'redscore', True, 0, 0, b_opcode.k_red_team_score, False ],\n [ 'redteamscore', True, 0, 0, b_opcode.k_red_team_score, False ],\n [ 'isoneflagctf', True, 0, 0, b_opcode.k_is_one_flag_ctf, False ],\n [ 'getinvasionwave', True, 0, 0, b_opcode.k_get_invasion_wave, False ],\n [ 'getinvasionstate', True, 0, 0, b_opcode.k_get_invasion_state, False ],\n [ 'music_change', False, 2, 2, b_opcode.k_music_change, False ],\n [ 'consolecommand', False, 3, 1, b_opcode.k_console_command, False ],\n [ 'singleplayer', True, 0, 0, b_opcode.k_single_player, False ],\n [ 'fixedmul', True, 2, 2, b_opcode.k_fixed_mul, False ],\n [ 'fixeddiv', True, 2, 2, b_opcode.k_fixed_div, False ],\n [ 'setgravity', False, 1, 1, b_opcode.k_set_gravity, False ],\n [ 'setaircontrol', False, 1, 1, b_opcode.k_set_air_control, False ],\n [ 'clearinventory', False, 0, 0, b_opcode.k_clear_inventory, False ],\n [ 'giveinventory', False, 2, 2, b_opcode.k_give_inventory, False ],\n [ 'takeinventory', False, 2, 2, b_opcode.k_take_inventory, False ],\n [ 'checkinventory', True, 1, 1, b_opcode.k_check_inventory, False ],\n [ 'spawn', True, 6, 4, b_opcode.k_spawn, False ],\n [ 'spawnspot', True, 4, 2, b_opcode.k_spawn_spot, False ],\n [ 'setmusic', False, 3, 1, b_opcode.k_set_music, False ],\n [ 'localsetmusic', False, 3, 1, b_opcode.k_local_set_music, False ],\n [ 'setfont', False, 1, 1, b_opcode.k_set_font, False ],\n [ 'setthingspecial', False, 7, 2, b_opcode.k_set_thing_special, False ],\n [ 'fadeto', False, 5, 5, b_opcode.k_fade_to, False ],\n [ 'faderange', False, 9, 9, b_opcode.k_fade_range, False ],\n [ 'cancelfade', False, 0, 0, b_opcode.k_cancel_fade, False ],\n [ 'playmovie', True, 1, 1, b_opcode.k_play_movie, False ],\n [ 'setfloortrigger', False, 8, 3, b_opcode.k_set_floor_trigger, False ],\n [ 'setceilingtrigger', False, 8, 3, b_opcode.k_set_ceiling_trigger,\n False ],\n [ 'getactorx', True, 1, 1, b_opcode.k_get_actor_x, False ],\n [ 'getactory', True, 1, 1, b_opcode.k_get_actor_y, False ],\n [ 'getactorz', True, 1, 1, b_opcode.k_get_actor_z, False ],\n [ 'sin', True, 1, 1, b_opcode.k_sin, False ],\n [ 'cos', True, 1, 1, b_opcode.k_cos, False ],\n [ 'vectorangle', True, 2, 2, b_opcode.k_vector_angle, False ],\n [ 'checkweapon', True, 1, 1, b_opcode.k_check_weapon, False ],\n [ 'setweapon', True, 1, 1, b_opcode.k_set_weapon, False ],\n [ 'setmarineweapon', False, 2, 2, b_opcode.k_set_marine_weapon, False ],\n [ 'setactorproperty', False, 3, 3, b_opcode.k_set_actor_property, False ],\n [ 'getactorproperty', True, 2, 2, b_opcode.k_get_actor_property, False ],\n [ 'playernumber', True, 0, 0, b_opcode.k_player_number, False ],\n [ 'activatortid', True, 0, 0, b_opcode.k_activator_tid, False ],\n [ 'setmarinesprite', False, 2, 2, b_opcode.k_set_marine_sprite, False ],\n [ 'getscreenwidth', True, 0, 0, b_opcode.k_get_screen_width, False ],\n [ 'getscreenheight', True, 0, 0, b_opcode.k_get_screen_height, False ],\n [ 'thing_projectile2', False, 7, 7, b_opcode.k_thing_projectile2, False ],\n [ 'strlen', True, 1, 1, b_opcode.k_strlen, False ],\n [ 'sethudsize', False, 3, 3, b_opcode.k_set_hud_size, False ],\n [ 'getcvar', True, 1, 1, b_opcode.k_get_cvar, False ],\n [ 'setresultvalue', False, 1, 1, b_opcode.k_set_result_value, False ],\n [ 'getlinerowoffset', True, 0, 0, b_opcode.k_get_line_row_offset, False ],\n [ 'getactorfloorz', True, 1, 1, b_opcode.k_get_actor_floor_z, False ],\n [ 'getactorangle', True, 1, 1, b_opcode.k_get_actor_angle, False ],\n [ 'getsectorfloorz', True, 3, 3, b_opcode.k_get_sector_floor_z, False ],\n [ 'getsectorceilingz', True, 3, 3, b_opcode.k_get_sector_ceiling_z,\n False ],\n [ 'getsigilpieces', True, 0, 0, b_opcode.k_get_sigil_pieces, False ],\n [ 'getlevelinfo', True, 1, 1, b_opcode.k_get_level_info, False ],\n [ 'changesky', False, 2, 2, b_opcode.k_change_sky, False ],\n [ 'playeringame', True, 1, 1, b_opcode.k_player_in_game, False ],\n [ 'playerisbot', True, 1, 1, b_opcode.k_player_is_bot, False ],\n [ 'setcameratotexture', False, 3, 3, b_opcode.k_set_camera_to_texture,\n False ],\n [ 'getammocapacity', True, 1, 1, b_opcode.k_get_ammo_capacity, False ],\n [ 'setammocapacity', False, 2, 2, b_opcode.k_set_ammo_capacity, False ],\n [ 'setactorangle', False, 2, 2, b_opcode.k_set_actor_angle, False ],\n [ 'spawnprojectile', False, 7, 7, b_opcode.k_spawn_projectile, False ],\n [ 'getsectorlightlevel', True, 1, 1, b_opcode.k_get_sector_light_level,\n False ],\n [ 'getactorceilingz', True, 1, 1, b_opcode.k_get_actor_ceiling_z, False ],\n [ 'clearactorinventory', False, 1, 1, b_opcode.k_clear_actor_inventory,\n False ],\n [ 'giveactorinventory', False, 3, 3, b_opcode.k_give_actor_inventory,\n False ],\n [ 'takeactorinventory', False, 3, 3, b_opcode.k_take_actor_inventory,\n False ],\n [ 'checkactorinventory', True, 2, 2, b_opcode.k_check_actor_inventory,\n False ],\n [ 'thingcountname', True, 2, 2, b_opcode.k_thing_count_name, False ],\n [ 'spawnspotfacing', True, 3, 2, b_opcode.k_spawn_spot_facing, False ],\n [ 'playerclass', True, 1, 1, b_opcode.k_player_class, False ],\n [ 'getplayerinfo', True, 2, 2, b_opcode.k_get_player_info, False ],\n [ 'changelevel', False, 4, 3, b_opcode.k_change_level, False ],\n [ 'sectordamage', False, 5, 5, b_opcode.k_sector_damage, False ],\n [ 'replacetextures', False, 3, 2, b_opcode.k_replace_textures, False ],\n [ 'getactorpitch', True, 1, 1, b_opcode.k_get_actor_pitch, False ],\n [ 'setactorpitch', False, 2, 2, b_opcode.k_set_actor_pitch, False ],\n [ 'setactorstate', True, 3, 2, b_opcode.k_set_actor_state, False ],\n [ 'thing_damage2', True, 3, 3, b_opcode.k_thing_damage2, False ],\n [ 'useinventory', True, 1, 1, b_opcode.k_use_inventory, False ],\n [ 'useactorinventory', True, 2, 2, b_opcode.k_use_actor_inventory,\n False ],\n [ 'checkactorceilingtexture', True, 2, 2,\n b_opcode.k_check_actor_ceiling_texture, False ],\n [ 'checkactorfloortexture', True, 2, 2,\n b_opcode.k_check_actor_floor_texture, False ],\n [ 'getactorlightlevel', True, 1, 1, b_opcode.k_get_actor_light_level,\n False ],\n [ 'setmugshotstate', False, 1, 1, b_opcode.k_set_mugshot_state, False ],\n [ 'thingcountsector', True, 3, 3, b_opcode.k_thing_count_sector, False ],\n [ 'thingcountnamesector', True, 3, 3, b_opcode.k_thing_count_name_sector,\n False ],\n [ 'checkplayercamera', True, 1, 1, b_opcode.k_check_player_camera,\n False ],\n [ 'morphactor', True, 7, 7, b_opcode.k_morph_actor, False ],\n [ 'unmorphactor', True, 2, 1, b_opcode.k_unmorph_actor, False ],\n [ 'getplayerinput', True, 2, 2, b_opcode.k_get_player_input, False ],\n [ 'classifyactor', True, 1, 1, b_opcode.k_classify_actor, False ] ]\n for template in ded:\n name, value, max_param, min_param, opcode, latent = template\n func = common.func_t()\n func.type = common.FUNC_DED\n func.value = value\n func.min_param = min_param\n func.max_param = max_param\n func.impl = {\n 'opcode' : opcode,\n 'latent' : latent }\n front.scope.names[ name ] = func\n format = [\n # Format: name / returns-value / parameter-count / parameter-minimum /\n # terminating-opcode.\n # Note: The format items together count as a single parameter.\n [ 'print', False, 1, 1, b_opcode.k_end_print ],\n [ 'printbold', False, 1, 1, b_opcode.k_end_print_bold ],\n [ 'hudmessage', False, 9, 7, b_opcode.k_end_hud_message ],\n [ 'hudmessagebold', False, 9, 7, b_opcode.k_end_hud_message_bold ],\n [ 'log', False, 1, 1, b_opcode.k_end_log ],\n [ 'strparam', True, 1, 1, b_opcode.k_save_string ] ]\n for template in format:\n name, value, max_param, min_param, opcode = template\n func = common.func_t()\n func.type = common.FUNC_FORMAT\n func.value = value\n func.min_param = min_param\n func.max_param = max_param\n func.impl = {\n 'opcode' : opcode }\n front.scope.names[ name ] = func", "repo_name": "positively-charged/bcc-python", "sub_path": "src/f_dec.py", "file_name": "f_dec.py", "file_ext": "py", "file_size_in_byte": 27855, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "24", "api": [{"api_name": "f_token.T_INT", "line_number": 23, "usage_type": "attribute"}, {"api_name": "f_token.T_STR", "line_number": 23, "usage_type": "attribute"}, {"api_name": "f_token.T_BOOL", "line_number": 23, "usage_type": "attribute"}, {"api_name": "f_token.T_VOID", "line_number": 23, "usage_type": "attribute"}, {"api_name": "f_token.T_FUNCTION", "line_number": 24, "usage_type": "attribute"}, {"api_name": "f_token.T_WORLD", "line_number": 24, "usage_type": "attribute"}, {"api_name": "f_token.T_GLOBAL", "line_number": 24, "usage_type": "attribute"}, {"api_name": "f_token.T_STATIC", "line_number": 24, "usage_type": "attribute"}, {"api_name": "common.STORAGE_LOCAL", "line_number": 30, "usage_type": "attribute"}, {"api_name": "f_token.T_FUNCTION", "line_number": 41, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 42, "usage_type": "call"}, {"api_name": "f_token.T_GLOBAL", "line_number": 47, "usage_type": "attribute"}, {"api_name": "common.STORAGE_GLOBAL", "line_number": 48, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 51, "usage_type": "call"}, {"api_name": "f_token.T_WORLD", "line_number": 52, "usage_type": "attribute"}, {"api_name": "common.STORAGE_WORLD", "line_number": 53, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 56, "usage_type": "call"}, {"api_name": "f_token.T_STATIC", "line_number": 57, "usage_type": "attribute"}, {"api_name": "common.STORAGE_MAP", "line_number": 58, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 61, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 61, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 61, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 61, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 61, "usage_type": "name"}, {"api_name": "f_main.diag", "line_number": 64, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 64, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 64, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 64, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 64, "usage_type": "name"}, {"api_name": "f_main.diag", "line_number": 67, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 67, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 67, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 67, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 67, "usage_type": "name"}, {"api_name": "f_token.read", "line_number": 69, "usage_type": "call"}, {"api_name": "common.STORAGE_MAP", "line_number": 71, "usage_type": "attribute"}, {"api_name": "f_token.T_INT", "line_number": 75, "usage_type": "attribute"}, {"api_name": "f_token.T_STR", "line_number": 75, "usage_type": "attribute"}, {"api_name": "f_token.T_BOOL", "line_number": 75, "usage_type": "attribute"}, {"api_name": "f_token.T_INT", "line_number": 78, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 79, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 79, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 79, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 79, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 79, "usage_type": "name"}, {"api_name": "f_token.read", "line_number": 82, "usage_type": "call"}, {"api_name": "f_token.T_VOID", "line_number": 83, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 84, "usage_type": "call"}, {"api_name": "f_main.diag", "line_number": 86, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 86, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 86, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 86, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 86, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 89, "usage_type": "call"}, {"api_name": "f_token.T_LIT_DECIMAL", "line_number": 94, "usage_type": "attribute"}, {"api_name": "f_expr.read_literal", "line_number": 96, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 97, "usage_type": "call"}, {"api_name": "f_token.T_COLON", "line_number": 97, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 98, "usage_type": "call"}, {"api_name": "common.STORAGE_WORLD", "line_number": 100, "usage_type": "attribute"}, {"api_name": "common.STORAGE_GLOBAL", "line_number": 101, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 104, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 104, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 104, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 104, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 104, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 107, "usage_type": "call"}, {"api_name": "f_main.diag", "line_number": 109, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 109, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 109, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 109, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 109, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 112, "usage_type": "call"}, {"api_name": "common.STORAGE_WORLD", "line_number": 115, "usage_type": "attribute"}, {"api_name": "common.STORAGE_GLOBAL", "line_number": 116, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 117, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 117, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 117, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 117, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 117, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 120, "usage_type": "call"}, {"api_name": "f_token.T_ID", "line_number": 123, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 129, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 129, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 129, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 129, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 129, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 131, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 134, "usage_type": "call"}, {"api_name": "f_token.T_PAREN_L", "line_number": 134, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 135, "usage_type": "call"}, {"api_name": "f_main.add_scope", "line_number": 136, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 138, "usage_type": "call"}, {"api_name": "f_token.T_PAREN_R", "line_number": 138, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 139, "usage_type": "call"}, {"api_name": "common.func_t", "line_number": 140, "usage_type": "call"}, {"api_name": "common.block_t", "line_number": 147, "usage_type": "call"}, {"api_name": "f_stmt.read_block", "line_number": 150, "usage_type": "call"}, {"api_name": "f_main.pop_scope", "line_number": 151, "usage_type": "call"}, {"api_name": "f_token.T_BRACKET_L", "line_number": 165, "usage_type": "attribute"}, {"api_name": "f_token.T_COMMA", "line_number": 177, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 178, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 180, "usage_type": "call"}, {"api_name": "f_token.T_SEMICOLON", "line_number": 180, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 181, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 185, "usage_type": "call"}, {"api_name": "f_token.T_ID", "line_number": 185, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 187, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 187, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 187, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 187, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 187, "usage_type": "name"}, {"api_name": "common.NODE_VAR", "line_number": 191, "usage_type": "attribute"}, {"api_name": "common.NODE_CONSTANT", "line_number": 192, "usage_type": "attribute"}, {"api_name": "common.NODE_FUNC", "line_number": 194, "usage_type": "attribute"}, {"api_name": "common.FUNC_DED", "line_number": 197, "usage_type": "attribute"}, {"api_name": "common.FUNC_FORMAT", "line_number": 198, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 201, "usage_type": "call"}, {"api_name": "f_main.DIAG_FILE", "line_number": 201, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 201, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 201, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 203, "usage_type": "call"}, {"api_name": "f_token.read", "line_number": 206, "usage_type": "call"}, {"api_name": "common.STORAGE_LOCAL", "line_number": 211, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 212, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 212, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 212, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 212, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 212, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 214, "usage_type": "call"}, {"api_name": "f_token.T_BRACKET_L", "line_number": 215, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 217, "usage_type": "call"}, {"api_name": "f_token.T_BRACKET_R", "line_number": 220, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 223, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 223, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 223, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 223, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 223, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 225, "usage_type": "call"}, {"api_name": "f_token.read", "line_number": 226, "usage_type": "call"}, {"api_name": "f_expr.read", "line_number": 228, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 229, "usage_type": "call"}, {"api_name": "f_token.T_BRACKET_R", "line_number": 229, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 230, "usage_type": "call"}, {"api_name": "f_main.diag", "line_number": 232, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 232, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 232, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 232, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 232, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 234, "usage_type": "call"}, {"api_name": "f_main.diag", "line_number": 236, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 236, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 236, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 236, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 236, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 238, "usage_type": "call"}, {"api_name": "common.dim_t", "line_number": 239, "usage_type": "call"}, {"api_name": "f_token.T_ASSIGN", "line_number": 255, "usage_type": "attribute"}, {"api_name": "common.STORAGE_WORLD", "line_number": 257, "usage_type": "attribute"}, {"api_name": "common.STORAGE_GLOBAL", "line_number": 258, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 260, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 260, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 260, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 260, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 260, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 262, "usage_type": "call"}, {"api_name": "common.STORAGE_WORLD", "line_number": 266, "usage_type": "attribute"}, {"api_name": "common.STORAGE_GLOBAL", "line_number": 267, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 269, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 269, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 269, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 269, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 269, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 272, "usage_type": "call"}, {"api_name": "f_token.read", "line_number": 273, "usage_type": "call"}, {"api_name": "f_token.T_BRACE_L", "line_number": 274, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 278, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 278, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 278, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 278, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 278, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 280, "usage_type": "call"}, {"api_name": "f_expr.read", "line_number": 281, "usage_type": "call"}, {"api_name": "common.STORAGE_MAP", "line_number": 282, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 283, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 283, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 283, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 283, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 283, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 285, "usage_type": "call"}, {"api_name": "common.initial_t", "line_number": 286, "usage_type": "call"}, {"api_name": "common.var_t", "line_number": 291, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 323, "usage_type": "call"}, {"api_name": "f_token.T_SCRIPT", "line_number": 323, "usage_type": "attribute"}, {"api_name": "common.script_t", "line_number": 324, "usage_type": "call"}, {"api_name": "f_token.read", "line_number": 326, "usage_type": "call"}, {"api_name": "f_token.T_SHIFT_L", "line_number": 329, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 330, "usage_type": "call"}, {"api_name": "f_token.T_LIT_DECIMAL", "line_number": 332, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 334, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 335, "usage_type": "call"}, {"api_name": "f_token.T_SHIFT_R", "line_number": 335, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 336, "usage_type": "call"}, {"api_name": "f_main.diag", "line_number": 338, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 338, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 338, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 338, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 338, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 340, "usage_type": "call"}, {"api_name": "f_expr.read", "line_number": 343, "usage_type": "call"}, {"api_name": "f_main.diag", "line_number": 346, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 346, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 346, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 346, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 346, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 348, "usage_type": "call"}, {"api_name": "f_main.diag", "line_number": 350, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 350, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 350, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 350, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 350, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 353, "usage_type": "call"}, {"api_name": "f_main.diag", "line_number": 355, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 355, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 355, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 355, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 355, "usage_type": "name"}, {"api_name": "f_main.bail", "line_number": 357, "usage_type": "call"}, {"api_name": "f_main.diag", "line_number": 364, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 364, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 364, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 364, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 364, "usage_type": "name"}, {"api_name": "f_main.diag", "line_number": 366, "usage_type": "call"}, {"api_name": "f_main.DIAG_FILE", "line_number": 366, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 366, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 366, "usage_type": "name"}, {"api_name": "f_main.add_scope", "line_number": 369, "usage_type": "call"}, {"api_name": "f_token.T_PAREN_L", "line_number": 372, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 374, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 376, "usage_type": "call"}, {"api_name": "f_token.T_PAREN_R", "line_number": 376, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 377, "usage_type": "call"}, {"api_name": "f_token.T_OPEN", "line_number": 380, "usage_type": "attribute"}, {"api_name": "f_token.T_RESPAWN", "line_number": 381, "usage_type": "attribute"}, {"api_name": "f_token.T_DEATH", "line_number": 382, "usage_type": "attribute"}, {"api_name": "f_token.T_ENTER", "line_number": 383, "usage_type": "attribute"}, {"api_name": "f_token.T_PICKUP", "line_number": 384, "usage_type": "attribute"}, {"api_name": "f_token.T_BLUE_RETURN", "line_number": 385, "usage_type": "attribute"}, {"api_name": "f_token.T_RED_RETURN", "line_number": 386, "usage_type": "attribute"}, {"api_name": "f_token.T_WHITE_RETURN", "line_number": 387, "usage_type": "attribute"}, {"api_name": "f_token.T_LIGHTNING", "line_number": 388, "usage_type": "attribute"}, {"api_name": "f_token.T_DISCONNECT", "line_number": 389, "usage_type": "attribute"}, {"api_name": "f_token.T_UNLOADING", "line_number": 390, "usage_type": "attribute"}, {"api_name": "f_token.T_RETURN", "line_number": 391, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_TYPE_OPEN", "line_number": 380, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_TYPE_RESPAWN", "line_number": 381, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_TYPE_DEATH", "line_number": 382, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_TYPE_ENTER", "line_number": 383, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_TYPE_PICKUP", "line_number": 384, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_TYPE_BLUE_RETURN", "line_number": 385, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_TYPE_RED_RETURN", "line_number": 386, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_TYPE_WHITE_RETURN", "line_number": 387, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_TYPE_LIGHTNING", "line_number": 388, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_TYPE_DISCONNECT", "line_number": 389, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_TYPE_UNLOADING", "line_number": 390, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_TYPE_RETURN", "line_number": 391, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 396, "usage_type": "call"}, {"api_name": "common.SCRIPT_TYPE_CLOSED", "line_number": 397, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 399, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 399, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 399, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 399, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 399, "usage_type": "name"}, {"api_name": "common.SCRIPT_TYPE_DISCONNECT", "line_number": 402, "usage_type": "attribute"}, {"api_name": "f_main.diag", "line_number": 406, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 406, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 406, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 406, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 406, "usage_type": "name"}, {"api_name": "f_main.diag", "line_number": 411, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 411, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 411, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 411, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 411, "usage_type": "name"}, {"api_name": "common.SCRIPT_FLAG_NET", "line_number": 415, "usage_type": "attribute"}, {"api_name": "f_token.T_NET", "line_number": 416, "usage_type": "attribute"}, {"api_name": "f_token.T_CLIENTSIDE", "line_number": 417, "usage_type": "attribute"}, {"api_name": "common.SCRIPT_FLAG_CLIENTSIDE", "line_number": 418, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 423, "usage_type": "call"}, {"api_name": "f_main.diag", "line_number": 425, "usage_type": "call"}, {"api_name": "f_main.DIAG_ERR", "line_number": 425, "usage_type": "name"}, {"api_name": "f_main.DIAG_FILE", "line_number": 425, "usage_type": "name"}, {"api_name": "f_main.DIAG_LINE", "line_number": 425, "usage_type": "name"}, {"api_name": "f_main.DIAG_COLUMN", "line_number": 425, "usage_type": "name"}, {"api_name": "f_token.read", "line_number": 427, "usage_type": "call"}, {"api_name": "common.block_t", "line_number": 429, "usage_type": "call"}, {"api_name": "f_stmt.read_block", "line_number": 432, "usage_type": "call"}, {"api_name": "f_main.pop_scope", "line_number": 435, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 439, "usage_type": "call"}, {"api_name": "f_token.T_SPECIAL", "line_number": 439, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 440, "usage_type": "call"}, {"api_name": "f_token.T_MINUS", "line_number": 443, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 444, "usage_type": "call"}, {"api_name": "f_expr.read_literal", "line_number": 446, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 447, "usage_type": "call"}, {"api_name": "f_token.T_COLON", "line_number": 447, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 448, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 451, "usage_type": "call"}, {"api_name": "f_token.T_PAREN_L", "line_number": 451, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 452, "usage_type": "call"}, {"api_name": "f_expr.read_literal", "line_number": 454, "usage_type": "call"}, {"api_name": "f_token.T_COMMA", "line_number": 455, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 457, "usage_type": "call"}, {"api_name": "f_expr.read_literal", "line_number": 458, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 459, "usage_type": "call"}, {"api_name": "f_token.T_PAREN_R", "line_number": 459, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 460, "usage_type": "call"}, {"api_name": "common.func_t", "line_number": 461, "usage_type": "call"}, {"api_name": "common.FUNC_EXT", "line_number": 468, "usage_type": "attribute"}, {"api_name": "common.FUNC_ASPEC", "line_number": 471, "usage_type": "attribute"}, {"api_name": "f_token.T_COMMA", "line_number": 473, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 474, "usage_type": "call"}, {"api_name": "f_token.test", "line_number": 476, "usage_type": "call"}, {"api_name": "f_token.T_SEMICOLON", "line_number": 476, "usage_type": "attribute"}, {"api_name": "f_token.read", "line_number": 477, "usage_type": "call"}, {"api_name": "b_opcode.k_delay", "line_number": 484, "usage_type": "attribute"}, {"api_name": "b_opcode.k_random", "line_number": 485, "usage_type": "attribute"}, {"api_name": "b_opcode.k_thing_count", "line_number": 486, "usage_type": "attribute"}, {"api_name": "b_opcode.k_tag_wait", "line_number": 487, "usage_type": "attribute"}, {"api_name": "b_opcode.k_poly_wait", "line_number": 488, "usage_type": "attribute"}, {"api_name": "b_opcode.k_change_floor", "line_number": 489, "usage_type": "attribute"}, {"api_name": "b_opcode.k_change_ceiling", "line_number": 490, "usage_type": "attribute"}, {"api_name": "b_opcode.k_line_side", "line_number": 491, "usage_type": "attribute"}, {"api_name": "b_opcode.k_script_wait", "line_number": 492, "usage_type": "attribute"}, {"api_name": "b_opcode.k_clear_line_special", "line_number": 493, "usage_type": "attribute"}, {"api_name": "b_opcode.k_player_count", "line_number": 494, "usage_type": "attribute"}, {"api_name": "b_opcode.k_game_type", "line_number": 495, "usage_type": "attribute"}, {"api_name": "b_opcode.k_game_skill", "line_number": 496, "usage_type": "attribute"}, {"api_name": "b_opcode.k_timer", "line_number": 497, "usage_type": "attribute"}, {"api_name": "b_opcode.k_sector_sound", "line_number": 498, "usage_type": "attribute"}, {"api_name": "b_opcode.k_ambient_sound", "line_number": 499, "usage_type": "attribute"}, {"api_name": "b_opcode.k_sound_sequence", "line_number": 500, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_line_texture", "line_number": 501, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_line_blocking", "line_number": 502, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_line_special", "line_number": 503, "usage_type": "attribute"}, {"api_name": "b_opcode.k_thing_sound", "line_number": 504, "usage_type": "attribute"}, {"api_name": "b_opcode.k_activator_sound", "line_number": 505, "usage_type": "attribute"}, {"api_name": "b_opcode.k_local_ambient_sound", "line_number": 506, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_line_monster_blocking", "line_number": 509, "usage_type": "attribute"}, {"api_name": "b_opcode.k_is_multiplayer", "line_number": 510, "usage_type": "attribute"}, {"api_name": "b_opcode.k_player_team", "line_number": 511, "usage_type": "attribute"}, {"api_name": "b_opcode.k_player_health", "line_number": 512, "usage_type": "attribute"}, {"api_name": "b_opcode.k_player_armor_points", "line_number": 513, "usage_type": "attribute"}, {"api_name": "b_opcode.k_player_frags", "line_number": 515, "usage_type": "attribute"}, {"api_name": "b_opcode.k_blue_team_count", "line_number": 516, "usage_type": "attribute"}, {"api_name": "b_opcode.k_blue_team_count", "line_number": 517, "usage_type": "attribute"}, {"api_name": "b_opcode.k_red_team_count", "line_number": 518, "usage_type": "attribute"}, {"api_name": "b_opcode.k_red_team_count", "line_number": 519, "usage_type": "attribute"}, {"api_name": "b_opcode.k_blue_team_score", "line_number": 520, "usage_type": "attribute"}, {"api_name": "b_opcode.k_blue_team_score", "line_number": 521, "usage_type": "attribute"}, {"api_name": "b_opcode.k_red_team_score", "line_number": 522, "usage_type": "attribute"}, {"api_name": "b_opcode.k_red_team_score", "line_number": 523, "usage_type": "attribute"}, {"api_name": "b_opcode.k_is_one_flag_ctf", "line_number": 524, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_invasion_wave", "line_number": 525, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_invasion_state", "line_number": 526, "usage_type": "attribute"}, {"api_name": "b_opcode.k_music_change", "line_number": 527, "usage_type": "attribute"}, {"api_name": "b_opcode.k_console_command", "line_number": 528, "usage_type": "attribute"}, {"api_name": "b_opcode.k_single_player", "line_number": 529, "usage_type": "attribute"}, {"api_name": "b_opcode.k_fixed_mul", "line_number": 530, "usage_type": "attribute"}, {"api_name": "b_opcode.k_fixed_div", "line_number": 531, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_gravity", "line_number": 532, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_air_control", "line_number": 533, "usage_type": "attribute"}, {"api_name": "b_opcode.k_clear_inventory", "line_number": 534, "usage_type": "attribute"}, {"api_name": "b_opcode.k_give_inventory", "line_number": 535, "usage_type": "attribute"}, {"api_name": "b_opcode.k_take_inventory", "line_number": 536, "usage_type": "attribute"}, {"api_name": "b_opcode.k_check_inventory", "line_number": 537, "usage_type": "attribute"}, {"api_name": "b_opcode.k_spawn", "line_number": 538, "usage_type": "attribute"}, {"api_name": "b_opcode.k_spawn_spot", "line_number": 539, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_music", "line_number": 540, "usage_type": "attribute"}, {"api_name": "b_opcode.k_local_set_music", "line_number": 541, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_font", "line_number": 542, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_thing_special", "line_number": 543, "usage_type": "attribute"}, {"api_name": "b_opcode.k_fade_to", "line_number": 544, "usage_type": "attribute"}, {"api_name": "b_opcode.k_fade_range", "line_number": 545, "usage_type": "attribute"}, {"api_name": "b_opcode.k_cancel_fade", "line_number": 546, "usage_type": "attribute"}, {"api_name": "b_opcode.k_play_movie", "line_number": 547, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_floor_trigger", "line_number": 548, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_ceiling_trigger", "line_number": 549, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_actor_x", "line_number": 551, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_actor_y", "line_number": 552, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_actor_z", "line_number": 553, "usage_type": "attribute"}, {"api_name": "b_opcode.k_sin", "line_number": 554, "usage_type": "attribute"}, {"api_name": "b_opcode.k_cos", "line_number": 555, "usage_type": "attribute"}, {"api_name": "b_opcode.k_vector_angle", "line_number": 556, "usage_type": "attribute"}, {"api_name": "b_opcode.k_check_weapon", "line_number": 557, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_weapon", "line_number": 558, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_marine_weapon", "line_number": 559, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_actor_property", "line_number": 560, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_actor_property", "line_number": 561, "usage_type": "attribute"}, {"api_name": "b_opcode.k_player_number", "line_number": 562, "usage_type": "attribute"}, {"api_name": "b_opcode.k_activator_tid", "line_number": 563, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_marine_sprite", "line_number": 564, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_screen_width", "line_number": 565, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_screen_height", "line_number": 566, "usage_type": "attribute"}, {"api_name": "b_opcode.k_thing_projectile2", "line_number": 567, "usage_type": "attribute"}, {"api_name": "b_opcode.k_strlen", "line_number": 568, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_hud_size", "line_number": 569, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_cvar", "line_number": 570, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_result_value", "line_number": 571, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_line_row_offset", "line_number": 572, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_actor_floor_z", "line_number": 573, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_actor_angle", "line_number": 574, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_sector_floor_z", "line_number": 575, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_sector_ceiling_z", "line_number": 576, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_sigil_pieces", "line_number": 578, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_level_info", "line_number": 579, "usage_type": "attribute"}, {"api_name": "b_opcode.k_change_sky", "line_number": 580, "usage_type": "attribute"}, {"api_name": "b_opcode.k_player_in_game", "line_number": 581, "usage_type": "attribute"}, {"api_name": "b_opcode.k_player_is_bot", "line_number": 582, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_camera_to_texture", "line_number": 583, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_ammo_capacity", "line_number": 585, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_ammo_capacity", "line_number": 586, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_actor_angle", "line_number": 587, "usage_type": "attribute"}, {"api_name": "b_opcode.k_spawn_projectile", "line_number": 588, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_sector_light_level", "line_number": 589, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_actor_ceiling_z", "line_number": 591, "usage_type": "attribute"}, {"api_name": "b_opcode.k_clear_actor_inventory", "line_number": 592, "usage_type": "attribute"}, {"api_name": "b_opcode.k_give_actor_inventory", "line_number": 594, "usage_type": "attribute"}, {"api_name": "b_opcode.k_take_actor_inventory", "line_number": 596, "usage_type": "attribute"}, {"api_name": "b_opcode.k_check_actor_inventory", "line_number": 598, "usage_type": "attribute"}, {"api_name": "b_opcode.k_thing_count_name", "line_number": 600, "usage_type": "attribute"}, {"api_name": "b_opcode.k_spawn_spot_facing", "line_number": 601, "usage_type": "attribute"}, {"api_name": "b_opcode.k_player_class", "line_number": 602, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_player_info", "line_number": 603, "usage_type": "attribute"}, {"api_name": "b_opcode.k_change_level", "line_number": 604, "usage_type": "attribute"}, {"api_name": "b_opcode.k_sector_damage", "line_number": 605, "usage_type": "attribute"}, {"api_name": "b_opcode.k_replace_textures", "line_number": 606, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_actor_pitch", "line_number": 607, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_actor_pitch", "line_number": 608, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_actor_state", "line_number": 609, "usage_type": "attribute"}, {"api_name": "b_opcode.k_thing_damage2", "line_number": 610, "usage_type": "attribute"}, {"api_name": "b_opcode.k_use_inventory", "line_number": 611, "usage_type": "attribute"}, {"api_name": "b_opcode.k_use_actor_inventory", "line_number": 612, "usage_type": "attribute"}, {"api_name": "b_opcode.k_check_actor_ceiling_texture", "line_number": 615, "usage_type": "attribute"}, {"api_name": "b_opcode.k_check_actor_floor_texture", "line_number": 617, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_actor_light_level", "line_number": 618, "usage_type": "attribute"}, {"api_name": "b_opcode.k_set_mugshot_state", "line_number": 620, "usage_type": "attribute"}, {"api_name": "b_opcode.k_thing_count_sector", "line_number": 621, "usage_type": "attribute"}, {"api_name": "b_opcode.k_thing_count_name_sector", "line_number": 622, "usage_type": "attribute"}, {"api_name": "b_opcode.k_check_player_camera", "line_number": 624, "usage_type": "attribute"}, {"api_name": "b_opcode.k_morph_actor", "line_number": 626, "usage_type": "attribute"}, {"api_name": "b_opcode.k_unmorph_actor", "line_number": 627, "usage_type": "attribute"}, {"api_name": "b_opcode.k_get_player_input", "line_number": 628, "usage_type": "attribute"}, {"api_name": "b_opcode.k_classify_actor", "line_number": 629, "usage_type": "attribute"}, {"api_name": "common.func_t", "line_number": 632, "usage_type": "call"}, {"api_name": "common.FUNC_DED", "line_number": 633, "usage_type": "attribute"}, {"api_name": "b_opcode.k_end_print", "line_number": 645, "usage_type": "attribute"}, {"api_name": "b_opcode.k_end_print_bold", "line_number": 646, "usage_type": "attribute"}, {"api_name": "b_opcode.k_end_hud_message", "line_number": 647, "usage_type": "attribute"}, {"api_name": "b_opcode.k_end_hud_message_bold", "line_number": 648, "usage_type": "attribute"}, {"api_name": "b_opcode.k_end_log", "line_number": 649, "usage_type": "attribute"}, {"api_name": "b_opcode.k_save_string", "line_number": 650, "usage_type": "attribute"}, {"api_name": "common.func_t", "line_number": 653, "usage_type": "call"}, {"api_name": "common.FUNC_FORMAT", "line_number": 654, "usage_type": "attribute"}]}
+{"seq_id": "2622303921", "text": "import pickle\nimport os\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport easygui as e\n\n\n# exports a single scan as a txt file\ndef print_to_file(name, mat):\n y = []\n path = \"C://Users//David//PycharmProjects//CalKit//exports\"\n filename = (name + \"_export.txt\")\n fullpath = os.path.join(path, filename)\n file = open(fullpath, \"w+\")\n # delimiter is set to space, this could be changed later\n file.writelines(\"labels\" + \", \" + name + \"\\n\")\n # generates the data for the file\n end = int(mat.x_end)\n start = int(mat.x_start)\n step = int(mat.step)\n steps = int((end - start) / step) + 1\n x = np.linspace(start, end, num=steps)\n for i in range(len(mat.spline)):\n spl = mat.spline[i]\n y.append(spl(x))\n # prints the data in procedurally\n for i in range(len(x)):\n file.write(str(x[i]) + \",\")\n for j in range(len(y)):\n temp = str(y[j][i])\n file.write(temp + \",\")\n file.write(\"\\n\")\n\n\n# saves to the .pkl file, functions via append\ndef save_file(library, ck, name):\n # adds kit to library\n if type(name) != str:\n ck.name = name.get()\n library.library = ck\n ck.print() # for testing\n dump_ck_list(library) # calls the library to dump to the pickle file\n for i in library.get_ck_list(): # testing function to validate what was printed\n i.print()\n\n\n# dumps the library to the .pkl\ndef dump_ck_list(self):\n with open(\"cklib.pkl\", \"wb\") as openfile:\n pickle.dump(self.library, openfile)\n\n\ndef plot_ck(mat, plt):\n if len(mat.spline) != 0:\n plt.clear()\n for i in range(len(mat.spline)):\n x = []\n y = []\n date = \"no date loaded\"\n if mat.date is not None:\n date = mat.date\n text_title = str(\"scan created date: \" + str(date))\n end = int(mat.x_end)\n start = int(mat.x_start)\n step = mat.step\n steps = int((end - start) / step) # determines the number of data-points\n x = np.linspace(mat.x_start, mat.x_end,\n num=steps) # extrapolates the X axis based on the start, stop and number of data-points\n try:\n spl = mat.spline[i]\n y = spl(x) # gets the Y axis data\n plt.set(xlabel=\"Wavelength, nm\", ylabel=\"Intensity, a.u.\", title=text_title)\n plt.plot(x, y) # plots it\n except TypeError:\n error_message(\"This calibration kit does not have a scan of the selected type stored\")\n plt.grid()\n else:\n error_message(\"This calibration kit does not have a scan of the selected type stored\")\n\n\ndef error_message(text):\n e.msgbox(text)\n\n\ndef plot_scans(scans):\n for scan in scans:\n plt.plot(scan[0][0], scan[0][1], label=scan[1])\n plt.legend()\n plt.show()\n", "repo_name": "dGakamsky/CalKit", "sub_path": "outputer.py", "file_name": "outputer.py", "file_ext": "py", "file_size_in_byte": 2882, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 22, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clear", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.set", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "easygui.msgbox", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}]}
+{"seq_id": "4060743827", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Apr 7 14:28:12 2020\n\n@author: MAYANK\n\"\"\"\nimport pandas as pd\nimport mysql.connector\nimport datetime\nfrom datetime import timedelta\nfrom sklearn.linear_model import LinearRegression\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport time\ndatabase = mysql.connector.connect(host=\"localhost\", database=\"medical\", \n passwd=\"test\", user = 'root')\n\ndb = database.cursor()\nwhile True:\n db.execute(\"select max(s_no) from bill\")\n s_no = db.fetchall()\n print(s_no)\n while True:\n database = mysql.connector.connect(host=\"localhost\", database=\"medical\", \n passwd=\"test\", user = 'root')\n db = database.cursor()\n time.sleep(2)\n db = database.cursor()\n db.execute(\"select max(s_no) from bill\")\n s_no_temp = db.fetchall()\n print(s_no_temp)\n if(s_no_temp != s_no):\n break\n \n db.execute(\"select medicine_name,category from bill where bill_no = \"+\n \"(select bill_no from bill where s_no = (select max(s_no) from bill))\")\n data = db.fetchall()\n print(data, \"This is data\")\n print(len(data))\n regressor = LinearRegression()\n \n while len(data)>0:\n print(\"here\")\n nameandcategory = data.pop()\n med_name, category = nameandcategory[0], nameandcategory[1]\n startdate= datetime.date.today() + timedelta(days=-60)\n \n db.execute(\"select sum(bill_quantity),date from bill where medicine_name='\"\n +med_name+\"'and category='\"+category+\"' and date>='\"\n +str(startdate)+\"' group by date \")\n \n \n df = pd.DataFrame(db.fetchall())\n print(df)\n if(df.empty): continue\n parsedate = startdate\n while parsedate<=datetime.date.today():\n if parsedate not in df[1].values:\n df=df.append({0:0, 1:parsedate}, ignore_index=True)\n parsedate += timedelta(days=1)\n \n df=df.sort_values(1)\n \n samples = [10, 7, 3]\n need = []\n while len(samples)>0:\n \n df3=df[df[1]>datetime.date.today()+timedelta(days=-samples.pop())]\n \n regressor.fit(pd.to_datetime(df3[1]).values.reshape(-1,1), df3[0])\n \n today = datetime.date.today()\n upcoming_dates = pd.to_datetime(np.asarray([today, today+timedelta(days=1), \n today+timedelta(days=2),\n today+timedelta(days=3), \n today+timedelta(days=4), \n today+timedelta(days=5), \n today+timedelta(days=6), \n today+timedelta(days=7),\n today+timedelta(days=8),\n today+timedelta(days=9),\n today+timedelta(days=10)]))\n \n \n upcoming_dates_prediction = regressor.predict(upcoming_dates.values.astype(float).reshape(-1,1))\n plt.figure(figsize=(10,6))\n plt.plot(df3[1], df3[0], color='red')\n \n \n plt.plot(df3[1], regressor.predict(pd.to_datetime(df3[1]).values.astype(float).reshape(-1,1))) \n plt.plot(upcoming_dates, upcoming_dates_prediction)\n plt.xticks(df3[1], rotation=\"70\")\n \n \n upcoming_dates_prediction = [max(x,0) for x in upcoming_dates_prediction]\n \n need.append(sum(upcoming_dates_prediction))\n print(max(need), \" need\")\n \n db.execute(\"Select sum(total_quantity) from medicine where name = '\"+\n med_name+\"' and category='\"+category+\"'\") \n quantity_left = db.fetchall()[0][0]\n print(quantity_left)\n if(quantity_left 0 and self.solvent is None:\n nextdir = \"tddft\"\n self.calc_all_excited_states(namelist,prevdir,nextdir,self.wrapper.excitations,\n self.calc_params,self.nroots,charges=self.charges)\n # Vibrational frequency calculations, if requested and no higher\n # level of theory will follow later\n if self.vibrations and self.solvent is None:\n nextdir = \"freq\"\n self.calc_vib_freq(namelist,prevdir,nextdir,self.wrapper.freq,self.calc_params,charges=self.charges)\n # Solvated calculations\n if self.solvent is not None:\n nextdir = \"is_opt_\"+self.solvstr(self.solvent)\n if self.vibrations:\n driver_tol = 'tight'\n # Geometry Optimisation\n self.geom_opt_all(namelist,prevdir,nextdir,self.wrapper.geom_opt,\n self.calc_params,driver_tol,solvent=self.solvent,\n charges=self.charges)\n prevdir = nextdir\n # TDDFT calculations\n if self.nroots > 0:\n nextdir = \"is_tddft_\"+self.solvstr(self.solvent)\n self.calc_all_excited_states(namelist,prevdir,nextdir,self.wrapper.excitations,\n self.calc_params,nroots=self.nroots,\n solvent=self.solvent,charges=self.charges,\n plot_homo=self.plot_homo,plot_lumo=self.plot_lumo,\n plot_trans_den=self.plot_trans_den)\n # Vibrational frequency calculations, if requested\n if self.vibrations:\n nextdir = \"is_freq_\"+self.solvstr(self.solvent)\n self.calc_vib_freq(namelist,prevdir,nextdir,self.wrapper.freq,\n self.calc_params,solvent=self.solvent,\n charges=self.charges)\n\n from os import makedirs, path\n\n # Download xyz files if they do not already exist\n def get_xyz_files(self,namelist,out_path):\n \"\"\"\n Downloads initial geometries from the NCI's webserver cactus, based on their IUPAC names.\n\n These geometries are usually not great, but are a reasonable starting point for optimisation.\n\n Visit https://cactus.nci.nih.gov/chemical/structure to see what works and check your names before use.\n\n Uses ``wget``, so if the machine you are using does not have access to this command,\n this routine will fail, in which case put starting point geometries in the directory\n 'xyz'.\n\n namelist: dict\n Keys are shortnames (eg \"cate\"), entries are full names (eg \"catechol\" or \"1,2-dihydroxybenzene\")\n out_path: str\n String for directory name where xyz files will be written (eg \"xyz\"). Created if not present\n \"\"\"\n from os import path, makedirs\n import subprocess\n\n # Download sdf file from cactus.nci.nih.gov\n if not path.exists(out_path):\n makedirs(out_path)\n for seed in namelist:\n # TODO: use urllib here\n wget_str = (\"wget -O \\\"\"+out_path+\"/\" + seed + \".xyz\\\" \" +\n \"\\\"https://cactus.nci.nih.gov/chemical/structure/\" + \n namelist[seed] + \"/file?format=xyz\\\"\")\n if not path.exists(out_path+\"/\"+seed+\".xyz\"):\n print(wget_str)\n errorcode = subprocess.call(wget_str, shell=True)\n if errorcode:\n raise RuntimeError('{} returned an error: {}'\n .format('wget', errorcode))\n else:\n print(\"Skipping download: \"+out_path+\"/\"+seed+\".xyz already exists\")\n\n # strip out long names and convert to list\n shortnames = [x for x in namelist]\n return shortnames\n\n from os import path, makedirs, getcwd, chdir\n from ase.io import read, write\n\n def solvstr(self,solvent):\n if isinstance(solvent,str):\n return solvent\n if isinstance(solvent,dict):\n return solvent['solvent']\n\n # Optimize geometries of solute selection\n def geom_opt_all(self,solute_names,in_path,out_path,geom_opt_func,calc_params,\n driver_tol='default',solvent=None,charges={}):\n \"\"\"\n Geometry optimise all of a list of solutes\n\n solute_names: list of str\n Short names of the solutes to be optimised\n in_path: str\n Directory where .xyz files are expected to be found. Any not present are skipped.\n out_path: str\n Directory where optimised structure .xyz files are written. Created if not present.\n geom_opt_func: function\n A function wrapping creation of an ASE calculator and using it to perform geometry optimisation.\n calc_params: dict\n Contents varies between different wrappers, but generally specifies basis, functional etc\n driver_tol: str\n Geometry optimisation tolerance level (eg in NWChem)\n target: int\n Excited state index, or None for ground state\n solvent: str\n Implicit solvent name, or None for gas-phase\n charges: dict\n Keys are strings corresponding to some or all of the entries in solute names, entries are net charges on each molecule\n \"\"\"\n from ase.io import read, write\n from os import path, makedirs, getcwd, chdir\n\n # Make directory for optimised structures\n if not path.exists(out_path):\n makedirs(out_path)\n sol_str = ''\n target = calc_params['target']\n if solvent is not None:\n sol_str = f'in {self.solvstr(solvent)} solvent '\n\n for seed in solute_names:\n if target is not None and target!=0:\n baseseed = seed\n seed = seed+\"_es\"+str(target)\n if seed in charges:\n charge = charges[seed]\n else:\n charge = 0\n infile = in_path+\"/\"+seed+\".xyz\"\n outfile = out_path+\"/\"+seed+\".xyz\"\n if not path.exists(infile):\n # Try basename without _esX\n if target is not None and target !=0:\n infile = in_path+\"/\"+baseseed+\".xyz\"\n if not path.exists(infile):\n print(f'Skipping geometry optimisation {sol_str}'+\n f' for: {seed} - no input file')\n continue\n solute_opt = read(infile)\n if path.exists(outfile):\n print(f'Skipping geometry optimisation {sol_str}'+\n f' for: {seed} - output file already present')\n continue\n print(f'Geometry optimization {sol_str}for: {seed}')\n label = seed\n origdir = getcwd()\n wdir = f'geom/{seed}'\n if not path.exists(wdir):\n makedirs(wdir)\n chdir(wdir)\n if solvent is not None:\n label = label+\"_\"+self.solvstr(solvent)\n try:\n geom_opt_func(solute_opt,label,calc_params,driver_tol,solvent,charge)\n except KeyboardInterrupt:\n raise Exception('Keyboard Interrupt')\n except SyntaxError:\n raise Exception('Syntax Error')\n\n chdir(origdir)\n print('Writing to ',outfile)\n if '' in solute_opt.info:\n del solute_opt.info['']\n write(outfile,solute_opt)\n\n # Attempt to find best rotamer for each solute\n\n def find_best_rotas(self,solute_names,in_path,out_path,singlepoint_func,\n geom_opt_func,calc_params,solvent=None,charges={}):\n \"\"\"\n Finds the lowest energy rotamer for each of a list of solutes.\n Proceeds by identifying -OH groups attached to C-C units, and tries 'flipping' the dihedral, then optimising\n the resulting geometry if it within a certain tolerance of the original energy. If any lower energy structure\n is found, this will be returned instead of the original one.\n\n solute_names: list of str\n Short names of the solutes to be tested\n in_path: str\n Directory where .xyz files are expected to be found. Any not present are skipped.\n out_path: str\n Directory where best rotamer structure .xyz files are written. Created if not present.\n singlepoint_func: function\n A function wrapping creation of an ASE calculator and using it to perform a singlepoint calculation.\n geom_opt_func: function\n A function wrapping creation of an ASE calculator and using it to perform geometry optimisation.\n calc_params: dict\n Contents varies between different wrappers, but generally specifies basis, functional etc\n solvent: str\n Implicit solvent name, or None for gas-phase\n charges: dict\n Keys are strings corresponding to some or all of the entries in solute names, entries are net charges on each molecule\n \"\"\"\n from os import path, makedirs\n from ase.io import read, write\n\n # Hard-coded logic for what constitutes a rotatable bond\n # Works OK for -OH groups in organic compounds but will need editing\n # for anything else. Assumes anything within 1.5A is a bond.\n rota_elem = ['H','O','C','C']\n rota_max_dist = [1.5,1.5,1.5]\n rota_dih_range = 40\n rota_opt_thresh = 0.2\n\n # Make directory for optimised structures\n if not path.exists(out_path):\n makedirs(out_path)\n target = calc_params['target']\n\n for seed in solute_names:\n if seed in charges:\n charge = charges[seed]\n else:\n charge = 0\n if target is not None and target!=0:\n seed = seed+\"_es\"+str(target)\n outfile = out_path+'/'+seed+'.xyz'\n infile = in_path+\"/\"+seed+\".xyz\"\n if not path.exists(infile):\n print('Skipping Rotamer Search for: ',seed,\n ' - input structure not found')\n continue\n if path.exists(outfile):\n print('Skipping Rotamer Search for: ',seed,\n ' - output file already present')\n continue\n\n print('Finding best rotamer for: ',seed)\n sol_opt = read(infile)\n # Load defaults from module\n elem = rota_elem\n max_dist = rota_max_dist\n dih_range = rota_dih_range\n opt_thresh = rota_opt_thresh\n nrot = 0\n ijkl = []\n sym = sol_opt.get_chemical_symbols()\n for i in range(len(sol_opt)):\n if sym[i]==elem[0]:\n for j in range(len(sol_opt)):\n if sym[j]==elem[1] and sol_opt.get_distance(i,j) < max_dist[0]:\n for k in range(len(sol_opt)):\n if sym[k]==elem[2] and sol_opt.get_distance(j,k) < max_dist[1]:\n for l in range(len(sol_opt)):\n if l!=k and sym[l]==elem[3] and sol_opt.get_distance(k,l) < max_dist[2]:\n dih = sol_opt.get_dihedral(l,k,j,i)\n if dih>180-dih_range and dih<180+dih_range:\n ijkl.append([i,j,k,l])\n nrot = nrot + 1\n break\n if nrot==0:\n print('No rotatable OH bonds found')\n if '' in sol_opt.info:\n del sol_opt.info['']\n write(outfile,sol_opt)\n else:\n print(nrot, 'rotatable bonds found, generating all',2**nrot,\n 'rotamers')\n origdir = getcwd()\n wdir = f'{out_path}/{seed}'\n if not path.exists(wdir):\n makedirs(wdir)\n chdir(wdir)\n flip = [0 for i in range(len(ijkl))]\n sol_opt_rota = []\n energy_rota = []\n for rota in range(2**len(ijkl)):\n sol_opt_rota.append(sol_opt.copy())\n flip = [(rota&(2**(len(ijkl)-i-1)))>>(len(ijkl)-i-1) for i in range(len(ijkl))]\n for oH in range(len(ijkl)):\n i = ijkl[oH][0]; j = ijkl[oH][1]; k = ijkl[oH][2]; l = ijkl[oH][3];\n if flip[oH]:\n sol_opt_rota[rota].set_dihedral(l,k,j,i,0)\n label = 'rota'+repr(rota).zfill(3)\n driver_tol = 'loose'\n opt = 0\n try:\n energy,_ = singlepoint_func(sol_opt_rota[rota],label,calc_params,charge)\n if rota==0:\n energy_rota.append(energy)\n if rota>0:\n if energy pivot]\n quick_compare_count += len(greater)\n return quick_sort(less) + [pivot] + quick_sort(greater)\n\n\ndef QuickSort(l): # 封装一下排序和输出性能指标\n global quick_compare_count\n quick_compare_count = 0 # 重置全局变量\n result = quick_sort(l)\n return quick_compare_count\n\n\n##归并排序\nmerge_compare_count = 0\n\n\ndef merge(li, low, mid, high):\n global merge_compare_count\n # 列表,最开始的值,中间值(第一个有序列表的最后一位),最后面的值\n\n # 将两个有序列表的开头标记出来\n i = low\n j = mid + 1 # 第二段有序数列的开头\n list_1 = []\n while i <= mid and j <= high: # 限制条件(开头小于结尾)两边都有数\n merge_compare_count+=1\n if li[i] < li[j]:\n list_1.append(li[i])\n i += 1\n else:\n list_1.append(li[j])\n j += 1\n # while执行完成,,肯定有一部分没数了\n while i <= mid: # 左列表还有数\n list_1.append(li[i])\n i += 1\n while j <= high: # 右列表还有数\n list_1.append(li[j])\n j += 1\n # 再将list_1里的数放回li中\n li[low:high + 1] = list_1\n\n\ndef merge_sort(li, low, high):\n global merge_compare_count\n # 终止条件 只有一个元素\n if low < high: # 至少有两个,递归终止条件(只剩一个的时候)\n mid = (low + high) // 2 # 二分查找中间值\n merge_sort(li, low, mid) # 递归左边,左边排序\n\n merge_sort(li, mid + 1, high) # 递归右边,右边排序\n\n merge(li, low, mid, high)\n\n\n\ndef MergeSort(l=list):\n global merge_compare_count\n merge_compare_count = 0\n result = merge_sort(l,0,len(l)-1)\n return merge_compare_count\n\n\n\n##图像绘制\nx_axis_data = [10, 100, 1000, 2000, 5000, 10000]\ny_axis_data1 = [BubbleSort(l1), BubbleSort(l2), BubbleSort(l3), BubbleSort(l4), BubbleSort(l5), BubbleSort(l6)]\ny_axis_data2 = [MergeSort(l1), MergeSort(l2), MergeSort(l3), MergeSort(l4), MergeSort(l5), MergeSort(l6)]\ny_axis_data3 = [QuickSort(l1), QuickSort(l2), QuickSort(l3), QuickSort(l4), QuickSort(l5), QuickSort(l6)]\n\n# 画图\nplt.plot(x_axis_data, y_axis_data1, 'b*--', alpha=0.5, linewidth=1, label='BubbleSort')\nplt.plot(x_axis_data, y_axis_data2, 'rs--', alpha=0.5, linewidth=1, label='MergeSort')\nplt.plot(x_axis_data, y_axis_data3, 'go--', alpha=0.5, linewidth=1, label='QuickSort')\n\nplt.legend() # 显示上面的label\nplt.xlabel('Data Size')\nplt.ylabel('Number of Comparisons')\n\n# plt.ylim(-1,1)#仅设置y轴坐标范围\nplt.show()\n", "repo_name": "weixing18/Homework", "sub_path": "分治算法实验/Code.py", "file_name": "Code.py", "file_ext": "py", "file_size_in_byte": 3539, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 1, "dataset": "github-code", "pt": "25", "api": [{"api_name": "sys.setrecursionlimit", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 8, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 13, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}]}
+{"seq_id": "34292757989", "text": "# _*_coding:utf-8_*_\nimport json\n\nfrom django.http import HttpResponse\nfrom django.shortcuts import render\nfrom django.views.generic.base import View\nfrom .models import Course,Lesson,Video\nfrom organization.models import CourseOrg\nfrom pure_pagination import Paginator, EmptyPage, PageNotAnInteger\nfrom operation.models import UserFavorite,CourseComment,UserCourse\nfrom utils.mixin_untils import LoginRequiredMixin\n# Create your views here.\n\n\nclass CourseList(View):\n def get(self,request):\n\n all_course = Course.objects.all()\n hot_course = all_course.order_by('-click_nums')[:5]\n # 按照最新,最热门,参与人数来排序,\n sort = request.GET.get('sort','')\n if sort:\n if sort == 'new':\n all_course = all_course.order_by('-add_time')\n if sort == 'hot':\n all_course = all_course.order_by('-click_nums')\n if sort == 'students':\n all_course = all_course.order_by('-students')\n\n try:\n page = request.GET.get('page', 1)\n except PageNotAnInteger:\n page = 1\n p = Paginator(all_course,3)\n course = p.page(page)\n return render(request,'course-list.html',\n {'all_course':course,\n 'hot_course':hot_course,\n })\n\n\nclass CourseDetail(View):\n def get(self,request,course_id):\n coursedetail = Course.objects.get(id=int(course_id))\n coursedetail.click_nums += 1\n coursedetail.save()\n lesson = coursedetail.lesson_set.all()\n # 获取对应id的机构对象\n org = CourseOrg.objects.get(id = int(course_id))\n # 机构老师对象\n org_teacher = org.teacher_set.all()\n #机构教师数量\n org_teacher_num = org_teacher.count()\n # 课程标签,用于做相关推荐\n tag = coursedetail.tag\n if tag:\n # 这是要展示的对象,标签和浏览的课程相同的话就会被推荐,这个可以自己在后台定义\n relate_course = Course.objects.filter(tag=tag)[:1]\n else:\n # 因为定了初始值是空,这里不设置空的就是空列表的话,html如果接收到空字符串,无法进行遍历\n relate_course = []\n # 课程收藏功能\n has_fav_c = False\n has_fav_o = False\n if request.user.is_authenticated():\n if UserFavorite.objects.filter(user=request.user, fav_id=coursedetail.id, fav_type=1):\n has_fav_c = True\n if UserFavorite.objects.filter(user=request.user, fav_id=coursedetail.courseorg.id, fav_type=2):\n has_fav_o = True\n return render(request,'course-detail.html',\n {'course_detail':coursedetail,\n 'lesson':lesson,\n 'org':org,\n 'org_teacher_num':org_teacher_num,\n 'relate_course':relate_course,\n 'has_fav_c': has_fav_c,\n 'has_fav_o': has_fav_o,\n })\n\n# 具体课程章节视频详情\n# @login_required\nclass CourseVideo(LoginRequiredMixin,View):\n def get(self,request,course_id):\n # 对应id课程的章节对象\n course_video = Course.objects.get(id=int(course_id))\n course_video.students += 1\n course_video.save()\n\n course_lesson = course_video.lesson_set.all()\n # 资源下载对象\n courseresource = course_video.courseresource_set.all()\n\n #查询用户是否关联了该课程\n user_course2 = UserCourse.objects.filter(user=request.user,course=course_video)\n if not user_course2:\n user_course3 = UserCourse(user=request.user,course=course_video)\n user_course3.save()\n # 筛选出用户的所有课程\n user_courses = UserCourse.objects.filter(course=course_video)\n # 用列表式取出遍历出所有学习用户的id\n user_ids = [user_course.user.id for user_course in user_courses]\n #获取所有课程\n all_user_courses = UserCourse.objects.filter(user_id__in=user_ids)\n # 取出所有课程id\n course_ids = [all_user_course.course.id for all_user_course in all_user_courses]\n # 获取学过该用户学过的其他课程\n relate_courses = Course.objects.filter(id__in=course_ids).order_by('-click_nums')[:5]\n return render(request,'course-video.html',\n {'course_video':course_video,\n 'course_lesson':course_lesson,\n 'courseresource':courseresource,\n 'relate_courses':relate_courses,\n })\n\n# 课程评论区2\nclass CourseComment2(LoginRequiredMixin,View):\n def get(self,request,course_id):\n course_comment = Course.objects.get(id=int(course_id))\n courseresource = course_comment.courseresource_set.all()\n comments = course_comment.coursecomment_set.all()\n return render(request,'course-comment.html',\n {'course_comment':course_comment,\n 'courseresource':courseresource,\n 'comments': comments,\n })\n\n# 添加评论功能2\nclass AddComment(View):\n def post(self,request):\n # 先监测是否登录\n if not request.user.is_authenticated():\n return HttpResponse(json.dumps({'status':'fail','msg':'用户未登录'}),content_type='application/json')\n\n course_id = request.POST.get('course_id',0)\n comments = request.POST.get('comments','')\n\n # 如果id和评论存在获取到了,把各项内容插入到数据库中\n if int(course_id) >0 and comments:\n course_comments = CourseComment()\n # 课程必须是对应id\n course = Course.objects.get(id=int(course_id))\n course_comments.course = course\n course_comments.comments = comments\n course_comments.user = request.user\n course_comments.save()\n return HttpResponse(json.dumps({'status':'success','msg':'发表评论成功'}),content_type='application/json')\n else:\n return HttpResponse(json.dumps({'status':'fail','msg':'评论失败'}),content_type='application/json')\n\n\nclass Video2(View):\n # 视频播放页面\n def get(self, request, video_id):\n # 对应id课程的章节对象\n video = Video.objects.get(id=int(video_id))\n course_video = video.lesson.course\n course_video.students += 1\n course_video.save()\n\n course_lesson = course_video.lesson_set.all()\n # 资源下载对象\n courseresource = course_video.courseresource_set.all()\n\n # 查询用户是否关联了该课程\n user_course2 = UserCourse.objects.filter(user=request.user, course=course_video)\n if not user_course2:\n user_course3 = UserCourse(user=request.user, course=course_video)\n user_course3.save()\n # 筛选出用户的所有课程\n user_courses = UserCourse.objects.filter(course=course_video)\n # 用列表式取出遍历出所有学习用户的id\n user_ids = [user_course.user.id for user_course in user_courses]\n # 获取所有课程\n all_user_courses = UserCourse.objects.filter(user_id__in=user_ids)\n # 取出所有课程id\n course_ids = [all_user_course.course.id for all_user_course in all_user_courses]\n # 获取学过该用户学过的其他课程\n relate_courses = Course.objects.filter(id__in=course_ids).order_by('-click_nums')[:5]\n return render(request, 'course-play.html',\n {'course_video': course_video,\n 'course_lesson': course_lesson,\n 'courseresource': courseresource,\n 'relate_courses': relate_courses,\n 'video' : course_video,\n 'videoplay': video,\n })\n\n\n", "repo_name": "wnbaed/selfproject", "sub_path": "MxOnline/apps/courses/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 8055, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "django.views.generic.base.View", "line_number": 15, "usage_type": "name"}, {"api_name": "models.Course.objects.all", "line_number": 18, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 18, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 18, "usage_type": "name"}, {"api_name": "pure_pagination.PageNotAnInteger", "line_number": 32, "usage_type": "name"}, {"api_name": "pure_pagination.Paginator", "line_number": 34, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 36, "usage_type": "call"}, {"api_name": "django.views.generic.base.View", "line_number": 42, "usage_type": "name"}, {"api_name": "models.Course.objects.get", "line_number": 44, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 44, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 44, "usage_type": "name"}, {"api_name": "organization.models.CourseOrg.objects.get", "line_number": 49, "usage_type": "call"}, {"api_name": "organization.models.CourseOrg.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "organization.models.CourseOrg", "line_number": 49, "usage_type": "name"}, {"api_name": "models.Course.objects.filter", "line_number": 58, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 58, "usage_type": "name"}, {"api_name": "operation.models.UserFavorite.objects.filter", "line_number": 66, "usage_type": "call"}, {"api_name": "operation.models.UserFavorite.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "operation.models.UserFavorite", "line_number": 66, "usage_type": "name"}, {"api_name": "operation.models.UserFavorite.objects.filter", "line_number": 68, "usage_type": "call"}, {"api_name": "operation.models.UserFavorite.objects", "line_number": 68, "usage_type": "attribute"}, {"api_name": "operation.models.UserFavorite", "line_number": 68, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 70, "usage_type": "call"}, {"api_name": "utils.mixin_untils.LoginRequiredMixin", "line_number": 82, "usage_type": "name"}, {"api_name": "django.views.generic.base.View", "line_number": 82, "usage_type": "name"}, {"api_name": "models.Course.objects.get", "line_number": 85, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 85, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 85, "usage_type": "name"}, {"api_name": "operation.models.UserCourse.objects.filter", "line_number": 94, "usage_type": "call"}, {"api_name": "operation.models.UserCourse.objects", "line_number": 94, "usage_type": "attribute"}, {"api_name": "operation.models.UserCourse", "line_number": 94, "usage_type": "name"}, {"api_name": "operation.models.UserCourse", "line_number": 96, "usage_type": "call"}, {"api_name": "operation.models.UserCourse.objects.filter", "line_number": 99, "usage_type": "call"}, {"api_name": "operation.models.UserCourse.objects", "line_number": 99, "usage_type": "attribute"}, {"api_name": "operation.models.UserCourse", "line_number": 99, "usage_type": "name"}, {"api_name": "operation.models.UserCourse.objects.filter", "line_number": 103, "usage_type": "call"}, {"api_name": "operation.models.UserCourse.objects", "line_number": 103, "usage_type": "attribute"}, {"api_name": "operation.models.UserCourse", "line_number": 103, "usage_type": "name"}, {"api_name": "models.Course.objects.filter", "line_number": 107, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 107, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 107, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 108, "usage_type": "call"}, {"api_name": "utils.mixin_untils.LoginRequiredMixin", "line_number": 116, "usage_type": "name"}, {"api_name": "django.views.generic.base.View", "line_number": 116, "usage_type": "name"}, {"api_name": "models.Course.objects.get", "line_number": 118, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 118, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 118, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 121, "usage_type": "call"}, {"api_name": "django.views.generic.base.View", "line_number": 128, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 132, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 132, "usage_type": "call"}, {"api_name": "operation.models.CourseComment", "line_number": 139, "usage_type": "call"}, {"api_name": "models.Course.objects.get", "line_number": 141, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 141, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 141, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 146, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 146, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 148, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 148, "usage_type": "call"}, {"api_name": "django.views.generic.base.View", "line_number": 151, "usage_type": "name"}, {"api_name": "models.Video.objects.get", "line_number": 155, "usage_type": "call"}, {"api_name": "models.Video.objects", "line_number": 155, "usage_type": "attribute"}, {"api_name": "models.Video", "line_number": 155, "usage_type": "name"}, {"api_name": "operation.models.UserCourse.objects.filter", "line_number": 165, "usage_type": "call"}, {"api_name": "operation.models.UserCourse.objects", "line_number": 165, "usage_type": "attribute"}, {"api_name": "operation.models.UserCourse", "line_number": 165, "usage_type": "name"}, {"api_name": "operation.models.UserCourse", "line_number": 167, "usage_type": "call"}, {"api_name": "operation.models.UserCourse.objects.filter", "line_number": 170, "usage_type": "call"}, {"api_name": "operation.models.UserCourse.objects", "line_number": 170, "usage_type": "attribute"}, {"api_name": "operation.models.UserCourse", "line_number": 170, "usage_type": "name"}, {"api_name": "operation.models.UserCourse.objects.filter", "line_number": 174, "usage_type": "call"}, {"api_name": "operation.models.UserCourse.objects", "line_number": 174, "usage_type": "attribute"}, {"api_name": "operation.models.UserCourse", "line_number": 174, "usage_type": "name"}, {"api_name": "models.Course.objects.filter", "line_number": 178, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 178, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 178, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 179, "usage_type": "call"}]}
+{"seq_id": "25118929842", "text": "import mongomock\nfrom datetime import datetime\nfrom aggregation_builder import AggregationQueryBuilder\nfrom aggregation_builder.operators import *\nimport unittest\n\n\ndef ISODate(str):\n return datetime.strptime(str, '%Y-%m-%dT%H:%M:%SZ')\n\n\nclass QueryDictTests(unittest.TestCase):\n def test_limit(self):\n query = [\n {\n '$limit': 5\n }\n ]\n generated_query = AggregationQueryBuilder().limit(5).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_skip(self):\n query = [\n {\n '$skip': 5\n }\n ]\n generated_query = AggregationQueryBuilder().skip(5).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_project(self):\n query = [{'$project': {'title': 1, 'author': 1}}]\n generated_query = AggregationQueryBuilder().project(title=1, author=1).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_project_computed_fields(self):\n query = [\n {\n '$project': {\n 'title': 1,\n 'isbn': {\n 'prefix': {'$substr': [\"$isbn\", 0, 3]},\n 'group': {'$substr': [\"$isbn\", 3, 2]},\n 'publisher': {'$substr': [\"$isbn\", 5, 4]},\n 'title': {'$substr': [\"$isbn\", 9, 3]},\n 'checkDigit': {'$substr': [\"$isbn\", 12, 1]}\n },\n 'lastName': \"$author.last\",\n 'copiesSold': \"$copies\"\n }\n }\n ]\n generated_query = AggregationQueryBuilder().project(\n title=1,\n isbn=dict(\n prefix=SUB_STR(\"$isbn\", 0, 3),\n group=SUB_STR(\"$isbn\", 3, 2),\n publisher=SUB_STR(\"$isbn\", 5, 4),\n title=SUB_STR(\"$isbn\", 9, 3),\n checkDigit=SUB_STR(\"$isbn\", 12, 1)\n ),\n lastName=\"$author.last\",\n copiesSold=\"$copies\"\n\n ).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_match(self):\n query = [\n {\n '$match': {'author': \"dave\"}\n }\n ]\n generated_query = AggregationQueryBuilder().match(\n author=\"dave\"\n ).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_match_count(self):\n query = [\n {'$match': {'$or': [{'$gt': [\"$score\", 70]}, {'$gte': [\"$views\", 1000]}]}},\n {'$group': {'_id': None, 'count': {'$sum': 1}}}\n ]\n generated_query = AggregationQueryBuilder().match(\n OR(GT('$score', 70), GTE('$views', 1000))\n ).group(id=None, count=SUM(1)).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_unwind(self):\n query = [{'$unwind': \"$sizes\"}]\n generated_query = AggregationQueryBuilder().unwind(\n \"$sizes\"\n ).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_unwind_with_params(self):\n query = [{'$unwind': \"$sizes\", 'preserveNullAndEmptyArrays': True, 'includeArrayIndex': \"arrayIndex\"}]\n generated_query = AggregationQueryBuilder().unwind(\n path=\"$sizes\",\n include_array_index=\"arrayIndex\",\n preserve_null_and_empty_arrays=True\n ).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_sort(self):\n query = [\n {'$sort': {'age': -1, 'posts': 1}}\n ]\n generated_query = AggregationQueryBuilder().sort(age=-1, posts=1).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_sort_metadata(self):\n query = [\n {'$match': {'$text': {'$search': \"operating\"}}},\n {'$sort': {'score': {'$meta': \"textScore\"}, 'posts': -1}}\n ]\n generated_query = AggregationQueryBuilder().match(\n TEXT_SEARCH(\"operating\")\n ).sort(\n score=TEXT_META,\n posts=-1\n ).get_query()\n self.assertListEqual(generated_query, query)\n\n def test_sample(self):\n query = [{'$sample': {'size': 3}}]\n generated_query = AggregationQueryBuilder().sample(size=3).get_query()\n self.assertListEqual(generated_query, query)\n\n def test_lookup(self):\n query = [\n {\n '$lookup':\n {\n 'from': \"inventory\",\n 'localField': \"item\",\n 'foreignField': \"sku\",\n 'as': \"inventory_docs\"\n }\n }\n ]\n\n generated_query = AggregationQueryBuilder().look_up(\n _from='inventory',\n _localField='item',\n _foreignField='sku',\n _as='inventory_docs'\n ).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_lookup_with_array(self):\n query = [\n {\n '$unwind': \"$specs\"\n },\n {\n '$lookup':\n {\n 'from': \"inventory\",\n 'localField': \"specs\",\n 'foreignField': \"size\",\n 'as': \"inventory_docs\"\n }\n },\n {\n '$match': {\"inventory_docs\": {'$ne': []}}\n }\n ]\n\n generated_query = AggregationQueryBuilder().unwind(\n \"$specs\"\n ).look_up(\n _from='inventory',\n _localField='specs',\n _foreignField='size',\n _as='inventory_docs'\n ).match(\n inventory_docs=NE()\n ).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_graph_look_up(self):\n query = [\n {\n '$graphLookup': {\n 'from': \"employees\",\n 'startWith': \"$reportsTo\",\n 'connectFromField': \"reportsTo\",\n 'connectToField': \"name\",\n 'as': \"reportingHierarchy\"\n }\n }\n ]\n generated_query = AggregationQueryBuilder().graph_look_up(\n _from='employees',\n _startWith=\"$reportsTo\",\n _connectFromField=\"reportsTo\",\n _connectToField=\"name\",\n _as=\"reportingHierarchy\"\n ).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_add_fields(self):\n query = [\n {\n '$addFields': {\n 'totalHomework': {'$sum': \"$homework\"},\n 'totalQuiz': {'$sum': \"$quiz\"}\n }\n },\n {\n '$addFields': {'totalScore':\n {'$add': [\"$totalHomework\", \"$totalQuiz\", \"$extraCredit\"]}}\n }\n ]\n generated_query = AggregationQueryBuilder().add_fields(\n totalHomework=SUM(\"$homework\"),\n totalQuiz=SUM(\"$quiz\")\n ).add_fields(\n totalScore=ADD(\"$totalHomework\", \"$totalQuiz\", \"$extraCredit\")\n ).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_group(self):\n query = [\n {\n '$group': {\n '_id': {'month': {'$month': \"$date\"}, 'day': {'$dayOfMonth': \"$date\"}, 'year': {'$year': \"$date\"}},\n 'totalPrice': {'$sum': {'$multiply': [\"$price\", \"$quantity\"]}},\n 'averageQuantity': {'$avg': \"$quantity\"},\n 'count': {'$sum': 1}\n }\n }\n ]\n\n generated_query = AggregationQueryBuilder().group(\n id=dict(\n month=MONTH(\"$date\"),\n day=DAY_OF_MONTH(\"$date\"),\n year=YEAR(\"$date\")\n ),\n totalPrice=SUM(MULTIPLY(\"$price\", \"$quantity\")),\n averageQuantity=AVG(\"$quantity\"),\n count=SUM(1)\n ).get_query()\n\n self.assertListEqual(generated_query, query)\n\n def test_null_group(self):\n query = [\n {\n '$group': {\n '_id': None,\n 'totalPrice': {'$sum': {'$multiply': [\"$price\", \"$quantity\"]}},\n 'averageQuantity': {'$avg': \"$quantity\"},\n 'count': {'$sum': 1}\n }\n }\n ]\n\n generated_query = AggregationQueryBuilder().group(\n id=None,\n totalPrice=SUM(MULTIPLY(\"$price\", \"$quantity\")),\n averageQuantity=AVG(\"$quantity\"),\n count=SUM(1)\n ).get_query()\n\n self.assertListEqual(generated_query, query)\n\n\nif __name__ == '__main__':\n unittest.main()\n", "repo_name": "MosesSymeonidis/aggregation_builder", "sub_path": "tests/query_builder.py", "file_name": "query_builder.py", "file_ext": "py", "file_size_in_byte": 8823, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 9, "dataset": "github-code", "pt": "25", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 9, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 12, "usage_type": "attribute"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 19, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 29, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 35, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 56, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 78, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 89, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 97, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 105, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 117, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 126, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 136, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 152, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 180, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 205, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 228, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 249, "usage_type": "call"}, {"api_name": "aggregation_builder.AggregationQueryBuilder", "line_number": 274, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 285, "usage_type": "call"}]}
+{"seq_id": "12622774415", "text": "from chess.figures.figure import Figure\n\n\nclass Pawn(Figure):\n def __init__(self, *args, **kwargs):\n super(Pawn, self).__init__(*args, **kwargs)\n self.mega_step = False\n\n def check_move(self, cell_to):\n from_x, from_y = self.cell.coordinate\n to_x, to_y = cell_to.coordinate\n\n if from_x == to_x and from_y == to_y - 1:\n return True\n\n if from_x == to_x and from_y == to_y - 2 and from_y == 2:\n return True\n\n if (from_x == to_x + 1 or from_x == to_x - 1) and from_y == to_y - 1 and \\\n cell_to.figure and cell_to.figure.user != self.user:\n return True\n\n return False\n", "repo_name": "MykhailoKlimchuk/chess_prototype", "sub_path": "figures/pawn.py", "file_name": "pawn.py", "file_ext": "py", "file_size_in_byte": 669, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "chess.figures.figure.Figure", "line_number": 4, "usage_type": "name"}]}
+{"seq_id": "28676150964", "text": "import urlopen\nimport requests \nfrom bs4 import BeautifulSoup\n\nimport csv\nimport pandas as pd\n\nimport regex as re\n\ncomplete_data, invalid_data = [], []\n\n\ndef preprocess_text(text):\n \n text = re.sub(r'[^0-9a-zA-Z]', ' ', text)\n text = re.sub(r'_', ' ', text)\n return \" \" .join(text.split())\n \n\ndef get_contents(data):\n \n modify_data = {}\n\n for name, url in data.items(): \n\n # Read the homepage of the teacher \n try:\n html = requests.get(url)\n html.raise_for_status()\n modify_data = {} \n soup = BeautifulSoup(html.text, 'html.parser')\n \n # Get phone number / email\n get_info, personal_info = ['Email', 'email'], []\n for string in soup.body.strings:\n if 'Email' in string or 'email' in string:\n personal_info.append(string.strip())\n \n # Save only relevant content for the teacher\n text_data = [preprocess_text(string) for string in soup.body.strings if preprocess_text(string)!='']\n text_data = ' '.join(text_data)\n\n # Add the data to a dict\n modify_data['name'] = name\n modify_data['personal_info'] = personal_info\n modify_data['content'] = text_data \n complete_data.append(modify_data)\n\n except:\n invalid_data.append((name, url))\n\n text_file = open(\"data/error_file.txt\",\"w\")\n for val in invalid_data:\n text_file.write(str(val) + \"\\n\")\n\n text_file = open(\"data/output.txt\",\"w\")\n text_file.write(str({'complete_data':complete_data}))\n\n\ndef main_csv():\n # Read csv file \n csv_file = pd.read_csv('data/csv_data/csrankings.csv') # ['name', 'affiliation', 'homepage', 'scholarid']\n data = dict(zip(csv_file.name, csv_file.homepage))\n \n # Use a sample size\n test_size = 20\n test_data = dict(list(data.items())[:test_size])\n #print(test_data)\n\n get_contents(test_data)\n\n", "repo_name": "AkashNagaraj/Teacher_search", "sub_path": "read_csv.py", "file_name": "read_csv.py", "file_ext": "py", "file_size_in_byte": 1918, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "regex.sub", "line_number": 15, "usage_type": "call"}, {"api_name": "regex.sub", "line_number": 16, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 28, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 31, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 62, "usage_type": "call"}]}
+{"seq_id": "18390941441", "text": "import pandas as pd\nfrom sklearn.tree import DecisionTreeClassifier,export_graphviz\nfrom sklearn import metrics\nfrom sklearn.cross_validation import cross_val_score\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport warnings\nwarnings.filterwarnings(\"ignore\")\n\ntrainX = pd.read_csv('P1_data/trainX.csv',header=None)\ntrainY = pd.read_csv('P1_data/trainY.csv',header=None)\n\ntestX = pd.read_csv('P1_data/testX.csv',header=None)\ntestY = pd.read_csv('P1_data/testY.csv',header=None)\n\n\nprint('Shape of training dataset '+ str(trainX.shape))\nprint('Shape of training label dataset '+ str(trainY.shape))\nprint('Shape of test dataset '+ str(testX.shape))\n\n\ndef MSE(predicted,Y):\n s=0\n for i in range(len(predicted)):\n s+=(predicted[i]-Y[i])**2\n s=s/len(predicted)\n return s\ndepth = []\nmaxnodes = list(range(10,101,10))\nvalidationX = trainX[0:int(0.3*len(trainX))]\nvalidationY = trainY[0:int(0.3*len(trainY))]\nvalidationY = validationY.values.tolist()\n\ntrainnewX = trainX[int(0.3*len(trainX)):]\ntrainnewY = trainY[int(0.3*len(trainY)):]\nfor i in range(3,10):\n clf = DecisionTreeClassifier(max_depth=i)\n# Perform 7-fold cross validation \n clf.fit(trainnewX,trainnewY)\n predicted = clf.predict(validationX)\n depth.append(MSE(predicted,validationY))\n# print(depth)\n\n\nmaxdepth = list(range(3,10))\n\nplt.plot(maxdepth,depth)\nplt.xlabel('Hyperparameter Maxdepth')\nplt.ylabel('Mean Sqaure Error of 10cv(MSE)')\nplt.title('Calculating Maxdepth hyperparameter with Least MSE ')\nprint('Hence, maxdepth is 7')\t\n\ndepth = []\nmaxnodes = list(range(10,101,10))\nfor i in maxnodes:\n clf = DecisionTreeClassifier(max_depth=7,max_leaf_nodes=i)\n# Perform 7-fold cross validation \n clf.fit(trainnewX,trainnewY)\n predicted = clf.predict(validationX)\n depth.append(MSE(predicted,validationY))\n# print(depth)\t\n\nmaxdepth = list(range(3,10))\n\nplt.plot(maxnodes,depth)\nplt.xlabel('Hyperparameter Maxdepth')\nplt.ylabel('Mean Sqaure Error of 10cv(MSE)')\nplt.title('Calculating Maxdepth hyperparameter with Least MSE ')\nprint('Hence, max_leaf_nodes are 90')\n\nmodel = DecisionTreeClassifier(max_depth=7,max_leaf_nodes=80)\nmodel.fit(trainX,trainY)\nprint(model)\t\n\nprint(' Classification Report')\npredicted = model.predict(testX)\nprint(metrics.classification_report(testY,predicted))\nprint('Confusion Matrix')\nprint(metrics.confusion_matrix(testY,predicted))\n\nprint('(b)Total no of nodes')\nprint(model.tree_.node_count)\n\nn_nodes = model.tree_.node_count\nchildren_left = model.tree_.children_left\nchildren_right = model.tree_.children_right\nfeature = model.tree_.feature\nthreshold = model.tree_.threshold\n\nnode_depth = np.zeros(shape=n_nodes, dtype=np.int64)\nis_leaves = np.zeros(shape=n_nodes, dtype=bool)\nstack = [(0, -1)] # seed is the root node id and its parent depth\nwhile len(stack) > 0:\n node_id, parent_depth = stack.pop()\n node_depth[node_id] = parent_depth + 1\n\n # If we have a test node\n if (children_left[node_id] != children_right[node_id]):\n stack.append((children_left[node_id], parent_depth + 1))\n stack.append((children_right[node_id], parent_depth + 1))\n else:\n is_leaves[node_id] = True\nprint('(c)Total Number of leaf nodes')\nprint(sum(is_leaves))\n\nfrom graphviz import Source\nfrom IPython.display import SVG\ngraph = Source(export_graphviz(model, out_file=None, feature_names=trainX.columns))\nSVG(graph.pipe(format='svg'))\n\ndef accuracy(predicted,trainY):\n s=0\n for i in range(len(trainY)):\n #print(predicted[i],trainY[i])\n if(predicted[i] == trainY[i]):\n s+=1\n return float(s)/float(len(trainY))\n\n\ntraininacc = []\ntestacc = []\ndatasetsize = list(range(1,11))\nfor i in range(1,11):\n trainnewX = trainX[0:int(i*0.1*len(trainX))]\n trainnewY = trainY[0:int(i*0.1*len(trainX))]\n model = DecisionTreeClassifier()\n model.fit(trainnewX,trainnewY)\n trainnewY = trainnewY.values.tolist()\n\n predicted = model.predict(trainnewX)\n\n traininacc.append(accuracy(predicted,trainnewY))\n predicted = model.predict(testX)\n testnewY = testY.values.tolist()\n testacc.append(accuracy(predicted,testnewY))\n\n\nplt.plot(datasetsize,traininacc)\nplt.plot(datasetsize,testacc)\nplt.xlabel('Dataset Size')\nplt.ylabel('Training/Test Accuracy')\nplt.title('Accuracy vs Dataset S')\n", "repo_name": "lavishm58/Decision-Tree-Classifier-And-Regression", "sub_path": "P1/P1.py", "file_name": "P1.py", "file_ext": "py", "file_size_in_byte": 4290, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "26", "api": [{"api_name": "warnings.filterwarnings", "line_number": 8, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 10, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 11, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 14, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 71, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 77, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 77, "usage_type": "name"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 79, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 79, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 90, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 91, "usage_type": "call"}, {"api_name": "graphviz.Source", "line_number": 108, "usage_type": "call"}, {"api_name": "sklearn.tree.export_graphviz", "line_number": 108, "usage_type": "call"}, {"api_name": "IPython.display.SVG", "line_number": 109, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}]}
+{"seq_id": "3577458269", "text": "# coding:utf-8\nimport requests, re\nfrom bs4 import BeautifulSoup\nimport lxml,sys,io,json\nsys.stdout = io.TextIOWrapper(sys.stdout.buffer,encoding='gb18030')\n\nwith open(r'C:\\Users\\Jhon117\\Desktop\\demo\\爬虫实战.sb.txt','w',encoding='utf-8')as F:\n for i in range(1,2):\n print('**********')\n url = 'http://search.chinahr.com/sh/job/pn' + str(i) + '/?key=python'\n r = requests.get(url)\n r.raise_for_status()\n r.encoding = r.apparent_encoding\n s = r.text\n bs = BeautifulSoup(s,'html.parser')\n tes = bs.find_all(name='div', attrs={'class':'jobList pc_search_listclick'})\n for t in tes:\n tname = t.find(name='li', attrs={'class': 'job-name'})\n tmp = t.find(name='li', attrs={'class': 'job-salary'})\n tc = t.find(name='li', attrs={'class': 'job-company'})\n tt = t.find_all(name='span')\n print(tname.get_text())\n F.write(json.dumps(tname.get_text(), ensure_ascii=False) + '\\n')\n print(tmp.get_text())\n F.write(json.dumps(tmp.get_text(), ensure_ascii=False) + '\\n')\n print(tc.get_text())\n F.write(json.dumps(tc.get_text(), ensure_ascii=False) + '\\n')\n for n in tt[1:]:\n print(n.get_text(), end=' ')\n F.write(json.dumps(n.get_text(), ensure_ascii=False) + '\\n')\n print('')\n print('********')\n\n\n", "repo_name": "YYN117/Demo", "sub_path": "爬虫实战/练习/职位.py", "file_name": "职位.py", "file_ext": "py", "file_size_in_byte": 1425, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "sys.stdout", "line_number": 5, "usage_type": "attribute"}, {"api_name": "io.TextIOWrapper", "line_number": 5, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 11, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 15, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 23, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 25, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 27, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 30, "usage_type": "call"}]}
+{"seq_id": "72510577984", "text": "\"\"\"\nCorrects costs in produce sales spreadsheet.\n\"\"\"\nfrom os.path import dirname\nfrom openpyxl import load_workbook\n\n\ndef wrapper():\n \"\"\"\n The produce types and their updated prices\n Loop through the rows and update the prices\n skip the first row\n \"\"\"\n path = dirname(__file__) + \"\\\\\"\n wbook = load_workbook(f\"{path}produce_sales.xlsx\")\n sheet = wbook[\"Sheet\"]\n price_updates = {\"Garlic\": 3.07, \"Celery\": 1.19, \"Lemon\": 1.27}\n for row_num in range(2, sheet.max_row):\n produce_name = sheet.cell(row=row_num, column=1).value\n if produce_name in price_updates:\n sheet.cell(row=row_num, column=2).value = price_updates[produce_name]\n wbook.save(f\"{path}updated_produce_sales.xlsx\")\n\n\nif __name__ == \"__main__\":\n wrapper()\n", "repo_name": "jgyy/py-automate", "sub_path": "13/update_produce.py", "file_name": "update_produce.py", "file_ext": "py", "file_size_in_byte": 779, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "os.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 15, "usage_type": "call"}]}
+{"seq_id": "34690826948", "text": "from __future__ import print_function, division\n\nimport os\nimport sys\nimport shlex\nimport subprocess\n\nimport dirhash\n\nimport pytest\n\n\nconsole_script = os.path.join(\n os.path.dirname(sys.executable),\n 'dirhash'\n)\n\n\ndef dirhash_run(argstring, add_env=None):\n assert os.path.isfile(console_script)\n assert os.access(console_script, os.X_OK)\n if add_env:\n env = os.environ.copy()\n env.update(add_env)\n else:\n env = None\n process = subprocess.Popen(\n [console_script] + shlex.split(argstring),\n stdout=subprocess.PIPE,\n stderr=subprocess.PIPE,\n env=env\n )\n output, error = process.communicate()\n\n # in python3 output and error are `bytes` as opposed to `str` in python2\n if isinstance(output, bytes):\n output = output.decode('utf-8')\n if isinstance(error, bytes):\n error = error.decode('utf-8')\n\n return output, error, process.returncode\n\n\ndef create_default_tree(tmpdir):\n \"\"\"\n tmpdir/\n |__.dir/\n | |__file\n |__.file\n |__dir/\n | |__file\n |__empty/\n |__file\n |__file.ext1\n |__file.ext2\n \"\"\"\n dotdir = tmpdir.mkdir('.dir')\n dotdir.join('file').write('file in hidden sub-directory')\n tmpdir.join(\".file\").write('hidden file')\n dir = tmpdir.mkdir('dir')\n dir.join('file').write('file in sub-directory')\n tmpdir.mkdir('empty')\n tmpdir.join(\"file\").write('file')\n tmpdir.join(\"file.ext1\").write('file with extension .ext1')\n tmpdir.join(\"file.ext2\").write('file with extension .ext2')\n\n\nclass TestCLI(object):\n @pytest.mark.parametrize(\n 'argstring, non_default_kwargs',\n [\n (\n '. -a md5',\n {}\n ),\n (\n '.. -a md5',\n {'directory': '..'}\n ),\n (\n 'target-dir -a md5',\n {'directory': 'target-dir'}\n ),\n (\n '. -a sha256',\n {'algorithm': 'sha256'}\n ),\n # Filtering options\n (\n '. -a md5 -m \"*\" \"!.*\"',\n {'match': ['*', '!.*']}\n ),\n (\n '. -a md5 --match \"d1/*\" \"d2/*\" --ignore \"*.txt\"',\n {'match': ['d1/*', 'd2/*'], 'ignore': ['*.txt']}\n ),\n (\n '. -a md5 --empty-dirs',\n {'empty_dirs': True}\n ),\n (\n '. -a md5 --no-linked-dirs',\n {'linked_dirs': False}\n ),\n (\n '. -a md5 --no-linked-files',\n {'linked_files': False}\n ),\n # Protocol options\n (\n '. -a md5 --allow-cyclic-links',\n {'allow_cyclic_links': True}\n\n ),\n (\n '. -a md5 --properties name',\n {'entry_properties': ['name']}\n\n ),\n (\n '. -a md5 --properties name data',\n {'entry_properties': ['name', 'data']}\n\n ),\n # Implementation\n (\n '. -a md5 -j 10',\n {'jobs': 10}\n ),\n (\n '. -a md5 -s 32000',\n {'chunk_size': 32000}\n ),\n ]\n )\n def test_get_kwargs(self, argstring, non_default_kwargs):\n from dirhash.cli import get_kwargs\n kwargs_expected = {\n 'list': False,\n 'directory': '.',\n 'algorithm': 'md5',\n 'match': ['*'],\n 'ignore': None,\n 'empty_dirs': False,\n 'linked_dirs': True,\n 'linked_files': True,\n 'entry_properties': ['data', 'name'],\n 'allow_cyclic_links': False,\n 'chunk_size': 2 ** 20,\n 'jobs': 1\n }\n kwargs_expected.update(non_default_kwargs)\n kwargs = get_kwargs(shlex.split(argstring))\n assert kwargs == kwargs_expected\n\n @pytest.mark.parametrize(\n 'description, argstrings, output',\n [\n ('ARGS WITHOUT EFFECT WHEN LISTING',\n ['. -l',\n '. --list',\n '. -a md5 --list',\n '. -a sha256 --list',\n '. --properties name --list',\n '. --jobs 2 --list',\n '. --chunk-size 2 --list'],\n ('.dir/file\\n'\n '.file\\n'\n 'dir/file\\n'\n 'file\\n'\n 'file.ext1\\n'\n 'file.ext2\\n')),\n ('IGNORE EXTENSION',\n ['. -i \"*.ext1\" --list',\n '. --ignore \"*.ext1\" --list',\n '. -m \"*\" \"!*.ext1\" --list',\n '. --match \"*\" \"!*.ext1\" --list'],\n ('.dir/file\\n'\n '.file\\n'\n 'dir/file\\n'\n 'file\\n'\n 'file.ext2\\n')),\n ('IGNORE MULTIPLE EXTENSIONS',\n ['. -i \"*.ext1\" \"*.ext2\" --list',\n '. -i \"*.ext*\" --list'],\n ('.dir/file\\n'\n '.file\\n'\n 'dir/file\\n'\n 'file\\n')),\n ('IGNORE HIDDEN',\n ['. -i \".*\" \".*/\" --list'],\n ('dir/file\\n'\n 'file\\n'\n 'file.ext1\\n'\n 'file.ext2\\n')),\n ('INCLUDE EMPTY',\n ['. --empty-dirs --list'],\n ('.dir/file\\n'\n '.file\\n'\n 'dir/file\\n'\n 'empty/.\\n'\n 'file\\n'\n 'file.ext1\\n'\n 'file.ext2\\n')),\n ]\n )\n def test_list(self, description, argstrings, output, tmpdir):\n create_default_tree(tmpdir)\n with tmpdir.as_cwd():\n for argstring in argstrings:\n o, error, returncode = dirhash_run(argstring)\n assert returncode == 0\n assert error == ''\n assert o == output\n\n @pytest.mark.parametrize(\n 'argstring, kwargs, expected_hashes',\n [\n ('. -a md5',\n {'algorithm': 'md5'},\n ['594c48dde0776b03eddeeb0232190be7',\n 'd8ab965636d48e407b73b9dbba4cb928',\n '050e7bc9ffcb09c15186c04e0f8026df']\n ),\n ('. -a sha256',\n {'algorithm': 'sha256'},\n ['23a04964149889e932ba3348fe22442f4f6a3b3fec616a386a70579ee857ab7b',\n '7b76bac43e963f9561f37b96b92d7a174094bff230c6efbf1d8bf650e8b40b7a',\n '7156da2b2e5a2926eb4b72e65f389343cb6aca0578f0aedcd6f7457abd67d8f5']),\n ]\n )\n def test_hash_result(self, argstring, kwargs, expected_hashes, tmpdir):\n # verify same result from cmdline and library + regression test of actual\n # hashes\n create_default_tree(tmpdir)\n with tmpdir.as_cwd():\n for add_argstring, add_kwargs, expected_hash in zip(\n ['', ' -p data', ' -p name'],\n [\n {},\n {'entry_properties': ['data']},\n {'entry_properties': ['name']},\n ],\n expected_hashes\n ):\n # run CLI\n full_argstring = argstring + add_argstring\n cli_out, error, returncode = dirhash_run(full_argstring)\n assert error == ''\n assert returncode == 0\n assert cli_out[-1] == '\\n'\n cli_hash = cli_out[:-1]\n\n # run CLI multiproc\n full_argstring_mp = argstring + add_argstring + ' --jobs 2'\n cli_out_mp, error_mp, returncode_mp = dirhash_run(full_argstring_mp)\n assert error_mp == ''\n assert returncode_mp == 0\n assert cli_out_mp[-1] == '\\n'\n cli_hash_mp = cli_out_mp[:-1]\n\n # run lib function\n full_kwargs = kwargs.copy()\n full_kwargs.update(add_kwargs)\n lib_hash = dirhash.dirhash(str(tmpdir), **full_kwargs)\n\n assert cli_hash == cli_hash_mp == lib_hash == expected_hash\n\n def test_error_bad_argument(self, tmpdir):\n with tmpdir.as_cwd():\n o, error, returncode = dirhash_run('. --chunk-size not_an_int')\n assert returncode > 0\n assert error != ''\n", "repo_name": "andhus/dirhash-python", "sub_path": "tests/test_cli.py", "file_name": "test_cli.py", "file_ext": "py", "file_size_in_byte": 8290, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 35, "dataset": "github-code", "pt": "25", "api": [{"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sys.executable", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.access", "line_number": 21, "usage_type": "call"}, {"api_name": "os.X_OK", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.environ.copy", "line_number": 23, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 23, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 27, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 28, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 29, "usage_type": "attribute"}, {"api_name": "subprocess.PIPE", "line_number": 30, "usage_type": "attribute"}, {"api_name": "dirhash.cli.get_kwargs", "line_number": 153, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 153, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 69, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 156, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 156, "usage_type": "attribute"}, {"api_name": "dirhash.dirhash", "line_number": 265, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 216, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 216, "usage_type": "attribute"}]}
+{"seq_id": "41361531981", "text": "import socket\nimport threading\nimport json\nimport numpy as np\nfrom naive_serial import Naive_serial\nfrom Constant import CommandEnum\nfrom Constant import EventEnum\nfrom Constant import frame_kind\nfrom Constant import Head\nfrom Constant import Status\n\nclass Gateway:\n \"\"\"\n 网关类\n \"\"\"\n\n def __init__(self):\n \"\"\"\n 构造\n \"\"\"\n self.__socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n self.__chat_connections = {}\n self.__gobang_connections = {}\n self.__nicknames = {}\n self.totol_connect_num = 0\n self.black_player = None\n self.white_player = None\n self.qizi = np.zeros((15,15),dtype=np.int)\n self.naive_serial = Naive_serial()\n self.naive_serial.run()\n\n def __user_thread(self, user_id):\n \"\"\"\n 用户子线程\n :param user_id: 用户id\n \"\"\"\n connection = self.__chat_connections[user_id]\n nickname = self.__nicknames[user_id]\n print('[Server] 用户', user_id, nickname, '加入聊天室')\n #self.__broadcast(message='用户 ' + str(nickname) + '(' + str(user_id) + ')' + '加入聊天室')\n\n # 侦听\n while True: \n # noinspection PyBroadException\n try:\n buffer = connection.recv(1024).decode()\n # 解析成json数据\n obj = json.loads(buffer)\n # 如果是广播指令\n if obj['type'] == 'broadcast':\n #self.__broadcast(obj['sender_id'], obj['message'])\n self.broadcastToPotato(obj['sender_id'],obj['message'])\n elif obj['type'] == 'logout':\n self.logoutToPotato(user_id)\n\n\n break\n else:\n print('[Server] 无法解析json数据包:', connection.getsockname(), connection.fileno())\n except Exception:\n print('[Server] 连接失效:', connection.getsockname(), connection.fileno())\n #self.__chat_connections[user_id].close()\n self.__chat_connections.pop(user_id)\n self.__nicknames.pop(user_id)\n break\n\n def __gobang_thread(self, user_id, identity): \n \"\"\"\n gobang子线程\n :param user_id: 用户id\n \"\"\"\n connection = self.__gobang_connections[user_id]\n nickname = self.__nicknames[user_id]\n print('[Server] 用户', user_id, nickname, '加入gobang')\n self.__gobang_broadcast(message= identity +'' + str(nickname) + '(' + str(user_id) + ')' + '加入游戏', type=\"info\")\n\n # 侦听\n while True: \n # noinspection PyBroadException\n try:\n buffer = connection.recv(1024).decode()\n # 解析成json数据\n obj = json.loads(buffer)\n # 如果是广播指令\n if obj['type'] == 'xiaqi':\n self.__gobang_broadcast(sender_id=obj['sender_id'], type='xiaqi',identity=obj['identity'], pos=obj['pos'])\n x, y = obj['pos']\n self.qizi[x][y] = 1 if obj['identity']==\"black\" else 2\n winner = self.__check_winner(x,y,self.qizi[x][y])\n if winner:\n winner = 'black' if winner ==1 else 'white'\n message = \"winner is \" + winner\n self.__gobang_broadcast(sender_id=0, type='win',identity=obj['identity'], message=message)\n elif obj['type'] == 'logout':\n print('[Server] 用户', user_id, nickname, '退出游戏')\n if obj['identity'] != \"audience\":\n self.__gobang_broadcast(sender_id=0, type='logout',identity=obj['identity'])\n if identity == 'black':\n self.black_player = None\n if identity == 'white':\n self.white_player = None\n else:\n self.__gobang_broadcast(sender_id=0,message= identity +'' + str(nickname) + '(' + str(user_id) + ')' + '退出观看游戏', type=\"info\")\n\n self.__gobang_connections.pop(user_id)\n self.__nicknames.pop(user_id)\n \n thread = threading.Thread(target=self.__waitForLogin, args=(connection,user_id))\n thread.setDaemon(True)\n thread.start()\n break\n else:\n print('[Server] 无法解析json数据包:', connection.getsockname(), connection.fileno())\n except Exception as e:\n print(e)\n print(e.__traceback__.tb_lineno)\n print('[Server] 连接失效:', connection.getsockname(), connection.fileno())\n #self.__chat_connections[user_id].close()\n self.__gobang_connections.pop(user_id)\n self.__nicknames.pop(user_id)\n break\n\n def __broadcast(self, sender_id=0, message='', type='info',identity='', pos=(0,0)):\n \"\"\"\n 广播\n :param user_id: 用户id(0为系统)\n :param message: 广播内容\n \"\"\"\n for user_id, conection in self.__chat_connections.items():\n if user_id != sender_id and conection:\n conection.send(json.dumps({\n 'sender_id': sender_id,\n 'sender_nickname': self.__nicknames[sender_id],\n 'message': message\n }).encode())\n\n def __check_winner(self, x, y, turn):\n #沿x方向\n cnt = 0\n curx,cury = x,y\n while (0<=curx and curx<15 and 0<=cury and cury<15 and self.qizi[curx][cury]==turn):\n cnt += 1\n curx -= 1\n curx,cury = x,y\n while(0<=curx and curx<15 and 0<=cury and cury<15 and self.qizi[curx][cury]==turn):\n cnt += 1\n curx += 1\n if(cnt>5):\n return turn\n #沿y方向\n cnt=0\n curx,cury = x,y\n while(0<=curx and curx<15 and 0<=cury and cury<15 and self.qizi[curx][cury]==turn) :\n cnt+=1\n cury-=1\n curx,cury = x,y\n while(0<=curx and curx<15 and 0<=cury and cury<15 and self.qizi[curx][cury]==turn) :\n cnt+=1\n cury+=1\n if(cnt>5):\n return turn\n #沿右上方向\n cnt=0\n curx,cury = x,y\n while (0<=curx and curx<15 and 0<=cury and cury<15 and self.qizi[curx][cury]==turn) :\n cnt+=1\n curx+=1\n cury-=1\n curx,cury = x,y\n while (0<=curx and curx<15 and 0<=cury and cury<15 and self.qizi[curx][cury]==turn) :\n cnt+=1\n curx-=1\n cury+=1\n if(cnt>5):\n return turn\n #沿左上方向\n cnt=0\n curx,cury = x,y\n while(0<=curx and curx<15 and 0<=cury and cury<15 and self.qizi[curx][cury]==turn) :\n cnt+=1\n curx-=1\n cury-=1\n curx,cury = x,y\n while(0<=curx and curx<15 and 0<=cury and cury<15 and self.qizi[curx][cury]==turn):\n cnt+=1\n curx+=1\n cury+=1\n if (cnt>5):\n return turn\n return 0 # 没赢\n\n def __gobang_broadcast(self, sender_id=0, message='', type='info',identity='', pos=(0,0)):\n \"\"\"\n 广播\n :param user_id: 用户id(0为系统)\n :param message: 广播内容\n \"\"\"\n if type == \"info\":\n for user_id, conection in self.__gobang_connections.items():\n if user_id != sender_id and conection:\n conection.send(json.dumps({\n 'sender_id': sender_id,\n 'sender_nickname': self.__nicknames[sender_id],\n 'type': type,\n 'message': message\n }).encode())\n elif type == \"xiaqi\":\n for user_id, conection in self.__gobang_connections.items():\n if user_id != sender_id and conection:\n conection.send(json.dumps({\n 'sender_id': sender_id,\n 'sender_nickname': self.__nicknames[sender_id],\n 'type': type,\n 'identity': identity,\n 'pos': pos\n }).encode())\n elif type == \"logout\": # player logout\n for user_id, conection in self.__gobang_connections.items():\n if user_id != sender_id and conection:\n conection.send(json.dumps({\n 'sender_id': sender_id,\n 'sender_nickname': self.__nicknames[sender_id],\n 'identity': identity,\n 'type': type,\n }).encode())\n elif type == \"win\": # player win\n for user_id, conection in self.__gobang_connections.items():\n if user_id != sender_id and conection:\n conection.send(json.dumps({\n 'sender_id': sender_id,\n 'sender_nickname': self.__nicknames[sender_id],\n 'message': message,\n 'identity': identity,\n 'type': type,\n }).encode())\n \n def __waitForLogin(self, connection, user_id):\n # 尝试接受数据\n # noinspection PyBroadException\n try:\n buffer = connection.recv(1024).decode()\n # 解析成json数据\n obj = json.loads(buffer)\n # 如果是连接指令,那么则返回一个新的用户编号,接收用户连接\n if obj['type'] == 'login':\n self.__chat_connections.update({user_id: connection})\n self.__nicknames.update({user_id: obj['nickname']})\n connection.send(json.dumps({\n 'id': user_id\n }).encode())\n # 给土豆发连接指令\n self.loginToPotato(user_id)\n\n # 开辟一个新的线程\n thread = threading.Thread(target=self.__user_thread, args=(user_id,))\n thread.setDaemon(True)\n thread.start()\n\n elif obj['type'] == 'gobang_login':\n if self.black_player == None:\n identity = \"black\"\n self.black_player = user_id\n elif self.white_player == None:\n identity = \"white\"\n self.white_player = user_id\n else:\n identity = \"audience\"\n self.__gobang_connections.update({user_id: connection})\n self.__nicknames.update({user_id: obj['nickname']})\n connection.send(json.dumps({\n 'id': user_id,\n 'identity': identity\n }).encode())\n # 开辟一个新的线程\n thread = threading.Thread(target=self.__gobang_thread, args=(user_id,identity))\n thread.setDaemon(True)\n thread.start()\n else:\n print('[Server] 无法解析json数据包:', connection.getsockname(), connection.fileno())\n except Exception as e:\n print(e)\n print(e.__traceback__.tb_lineno)\n print('[Server] 无法接受数据:', connection.getsockname(), connection.fileno())\n\n def __uart2tcp(self):\n while True:\n data = self.naive_serial.queue_from_uart.get(block=True)\n datalen = (data[0]<<8)+data[1]-1 # 除掉指令位的长度\n cmd =data[2]\n data = data[3:]\n if cmd == CommandEnum.BROADCAST:\n id = data[0]\n message = data[3:]\n message = message.decode(encoding=\"utf8\")\n connection = self.__chat_connections[id]\n connection.send(json.dumps({\n 'sender_id': id,\n 'sender_nickname': self.__nicknames[id],\n 'message': message\n }).encode())\n\n elif cmd == CommandEnum.LOGOUT:\n id = data[0]\n connection = self.__chat_connections[id]\n print('[Server] 用户', id, '退出聊天室')\n self.__chat_connections.pop(id)\n self.__nicknames.pop(id)\n thread = threading.Thread(target=self.__waitForLogin, args=(connection,id))\n thread.setDaemon(True)\n thread.start()\n \n elif cmd == CommandEnum.PRINT:\n print(data.decode(encoding=\"utf-8\",errors = \"replace\"))\n\n\n ### potato methods begin\n def loginToPotato(self, id):\n a=b'\\x11\\x22\\x33'\n a=bytearray(a)\n datalen = 2\n datalen_high8 = datalen//256\n datalen_low8 = datalen%256\n datalen_checksum = datalen_high8 + datalen_low8\n a.append(datalen_high8)\n a.append(datalen_low8)\n a.append(datalen_checksum)\n checksum = 0\n cmd = CommandEnum.LOGIN\n a.append(cmd)\n checksum += cmd\n a.append(id)\n checksum += id\n a.append(checksum%256)\n self.naive_serial.queue_to_uart.put(a)\n\n def broadcastToPotato(self, sender_id, message):\n # message是str\n a=b'\\x11\\x22\\x33'\n a=bytearray(a)\n messagelen = len(message)\n messagelen_high8 = messagelen//256\n messagelen_low8 = messagelen%256\n datalen = 2 + 2 + messagelen\n datalen_high8 = datalen//256\n datalen_low8 = datalen%256\n datalen_checksum = datalen_high8 + datalen_low8\n a.append(datalen_high8)\n a.append(datalen_low8)\n a.append(datalen_checksum)\n checksum = 0\n cmd = CommandEnum.BROADCAST\n a.append(cmd)\n checksum += cmd\n a.append(sender_id)\n checksum += sender_id\n a.append(messagelen_high8)\n checksum += messagelen_high8\n a.append(messagelen_low8)\n checksum += messagelen_low8\n message_bytes = message.encode(encoding = \"utf8\")\n for byte in message_bytes:\n a.append(byte)\n checksum += byte\n a.append(checksum%256)\n self.naive_serial.queue_to_uart.put(a)\n\n def logoutToPotato(self, id):\n a=b'\\x11\\x22\\x33'\n a=bytearray(a)\n datalen = 2\n datalen_high8 = datalen//256\n datalen_low8 = datalen%256\n datalen_checksum = datalen_high8 + datalen_low8\n a.append(datalen_high8)\n a.append(datalen_low8)\n a.append(datalen_checksum)\n checksum = 0\n cmd = CommandEnum.LOGOUT\n a.append(cmd)\n checksum += cmd\n a.append(id)\n checksum += id\n a.append(checksum%256)\n self.naive_serial.queue_to_uart.put(a)\n\n ### potato methods end\n\n def start(self):\n \"\"\"\n 启动服务器\n \"\"\"\n # 绑定端口\n self.__socket.bind(('127.0.0.1', 12345))\n # 启用监听\n self.__socket.listen(10)\n print('[Server] 服务器正在运行......')\n\n # 清空连接\n self.__chat_connections.clear()\n self.__nicknames.clear()\n\n self.__chat_connections.update({0: None})\n self.__nicknames.update({0: \"System\"})\n self.totol_connect_num += 1\n\n thread = threading.Thread(target=self.__uart2tcp, args=())\n thread.setDaemon(True)\n thread.start()\n\n # 开始侦听\n while True:\n connection, address = self.__socket.accept()\n print('[Server] 收到一个新连接', connection.getsockname(), connection.fileno())\n user_id = self.totol_connect_num \n self.totol_connect_num += 1\n\n thread = threading.Thread(target=self.__waitForLogin, args=(connection,user_id))\n thread.setDaemon(True)\n thread.start()\n\nif __name__ == '__main__':\n gateway = Gateway()\n gateway.start()", "repo_name": "SimonLiu1999/Potatotype", "sub_path": "gateway_python_pc/gateway.py", "file_name": "gateway.py", "file_ext": "py", "file_size_in_byte": 16144, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "socket.socket", "line_number": 21, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 21, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 28, "usage_type": "attribute"}, {"api_name": "naive_serial.Naive_serial", "line_number": 29, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 48, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 83, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 108, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 131, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 201, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 210, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 220, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 229, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 243, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 248, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 255, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 270, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 275, "usage_type": "call"}, {"api_name": "Constant.CommandEnum.BROADCAST", "line_number": 291, "usage_type": "attribute"}, {"api_name": "Constant.CommandEnum", "line_number": 291, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 296, "usage_type": "call"}, {"api_name": "Constant.CommandEnum.LOGOUT", "line_number": 302, "usage_type": "attribute"}, {"api_name": "Constant.CommandEnum", "line_number": 302, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 308, "usage_type": "call"}, {"api_name": "Constant.CommandEnum.PRINT", "line_number": 312, "usage_type": "attribute"}, {"api_name": "Constant.CommandEnum", "line_number": 312, "usage_type": "name"}, {"api_name": "Constant.CommandEnum.LOGIN", "line_number": 328, "usage_type": "attribute"}, {"api_name": "Constant.CommandEnum", "line_number": 328, "usage_type": "name"}, {"api_name": "Constant.CommandEnum.BROADCAST", "line_number": 351, "usage_type": "attribute"}, {"api_name": "Constant.CommandEnum", "line_number": 351, "usage_type": "name"}, {"api_name": "Constant.CommandEnum.LOGOUT", "line_number": 378, "usage_type": "attribute"}, {"api_name": "Constant.CommandEnum", "line_number": 378, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 406, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 417, "usage_type": "call"}]}
+{"seq_id": "71837522945", "text": "from sqlalchemy.orm import DeclarativeBase\nfrom sqlalchemy import String, ForeignKey\nfrom sqlalchemy.orm import Mapped, mapped_column, relationship\n\n\nclass Base(DeclarativeBase):\n pass\n\n\nclass Portfolio(Base):\n __tablename__ = \"portfolios\"\n\n slug: Mapped[str] = mapped_column(String, primary_key=True, unique=True)\n title: Mapped[str]\n html_description_url: Mapped[str]\n image_url: Mapped[str]\n markdown_description_filename: Mapped[str]\n image_filename: Mapped[str]\n\n projects: Mapped[list[\"Project\"]] = relationship()\n\n\nclass Project(Base):\n __tablename__ = \"projects\"\n\n slug: Mapped[str] = mapped_column(String, primary_key=True, unique=True)\n portfolio_slug: Mapped[str] = mapped_column(\n ForeignKey(\"portfolios.slug\", ondelete=\"CASCADE\")\n )\n title: Mapped[str]\n html_description_url: Mapped[str]\n markdown_description_filename: Mapped[str]\n", "repo_name": "Uoyroem/Digit-Alem", "sub_path": "server/app/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 900, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "sqlalchemy.orm.DeclarativeBase", "line_number": 6, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.Mapped", "line_number": 13, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.mapped_column", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 13, "usage_type": "argument"}, {"api_name": "sqlalchemy.orm.Mapped", "line_number": 14, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.Mapped", "line_number": 15, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.Mapped", "line_number": 16, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.Mapped", "line_number": 17, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.Mapped", "line_number": 18, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.Mapped", "line_number": 20, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.Mapped", "line_number": 26, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.mapped_column", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 26, "usage_type": "argument"}, {"api_name": "sqlalchemy.orm.Mapped", "line_number": 27, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.mapped_column", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.Mapped", "line_number": 30, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.Mapped", "line_number": 31, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.Mapped", "line_number": 32, "usage_type": "name"}]}
+{"seq_id": "11893633713", "text": "from multiprocessing import Process\nimport os\n\n\ndef target_function() -> None:\n\n\tprint(f'\\niteration {os.getpid()}')\t\n\ndef main() -> None:\n\n\tfor i in range(100):\n\n\t\t# The constructor should always be called with keyword arguments. group \n\t\t# should always be None\n\n\t\tprocess = Process(\n\t\t\ttarget=target_function)\n\n\t\tprocess.start()\n\nif __name__ == '__main__':\n\n\tmain()\n", "repo_name": "software-foundations/learning-distributed-systems", "sub_path": "documentation_multiprocessing/08_reference/01_process_and_exceptions/process_and_exceptions_01.py", "file_name": "process_and_exceptions_01.py", "file_ext": "py", "file_size_in_byte": 369, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "os.getpid", "line_number": 7, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 16, "usage_type": "call"}]}
+{"seq_id": "7564683960", "text": "import os\nimport random\nimport bpy\nimport math\nimport numpy as np\nfrom mathutils import Vector, Matrix\nfrom bpy_extras.object_utils import world_to_camera_view\nfrom rd.modify_material import set_modify_material, set_modify_raw_material, set_modify_table_material, set_modify_floor_material\n\n# render parameter \nRENDERING_PATH = os.getcwd()\nLIGHT_EMITTER_ENERGY = 5\nLIGHT_ENV_MAP_ENERGY_IR = 0.035 \nLIGHT_ENV_MAP_ENERGY_RGB = 1.0 \nCYCLES_SAMPLE = 32\nCAMERA_TYPE = \"realsense\"\nNUM_FRAME_PER_SCENE = 24\n\n# view point parameter\nlook_at_shift = np.array([0,0,0])\nnum_point_ver = 6\nnum_point_hor = 4\nbeta_range = (15*math.pi/180, 45*math.pi/180)\nr = 0.5\ntsdf2blender_coord_T_shift = np.array([-0.15, -0.15, -0.0503])\nTABLE_CAD_MODEL_HEIGHT = 0.75\n\n# material randomization mode (transparent, specular, mixed, raw)\nmy_material_randomize_mode = 'mixed'\n\n# set depth sensor parameter\ncamera_width = 640\ncamera_height = 360\nbaseline_distance = 0.055\n\n# set background parameter\nbackground_size = 3.\nbackground_position = (0., 0., 0.)\nbackground_scale = (1., 1., 1.)\n\n\n# set camera randomize paramater\nstart_point_range = ((0.5, 0.95), (-0.6, 0.6, -0.6, 0.6))\nup_range = (-0.18, -0.18, -0.18, 0.18)\nlook_at_range = (background_position[0] - 0.05, background_position[0] + 0.05, \n background_position[1] - 0.05, background_position[1] + 0.05,\n background_position[2] - 0.05, background_position[2] + 0.05)\n\n\ng_syn_light_num_lowbound = 4\ng_syn_light_num_highbound = 6\ng_syn_light_dist_lowbound = 8\ng_syn_light_dist_highbound = 12\ng_syn_light_azimuth_degree_lowbound = 0\ng_syn_light_azimuth_degree_highbound = 360\ng_syn_light_elevation_degree_lowbound = 0\ng_syn_light_elevation_degree_highbound = 90\ng_syn_light_energy_mean = 3\ng_syn_light_energy_std = 0.5\ng_syn_light_environment_energy_lowbound = 0\ng_syn_light_environment_energy_highbound = 1\n\n\ng_shape_synset_name_pairs_all = {'02691156': 'aeroplane',\n '02747177': 'ashtray',\n '02773838': 'backpack',\n '02801938': 'basket',\n '02808440': 'tub', # bathtub\n '02818832': 'bed',\n '02828884': 'bench',\n '02834778': 'bicycle',\n '02843684': 'mailbox', # missing in objectnet3d, birdhouse, use view distribution of mailbox\n '02858304': 'boat',\n '02871439': 'bookshelf',\n '02876657': 'bottle',\n '02880940': 'bowl', # missing in objectnet3d, bowl, use view distribution of plate\n '02924116': 'bus',\n '02933112': 'cabinet',\n '02942699': 'camera',\n '02946921': 'can',\n '02954340': 'cap',\n '02958343': 'car',\n '02992529': 'cellphone',\n '03001627': 'chair',\n '03046257': 'clock',\n '03085013': 'keyboard',\n '03207941': 'dishwasher',\n '03211117': 'tvmonitor',\n '03261776': 'headphone',\n '03325088': 'faucet',\n '03337140': 'filing_cabinet',\n '03467517': 'guitar',\n '03513137': 'helmet',\n '03593526': 'jar',\n '03624134': 'knife',\n '03636649': 'lamp',\n '03642806': 'laptop',\n '03691459': 'speaker',\n '03710193': 'mailbox',\n '03759954': 'microphone',\n '03761084': 'microwave',\n '03790512': 'motorbike',\n '03797390': 'mug', # missing in objectnet3d, mug, use view distribution of cup\n '03928116': 'piano',\n '03938244': 'pillow',\n '03948459': 'rifle', # missing in objectnet3d, pistol, use view distribution of rifle\n '03991062': 'pot',\n '04004475': 'printer',\n '04074963': 'remote_control',\n '04090263': 'rifle',\n '04099429': 'road_pole', # missing in objectnet3d, rocket, use view distribution of road_pole\n '04225987': 'skateboard',\n '04256520': 'sofa',\n '04330267': 'stove',\n '04379243': 'diningtable', # use view distribution of dining_table\n '04401088': 'telephone',\n '04460130': 'road_pole', # missing in objectnet3d, tower, use view distribution of road_pole\n '04468005': 'train',\n '04530566': 'washing_machine',\n '04554684': 'dishwasher'} # washer, use view distribution of dishwasher\n\ng_synset_name_label_pairs = {#'aeroplane': 7,\n #'bottle': 1,\n #'bowl': 2, \n #'camera': 3,\n #'can': 4,\n #'car': 5,\n #'mug': 6, \n 'other': 0} \n\nmaterial_class_instance_pairs = {'specular': ['metal', 'paintsp'], # 'porcelain','plasticsp',\n 'transparent': ['glass'],\n 'diffuse': ['plastic','rubber','paper','leather','wood','clay','fabric'],\n 'background': ['background']}\n\n\n# material list\nclass_material_pairs = {'specular': ['other'],\n 'transparent': ['other'],\n 'diffuse': ['other']}\n\ninstance_material_except_pairs = {'metal': [],\n 'porcelain': [],\n 'plasticsp': [],\n 'paintsp':[],\n\n 'glass': [],#[8,9,18,19,20,24,25,26,27,28,29,30,31,32,34,43,59,72],\n \n 'plastic': [],\n 'rubber': [], \n 'leather': [],\n 'wood':[],\n 'paper':[],\n 'fabric':[],\n 'clay':[], \n }\ninstance_material_include_pairs = {\n }\n\nmaterial_class_id_dict = {'raw': 0,\n 'diffuse': 1,\n 'transparent': 2,\n 'specular': 3}\n\nmaterial_type_id_dict = {'raw': 0,\n 'metal': 1,\n 'porcelain': 2,\n 'plasticsp': 3,\n 'paintsp':4,\n 'glass': 5, \n 'plastic': 6,\n 'rubber': 7, \n 'leather': 8,\n 'wood':9,\n 'paper':10,\n 'fabric':11,\n 'clay':12, \n }\n\n\ndef obj_centered_camera_pos(dist, azimuth_deg, elevation_deg):\n phi = float(elevation_deg) / 180 * math.pi\n theta = float(azimuth_deg) / 180 * math.pi\n x = (dist * math.cos(theta) * math.cos(phi))\n y = (dist * math.sin(theta) * math.cos(phi))\n z = (dist * math.sin(phi))\n return (x, y, z)\n\ndef quaternionFromYawPitchRoll(yaw, pitch, roll):\n c1 = math.cos(yaw / 2.0)\n c2 = math.cos(pitch / 2.0)\n c3 = math.cos(roll / 2.0) \n s1 = math.sin(yaw / 2.0)\n s2 = math.sin(pitch / 2.0)\n s3 = math.sin(roll / 2.0) \n q1 = c1 * c2 * c3 + s1 * s2 * s3\n q2 = c1 * c2 * s3 - s1 * s2 * c3\n q3 = c1 * s2 * c3 + s1 * c2 * s3\n q4 = s1 * c2 * c3 - c1 * s2 * s3\n return (q1, q2, q3, q4)\n\ndef camPosToQuaternion(cx, cy, cz):\n q1a = 0\n q1b = 0\n q1c = math.sqrt(2) / 2\n q1d = math.sqrt(2) / 2\n camDist = math.sqrt(cx * cx + cy * cy + cz * cz)\n cx = cx / camDist\n cy = cy / camDist\n cz = cz / camDist \n t = math.sqrt(cx * cx + cy * cy) \n tx = cx / t\n ty = cy / t\n yaw = math.acos(ty)\n if tx > 0:\n yaw = 2 * math.pi - yaw\n pitch = 0\n tmp = min(max(tx*cx + ty*cy, -1),1)\n #roll = math.acos(tx * cx + ty * cy)\n roll = math.acos(tmp)\n if cz < 0:\n roll = -roll \n print(\"%f %f %f\" % (yaw, pitch, roll))\n q2a, q2b, q2c, q2d = quaternionFromYawPitchRoll(yaw, pitch, roll) \n q1 = q1a * q2a - q1b * q2b - q1c * q2c - q1d * q2d\n q2 = q1b * q2a + q1a * q2b + q1d * q2c - q1c * q2d\n q3 = q1c * q2a - q1d * q2b + q1a * q2c + q1b * q2d\n q4 = q1d * q2a + q1c * q2b - q1b * q2c + q1a * q2d\n return (q1, q2, q3, q4)\n\ndef camRotQuaternion(cx, cy, cz, theta): \n theta = theta / 180.0 * math.pi\n camDist = math.sqrt(cx * cx + cy * cy + cz * cz)\n cx = -cx / camDist\n cy = -cy / camDist\n cz = -cz / camDist\n q1 = math.cos(theta * 0.5)\n q2 = -cx * math.sin(theta * 0.5)\n q3 = -cy * math.sin(theta * 0.5)\n q4 = -cz * math.sin(theta * 0.5)\n return (q1, q2, q3, q4)\n\ndef quaternionProduct(qx, qy): \n a = qx[0]\n b = qx[1]\n c = qx[2]\n d = qx[3]\n e = qy[0]\n f = qy[1]\n g = qy[2]\n h = qy[3]\n q1 = a * e - b * f - c * g - d * h\n q2 = a * f + b * e + c * h - d * g\n q3 = a * g - b * h + c * e + d * f\n q4 = a * h + b * g - c * f + d * e \n return (q1, q2, q3, q4)\n\ndef quaternionToRotation(q):\n w, x, y, z = q\n r00 = 1 - 2 * y ** 2 - 2 * z ** 2\n r01 = 2 * x * y + 2 * w * z\n r02 = 2 * x * z - 2 * w * y\n\n r10 = 2 * x * y - 2 * w * z\n r11 = 1 - 2 * x ** 2 - 2 * z ** 2\n r12 = 2 * y * z + 2 * w * x\n\n r20 = 2 * x * z + 2 * w * y\n r21 = 2 * y * z - 2 * w * x\n r22 = 1 - 2 * x ** 2 - 2 * y ** 2\n r = [[r00, r01, r02], [r10, r11, r12], [r20, r21, r22]]\n return r\n\ndef quaternionToRotation_xyzw(q):\n x, y, z, w = q\n r00 = 1 - 2 * y ** 2 - 2 * z ** 2\n r01 = 2 * x * y + 2 * w * z\n r02 = 2 * x * z - 2 * w * y\n\n r10 = 2 * x * y - 2 * w * z\n r11 = 1 - 2 * x ** 2 - 2 * z ** 2\n r12 = 2 * y * z + 2 * w * x\n\n r20 = 2 * x * z + 2 * w * y\n r21 = 2 * y * z - 2 * w * x\n r22 = 1 - 2 * x ** 2 - 2 * y ** 2\n r = [[r00, r01, r02], [r10, r11, r12], [r20, r21, r22]]\n return r\n\ndef quaternionFromRotMat(rotation_matrix):\n rotation_matrix = np.reshape(rotation_matrix, (1, 9))[0]\n w = math.sqrt(rotation_matrix[0]+rotation_matrix[4]+rotation_matrix[8]+1 + 1e-6)/2\n x = math.sqrt(rotation_matrix[0]-rotation_matrix[4]-rotation_matrix[8]+1 + 1e-6)/2\n y = math.sqrt(-rotation_matrix[0]+rotation_matrix[4]-rotation_matrix[8]+1 + 1e-6)/2\n z = math.sqrt(-rotation_matrix[0]-rotation_matrix[4]+rotation_matrix[8]+1 + 1e-6)/2\n a = [w,x,y,z]\n m = a.index(max(a))\n if m == 0:\n x = (rotation_matrix[7]-rotation_matrix[5])/(4*w)\n y = (rotation_matrix[2]-rotation_matrix[6])/(4*w)\n z = (rotation_matrix[3]-rotation_matrix[1])/(4*w)\n if m == 1:\n w = (rotation_matrix[7]-rotation_matrix[5])/(4*x)\n y = (rotation_matrix[1]+rotation_matrix[3])/(4*x)\n z = (rotation_matrix[6]+rotation_matrix[2])/(4*x)\n if m == 2:\n w = (rotation_matrix[2]-rotation_matrix[6])/(4*y)\n x = (rotation_matrix[1]+rotation_matrix[3])/(4*y)\n z = (rotation_matrix[5]+rotation_matrix[7])/(4*y)\n if m == 3:\n w = (rotation_matrix[3]-rotation_matrix[1])/(4*z)\n x = (rotation_matrix[6]+rotation_matrix[2])/(4*z)\n y = (rotation_matrix[5]+rotation_matrix[7])/(4*z)\n quaternion = (w,x,y,z)\n return quaternion\n\ndef quaternionFromRotMat_xyzw(rotation_matrix):\n rotation_matrix = np.reshape(rotation_matrix, (1, 9))[0]\n w = math.sqrt(rotation_matrix[0]+rotation_matrix[4]+rotation_matrix[8]+1 + 1e-6)/2\n x = math.sqrt(rotation_matrix[0]-rotation_matrix[4]-rotation_matrix[8]+1 + 1e-6)/2\n y = math.sqrt(-rotation_matrix[0]+rotation_matrix[4]-rotation_matrix[8]+1 + 1e-6)/2\n z = math.sqrt(-rotation_matrix[0]-rotation_matrix[4]+rotation_matrix[8]+1 + 1e-6)/2\n a = [x,y,z,w]\n m = a.index(max(a))\n if m == 0:\n x = (rotation_matrix[7]-rotation_matrix[5])/(4*w)\n y = (rotation_matrix[2]-rotation_matrix[6])/(4*w)\n z = (rotation_matrix[3]-rotation_matrix[1])/(4*w)\n if m == 1:\n w = (rotation_matrix[7]-rotation_matrix[5])/(4*x)\n y = (rotation_matrix[1]+rotation_matrix[3])/(4*x)\n z = (rotation_matrix[6]+rotation_matrix[2])/(4*x)\n if m == 2:\n w = (rotation_matrix[2]-rotation_matrix[6])/(4*y)\n x = (rotation_matrix[1]+rotation_matrix[3])/(4*y)\n z = (rotation_matrix[5]+rotation_matrix[7])/(4*y)\n if m == 3:\n w = (rotation_matrix[3]-rotation_matrix[1])/(4*z)\n x = (rotation_matrix[6]+rotation_matrix[2])/(4*z)\n y = (rotation_matrix[5]+rotation_matrix[7])/(4*z)\n quaternion = (x,y,z,w)\n return quaternion\n\ndef rotVector(q, vector_ori):\n r = quaternionToRotation(q)\n x_ori = vector_ori[0]\n y_ori = vector_ori[1]\n z_ori = vector_ori[2]\n x_rot = r[0][0] * x_ori + r[1][0] * y_ori + r[2][0] * z_ori\n y_rot = r[0][1] * x_ori + r[1][1] * y_ori + r[2][1] * z_ori\n z_rot = r[0][2] * x_ori + r[1][2] * y_ori + r[2][2] * z_ori\n return (x_rot, y_rot, z_rot)\n\ndef cameraLPosToCameraRPos(q_l, pos_l, baseline_dis):\n vector_camera_l_y = (1, 0, 0)\n vector_rot = rotVector(q_l, vector_camera_l_y)\n pos_r = (pos_l[0] + vector_rot[0] * baseline_dis,\n pos_l[1] + vector_rot[1] * baseline_dis,\n pos_l[2] + vector_rot[2] * baseline_dis)\n return pos_r\n\ndef getRTFromAToB(pointCloudA, pointCloudB):\n muA = np.mean(pointCloudA, axis=0)\n muB = np.mean(pointCloudB, axis=0)\n\n zeroMeanA = pointCloudA - muA\n zeroMeanB = pointCloudB - muB\n\n covMat = np.matmul(np.transpose(zeroMeanA), zeroMeanB)\n U, S, Vt = np.linalg.svd(covMat)\n R = np.matmul(Vt.T, U.T)\n\n if np.linalg.det(R) < 0:\n print(\"[V]getRTFromAToB: Reflection detected\")\n Vt[2, :] *= -1\n R = Vt.T * U.T\n T = (-np.matmul(R, muA.T) + muB.T).reshape(3, 1)\n return R, T\n\ndef cameraPositionRandomize(start_point_range, look_at_range, up_range):\n r_range, vector_range = start_point_range\n r_min, r_max = r_range\n x_min, x_max, y_min, y_max = vector_range\n r = random.uniform(r_min, r_max)\n x = random.uniform(x_min, x_max)\n y = random.uniform(y_min, y_max)\n z = math.sqrt(1 - x**2 - y**2)\n vector_camera_axis = np.array([x, y, z])\n\n x_min, x_max, y_min, y_max = up_range\n x = random.uniform(x_min, x_max)\n y = random.uniform(y_min, y_max) \n z = math.sqrt(1 - x**2 - y**2)\n up = np.array([x, y, z])\n\n x_min, x_max, y_min, y_max, z_min, z_max = look_at_range\n look_at = np.array([random.uniform(x_min, x_max),\n random.uniform(y_min, y_max),\n random.uniform(z_min, z_max)])\n position = look_at + r * vector_camera_axis\n\n vectorZ = - (look_at - position)/np.linalg.norm(look_at - position)\n vectorX = np.cross(up, vectorZ)/np.linalg.norm(np.cross(up, vectorZ))\n vectorY = np.cross(vectorZ, vectorX)/np.linalg.norm(np.cross(vectorX, vectorZ))\n\n # points in camera coordinates\n pointSensor= np.array([[0., 0., 0.], [1., 0., 0.], [0., 2., 0.], [0., 0., 3.]])\n\n # points in world coordinates \n pointWorld = np.array([position,\n position + vectorX,\n position + vectorY * 2,\n position + vectorZ * 3])\n\n resR, resT = getRTFromAToB(pointSensor, pointWorld)\n resQ = quaternionFromRotMat(resR)\n return resQ, resT \n\n\ndef genCameraPosition(look_at):\n quat_list = []\n rot_list = []\n trans_list = []\n position_list = []\n \n # alpha: \n alpha = 0\n alpha_delta = (2 * math.pi) / num_point_ver\n for i in range(num_point_ver):\n alpha = alpha + alpha_delta\n flag_x = 1\n flag_y = 1\n alpha1 = alpha\n if alpha > math.pi/2 and alpha <= math.pi: \n alpha1 = math.pi - alpha\n flag_x = -1\n flag_y = 1\n elif alpha > math.pi and alpha <= math.pi*(3/2):\n alpha1 = alpha - math.pi\n flag_x = -1\n flag_y = -1\n elif alpha > math.pi*(3/2):\n alpha1 = math.pi*2 - alpha\n flag_x = 1\n flag_y = -1\n \n beta = beta_range[0]\n beta_delta = (beta_range[1]-beta_range[0])/(num_point_hor-1)\n for j in range(num_point_hor):\n if j != 0:\n beta = beta + beta_delta\n\n x = flag_x * (r * math.sin(beta)) * math.cos(alpha1)\n y = flag_y * (r * math.sin(beta)) * math.sin(alpha1)\n z = r * math.cos(beta)\n position = np.array([x, y, z]) + look_at\n look_at = look_at\n up = np.array([0, 0, 1])\n\n vectorZ = - (look_at - position)/np.linalg.norm(look_at - position)\n vectorX = np.cross(up, vectorZ)/np.linalg.norm(np.cross(up, vectorZ))\n vectorY = np.cross(vectorZ, vectorX)/np.linalg.norm(np.cross(vectorX, vectorZ))\n\n # points in camera coordinates\n pointSensor= np.array([[0., 0., 0.], [1., 0., 0.], [0., 2., 0.], [0., 0., 3.]])\n\n # points in world coordinates \n pointWorld = np.array([position,\n position + vectorX,\n position + vectorY * 2,\n position + vectorZ * 3])\n\n resR, resT = getRTFromAToB(pointSensor, pointWorld)\n resQ = quaternionFromRotMat(resR)\n\n quat_list.append(resQ)\n rot_list.append(resR)\n trans_list.append(resT)\n position_list.append(position)\n return quat_list, trans_list, rot_list \n\n\ndef quanternion_mul(q1, q2):\n s1 = q1[0]\n v1 = np.array(q1[1:])\n s2 = q2[0]\n v2 = np.array(q2[1:])\n s = s1 * s2 - np.dot(v1, v2)\n v = s1 * v2 + s2 * v1 + np.cross(v1, v2)\n return (s, v[0], v[1], v[2])\n\nclass BlenderRenderer(object):\n def __init__(self, viewport_size_x=640, viewport_size_y=360, DEVICE_LIST=None):\n '''\n viewport_size_x, viewport_size_y: rendering viewport resolution\n '''\n\n self.DEVICE_LIST = DEVICE_LIST\n\n # remove all objects, cameras and lights\n for obj in bpy.data.meshes:\n bpy.data.meshes.remove(obj)\n\n for cam in bpy.data.cameras:\n bpy.data.cameras.remove(cam)\n\n for light in bpy.data.lights:\n bpy.data.lights.remove(light)\n\n for obj in bpy.data.objects:\n bpy.data.objects.remove(obj, do_unlink=True)\n\n render_context = bpy.context.scene.render\n\n # add left camera\n camera_l_data = bpy.data.cameras.new(name=\"camera_l\")\n camera_l_object = bpy.data.objects.new(name=\"camera_l\", object_data=camera_l_data)\n bpy.context.collection.objects.link(camera_l_object)\n\n # add right camera\n camera_r_data = bpy.data.cameras.new(name=\"camera_r\")\n camera_r_object = bpy.data.objects.new(name=\"camera_r\", object_data=camera_r_data)\n bpy.context.collection.objects.link(camera_r_object)\n\n camera_l = bpy.data.objects[\"camera_l\"]\n camera_r = bpy.data.objects[\"camera_r\"]\n\n # set the camera postion and orientation so that it is in\n # the front of the object\n camera_l.location = (1, 0, 0)\n camera_r.location = (1, 0, 0)\n\n # add emitter light\n light_emitter_data = bpy.data.lights.new(name=\"light_emitter\", type='SPOT')\n light_emitter_object = bpy.data.objects.new(name=\"light_emitter\", object_data=light_emitter_data)\n bpy.context.collection.objects.link(light_emitter_object)\n\n light_emitter = bpy.data.objects[\"light_emitter\"]\n light_emitter.location = (1, 0, 0)\n light_emitter.data.energy = LIGHT_EMITTER_ENERGY\n\n # render setting\n render_context.resolution_percentage = 100\n self.render_context = render_context\n\n self.camera_l = camera_l\n self.camera_r = camera_r\n\n self.light_emitter = light_emitter\n\n self.model_loaded = False\n self.background_added = None\n\n self.render_context.resolution_x = viewport_size_x\n self.render_context.resolution_y = viewport_size_y\n\n self.my_material = {}\n self.render_mode = 'IR'\n\n # output setting \n self.render_context.image_settings.file_format = 'PNG'\n self.render_context.image_settings.compression = 0\n self.render_context.image_settings.color_mode = 'BW'\n self.render_context.image_settings.color_depth = '8'\n\n # cycles setting\n self.render_context.engine = 'CYCLES'\n bpy.context.scene.cycles.progressive = 'BRANCHED_PATH'\n bpy.context.scene.cycles.use_denoising = True\n bpy.context.scene.cycles.denoiser = 'NLM'\n bpy.context.scene.cycles.film_exposure = 0.5\n\n # self.render_context.use_antialiasing = False\n ##########\n bpy.context.scene.view_layers[\"View Layer\"].use_sky = True\n ##########\n\n # switch on nodes\n bpy.context.scene.use_nodes = True\n tree = bpy.context.scene.node_tree\n links = tree.links\n \n # clear default nodes\n for n in tree.nodes:\n tree.nodes.remove(n)\n \n # create input render layer node\n rl = tree.nodes.new('CompositorNodeRLayers')\n\n # create output node\n self.fileOutput = tree.nodes.new(type=\"CompositorNodeOutputFile\")\n self.fileOutput.base_path = \"./new_data/0000\"\n self.fileOutput.format.file_format = 'OPEN_EXR'\n self.fileOutput.format.color_depth= '32'\n self.fileOutput.file_slots[0].path = 'depth#'\n links.new(rl.outputs[2], self.fileOutput.inputs[0])\n\n # depth sensor pattern\n self.pattern = []\n # environment map\n self.env_map = []\n self.realtable_img_list = []\n self.realfloor_img_list = []\n self.obj_texture_img_list = []\n\n self.src_energy_for_rgb_render = 0\n\n def loadImages(self, pattern_path, env_map_path, real_table_image_root_path, real_floor_image_root_path, obj_texture_image_root_path, obj_texture_image_idxfile, check_seen_scene):\n # load pattern image\n self.pattern = bpy.data.images.load(filepath=pattern_path)\n if check_seen_scene:\n env_map_path_list = os.listdir(env_map_path)\n real_table_image_root_path_list = os.listdir(real_table_image_root_path)\n real_floor_image_root_path_list = os.listdir(real_floor_image_root_path)\n else:\n env_map_path_list = sorted(os.listdir(env_map_path))\n real_table_image_root_path_list = sorted(os.listdir(real_table_image_root_path))\n real_floor_image_root_path_list = sorted(os.listdir(real_floor_image_root_path))\n # load env map\n for item in env_map_path_list:\n if item.split('.')[-1] == 'hdr':\n self.env_map.append(bpy.data.images.load(filepath=os.path.join(env_map_path, item)))\n # load real table images\n for item in real_table_image_root_path_list:\n if item.split('.')[-1] == 'jpg':\n self.realtable_img_list.append(bpy.data.images.load(filepath=os.path.join(real_table_image_root_path, item)))\n # load real floor images\n for item in real_floor_image_root_path_list:\n if item.split('.')[-1] == 'jpg':\n self.realfloor_img_list.append(bpy.data.images.load(filepath=os.path.join(real_floor_image_root_path, item)))\n # load obj texture images\n f_teximg_idx = open(os.path.join(obj_texture_image_root_path, obj_texture_image_idxfile),\"r\")\n lines = f_teximg_idx.readlines() \n for item in lines:\n item = item[:-1] \n self.obj_texture_img_list.append(bpy.data.images.load(filepath=os.path.join(obj_texture_image_root_path, \"images\", item)))\n\n\n def addEnvMap(self):\n # Get the environment node tree of the current scene\n node_tree = bpy.context.scene.world.node_tree\n tree_nodes = node_tree.nodes\n\n # Clear all nodes\n tree_nodes.clear()\n\n # Add Background node\n node_background = tree_nodes.new(type='ShaderNodeBackground')\n\n # Add Environment Texture node\n node_environment = tree_nodes.new('ShaderNodeTexEnvironment')\n # Load and assign the image to the node property\n # node_environment.image = bpy.data.images.load(\"/Users/zhangjiyao/Desktop/test_addon/envmap_lib/autoshop_01_1k.hdr\") # Relative path\n node_environment.location = -300,0\n\n node_tex_coord = tree_nodes.new(type='ShaderNodeTexCoord')\n node_tex_coord.location = -700,0\n\n node_mapping = tree_nodes.new(type='ShaderNodeMapping')\n node_mapping.location = -500,0\n\n # Add Output node\n node_output = tree_nodes.new(type='ShaderNodeOutputWorld') \n node_output.location = 200,0\n\n # Link all nodes\n links = node_tree.links\n links.new(node_environment.outputs[\"Color\"], node_background.inputs[\"Color\"])\n links.new(node_background.outputs[\"Background\"], node_output.inputs[\"Surface\"])\n links.new(node_tex_coord.outputs[\"Generated\"], node_mapping.inputs[\"Vector\"])\n links.new(node_mapping.outputs[\"Vector\"], node_environment.inputs[\"Vector\"])\n\n #### bpy.data.worlds[\"World\"].node_tree.nodes[\"Background\"].inputs[1].default_value = 1.0\n random_energy = random.uniform(LIGHT_ENV_MAP_ENERGY_RGB * 0.8, LIGHT_ENV_MAP_ENERGY_RGB * 1.2)\n bpy.data.worlds[\"World\"].node_tree.nodes[\"Background\"].inputs[1].default_value = random_energy\n ####\n\n\n def setEnvMap(self, env_map_id, rotation_elur_z):\n # Get the environment node tree of the current scene\n node_tree = bpy.context.scene.world.node_tree\n\n # Get Environment Texture node\n node_environment = node_tree.nodes['Environment Texture']\n # Load and assign the image to the node property\n node_environment.image = self.env_map[env_map_id]\n\n node_mapping = node_tree.nodes['Mapping']\n node_mapping.inputs[2].default_value[2] = rotation_elur_z\n\n\n def addMaskMaterial(self, num=20):\n background_material_name_list = [\"mask_background\", \"mask_table\", \"mask_tableplane\"]\n for material_name in background_material_name_list:\n material_class = (bpy.data.materials.get(material_name) or bpy.data.materials.new(material_name)) # test if material exists, if it does not exist, create it:\n\n # enable 'Use nodes'\n material_class.use_nodes = True\n node_tree = material_class.node_tree\n\n # remove default nodes\n material_class.node_tree.nodes.clear()\n\n # add new nodes \n node_1 = node_tree.nodes.new('ShaderNodeOutputMaterial')\n node_2= node_tree.nodes.new('ShaderNodeBrightContrast')\n\n # link nodes\n node_tree.links.new(node_1.inputs[0], node_2.outputs[0])\n node_2.inputs[0].default_value = (1, 1, 1, 1)\n self.my_material[material_name] = material_class\n\n\n for i in range(num):\n class_name = str(i + 1)\n # set the material of background \n material_name = \"mask_\" + class_name\n\n # test if material exists\n # if it does not exist, create it:\n material_class = (bpy.data.materials.get(material_name) or \n bpy.data.materials.new(material_name))\n\n # enable 'Use nodes'\n material_class.use_nodes = True\n node_tree = material_class.node_tree\n\n # remove default nodes\n material_class.node_tree.nodes.clear()\n\n # add new nodes \n node_1 = node_tree.nodes.new('ShaderNodeOutputMaterial')\n node_2= node_tree.nodes.new('ShaderNodeBrightContrast')\n\n # link nodes\n node_tree.links.new(node_1.inputs[0], node_2.outputs[0])\n\n if class_name.split('_')[0] == 'background' or class_name.split('_')[0] == 'table' or class_name.split('_')[0] == 'tableplane':\n node_2.inputs[0].default_value = (1, 1, 1, 1)\n else:\n node_2.inputs[0].default_value = ((i + 1)/255., 0., 0., 1)\n\n self.my_material[material_name] = material_class\n\n\n def addNOCSMaterial(self):\n material_name = 'coord_color'\n mat = (bpy.data.materials.get(material_name) or bpy.data.materials.new(material_name))\n\n mat.use_nodes = True\n node_tree = mat.node_tree\n nodes = node_tree.nodes\n nodes.clear() \n\n links = node_tree.links\n links.clear()\n\n vcol_R = nodes.new(type=\"ShaderNodeVertexColor\")\n vcol_R.layer_name = \"Col_R\" # the vertex color layer name\n vcol_G = nodes.new(type=\"ShaderNodeVertexColor\")\n vcol_G.layer_name = \"Col_G\" # the vertex color layer name\n vcol_B = nodes.new(type=\"ShaderNodeVertexColor\")\n vcol_B.layer_name = \"Col_B\" # the vertex color layer name\n\n node_Output = node_tree.nodes.new('ShaderNodeOutputMaterial')\n node_Emission = node_tree.nodes.new('ShaderNodeEmission')\n node_LightPath = node_tree.nodes.new('ShaderNodeLightPath')\n node_Mix = node_tree.nodes.new('ShaderNodeMixShader')\n node_Combine = node_tree.nodes.new(type=\"ShaderNodeCombineRGB\")\n\n\n # make links\n node_tree.links.new(vcol_R.outputs[1], node_Combine.inputs[0])\n node_tree.links.new(vcol_G.outputs[1], node_Combine.inputs[1])\n node_tree.links.new(vcol_B.outputs[1], node_Combine.inputs[2])\n node_tree.links.new(node_Combine.outputs[0], node_Emission.inputs[0])\n\n node_tree.links.new(node_LightPath.outputs[0], node_Mix.inputs[0])\n node_tree.links.new(node_Emission.outputs[0], node_Mix.inputs[2])\n node_tree.links.new(node_Mix.outputs[0], node_Output.inputs[0])\n\n self.my_material[material_name] = mat\n\n\n def addNormalMaterial(self):\n material_name = 'normal'\n mat = (bpy.data.materials.get(material_name) or bpy.data.materials.new(material_name))\n mat.use_nodes = True\n node_tree = mat.node_tree\n nodes = node_tree.nodes\n nodes.clear()\n \n links = node_tree.links\n links.clear()\n \n # Nodes :\n new_node = nodes.new(type='ShaderNodeMath')\n new_node.active_preview = False\n new_node.color = (0.6079999804496765, 0.6079999804496765, 0.6079999804496765)\n new_node.location = (151.59744262695312, 854.5482177734375)\n new_node.name = 'Math'\n new_node.operation = 'MULTIPLY'\n new_node.select = False\n new_node.use_clamp = False\n new_node.width = 140.0\n new_node.inputs[0].default_value = 0.5\n new_node.inputs[1].default_value = 1.0\n new_node.inputs[2].default_value = 0.0\n new_node.outputs[0].default_value = 0.0\n\n new_node = nodes.new(type='ShaderNodeLightPath')\n new_node.active_preview = False\n new_node.color = (0.6079999804496765, 0.6079999804496765, 0.6079999804496765)\n new_node.location = (602.9912719726562, 1046.660888671875)\n new_node.name = 'Light Path'\n new_node.select = False\n new_node.width = 140.0\n new_node.outputs[0].default_value = 0.0\n new_node.outputs[1].default_value = 0.0\n new_node.outputs[2].default_value = 0.0\n new_node.outputs[3].default_value = 0.0\n new_node.outputs[4].default_value = 0.0\n new_node.outputs[5].default_value = 0.0\n new_node.outputs[6].default_value = 0.0\n new_node.outputs[7].default_value = 0.0\n new_node.outputs[8].default_value = 0.0\n new_node.outputs[9].default_value = 0.0\n new_node.outputs[10].default_value = 0.0\n new_node.outputs[11].default_value = 0.0\n new_node.outputs[12].default_value = 0.0\n\n new_node = nodes.new(type='ShaderNodeOutputMaterial')\n new_node.active_preview = False\n new_node.color = (0.6079999804496765, 0.6079999804496765, 0.6079999804496765)\n new_node.is_active_output = True\n new_node.location = (1168.93017578125, 701.84033203125)\n new_node.name = 'Material Output'\n new_node.select = False\n new_node.target = 'ALL'\n new_node.width = 140.0\n new_node.inputs[2].default_value = [0.0, 0.0, 0.0]\n\n new_node = nodes.new(type='ShaderNodeBsdfTransparent')\n new_node.active_preview = False\n new_node.color = (0.6079999804496765, 0.6079999804496765, 0.6079999804496765)\n new_node.location = (731.72900390625, 721.4832763671875)\n new_node.name = 'Transparent BSDF'\n new_node.select = False\n new_node.width = 140.0\n new_node.inputs[0].default_value = [1.0, 1.0, 1.0, 1.0]\n\n new_node = nodes.new(type='ShaderNodeCombineXYZ')\n new_node.active_preview = False\n new_node.color = (0.6079999804496765, 0.6079999804496765, 0.6079999804496765)\n new_node.location = (594.4229736328125, 602.9271240234375)\n new_node.name = 'Combine XYZ'\n new_node.select = False\n new_node.width = 140.0\n new_node.inputs[0].default_value = 0.0\n new_node.inputs[1].default_value = 0.0\n new_node.inputs[2].default_value = 0.0\n new_node.outputs[0].default_value = [0.0, 0.0, 0.0]\n\n new_node = nodes.new(type='ShaderNodeMixShader')\n new_node.active_preview = False\n new_node.color = (0.6079999804496765, 0.6079999804496765, 0.6079999804496765)\n new_node.location = (992.7239990234375, 707.2142333984375)\n new_node.name = 'Mix Shader'\n new_node.select = False\n new_node.width = 140.0\n new_node.inputs[0].default_value = 0.5\n\n new_node = nodes.new(type='ShaderNodeEmission')\n new_node.active_preview = False\n new_node.color = (0.6079999804496765, 0.6079999804496765, 0.6079999804496765)\n new_node.location = (774.0802612304688, 608.2547607421875)\n new_node.name = 'Emission'\n new_node.select = False\n new_node.width = 140.0\n new_node.inputs[0].default_value = [1.0, 1.0, 1.0, 1.0]\n new_node.inputs[1].default_value = 1.0\n\n new_node = nodes.new(type='ShaderNodeSeparateXYZ')\n new_node.active_preview = False\n new_node.color = (0.6079999804496765, 0.6079999804496765, 0.6079999804496765)\n new_node.location = (-130.12167358398438, 558.1497802734375)\n new_node.name = 'Separate XYZ'\n new_node.select = False\n new_node.width = 140.0\n new_node.inputs[0].default_value = [0.0, 0.0, 0.0]\n new_node.outputs[0].default_value = 0.0\n new_node.outputs[1].default_value = 0.0\n new_node.outputs[2].default_value = 0.0\n\n new_node = nodes.new(type='ShaderNodeMath')\n new_node.active_preview = False\n new_node.color = (0.6079999804496765, 0.6079999804496765, 0.6079999804496765)\n new_node.location = (162.43240356445312, 618.8094482421875)\n new_node.name = 'Math.002'\n new_node.operation = 'MULTIPLY'\n new_node.select = False\n new_node.use_clamp = False\n new_node.width = 140.0\n new_node.inputs[0].default_value = 0.5\n new_node.inputs[1].default_value = 1.0\n new_node.inputs[2].default_value = 0.0\n new_node.outputs[0].default_value = 0.0\n\n new_node = nodes.new(type='ShaderNodeMath')\n new_node.active_preview = False\n new_node.color = (0.6079999804496765, 0.6079999804496765, 0.6079999804496765)\n new_node.location = (126.8158187866211, 364.5539855957031)\n new_node.name = 'Math.001'\n new_node.operation = 'MULTIPLY'\n new_node.select = False\n new_node.use_clamp = False\n new_node.width = 140.0\n new_node.inputs[0].default_value = 0.5\n new_node.inputs[1].default_value = -1.0\n new_node.inputs[2].default_value = 0.0\n new_node.outputs[0].default_value = 0.0\n\n new_node = nodes.new(type='ShaderNodeVectorTransform')\n new_node.active_preview = False\n new_node.color = (0.6079999804496765, 0.6079999804496765, 0.6079999804496765)\n new_node.convert_from = 'WORLD'\n new_node.convert_to = 'CAMERA'\n new_node.location = (-397.0209045410156, 594.7037353515625)\n new_node.name = 'Vector Transform'\n new_node.select = False\n new_node.vector_type = 'VECTOR'\n new_node.width = 140.0\n new_node.inputs[0].default_value = [0.5, 0.5, 0.5]\n new_node.outputs[0].default_value = [0.0, 0.0, 0.0]\n\n new_node = nodes.new(type='ShaderNodeNewGeometry')\n new_node.active_preview = False\n new_node.color = (0.6079999804496765, 0.6079999804496765, 0.6079999804496765)\n new_node.location = (-651.8067016601562, 593.0455932617188)\n new_node.name = 'Geometry'\n new_node.width = 140.0\n new_node.outputs[0].default_value = [0.0, 0.0, 0.0]\n new_node.outputs[1].default_value = [0.0, 0.0, 0.0]\n new_node.outputs[2].default_value = [0.0, 0.0, 0.0]\n new_node.outputs[3].default_value = [0.0, 0.0, 0.0]\n new_node.outputs[4].default_value = [0.0, 0.0, 0.0]\n new_node.outputs[5].default_value = [0.0, 0.0, 0.0]\n new_node.outputs[6].default_value = 0.0\n new_node.outputs[7].default_value = 0.0\n new_node.outputs[8].default_value = 0.0\n\n # Links :\n\n links.new(nodes[\"Light Path\"].outputs[0], nodes[\"Mix Shader\"].inputs[0]) \n links.new(nodes[\"Separate XYZ\"].outputs[0], nodes[\"Math\"].inputs[0]) \n links.new(nodes[\"Separate XYZ\"].outputs[1], nodes[\"Math.002\"].inputs[0]) \n links.new(nodes[\"Separate XYZ\"].outputs[2], nodes[\"Math.001\"].inputs[0]) \n links.new(nodes[\"Vector Transform\"].outputs[0], nodes[\"Separate XYZ\"].inputs[0]) \n links.new(nodes[\"Combine XYZ\"].outputs[0], nodes[\"Emission\"].inputs[0]) \n links.new(nodes[\"Math\"].outputs[0], nodes[\"Combine XYZ\"].inputs[0]) \n links.new(nodes[\"Math.002\"].outputs[0], nodes[\"Combine XYZ\"].inputs[1]) \n links.new(nodes[\"Math.001\"].outputs[0], nodes[\"Combine XYZ\"].inputs[2]) \n links.new(nodes[\"Transparent BSDF\"].outputs[0], nodes[\"Mix Shader\"].inputs[1]) \n links.new(nodes[\"Emission\"].outputs[0], nodes[\"Mix Shader\"].inputs[2]) \n links.new(nodes[\"Mix Shader\"].outputs[0], nodes[\"Material Output\"].inputs[0]) \n links.new(nodes[\"Geometry\"].outputs[1], nodes[\"Vector Transform\"].inputs[0]) \n\n self.my_material[material_name] = mat\n\n def addMaterialLib(self, material_class_instance_pairs):\n for mat in bpy.data.materials:\n name = mat.name\n name_class = str(name.split('_')[0])\n if name_class != 'Dots Stroke' and name_class != 'default': \n if name_class not in self.my_material:\n self.my_material[name_class] = [mat]\n else:\n self.my_material[name_class].append(mat) # e.g. self.my_material['metal'] = [.....]\n ###\n\n def setCamera(self, quaternion, translation, fov, baseline_distance):\n self.camera_l.data.angle = fov\n self.camera_r.data.angle = self.camera_l.data.angle\n cx = translation[0]\n cy = translation[1]\n cz = translation[2]\n\n self.camera_l.location[0] = cx\n self.camera_l.location[1] = cy \n self.camera_l.location[2] = cz\n\n self.camera_l.rotation_mode = 'QUATERNION'\n self.camera_l.rotation_quaternion[0] = quaternion[0]\n self.camera_l.rotation_quaternion[1] = quaternion[1]\n self.camera_l.rotation_quaternion[2] = quaternion[2]\n self.camera_l.rotation_quaternion[3] = quaternion[3]\n\n self.camera_r.rotation_mode = 'QUATERNION'\n self.camera_r.rotation_quaternion[0] = quaternion[0]\n self.camera_r.rotation_quaternion[1] = quaternion[1]\n self.camera_r.rotation_quaternion[2] = quaternion[2]\n self.camera_r.rotation_quaternion[3] = quaternion[3]\n cx, cy, cz = cameraLPosToCameraRPos(quaternion, (cx, cy, cz), baseline_distance)\n self.camera_r.location[0] = cx\n self.camera_r.location[1] = cy \n self.camera_r.location[2] = cz\n\n\n def setLighting(self):\n # emitter \n #self.light_emitter.location = self.camera_r.location\n self.light_emitter.location = self.camera_l.location + 0.51 * (self.camera_r.location - self.camera_l.location)\n self.light_emitter.rotation_mode = 'QUATERNION'\n self.light_emitter.rotation_quaternion = self.camera_r.rotation_quaternion\n\n # emitter setting\n bpy.context.view_layer.objects.active = None\n # bpy.ops.object.select_all(action=\"DESELECT\")\n self.render_context.engine = 'CYCLES'\n self.light_emitter.select_set(True)\n self.light_emitter.data.use_nodes = True\n self.light_emitter.data.type = \"POINT\"\n self.light_emitter.data.shadow_soft_size = 0.001\n random_energy = random.uniform(LIGHT_EMITTER_ENERGY * 0.9, LIGHT_EMITTER_ENERGY * 1.1)\n self.light_emitter.data.energy = random_energy\n\n # remove default node\n light_emitter = bpy.data.objects[\"light_emitter\"].data\n light_emitter.node_tree.nodes.clear()\n\n # add new nodes\n light_output = light_emitter.node_tree.nodes.new(\"ShaderNodeOutputLight\")\n node_1 = light_emitter.node_tree.nodes.new(\"ShaderNodeEmission\")\n node_2 = light_emitter.node_tree.nodes.new(\"ShaderNodeTexImage\")\n node_3 = light_emitter.node_tree.nodes.new(\"ShaderNodeMapping\")\n node_4 = light_emitter.node_tree.nodes.new(\"ShaderNodeVectorMath\")\n node_5 = light_emitter.node_tree.nodes.new(\"ShaderNodeSeparateXYZ\")\n node_6 = light_emitter.node_tree.nodes.new(\"ShaderNodeTexCoord\")\n\n # link nodes\n light_emitter.node_tree.links.new(light_output.inputs[0], node_1.outputs[0])\n light_emitter.node_tree.links.new(node_1.inputs[0], node_2.outputs[0])\n light_emitter.node_tree.links.new(node_2.inputs[0], node_3.outputs[0])\n light_emitter.node_tree.links.new(node_3.inputs[0], node_4.outputs[0])\n light_emitter.node_tree.links.new(node_4.inputs[0], node_6.outputs[1])\n light_emitter.node_tree.links.new(node_4.inputs[1], node_5.outputs[2])\n light_emitter.node_tree.links.new(node_5.inputs[0], node_6.outputs[1])\n\n # set parameter of nodes\n node_1.inputs[1].default_value = 1.0 # scale\n node_2.extension = 'CLIP'\n # node_2.interpolation = 'Cubic'\n\n node_3.inputs[1].default_value[0] = 0.5\n node_3.inputs[1].default_value[1] = 0.5\n node_3.inputs[1].default_value[2] = 0\n node_3.inputs[2].default_value[0] = 0\n node_3.inputs[2].default_value[1] = 0\n node_3.inputs[2].default_value[2] = 0.05\n\n # scale of pattern\n node_3.inputs[3].default_value[0] = 0.6\n node_3.inputs[3].default_value[1] = 0.85\n node_3.inputs[3].default_value[2] = 0\n node_4.operation = 'DIVIDE'\n\n # pattern path\n node_2.image = self.pattern\n\n\n def lightModeSelect(self, light_mode):\n if light_mode == \"RGB\":\n self.light_emitter.hide_render = True\n ###\n bpy.data.worlds[\"World\"].node_tree.nodes[\"Background\"].inputs[1].default_value = self.src_energy_for_rgb_render\n\n elif light_mode == \"IR\":\n self.light_emitter.hide_render = False\n # set the environment map energy\n random_energy = random.uniform(LIGHT_ENV_MAP_ENERGY_IR * 0.8, LIGHT_ENV_MAP_ENERGY_IR * 1.2)\n bpy.data.worlds[\"World\"].node_tree.nodes[\"Background\"].inputs[1].default_value = random_energy\n \n elif light_mode == \"Mask\" or light_mode == \"NOCS\" or light_mode == \"Normal\":\n self.light_emitter.hide_render = True\n bpy.data.worlds[\"World\"].node_tree.nodes[\"Background\"].inputs[1].default_value = 0\n else:\n raise NotImplementedError \n\n\n def outputModeSelect(self, output_mode):\n if output_mode == \"RGB\":\n self.render_context.image_settings.file_format = 'PNG'\n self.render_context.image_settings.compression = 0\n self.render_context.image_settings.color_mode = 'RGB'\n self.render_context.image_settings.color_depth = '8'\n bpy.context.scene.view_settings.view_transform = 'Filmic'\n bpy.context.scene.render.filter_size = 1.5\n self.render_context.resolution_x = 640 ### 1280\n self.render_context.resolution_y = 360 ### 720\n elif output_mode == \"IR\":\n self.render_context.image_settings.file_format = 'PNG'\n self.render_context.image_settings.compression = 0\n self.render_context.image_settings.color_mode = 'BW'\n self.render_context.image_settings.color_depth = '8'\n bpy.context.scene.view_settings.view_transform = 'Filmic'\n bpy.context.scene.render.filter_size = 1.5\n self.render_context.resolution_x = 640 ### 1280\n self.render_context.resolution_y = 360 ### 720\n elif output_mode == \"Mask\":\n self.render_context.image_settings.file_format = 'OPEN_EXR'\n self.render_context.image_settings.color_mode = 'RGB'\n bpy.context.scene.view_settings.view_transform = 'Raw'\n bpy.context.scene.render.filter_size = 0\n self.render_context.resolution_x = 640\n self.render_context.resolution_y = 360\n elif output_mode == \"NOCS\":\n # self.render_context.image_settings.file_format = 'OPEN_EXR'\n self.render_context.image_settings.file_format = 'PNG' \n self.render_context.image_settings.color_mode = 'RGB'\n self.render_context.image_settings.color_depth = '8'\n bpy.context.scene.view_settings.view_transform = 'Raw'\n bpy.context.scene.render.filter_size = 0\n self.render_context.resolution_x = 640\n self.render_context.resolution_y = 360\n elif output_mode == \"Normal\":\n self.render_context.image_settings.file_format = 'OPEN_EXR'\n self.render_context.image_settings.color_mode = 'RGB'\n bpy.context.scene.view_settings.view_transform = 'Raw'\n bpy.context.scene.render.filter_size = 1.5\n self.render_context.resolution_x = 640\n self.render_context.resolution_y = 360\n else:\n raise NotImplementedError\n\n def renderEngineSelect(self, engine_mode):\n\n if engine_mode == \"CYCLES\":\n self.render_context.engine = 'CYCLES'\n bpy.context.scene.cycles.progressive = 'BRANCHED_PATH'\n bpy.context.scene.cycles.use_denoising = True\n bpy.context.scene.cycles.denoiser = 'NLM'\n bpy.context.scene.cycles.film_exposure = 1.0\n bpy.context.scene.cycles.aa_samples = CYCLES_SAMPLE\n\n ## Set the device_type\n bpy.context.preferences.addons[\"cycles\"].preferences.compute_device_type = \"CUDA\" # or \"OPENCL\"\n\n ## get_devices() to let Blender detects GPU device\n cuda_devices, _ = bpy.context.preferences.addons[\"cycles\"].preferences.get_devices()\n #print(bpy.context.preferences.addons[\"cycles\"].preferences.compute_device_type)\n for d in bpy.context.preferences.addons[\"cycles\"].preferences.devices:\n d[\"use\"] = 1 # Using all devices, include GPU and CPU\n #print(d[\"name\"], d[\"use\"])\n device_list = self.DEVICE_LIST\n activated_gpus = []\n for i, device in enumerate(cuda_devices):\n if (i in device_list):\n device.use = True\n activated_gpus.append(device.name)\n else:\n device.use = False\n\n\n elif engine_mode == \"EEVEE\":\n bpy.context.scene.render.engine = 'BLENDER_EEVEE'\n else:\n print(\"Not support the mode!\") \n\n\n def addBackground(self, size, position, scale, default_background_texture_path):\n # set the material of background \n material_name = \"default_background\"\n\n # test if material exists\n # if it does not exist, create it:\n material_background = (bpy.data.materials.get(material_name) or \n bpy.data.materials.new(material_name))\n\n # enable 'Use nodes'\n material_background.use_nodes = True\n node_tree = material_background.node_tree\n\n # remove default nodes\n material_background.node_tree.nodes.clear()\n\n # add new nodes \n node_1 = node_tree.nodes.new('ShaderNodeOutputMaterial')\n node_2 = node_tree.nodes.new('ShaderNodeBsdfPrincipled')\n node_3 = node_tree.nodes.new('ShaderNodeTexImage')\n\n # link nodes\n node_tree.links.new(node_1.inputs[0], node_2.outputs[0])\n node_tree.links.new(node_2.inputs[0], node_3.outputs[0])\n\n # add texture image\n node_3.image = bpy.data.images.load(filepath=default_background_texture_path)\n self.my_material['default_background'] = material_background\n\n # add background plane\n for i in range(-2, 3, 1):\n for j in range(-2, 3, 1):\n position_i_j = (i * size + position[0], j * size + position[1], position[2] - TABLE_CAD_MODEL_HEIGHT)\n bpy.ops.mesh.primitive_plane_add(size=size, enter_editmode=False, align='WORLD', location=position_i_j, scale=scale)\n bpy.ops.rigidbody.object_add()\n bpy.context.object.rigid_body.type = 'PASSIVE'\n bpy.context.object.rigid_body.collision_shape = 'BOX'\n for i in range(-2, 3, 1):\n for j in [-2, 2]:\n position_i_j = (i * size + position[0], j * size + position[1], position[2] - 0.25)# - TABLE_CAD_MODEL_HEIGHT)\n rotation_elur = (math.pi / 2., 0., 0.)\n bpy.ops.mesh.primitive_plane_add(size=size, enter_editmode=False, align='WORLD', location=position_i_j, rotation = rotation_elur)\n bpy.ops.rigidbody.object_add()\n bpy.context.object.rigid_body.type = 'PASSIVE'\n bpy.context.object.rigid_body.collision_shape = 'BOX' \n for j in range(-2, 3, 1):\n for i in [-2, 2]:\n position_i_j = (i * size + position[0], j * size + position[1], position[2] - 0.25)# - TABLE_CAD_MODEL_HEIGHT)\n rotation_elur = (0, math.pi / 2, 0)\n bpy.ops.mesh.primitive_plane_add(size=size, enter_editmode=False, align='WORLD', location=position_i_j, rotation = rotation_elur)\n bpy.ops.rigidbody.object_add()\n bpy.context.object.rigid_body.type = 'PASSIVE'\n bpy.context.object.rigid_body.collision_shape = 'BOX' \n count = 0\n for obj in bpy.data.objects:\n if obj.type == \"MESH\":\n obj.name = \"background_\" + str(count)\n obj.data.name = \"background_\" + str(count)\n obj.active_material = material_background\n count += 1\n\n self.background_added = True\n\n\n def clearModel(self):\n '''\n # delete all meshes\n for item in bpy.data.meshes:\n bpy.data.meshes.remove(item)\n for item in bpy.data.materials:\n bpy.data.materials.remove(item)\n '''\n\n # remove all objects except background\n for obj in bpy.data.objects:\n if obj.type == 'MESH' and not obj.name.split('_')[0] == 'background':\n bpy.data.meshes.remove(obj.data)\n for obj in bpy.data.objects:\n if obj.type == 'MESH' and not obj.name.split('_')[0] == 'background':\n bpy.data.objects.remove(obj, do_unlink=True)\n\n # remove all default material\n for mat in bpy.data.materials:\n name = mat.name.split('.')\n if name[0] == 'Material':\n bpy.data.materials.remove(mat)\n\n\n def loadModel(self, file_path):\n self.model_loaded = True\n try:\n if file_path.endswith('obj'):\n bpy.ops.import_scene.obj(filepath=file_path)\n elif file_path.endswith('3ds'):\n bpy.ops.import_scene.autodesk_3ds(filepath=file_path)\n elif file_path.endswith('dae'):\n # Must install OpenCollada. Please read README.md\n bpy.ops.wm.collada_import(filepath=file_path)\n else:\n self.model_loaded = False\n raise Exception(\"Loading failed: %s\" % (file_path))\n except Exception:\n self.model_loaded = False\n\n\n def render(self, image_name=\"tmp\", image_path=RENDERING_PATH):\n # Render the object\n if not self.model_loaded:\n print(\"[W]render: Model not loaded.\")\n return \n\n if self.render_mode == \"IR\":\n bpy.context.scene.use_nodes = False\n # set light and render mode\n self.lightModeSelect(\"IR\")\n self.outputModeSelect(\"IR\")\n self.renderEngineSelect(\"CYCLES\")\n\n elif self.render_mode == 'RGB':\n bpy.context.scene.use_nodes = False\n # set light and render mode\n self.lightModeSelect(\"RGB\")\n self.outputModeSelect(\"RGB\")\n self.renderEngineSelect(\"CYCLES\")\n\n elif self.render_mode == \"Mask\":\n bpy.context.scene.use_nodes = False\n # set light and render mode\n self.lightModeSelect(\"Mask\")\n self.outputModeSelect(\"Mask\")\n # self.renderEngineSelect(\"EEVEE\")\n self.renderEngineSelect(\"CYCLES\")\n bpy.context.scene.cycles.use_denoising = False\n bpy.context.scene.cycles.aa_samples = 1\n\n elif self.render_mode == \"NOCS\":\n bpy.context.scene.use_nodes = False\n # set light and render mode\n self.lightModeSelect(\"NOCS\")\n self.outputModeSelect(\"NOCS\")\n # self.renderEngineSelect(\"EEVEE\")\n self.renderEngineSelect(\"CYCLES\")\n bpy.context.scene.cycles.use_denoising = False\n bpy.context.scene.cycles.aa_samples = 1\n\n elif self.render_mode == \"Normal\":\n bpy.context.scene.use_nodes = True\n self.fileOutput.base_path = image_path.replace(\"normal\",\"depth\")\n self.fileOutput.file_slots[0].path = image_name[:4]+\"_#\"# + 'depth_#'\n\n # set light and render mode\n self.lightModeSelect(\"Normal\")\n self.outputModeSelect(\"Normal\")\n # self.renderEngineSelect(\"EEVEE\")\n self.renderEngineSelect(\"CYCLES\")\n bpy.context.scene.cycles.use_denoising = False\n bpy.context.scene.cycles.aa_samples = 32\n\n else:\n print(\"[W]render: The render mode is not supported\")\n return \n\n bpy.context.scene.render.filepath = os.path.join(image_path, image_name)\n bpy.ops.render.render(write_still=True) # save straight to file\n\n\n def set_material_randomize_mode(self, class_material_pairs, mat_randomize_mode, instance, material_type_in_mixed_mode):\n if mat_randomize_mode in ['mixed','diffuse','transparent','specular_tex','specular_texmix','specular_and_transparent']:\n if material_type_in_mixed_mode == 'raw':\n print(\"[V]set_material_randomize_mode\", instance.name, 'material type: raw')\n set_modify_raw_material(instance)\n else:\n material = random.sample(self.my_material[material_type_in_mixed_mode], 1)[0]\n print(\"[V]set_material_randomize_mode\", instance.name, 'material type: ', material_type_in_mixed_mode)\n ## graspnet\n set_modify_material(instance, material, self.obj_texture_img_list, mat_randomize_mode=mat_randomize_mode)\n elif mat_randomize_mode == 'specular':\n material = random.sample(self.my_material[material_type_in_mixed_mode], 1)[0]\n print(\"[V]set_material_randomize_mode\", instance.name, 'material type: ', material_type_in_mixed_mode)\n set_modify_material(instance, material, self.obj_texture_img_list, mat_randomize_mode=mat_randomize_mode,\n is_transfer=False)\n else:\n raise NotImplementedError(\"No such mat_randomize_mode!\")\n\n\n def get_instance_pose(self):\n instance_pose = {}\n bpy.context.view_layer.update()\n cam = self.camera_l\n mat_rot_x = Matrix.Rotation(math.radians(180.0), 4, 'X')\n for obj in bpy.data.objects:\n if obj.type == 'MESH' and not obj.name.split('_')[0] == 'background':\n instance_id = obj.name.split('_')[0]\n mat_rel = cam.matrix_world.inverted() @ obj.matrix_world\n # location\n relative_location = [mat_rel.translation[0],\n - mat_rel.translation[1],\n - mat_rel.translation[2]]\n # rotation\n # relative_rotation_euler = mat_rel.to_euler() # must be converted from radians to degrees\n relative_rotation_quat = [mat_rel.to_quaternion()[0],\n mat_rel.to_quaternion()[1],\n mat_rel.to_quaternion()[2],\n mat_rel.to_quaternion()[3]]\n quat_x = [0, 1, 0, 0]\n quat = quanternion_mul(quat_x, relative_rotation_quat)\n quat = [quat[0], - quat[1], - quat[2], - quat[3]]\n instance_pose[str(instance_id)] = [quat, relative_location]\n\n return instance_pose\n\n\n def check_visible(self, threshold=(0.1, 0.9, 0.1, 0.9)):\n w_min, x_max, h_min, h_max = threshold\n visible_objects_list = []\n bpy.context.view_layer.update()\n cs, ce = self.camera_l.data.clip_start, self.camera_l.data.clip_end\n for obj in bpy.data.objects:\n if obj.type == 'MESH' and not obj.name.split('_')[0] == 'background':\n obj_center = obj.matrix_world.translation\n co_ndc = world_to_camera_view(scene, self.camera_l, obj_center)\n if (w_min < co_ndc.x < x_max and\n h_min < co_ndc.y < h_max and\n cs < co_ndc.z < ce):\n obj.select_set(True)\n visible_objects_list.append(obj)\n else:\n obj.select_set(False)\n return visible_objects_list\n\n\ndef setModelPosition(instance, location, quaternion):\n instance.rotation_mode = 'QUATERNION'\n instance.rotation_quaternion[0] = quaternion[0]\n instance.rotation_quaternion[1] = quaternion[1]\n instance.rotation_quaternion[2] = quaternion[2]\n instance.rotation_quaternion[3] = quaternion[3]\n instance.location = location\n###\n\ndef setRigidBody(instance):\n bpy.context.view_layer.objects.active = instance \n object_single = bpy.context.active_object\n\n # add rigid body constraints to cube\n bpy.ops.rigidbody.object_add()\n bpy.context.object.rigid_body.mass = 1\n bpy.context.object.rigid_body.kinematic = True\n bpy.context.object.rigid_body.collision_shape = 'CONVEX_HULL'\n bpy.context.object.rigid_body.restitution = 0.01\n bpy.context.object.rigid_body.angular_damping = 0.8\n bpy.context.object.rigid_body.linear_damping = 0.99\n\n bpy.context.object.rigid_body.kinematic = False\n object_single.keyframe_insert(data_path='rigid_body.kinematic', frame=0)\n\n\ndef set_visiable_objects(visible_objects_list):\n for obj in bpy.data.objects:\n if obj.type == 'MESH' and not obj.name.split('_')[0] == 'background':\n if obj in visible_objects_list:\n obj.hide_render = False\n else:\n obj.hide_render = True\n\n\ndef generate_CAD_model_list(scene_type, urdf_path_list, obj_uid_list):\n CAD_model_list = {}\n ###\n for idx in range(len(urdf_path_list)):\n urdf_path = urdf_path_list[idx]\n obj_uid = obj_uid_list[idx]\n class_name = 'other'\n urdf_path = str(urdf_path).replace(\"\\\\\",\"/\").split(\"/\")\n if scene_type == \"blocks\":\n instance_path = \"/\".join(urdf_path[:-1]) + \"/\" + urdf_path[-1][:-5]+\".obj\"\n else:\n instance_path = \"/\".join(urdf_path[:-1])+\"/\"+urdf_path[-1][:-5]+\"_visual.obj\"\n class_list = []\n class_list.append([instance_path, class_name, obj_uid])\n if class_name == 'other' and 'other' in CAD_model_list:\n CAD_model_list[class_name] = CAD_model_list[class_name] + class_list\n else:\n CAD_model_list[class_name] = class_list\n \n return CAD_model_list\n\n\ndef generate_material_type(obj_name, class_material_pairs, instance_material_except_pairs, instance_material_include_pairs, material_class_instance_pairs, material_type):\n ###\n specular_type_for_ins_list = []\n transparent_type_for_ins_list = []\n diffuse_type_for_ins_list = []\n for key in instance_material_except_pairs:\n if key in material_class_instance_pairs['specular']:\n specular_type_for_ins_list.append(key)\n elif key in material_class_instance_pairs['transparent']:\n transparent_type_for_ins_list.append(key)\n elif key in material_class_instance_pairs['diffuse']:\n diffuse_type_for_ins_list.append(key)\n for key in instance_material_include_pairs:\n if key in material_class_instance_pairs['specular']:\n specular_type_for_ins_list.append(key)\n elif key in material_class_instance_pairs['transparent']:\n transparent_type_for_ins_list.append(key)\n elif key in material_class_instance_pairs['diffuse']:\n diffuse_type_for_ins_list.append(key)\n\n if material_type == \"transparent\":\n return random.sample(transparent_type_for_ins_list, 1)[0]\n elif material_type == \"diffuse\":\n return random.sample(diffuse_type_for_ins_list, 1)[0]\n elif material_type == \"specular\" or material_type == \"specular_tex\" or material_type == \"specular_texmix\":\n return random.sample(specular_type_for_ins_list, 1)[0]\n elif material_type == \"specular_and_transparent\":\n flag = random.randint(0, 2)\n if flag == 0:\n return random.sample(specular_type_for_ins_list, 1)[0] ### 'specular'\n else:\n return random.sample(transparent_type_for_ins_list, 1)[0] ### 'transparent'\n elif material_type == \"mixed\":\n # randomly pick material class\n flag = random.randint(0, 2) # D:S:T=1:2:2\n # select the raw material\n if flag == 0:\n return random.sample(diffuse_type_for_ins_list, 1)[0] ### 'diffuse'\n # select one from specular and transparent\n elif flag == 1:\n return random.sample(specular_type_for_ins_list, 1)[0] ### 'specular'\n else:\n return random.sample(transparent_type_for_ins_list, 1)[0] ### 'transparent'\n else:\n raise ValueError(f\"Material type error: {material_type}\")", "repo_name": "PKU-EPIC/GraspNeRF", "sub_path": "src/rd/render_utils.py", "file_name": "render_utils.py", "file_ext": "py", "file_size_in_byte": 64167, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 92, "dataset": "github-code", "pt": "25", "api": [{"api_name": "os.getcwd", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 20, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 25, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 182, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 183, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 184, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 185, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 185, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 186, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 190, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 191, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 192, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 193, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 194, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 195, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 205, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 206, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 207, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 211, "usage_type": "call"}, {"api_name": "math.acos", "line_number": 214, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 216, "usage_type": "attribute"}, {"api_name": "math.acos", "line_number": 220, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 232, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 233, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 237, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 238, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 239, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 291, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 292, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 293, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 294, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 295, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 318, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 319, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 320, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 321, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 363, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 364, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 369, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 369, "usage_type": "call"}, {"api_name": "numpy.linalg.svd", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 370, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 371, "usage_type": "call"}, {"api_name": "numpy.linalg.det", "line_number": 373, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 373, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 377, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 384, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 385, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 386, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 387, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 388, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 391, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 392, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 394, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 397, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 397, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 398, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 402, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 402, "usage_type": "attribute"}, {"api_name": "numpy.cross", "line_number": 403, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 403, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 403, "usage_type": "attribute"}, {"api_name": "numpy.cross", "line_number": 404, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 404, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 404, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 407, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 410, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 428, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 434, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 435, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 438, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 439, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 442, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 443, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 453, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 453, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 454, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 455, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 456, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 458, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 460, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 460, "usage_type": "attribute"}, {"api_name": "numpy.cross", "line_number": 461, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 461, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 461, "usage_type": "attribute"}, {"api_name": "numpy.cross", "line_number": 462, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 462, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 462, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 465, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 468, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 485, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 487, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 488, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 489, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 501, "usage_type": "attribute"}, {"api_name": "bpy.data.meshes.remove", "line_number": 502, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 502, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 504, "usage_type": "attribute"}, {"api_name": "bpy.data.cameras.remove", "line_number": 505, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 505, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 507, "usage_type": "attribute"}, {"api_name": "bpy.data.lights.remove", "line_number": 508, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 508, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 510, "usage_type": "attribute"}, {"api_name": "bpy.data.objects.remove", "line_number": 511, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 511, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 513, "usage_type": "attribute"}, {"api_name": "bpy.data.cameras.new", "line_number": 516, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 516, "usage_type": "attribute"}, {"api_name": "bpy.data.objects.new", "line_number": 517, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 517, "usage_type": "attribute"}, {"api_name": "bpy.context.collection.objects.link", "line_number": 518, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 518, "usage_type": "attribute"}, {"api_name": "bpy.data.cameras.new", "line_number": 521, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 521, "usage_type": "attribute"}, {"api_name": "bpy.data.objects.new", "line_number": 522, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 522, "usage_type": "attribute"}, {"api_name": "bpy.context.collection.objects.link", "line_number": 523, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 523, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 525, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 526, "usage_type": "attribute"}, {"api_name": "bpy.data.lights.new", "line_number": 534, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 534, "usage_type": "attribute"}, {"api_name": "bpy.data.objects.new", "line_number": 535, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 535, "usage_type": "attribute"}, {"api_name": "bpy.context.collection.objects.link", "line_number": 536, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 536, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 538, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 568, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 569, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 570, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 571, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 575, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 579, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 580, "usage_type": "attribute"}, {"api_name": "bpy.data.images.load", "line_number": 610, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 610, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 612, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 613, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 614, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 616, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 617, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 618, "usage_type": "call"}, {"api_name": "bpy.data.images.load", "line_number": 622, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 622, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 622, "usage_type": "call"}, {"api_name": "os.path", "line_number": 622, "usage_type": "attribute"}, {"api_name": "bpy.data.images.load", "line_number": 626, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 626, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 626, "usage_type": "call"}, {"api_name": "os.path", "line_number": 626, "usage_type": "attribute"}, {"api_name": "bpy.data.images.load", "line_number": 630, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 630, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 630, "usage_type": "call"}, {"api_name": "os.path", "line_number": 630, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 632, "usage_type": "call"}, {"api_name": "os.path", "line_number": 632, "usage_type": "attribute"}, {"api_name": "bpy.data.images.load", "line_number": 636, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 636, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 636, "usage_type": "call"}, {"api_name": "os.path", "line_number": 636, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 641, "usage_type": "attribute"}, {"api_name": "random.uniform", "line_number": 674, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 675, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 681, "usage_type": "attribute"}, {"api_name": "bpy.data.materials.get", "line_number": 695, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 695, "usage_type": "attribute"}, {"api_name": "bpy.data.materials.new", "line_number": 695, "usage_type": "call"}, {"api_name": "bpy.data.materials.get", "line_number": 721, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 721, "usage_type": "attribute"}, {"api_name": "bpy.data.materials.new", "line_number": 722, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 722, "usage_type": "attribute"}, {"api_name": "bpy.data.materials.get", "line_number": 748, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 748, "usage_type": "attribute"}, {"api_name": "bpy.data.materials.new", "line_number": 748, "usage_type": "call"}, {"api_name": "bpy.data.materials.get", "line_number": 787, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 787, "usage_type": "attribute"}, {"api_name": "bpy.data.materials.new", "line_number": 787, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 971, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1017, "usage_type": "attribute"}, {"api_name": "random.uniform", "line_number": 1024, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 1028, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 1075, "usage_type": "attribute"}, {"api_name": "random.uniform", "line_number": 1080, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 1081, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 1085, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1096, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1097, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1105, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1106, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1112, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1113, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1121, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1122, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1128, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1129, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1139, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1140, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1141, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1142, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1143, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1146, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1149, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1151, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1165, "usage_type": "attribute"}, {"api_name": "bpy.data.materials.get", "line_number": 1176, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 1176, "usage_type": "attribute"}, {"api_name": "bpy.data.materials.new", "line_number": 1177, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 1177, "usage_type": "attribute"}, {"api_name": "bpy.data.images.load", "line_number": 1196, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 1196, "usage_type": "attribute"}, {"api_name": "bpy.ops.mesh.primitive_plane_add", "line_number": 1203, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 1203, "usage_type": "attribute"}, {"api_name": "bpy.ops.rigidbody.object_add", "line_number": 1204, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 1204, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1205, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1206, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 1210, "usage_type": "attribute"}, {"api_name": "bpy.ops.mesh.primitive_plane_add", "line_number": 1211, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 1211, "usage_type": "attribute"}, {"api_name": "bpy.ops.rigidbody.object_add", "line_number": 1212, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 1212, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1213, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1214, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 1218, "usage_type": "attribute"}, {"api_name": "bpy.ops.mesh.primitive_plane_add", "line_number": 1219, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 1219, "usage_type": "attribute"}, {"api_name": "bpy.ops.rigidbody.object_add", "line_number": 1220, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 1220, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1221, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1222, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 1224, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 1244, "usage_type": "attribute"}, {"api_name": "bpy.data.meshes.remove", "line_number": 1246, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 1246, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 1247, "usage_type": "attribute"}, {"api_name": "bpy.data.objects.remove", "line_number": 1249, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 1249, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 1252, "usage_type": "attribute"}, {"api_name": "bpy.data.materials.remove", "line_number": 1255, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 1255, "usage_type": "attribute"}, {"api_name": "bpy.ops.import_scene.obj", "line_number": 1262, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 1262, "usage_type": "attribute"}, {"api_name": "bpy.ops.import_scene.autodesk_3ds", "line_number": 1264, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 1264, "usage_type": "attribute"}, {"api_name": "bpy.ops.wm.collada_import", "line_number": 1267, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 1267, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1282, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1289, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1296, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1302, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1303, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1306, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1312, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1313, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1316, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1325, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1326, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1332, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1332, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1332, "usage_type": "attribute"}, {"api_name": "bpy.ops.render.render", "line_number": 1333, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 1333, "usage_type": "attribute"}, {"api_name": "rd.modify_material.set_modify_raw_material", "line_number": 1340, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 1342, "usage_type": "call"}, {"api_name": "rd.modify_material.set_modify_material", "line_number": 1345, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 1347, "usage_type": "call"}, {"api_name": "rd.modify_material.set_modify_material", "line_number": 1349, "usage_type": "call"}, {"api_name": "bpy.context.view_layer.update", "line_number": 1357, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 1357, "usage_type": "attribute"}, {"api_name": "mathutils.Matrix.Rotation", "line_number": 1359, "usage_type": "call"}, {"api_name": "mathutils.Matrix", "line_number": 1359, "usage_type": "name"}, {"api_name": "math.radians", "line_number": 1359, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 1360, "usage_type": "attribute"}, {"api_name": "bpy.context.view_layer.update", "line_number": 1385, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 1385, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 1387, "usage_type": "attribute"}, {"api_name": "bpy_extras.object_utils.world_to_camera_view", "line_number": 1390, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 1411, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1412, "usage_type": "attribute"}, {"api_name": "bpy.ops.rigidbody.object_add", "line_number": 1415, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 1415, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1416, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1417, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1418, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1419, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1420, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1421, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 1423, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 1428, "usage_type": "attribute"}, {"api_name": "random.sample", "line_number": 1479, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 1481, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 1483, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 1485, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 1487, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 1489, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 1492, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 1495, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 1498, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 1500, "usage_type": "call"}]}
+{"seq_id": "27897716078", "text": "import numpy as np\nfrom math import log\nfrom scipy.special import erf\nimport matplotlib.pyplot as plt\n\n\ndef modulate(x, constel):\n return constel[x]\n\n\ndef demodulate(x, constel):\n samples = len(x)\n result = np.zeros(x.shape)\n for i in range(samples):\n # Медленно, но работает для любой модуляции\n j = np.argmin(np.abs(constel - x[i]))\n result[i] = j\n \n return result.astype(int)\n\n\ndef norm(constel):\n dot = np.dot(constel, np.conj(constel))\n return constel / np.sqrt(dot / len(constel))\n\n\ndef constellation(M):\n if np.fix(log(M, 4)) != log(M,4):\n raise ValueError(\"M must be power of 4!\")\n\n nbits = int(np.log2(M))\n x = np.arange(M)\n\n nbitsBy2 = nbits >> 1\n symbolI = x >> nbitsBy2\n symbolQ = x & ((M-1) >> nbitsBy2)\n\n i = 1\n while i < nbitsBy2:\n tmpI = symbolI\n tmpI = tmpI >> i\n symbolI = symbolI ^ tmpI\n\n tmpQ = symbolQ\n tmpQ = tmpQ >> i\n symbolQ = symbolQ ^ tmpQ\n i = i + i\n\n gray = (symbolI << nbitsBy2) + symbolQ\n\n x = x[gray]\n c = int(np.sqrt(M))\n I = -2 * np.mod(x, c) + c - 1\n Q = 2 * np.floor(x / c) - c + 1\n IQ = I + 1j*Q\n IQ = -np.transpose(np.reshape(IQ, (c, c)))\n return norm(IQ.flatten())\n\n\ndef qfunc(x):\n return 0.5 - 0.5 * erf(x / np.sqrt(2))\n\n\n# Only for square QAM\n# Source: https://www.mathworks.com/help/comm/ug/analytical-expressions-used-in-berawgn-function-and-bit-error-rate-analysis-app.html \ndef theory_ber(EbN0, M):\n if np.fix(log(M, 4)) != log(M,4):\n raise ValueError(\"M must be power of 4!\")\n \n if M == 4:\n ber = qfunc(np.sqrt(2*EbN0))\n elif M == 16:\n ber = 3/4*qfunc(np.sqrt(4/5*EbN0)) \n + 1/2*qfunc(3*np.sqrt(4/5*EbN0)) \n - 1/4*qfunc(5*np.sqrt(4/5*EbN0))\n elif M == 64:\n ber = 7/12*qfunc(np.sqrt(2/7*EbN0)) \n + 1/2*qfunc(3*np.sqrt(2/7*EbN0)) \n - 1/12*qfunc(5*np.sqrt(2/7*EbN0)) \n + 1/12*qfunc(9*np.sqrt(2/7*EbN0)) \n - 1/12*qfunc(13*np.sqrt(2/7*EbN0))\n else:\n k = np.log2(M)\n c = np.sqrt(M)\n ber = np.zeros(EbN0.shape)\n for i in range(1, round(np.log2(c)) + 1):\n berk = np.zeros(EbN0.shape)\n for j in range(0,round((1-2**(-i))*c)):\n berk = berk + (-1)**(np.floor(j*2**(i-1)/c)) * (2**(i-1) \n - np.floor(j*2**(i-1)/c+1/2)) * qfunc((2*j+1) * np.sqrt(6*k*EbN0/(2*(M-1))))\n berk = berk * 2 / c\n ber = ber + berk\n\n ber = ber / np.log2(c)\n\n return ber\n\n\ndef plot_constel(iq):\n count = len(iq)\n bits = np.log2(count)\n spec = '#0{}b'.format((bits + 2).astype(int))\n\n plt.figure(dpi = 80)\n for n in range(count):\n factor = np.log2(np.sqrt(count))\n d = 0.04 / factor\n i = np.real(iq[n])\n q = np.imag(iq[n])\n label = format(n, spec)\n plt.text(i + d, q + d, label[2::], fontsize = 30 / factor)\n\n scale = np.max(np.abs(iq))\n plt.scatter(np.real(iq), np.imag(iq), s = 120 / factor)\n plt.title('QAM{} constellation'.format(count))\n plt.xlim([-scale, scale])\n plt.ylim([-scale, scale])\n plt.xlabel('I')\n plt.ylabel('Q')\n plt.grid()\n plt.show()\n\n\ndef add_ber_plot(EbN0, OSNR, BER, spec='--'):\n plt.subplot(1,2,1)\n plt.semilogy(EbN0, BER, spec, markersize=8, linewidth=3)\n plt.xlabel('Eb/N0, dB')\n plt.ylabel('BER')\n plt.grid(b=True, which='major', color='k', linestyle='-')\n plt.grid(b=True, which='minor', color='k', linestyle='--')\n\n plt.subplot(1,2,2)\n plt.semilogy(OSNR, BER, spec, markersize=8, linewidth=3)\n plt.xlabel('OSNR, dB')\n plt.ylabel('BER')\n plt.grid(b=True, which='major', color='k', linestyle='-')\n plt.grid(b=True, which='minor', color='k', linestyle='--')\n\n\ndef hamming(str1,str2):\n result=0\n for _,(i,j) in enumerate(zip(str1, str2)):\n if i!=j:\n result+=1\n return result\n\n\ndef ber(input, output, M):\n bits_per_symbol = np.log2(M)\n symbols = len(input)\n diff = []\n for i in range(symbols):\n in_bits = format(input[i], '016b')\n out_bits = format(output[i], '016b')\n diff.append(hamming(in_bits, out_bits))\n\n return np.sum(diff) / (symbols * bits_per_symbol)\n\n\ndef save_ber(OSNR_dB, EbN0_db, BER, M):\n data = {\n 'M' : M,\n 'OSNR_dB' : OSNR_dB,\n 'EbN0_db' : EbN0_db,\n 'BER' : BER\n }\n np.save('QAM{}.npy'.format(M), data)", "repo_name": "Dirog/qam", "sub_path": "qam.py", "file_name": "qam.py", "file_ext": "py", "file_size_in_byte": 4440, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "numpy.zeros", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.conj", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.fix", "line_number": 28, "usage_type": "call"}, {"api_name": "math.log", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.log2", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.mod", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 56, "usage_type": "call"}, {"api_name": "scipy.special.erf", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.fix", "line_number": 67, "usage_type": "call"}, {"api_name": "math.log", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.log2", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.log2", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.log2", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.log2", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "numpy.log2", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.real", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.imag", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.text", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "numpy.real", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.imag", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 119, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.semilogy", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.semilogy", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "numpy.log2", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 167, "usage_type": "call"}]}
+{"seq_id": "8276734940", "text": "import praw\nimport pandas as pd\nimport requests\nfrom textblob import TextBlob\nimport plotly.express as px\n\ndef get_sentiment(text):\n blob = TextBlob(text)\n sentiment = blob.sentiment.polarity\n return sentiment\n\n\ndef get_posts(subreddit, limit):\n posts = []\n for post in subreddit.hot(limit=limit):\n posts.append([post.title, post.score, post.id, post.subreddit, post.url, post.num_comments, post.selftext, post.created])\n return pd.DataFrame(posts, columns=['title', 'score', 'id', 'subreddit', 'url', 'num_comments', 'body', 'created'])\n\n\nurl = \"https://www.sec.gov/files/company_tickers.json\"\nresponse = requests.get(url)\n\ndata = response.json()\ntickers = [info['ticker'] for info in data.values()]\ncompanies = {info['ticker']: info['title'] for info in data.values()}\n\n\nreddit = praw.Reddit(client_id='######',\n client_secret='########',\n user_agent='#######')\n\nsubreddit = reddit.subreddit('wallstreetbets')\nposts = get_posts(subreddit, 1000)\n\nposts['sentiment'] = posts['body'].apply(get_sentiment)\n\nsentiments = {ticker: [] for ticker in tickers}\nfor ticker in tickers:\n relevant_posts = posts[posts['body'].str.contains(ticker)]\n sentiments[ticker] = list(relevant_posts['sentiment'])\n\nsentiments_df = pd.DataFrame(list(sentiments.items()), columns=['Ticker', 'SentimentList'])\n\nsentiments_df['NumPosts'] = sentiments_df['SentimentList'].apply(len)\nsentiments_df = sentiments_df[sentiments_df['NumPosts'] >= 5]\n\nsentiments_df['Sentiment'] = sentiments_df['SentimentList'].apply(lambda x: sum(x) / len(x))\n\nsentiments_df = sentiments_df.sort_values('Sentiment')\n\nfig = px.bar(sentiments_df, x='Ticker', y='Sentiment', \n title='Sentiment Analysis of Stocks on r/wallstreetbets',\n labels={'Sentiment': 'Sentiment', 'Ticker': 'Stock'},\n hover_data=['Ticker'])\n\nfig.show()", "repo_name": "dhruvnanavati/Portfolio", "sub_path": "College/Python/wsb_crawler.py", "file_name": "wsb_crawler.py", "file_ext": "py", "file_size_in_byte": 1881, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "textblob.TextBlob", "line_number": 8, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 17, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 21, "usage_type": "call"}, {"api_name": "praw.Reddit", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 42, "usage_type": "call"}, {"api_name": "plotly.express.bar", "line_number": 51, "usage_type": "call"}, {"api_name": "plotly.express", "line_number": 51, "usage_type": "name"}]}
+{"seq_id": "21590752747", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport cv2\nfrom matplotlib.patches import Circle\nimport cv22_lab1_part3_utils as p3\n\ndef LoG(x, y, s):\n nom = y**2 + x**2 - 2*s**2 \n denom = 2*np.pi*(s**6)\n expo = np.exp(-(x**2 + y**2)/(2*s**2))\n return nom*expo/denom\n\ndef disk_strel(n):\n r = int(np.round(n))\n d = 2*r+1\n x = np.arange(d) - r\n y = np.arange(d) - r\n x, y = np.meshgrid(x,y)\n strel = x**2 + y**2 <= r**2\n return strel.astype(np.uint8)\n\ndef interest_points_visualization(I_, kp_data_, ax=None):\n I = np.array(I_)\n kp_data = np.array(kp_data_)\n assert(len(I.shape) == 2 or (len(I.shape) == 3 and I.shape[2] == 3))\n assert(len(kp_data.shape) == 2 and kp_data.shape[1] == 3)\n\n if ax is None:\n _, ax = plt.subplots()\n\n ax.set_aspect('equal')\n ax.imshow(I, 'gray')\n ax.tick_params(bottom=False, left=False, labelbottom=False, labelleft=False)\n\n for i in range(len(kp_data)):\n x, y, sigma = kp_data[i]\n circ = Circle((x, y), 3*sigma, edgecolor='g', fill=False, linewidth=1)\n ax.add_patch(circ)\n\n return ax\n\ndef EdgeDetect(img, s = 2, theta = 0.1, linear = False):\n n = int(np.ceil(3*s)*2 + 1)\n B = np.array([\n [0,1,0],\n [1,1,1],\n [0,1,0]\n ], dtype=np.uint8)\n G1D = cv2.getGaussianKernel(n, s)\n G2D = G1D @ G1D.T\n smoothed_img = cv2.filter2D(img, -1, G2D)\n \n if linear == True:\n kernel = np.zeros((n,n))\n x = y = np.linspace(-(n-1)/2, (n-1)/2, n)\n x, y = np.meshgrid(x, y)\n kernel = LoG(x, y, s)\n \n# =============================================================================\n# print (kernel)\n# plt.figure()\n# ax = plt.axes(projection='3d')\n# ax.plot_surface(x, y, kernel)\n# ax.set_xlabel('x')\n# ax.set_ylabel('y')\n# ax.set_zlabel('z')\n# =============================================================================\n \n laplacian = cv2.filter2D(img, -1, kernel)\n \n# =============================================================================\n# plt.figure()\n# plt.imshow(laplacian, 'gray')\n# =============================================================================\n \n else:\n dilated_img = cv2.dilate(smoothed_img, B)\n eroded_img = cv2.erode(smoothed_img, B)\n \n laplacian = dilated_img + eroded_img - 2*smoothed_img\n \n# =============================================================================\n# plt.figure()\n# plt.imshow(laplacian, 'gray')\n# =============================================================================\n \n _, binary_img = cv2.threshold(laplacian, 0, 1, cv2.THRESH_BINARY)\n y = cv2.dilate(binary_img, B) - cv2.erode(binary_img, B)\n ix, iy = np.gradient(smoothed_img)\n ig = np.sqrt(ix**2 + iy**2)\n max_g = ig.max()\n y_final = np.zeros((y.shape[0], y.shape[1]))\n \n for i in range(y.shape[0]):\n for j in range(y.shape[1]):\n if y[i][j] == 1 and ig[i][j] > theta*max_g:\n y_final[i][j] = 1\n\n return y_final\n\ndef CornerDetect(img, s = 2, r = 2.5, k = 0.05, theta = 0.005):\n n = int(np.ceil(3*s)*2 + 1)\n \n G1D = cv2.getGaussianKernel(n, s)\n G2D = G1D @ G1D.T\n r1D = cv2.getGaussianKernel(n, r)\n r2D = r1D @ r1D.T\n \n smoothed_img = cv2.filter2D(img, -1, G2D)\n ix, iy = np.gradient(smoothed_img)\n \n j1 = cv2.filter2D(ix*ix, -1, r2D)\n j2 = cv2.filter2D(ix*iy, -1, r2D)\n j3 = cv2.filter2D(iy*iy, -1, r2D)\n\n l1 = (j1 + j3 + np.sqrt((j1 - j3)*(j1 - j3) + 4*j2*j2))/2\n l2 = (j1 + j3 - np.sqrt((j1 - j3)*(j1 - j3) + 4*j2*j2))/2\n \n# =============================================================================\n# plt.figure()\n# plt.imshow(l1, 'gray')\n# plt.figure()\n# plt.imshow(l2, 'gray')\n# =============================================================================\n \n R = l1*l2 - k*(l1 + l2)*(l1 + l2)\n B_sq = disk_strel(n)\n Cond1 = (R==cv2.dilate(R,B_sq))\n Cond2 = (R > theta*R.max())\n\n R = Cond1*Cond2\n \n x = []\n y = []\n \n for i in range(R.shape[0]):\n for j in range(R.shape[1]):\n if R[i][j]:\n x.append(int(j))\n y.append(int(i))\n \n S = len(x)*[s]\n interest_points = np.array([x, y, S]).T\n \n return interest_points\n\ndef MultiscaleCornerDetect(img, s1 = 2, r = 2.5, k = 0.05, theta = 0.005, s2 = 1.5, N = 4):\n LoG = []\n i_points = []\n s = []\n for i in range(N):\n s.append(s1*s2**i)\n r = r*s2**i\n \n interest_points = CornerDetect(img, s[i], r, k, theta)\n i_points.append(interest_points)\n \n n = int(np.ceil(3*s[i])*2 + 1)\n \n G1D = cv2.getGaussianKernel(n, s[i])\n G2D = G1D @ G1D.T\n \n smoothed_img = cv2.filter2D(img, -1, G2D)\n ix, iy = np.gradient(smoothed_img)\n ixx, _ = np.gradient(ix)\n _, iyy = np.gradient(iy)\n LoG.append(s[i]**2*np.abs(ixx + iyy))\n \n interest_points = []\n for i, scale_points in enumerate(i_points):\n for points in scale_points:\n x = int(points[0])\n y = int(points[1])\n s = points[2]\n if i == 0 and LoG[i][y][x] > LoG[i+1][y][x]:\n interest_points.append([x, y, s])\n elif i == N-1 and LoG[i][y][x] > LoG[i-1][y][x]:\n interest_points.append([x, y, s])\n elif LoG[i][y][x] > LoG[i-1][y][x] and LoG[i][y][x] > LoG[i+1][y][x]:\n interest_points.append([x, y, s])\n return np.array(interest_points)\n \ndef BlobDetect(img, s = 2, theta = 0.005):\n n = int(np.ceil(3*s)*2 + 1)\n \n G1D = cv2.getGaussianKernel(n, s)\n G2D = G1D @ G1D.T\n \n smoothed_img = cv2.filter2D(img, -1, G2D)\n ix, iy = np.gradient(smoothed_img)\n ixx, ixy = np.gradient(ix)\n _, iyy = np.gradient(iy)\n \n R = ixx*iyy - ixy*ixy\n \n B_sq = disk_strel(n)\n Cond1 = (R==cv2.dilate(R,B_sq))\n Cond2 = (R > theta*R.max())\n \n R = Cond1*Cond2\n \n x = []\n y = []\n \n for i in range(R.shape[0]):\n for j in range(R.shape[1]):\n if R[i][j]:\n x.append(int(j))\n y.append(int(i))\n \n S = len(x)*[s]\n interest_points = np.array([x, y, S]).T\n \n return interest_points\n\ndef MultiscaleBlobDetect(img, s1 = 2, theta = 0.005, s2 = 1.5, N = 4):\n LoG = []\n i_points = []\n s = []\n for i in range(N):\n s.append(s1*s2**i)\n \n blobs = BlobDetect(img, s[i], theta)\n i_points.append(blobs)\n \n n = int(np.ceil(3*s[i])*2 + 1)\n \n G1D = cv2.getGaussianKernel(n, s[i])\n G2D = G1D @ G1D.T\n \n smoothed_img = cv2.filter2D(img, -1, G2D)\n ix, iy = np.gradient(smoothed_img)\n ixx, _ = np.gradient(ix)\n _, iyy = np.gradient(iy)\n \n LoG.append(s[i]**2*np.abs(ixx + iyy))\n \n blobs = []\n for i, scale_points in enumerate(i_points):\n for points in scale_points:\n x = int(points[0])\n y = int(points[1])\n s = points[2]\n if i == 0 and LoG[i][y][x] > LoG[i+1][y][x]:\n blobs.append([x, y, s])\n elif i == N-1 and LoG[i][y][x] > LoG[i-1][y][x]:\n blobs.append([x, y, s])\n elif LoG[i][y][x] > LoG[i-1][y][x] and LoG[i][y][x] > LoG[i+1][y][x]:\n blobs.append([x, y, s])\n return np.array(blobs)\n\ndef BoxFilter(int_img, s):\n n = int(np.ceil(3*s)*2 + 1)\n \n h = 4*np.floor(n/6) + 1\n \n if h%3 != 0:\n h = h + 3 - h%3\n if (h/3) % 2 == 0:\n h += 3\n \n w = 2*np.floor(n/6) + 1\n \n padded_int_img = np.pad(int_img, (int((h-1)/2), int((h-1)/2))) \n \n x = padded_int_img - np.roll(padded_int_img, int((h-1)/2-1), axis = 0) - np.roll(padded_int_img, int((w-1)/2 - 1), axis = 1) + np.roll(padded_int_img, (int((h-1)/2-1), int((w-1)/2 - 1)), axis = (0, 1))\n lxx = x + np.roll(x, int(2*h/3), axis=0) - 2*np.roll(x, int(h/3), axis=0)\n\n y = padded_int_img - np.roll(padded_int_img, int((h-1)/2 - 1), axis = 1) - np.roll(padded_int_img, int((w-1)/2 - 1), axis = 0) + np.roll(padded_int_img, (int((w-1)/2 - 1), int((h-1)/2 - 1)), axis = (0, 1))\n lyy = y + np.roll(y, int(2*h/3), axis=1) - 2*np.roll(y, int(h/3), axis=1)\n \n xy = padded_int_img - np.roll(padded_int_img, int((w-1)/2 - 1), axis = 0) - np.roll(padded_int_img, int((w-1)/2 - 1), axis = 1) + np.roll(padded_int_img, (int((w-1)/2 - 1), int((w-1)/2 - 1)), axis = (0, 1))\n lxy = xy + np.roll(xy, (int((w-1)/2 - 1), int((w-1)/2 - 1)), axis = (0, 1)) - np.roll(xy, int((w-1)/2 - 1), axis = 0) - np.roll(xy, int((w-1)/2 - 1), axis = 1) \n \n return lxx, lyy, lxy\n\ndef BlobDetectWithBoxFilters(img, s = 2, theta = 0.005):\n n = int(np.ceil(3*s)*2 + 1)\n G1D = cv2.getGaussianKernel(n, s)\n G2D = G1D @ G1D.T\n \n smoothed_img = cv2.filter2D(img, -1, G2D)\n \n int_img = np.cumsum(np.cumsum(smoothed_img, axis = 1), axis = 0)\n \n lxx, lyy, lxy = BoxFilter(int_img, s)\n ix, iy = np.gradient(smoothed_img)\n \n f = 15\n lxx = lxx[f:-f, f:-f]\n lyy = lyy[f:-f, f:-f]\n lxy = lxy[f:-f, f:-f]\n \n R = lxx*lyy - 0.81*lxy*lxy\n \n padx = img.shape[0] - R.shape[0]\n pady = img.shape[1] - R.shape[1]\n \n B_sq = disk_strel(n)\n Cond1 = (R ==cv2.dilate(R ,B_sq))\n Cond2 = (R > theta*R.max())\n \n R = Cond1*Cond2\n \n if padx>0 and pady>0:\n R = np.pad(R, ((padx//2 , padx//2), (pady//2, pady//2)))\n else:\n R = R[padx//2:-padx//2, pady//2:-pady//2]\n \n x = []\n y = []\n \n for i in range(R.shape[0]):\n for j in range(R.shape[1]):\n if R[i][j]:\n x.append(int(j))\n y.append(int(i))\n \n S = len(x)*[s]\n interest_points = np.array([x, y, S]).T\n \n return interest_points\n\ndef MultiscaleBlobDetectWithBoxFilters(img, s1 = 1.7, theta = 0.005, s2 = 1.3, N = 4):\n LoG = []\n i_points = []\n s = []\n for i in range(N):\n s.append(s1*s2**i)\n \n blobs = BlobDetectWithBoxFilters(img, s[i], theta)\n i_points.append(blobs)\n \n n = int(np.ceil(3*s[i])*2 + 1)\n G1D = cv2.getGaussianKernel(n, s[i])\n G2D = G1D @ G1D.T\n smoothed_img = cv2.filter2D(img, -1, G2D)\n \n ix, iy = np.gradient(smoothed_img)\n ixx, _ = np.gradient(ix)\n _, iyy = np.gradient(iy)\n LoG.append(s[i]**2*np.abs(ixx + iyy))\n \n blobs = []\n for i, scale_points in enumerate(i_points):\n for points in scale_points:\n x = int(points[0])\n y = int(points[1])\n s = points[2]\n if i == 0 and LoG[i][y][x] > LoG[i+1][y][x]:\n blobs.append([x, y, s])\n elif i == N-1 and LoG[i][y][x] > LoG[i-1][y][x]:\n blobs.append([x, y, s])\n elif LoG[i][y][x] > LoG[i-1][y][x] and LoG[i][y][x] > LoG[i+1][y][x]:\n blobs.append([x, y, s])\n \n return np.array(blobs)\n\ndef MatchingEvaluation(detect_function, SURF):\n detect_fun = lambda I: detect_function(I)\n \n if SURF:\n desc_fun = lambda I, kp: p3.featuresSURF(I, kp)\n else:\n desc_fun = lambda I, kp: p3.featuresHOG(I, kp) \n \n avg_scale_errors, avg_theta_errors = p3.matching_evaluation(detect_fun, desc_fun)\n\n for i in range(3):\n print('Avg. Scale Error for Image {}: {:.3f}'.format(i, avg_scale_errors[i]))\n print('Avg. Theta Error for Image {}: {:.3f}'.format(i, avg_theta_errors[i]))\n \ndef Scale_Theta_Errors_for_every_combination():\n detect_functions = [CornerDetect, MultiscaleCornerDetect, BlobDetect, MultiscaleBlobDetect, MultiscaleBlobDetectWithBoxFilters]\n flags = [True, False]\n for flag in flags:\n if flag:\n print(\"For SURF as local desciptor:\")\n print()\n else:\n print(\"For HOG as local desciptor:\")\n print()\n for detect_function in detect_functions:\n print('With ' + str(detect_function.__name__) + ' as a detect function:')\n MatchingEvaluation(detect_function, flag)\n print()\n \ndef Model_Training_and_Evaluation(detect_function, SURF):\n detect_fun = lambda I: detect_function(I)\n if SURF:\n a = 'SURF'\n desc_fun = lambda I, kp: p3.featuresSURF(I, kp)\n else:\n a = 'HOG'\n desc_fun = lambda I, kp: p3.featuresHOG(I, kp)\n \n feats = p3.FeatureExtraction(detect_fun, desc_fun)\n \n accs = []\n for k in range(5):\n x_train, y_train, x_test, y_test = p3.createTrainTest(feats, k)\n \n BOF_train, BOF_test = p3.BagOfWords(x_train, x_test)\n acc, preds, probas = p3.svm(BOF_train, y_train, BOF_test, y_test)\n accs.append(acc)\n \n print('Mean accuracy for ' + str(detect_function.__name__) + ' with ' + a + ' descriptors: {:.3f}%'.format(100.0*np.mean(accs)))\n \ndef Metric_Extraction_for_every_multiscale_detector():\n detect_functions = [MultiscaleCornerDetect, MultiscaleBlobDetect, MultiscaleBlobDetectWithBoxFilters]\n flags = [True, False]\n for flag in flags:\n if flag:\n print(\"For SURF as local desciptor:\")\n print()\n else:\n print(\"For HOG as local desciptor:\")\n print()\n for detect_function in detect_functions:\n print('With ' + str(detect_function.__name__) + ' as a detect function:')\n Model_Training_and_Evaluation(detect_function, flag)\n print()\n \nimg = cv2.imread(\"data_part12/edgetest_22.png\", cv2.IMREAD_GRAYSCALE)\nimg = img.astype(np.float64)/255\n# =============================================================================\n# plt.imshow(img, cmap='gray')\n# print(img.shape)\n# \n# =============================================================================\n# =============================================================================\n# Imax = img.max()\n# Imin = img.min()\n# PSNR1 = 20\n# PSNR2 = 10\n# std1 = (Imax - Imin)/10**(PSNR1/20)\n# std2 = (Imax - Imin)/10**(PSNR2/20)\n# n1 = np.random.normal(0, std1, size = (img.shape[0], img.shape[1]))\n# n2 = np.random.normal(0, std2, size = (img.shape[0], img.shape[1]))\n# =============================================================================\n# =============================================================================\n# plt.figure()\n# plt.imshow(img+n1, cmap = 'gray')\n# plt.figure()\n# plt.imshow(img+n2, cmap = 'gray')\n# =============================================================================\n\n# =============================================================================\n# n_img = img + n1\n# D = EdgeDetect(n_img, 2, linear = False)\n# =============================================================================\n# =============================================================================\n# plt.figure()\n# plt.imshow(D, 'gray')\n# =============================================================================\n\n# =============================================================================\n# B = np.array([\n# [0,1,0],\n# [1,1,1],\n# [0,1,0]\n# ], dtype=np.uint8)\n# \n# M = cv2.dilate(img, B) - cv2.erode(img, B)\n# #plt.figure()\n# #plt.imshow(M, 'gray')\n# T = M > 0.1\n# # =============================================================================\n# # plt.figure()\n# # plt.imshow(T, 'gray')\n# # =============================================================================\n# \n# counter = 0\n# \n# for i in range(D.shape[0]):\n# for j in range(D.shape[1]):\n# if D[i][j] == 1 and T[i][j] == 1:\n# counter += 1\n# \n# precision = counter/T.sum()\n# recall = counter/D.sum()\n# C = (precision + recall)/2\n# print (precision, recall, C)\n# =============================================================================\n\ncolored_duomo = cv2.imread(\"data_part12/duomo_edges.jpg\")\ncolored_duomo = cv2.cvtColor(colored_duomo, cv2.COLOR_BGR2RGB)\n\nduomo = cv2.cvtColor(colored_duomo, cv2.COLOR_RGB2GRAY)\nduomo = duomo.astype(np.float64)/255\n\n# =============================================================================\n# plt.figure()\n# plt.imshow(duomo, 'gray')\n# =============================================================================\n\n# =============================================================================\n# D2 = EdgeDetect(duomo, 2, linear = True)\n# plt.figure()\n# plt.imshow(D2, 'gray')\n# \n# M2 = cv2.dilate(duomo, B) - cv2.erode(duomo, B)\n# #plt.figure()\n# #plt.imshow(M, 'gray')\n# T2 = M2 > 0.1\n# plt.figure()\n# plt.imshow(T2, 'gray')\n# \n# counter = 0\n# \n# for i in range(D2.shape[0]):\n# for j in range(D2.shape[1]):\n# if D2[i][j] == 1 and T2[i][j] == 1:\n# counter += 1\n# \n# precision = counter/T2.sum()\n# recall = counter/D2.sum()\n# C2 = (precision + recall)/2\n# print (precision, recall, C2)\n# =============================================================================\n\ncolored_donuts = cv2.imread(\"data_part12/donuts.jpg\")\ncolored_donuts = cv2.cvtColor(colored_donuts, cv2.COLOR_BGR2RGB)\n\ndonuts = cv2.cvtColor(colored_donuts, cv2.COLOR_RGB2GRAY)\ndonuts = donuts.astype(np.float64)/255\n\ncolored_cells = cv2.imread(\"data_part12/cells.jpg\")\ncolored_cells = cv2.cvtColor(colored_cells, cv2.COLOR_BGR2RGB)\n\ncells = cv2.cvtColor(colored_cells, cv2.COLOR_RGB2GRAY)\ncells = cells.astype(np.float64)/255\n\n# =============================================================================\n# interest_points = CornerDetect(duomo, s = 2)\n# interest_points_visualization(colored_duomo, interest_points)\n# \n# multiscale_interest_points = MultiscaleCornerDetect(duomo)\n# interest_points_visualization(colored_duomo, multiscale_interest_points)\n# =============================================================================\n\n# =============================================================================\n# blobs = BlobDetect(cells, s = 2)\n# interest_points_visualization(colored_cells, blobs)\n# \n# multiscale_blobs = MultiscaleBlobDetect(cells)\n# interest_points_visualization(colored_cells, multiscale_blobs)\n# =============================================================================\n\n# =============================================================================\n# box_blobs = BlobDetectWithBoxFilters(donuts)\n# interest_points_visualization(colored_donuts, box_blobs)\n# \n# multiscale_box_blobs = MultiscaleBlobDetectWithBoxFilters(donuts)\n# interest_points_visualization(colored_donuts, multiscale_box_blobs)\n# =============================================================================\n\n#Scale_Theta_Errors_for_every_combination()\n\n#Metric_Extraction_for_every_multiscale_detector()\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "repo_name": "dbakalis8/Computer_Vision_NTUA_2021-2022", "sub_path": "Lab1/cv_lab1.py", "file_name": "cv_lab1.py", "file_ext": "py", "file_size_in_byte": 18818, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "numpy.pi", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 20, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.patches.Circle", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 48, "usage_type": "attribute"}, {"api_name": "cv2.getGaussianKernel", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 56, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 69, "usage_type": "call"}, {"api_name": "cv2.dilate", "line_number": 77, "usage_type": "call"}, {"api_name": "cv2.erode", "line_number": 78, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 87, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 87, "usage_type": "attribute"}, {"api_name": "cv2.dilate", "line_number": 88, "usage_type": "call"}, {"api_name": "cv2.erode", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 102, "usage_type": "call"}, {"api_name": "cv2.getGaussianKernel", "line_number": 104, "usage_type": "call"}, {"api_name": "cv2.getGaussianKernel", "line_number": 106, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 110, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 112, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 113, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 117, "usage_type": "call"}, {"api_name": "cv2.dilate", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 158, "usage_type": "call"}, {"api_name": "cv2.getGaussianKernel", "line_number": 160, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 184, "usage_type": "call"}, {"api_name": "cv2.getGaussianKernel", "line_number": 186, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 192, "usage_type": "call"}, {"api_name": "cv2.dilate", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 226, "usage_type": "call"}, {"api_name": "cv2.getGaussianKernel", "line_number": 228, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 236, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 250, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 267, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 269, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 270, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 272, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 278, "usage_type": "call"}, {"api_name": "cv2.getGaussianKernel", "line_number": 279, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 287, "usage_type": "call"}, {"api_name": "cv2.dilate", "line_number": 300, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 306, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 334, "usage_type": "call"}, {"api_name": "cv2.getGaussianKernel", "line_number": 335, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 337, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 339, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 341, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 357, "usage_type": "call"}, {"api_name": "cv22_lab1_part3_utils.featuresSURF", "line_number": 363, "usage_type": "call"}, {"api_name": "cv22_lab1_part3_utils.featuresHOG", "line_number": 365, "usage_type": "call"}, {"api_name": "cv22_lab1_part3_utils.matching_evaluation", "line_number": 367, "usage_type": "call"}, {"api_name": "cv22_lab1_part3_utils.featuresSURF", "line_number": 392, "usage_type": "call"}, {"api_name": "cv22_lab1_part3_utils.featuresHOG", "line_number": 395, "usage_type": "call"}, {"api_name": "cv22_lab1_part3_utils.FeatureExtraction", "line_number": 397, "usage_type": "call"}, {"api_name": "cv22_lab1_part3_utils.createTrainTest", "line_number": 401, "usage_type": "call"}, {"api_name": "cv22_lab1_part3_utils.BagOfWords", "line_number": 403, "usage_type": "call"}, {"api_name": "cv22_lab1_part3_utils.svm", "line_number": 404, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 407, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 424, "usage_type": "call"}, {"api_name": "cv2.IMREAD_GRAYSCALE", "line_number": 424, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 425, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 486, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 487, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 487, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 489, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2GRAY", "line_number": 489, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 490, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 522, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 523, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 523, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 525, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2GRAY", "line_number": 525, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 526, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 528, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 529, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 529, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 531, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2GRAY", "line_number": 531, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 532, "usage_type": "attribute"}]}
+{"seq_id": "19464114157", "text": "from django.contrib import admin\nfrom django.contrib.auth.admin import UserAdmin as DefaultUserAdmin\n\nfrom .models import Competition, Submission, User\n\n\nclass UserAdmin(DefaultUserAdmin):\n model = User\n list_display = [\n 'username',\n 'is_staff',\n 'is_active',\n 'last_login',\n ]\n list_filter = [\n 'is_staff',\n 'is_active',\n 'last_login',\n ]\n fieldsets = [\n (\n None,\n {'fields': ('username', 'password')},\n ),\n ('Permissions', {'fields': ['is_staff', 'is_superuser', 'is_active']}),\n ]\n add_fieldsets = [\n (\n 'User Details',\n {\n 'classes': ('wide',),\n 'fields': ('username'),\n },\n ),\n (\n 'Password Details',\n {\n 'classes': ('wide',),\n 'fields': ('password1', 'password2'),\n },\n ),\n ('Permissions', {'fields': ('is_staff', 'is_superuser', 'is_active')}),\n ]\n search_fields = ['id', 'username']\n ordering = ['username']\n\n\nclass CompetitionAdmin(admin.ModelAdmin):\n model = Competition\n list_display = [\n 'id',\n 'name',\n ]\n search_fields = ['id', 'name']\n ordering = ['name']\n\n\nclass SubmissionAdmin(admin.ModelAdmin):\n model = Submission\n list_display = ['id', 'name', 'score', 'user', 'competition']\n list_filter = [\n 'user',\n 'competition',\n ]\n search_fields = ['name', 'user', 'competition']\n ordering = ['name']\n\n\nadmin.site.register(User, UserAdmin)\nadmin.site.register(Competition, CompetitionAdmin)\nadmin.site.register(Submission, SubmissionAdmin)\n", "repo_name": "DurzoB5/Photocrowd-TT", "sub_path": "leaderboard/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 1699, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "django.contrib.auth.admin.UserAdmin", "line_number": 7, "usage_type": "name"}, {"api_name": "models.User", "line_number": 8, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 48, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 48, "usage_type": "name"}, {"api_name": "models.Competition", "line_number": 49, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 58, "usage_type": "name"}, {"api_name": "models.Submission", "line_number": 59, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 69, "usage_type": "call"}, {"api_name": "models.User", "line_number": 69, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 69, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 69, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 70, "usage_type": "call"}, {"api_name": "models.Competition", "line_number": 70, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 70, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 70, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 71, "usage_type": "call"}, {"api_name": "models.Submission", "line_number": 71, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 71, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 71, "usage_type": "name"}]}
+{"seq_id": "27569215964", "text": "from __future__ import annotations\n\nfrom typing import Optional, List, Tuple, TYPE_CHECKING\n\nimport app.engine.config as cf\nfrom app.engine import engine\nfrom app.engine.movement.movement_component import MovementComponent\nfrom app.engine.movement.unit_path_movement_component import UnitPathMovementComponent\nfrom app.utilities import utils\n\nif TYPE_CHECKING:\n from app.engine import camera, cursor\n\nimport logging\n\nclass MovementSystem:\n \"\"\"\n Operates upon MovementComponents and handles moving the camera around\n \"\"\"\n def __init__(self, cursor: Optional[cursor.BaseCursor], camera: Optional[camera.Camera]):\n self.cursor = cursor\n self.camera = camera\n\n self.moving_entities: List[MovementComponent] = []\n self.camera_follow: Tuple[int, int] = None # What position to be over\n self.camera_center: bool = False # Whether to center on the position\n\n def __len__(self):\n return len(self.moving_entities)\n\n def add(self, mc: MovementComponent):\n self.moving_entities.append(mc)\n\n def check_if_occupied_in_future(self, pos: Tuple[int, int]):\n for movement_component in self.moving_entities:\n if movement_component.get_end_goal() == pos:\n return movement_component.unit\n return None\n\n def is_moving(self, unit) -> bool:\n for movement_component in self.moving_entities:\n if movement_component.unit == unit:\n return True\n return False\n\n def stop(self, unit):\n \"\"\"\n # Stop all movement components associated with the given unit\n \"\"\"\n for movement_component in self.moving_entities:\n if movement_component.unit == unit:\n movement_component.finish() \n\n def begin_move(self, unit, path: List[Tuple[int, int]], \n event=False, follow=True, speed=0):\n \"\"\"\n # Used for simple movement of a unit in the normal way\n \"\"\"\n\n logging.info(\"Unit %s to begin moving\", unit)\n speed = speed or cf.SETTINGS['unit_speed']\n movement_component = \\\n UnitPathMovementComponent(unit, path, event, follow, speed=speed)\n self.moving_entities.append(movement_component)\n\n def update(self):\n current_time = engine.get_time()\n old_follow = self.camera_follow\n\n # Update all remaining entities\n for entity in self.moving_entities[:]:\n entity.update(current_time)\n if entity.follow:\n self.camera_follow = entity.get_camera_position()\n self.camera_center = entity.should_camera_center()\n # Remove inactive entities\n if not entity.active:\n self.moving_entities.remove(entity)\n\n # Update camera follow only if it's changed\n if self.camera_follow and old_follow != self.camera_follow:\n if self.cursor:\n self.cursor.set_pos(utils.round_pos(self.camera_follow))\n if self.camera and self.camera_center:\n self.camera.set_center(*self.camera_follow)\n", "repo_name": "ViolaBuddy/UnderTheShadowOfGrima", "sub_path": "lex-talionis/app/engine/movement/movement_system.py", "file_name": "movement_system.py", "file_ext": "py", "file_size_in_byte": 3089, "program_lang": "python", "lang": "en", "doc_type": "code", "stars": 0, "dataset": "github-code", "pt": "25", "api": [{"api_name": "typing.TYPE_CHECKING", "line_number": 11, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 20, "usage_type": "name"}, {"api_name": "app.engine.cursor.BaseCursor", "line_number": 20, "usage_type": "attribute"}, {"api_name": "app.engine.cursor", "line_number": 20, "usage_type": "name"}, {"api_name": "app.engine.camera.Camera", "line_number": 20, "usage_type": "attribute"}, {"api_name": "app.engine.camera", "line_number": 20, "usage_type": "name"}, {"api_name": "app.engine.cursor", "line_number": 21, "usage_type": "name"}, {"api_name": "app.engine.camera", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 24, "usage_type": "name"}, {"api_name": "app.engine.movement.movement_component.MovementComponent", "line_number": 24, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 25, "usage_type": "name"}, {"api_name": "app.engine.movement.movement_component.MovementComponent", "line_number": 31, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 34, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 54, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 54, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 60, "usage_type": "call"}, {"api_name": "app.engine.config.SETTINGS", "line_number": 61, "usage_type": "attribute"}, {"api_name": "app.engine.config", "line_number": 61, "usage_type": "name"}, {"api_name": "app.engine.movement.unit_path_movement_component.UnitPathMovementComponent", "line_number": 63, "usage_type": "call"}, {"api_name": "app.engine.engine.get_time", "line_number": 67, "usage_type": "call"}, {"api_name": "app.engine.engine", "line_number": 67, "usage_type": "name"}, {"api_name": "app.utilities.utils.round_pos", "line_number": 83, "usage_type": "call"}, {"api_name": "app.utilities.utils", "line_number": 83, "usage_type": "name"}]}
+{"seq_id": "38218223787", "text": "import logging\nimport subprocess # nosec\n\nfrom tern.analyze.default.command_lib import command_lib\nfrom tern.report import errors\nfrom tern.utils import constants\nfrom tern.utils import rootfs\n\n# global logger\nlogger = logging.getLogger(constants.logger_name)\n\n\ndef get_snippet_list(invoke_step, work_dir=None, envs=None):\n \"\"\"Given the invoke step dictionary i.e. steps of commands to run either\n in the chroot or on the host environment, get a list of command snippets\n invoke_dict: the value from the 'invoke' key\n workdir: If WORKDIR was set to anything, provide it here\n envs: If any environment variables were set, enter the key-value pairs\n here\"\"\"\n if 'container' in invoke_step.keys():\n snippet_list = invoke_step.get('container')\n # If environment variables exist, set them\n if envs:\n for var in envs:\n snippet_list.insert(\n 0, 'export ' + var.split('=')[0] + '=' + var.split('=')[1])\n # If work_dir exist cd into it\n if work_dir is not None:\n snippet_list.insert(0, 'cd ' + work_dir)\n return 'container', snippet_list\n return '', []\n\n\ndef invoke_in_rootfs(snippet_list, shell, package=''):\n '''Invoke the commands from the invoke dictionary in a root filesystem\n assuming the root filesystem is ready to accept commands'''\n # construct the full command\n full_cmd = command_lib.collate_snippets(snippet_list, package)\n try:\n result = rootfs.run_chroot_command(full_cmd, shell)\n try:\n result = result.decode('utf-8')\n except AttributeError:\n pass\n return result\n except subprocess.CalledProcessError as error:\n logger.warning('Error executing snippets: %s', error)\n raise\n\n\ndef get_pkg_attrs(attr_dict, shell, work_dir=None, envs=None, package_name=''):\n \"\"\"Given the dictionary containing the steps to invoke either in\n the container or on the host, invoke the steps and return the results\n either in list form or in raw form\"\"\"\n error_msgs = ''\n # the invoke dictionary contains steps\n # for each step make a command to invoke\n # currently we only deal with 1 step\n # so we just return the last step's results\n result = \"\"\n if 'invoke' in attr_dict.keys():\n for step in range(1, len(attr_dict['invoke'].keys()) + 1):\n method, snippet_list = get_snippet_list(\n attr_dict['invoke'][step], work_dir, envs)\n if method == 'container':\n # invoke the snippet list in a chroot environment\n try:\n result = invoke_in_rootfs(\n snippet_list, shell, package=package_name)\n result = result[:-1]\n except subprocess.CalledProcessError as error:\n error_msgs = error_msgs + error.stderr\n if 'delimiter' in attr_dict.keys():\n res_list = result.split(attr_dict['delimiter'])\n if res_list[-1] == '':\n res_list.pop()\n return res_list, error_msgs\n return res_list, error_msgs\n # if there is no delimiter, return the result string\n return result, error_msgs\n\n\ndef collect_list_metadata(shell, listing, work_dir=None, envs=None):\n '''Given the shell and the listing for the package manager, collect\n metadata that gets returned as a list'''\n pkg_dict = {}\n msgs = ''\n warnings = ''\n # a valid shell needs to exist in the filesystem for this to work\n for item in command_lib.base_keys:\n # check if the supported items exist in the given listing\n if item in listing.keys():\n items, msg = get_pkg_attrs(listing[item], shell, work_dir, envs)\n msgs = msgs + msg\n if item == 'files':\n # convert this data into a list before adding it to the\n # package dictionary\n file_list = []\n for files_str in items:\n # convert the string into a list\n files = []\n for filepath in filter(bool, files_str.split('\\n')):\n files.append(filepath.lstrip('/'))\n file_list.append(files)\n pkg_dict.update({item: file_list})\n else:\n pkg_dict.update({item: items})\n else:\n warnings = warnings + errors.no_listing_for_base_key.format(\n listing_key=item)\n return pkg_dict, msgs, warnings\n", "repo_name": "m1-key/tern", "sub_path": "tern/analyze/default/collect.py", "file_name": "collect.py", "file_ext": "py", "file_size_in_byte": 4523, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "github-code", "pt": "25", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "tern.utils.constants.logger_name", "line_number": 10, "usage_type": "attribute"}, {"api_name": "tern.utils.constants", "line_number": 10, "usage_type": "name"}, {"api_name": "tern.analyze.default.command_lib.command_lib.collate_snippets", "line_number": 38, "usage_type": "call"}, {"api_name": "tern.analyze.default.command_lib.command_lib", "line_number": 38, "usage_type": "name"}, {"api_name": "tern.utils.rootfs.run_chroot_command", "line_number": 40, "usage_type": "call"}, {"api_name": "tern.utils.rootfs", "line_number": 40, "usage_type": "name"}, {"api_name": "subprocess.CalledProcessError", "line_number": 46, "usage_type": "attribute"}, {"api_name": "subprocess.CalledProcessError", "line_number": 71, "usage_type": "attribute"}, {"api_name": "tern.analyze.default.command_lib.command_lib.base_keys", "line_number": 90, "usage_type": "attribute"}, {"api_name": "tern.analyze.default.command_lib.command_lib", "line_number": 90, "usage_type": "name"}, {"api_name": "tern.report.errors.no_listing_for_base_key.format", "line_number": 109, "usage_type": "call"}, {"api_name": "tern.report.errors.no_listing_for_base_key", "line_number": 109, "usage_type": "attribute"}, {"api_name": "tern.report.errors", "line_number": 109, "usage_type": "name"}]}
+{"seq_id": "24651098592", "text": "from imblearn.over_sampling import SMOTE as OrigModel\nimport lale.helpers\nimport lale.operators\nfrom typing import Any, Dict, Optional\n\nclass SMOTEImpl():\n\n def __init__(self, sampling_strategy='auto', random_state=None, k_neighbors=5, n_jobs=1):\n self._hyperparams = {\n 'sampling_strategy': sampling_strategy,\n 'random_state': random_state,\n 'k_neighbors': k_neighbors,\n 'n_jobs': n_jobs}\n \n self._sklearn_model = OrigModel(**self._hyperparams) #calling it _sklearn_model due to legacy :)\n\n def fit(self, X, y=None):\n if (y is not None):\n self._sklearn_model.fit(X, y)\n else:\n self._sklearn_model.fit(X)\n return self\n\n def transform(self, X, y=None):\n if y is None:\n #If y is not passed, or it is passed as None, we assume this means resampling is not to be applied.\n return X, y\n else:\n #If a not None value is passed for y, this would mean a call during fit, and hence resampling to be done.\n return self._sklearn_model.fit_resample(X, y)\n\n_input_fit_schema = {\n '$schema': 'http://json-schema.org/draft-04/schema#',\n 'type': 'object',\n 'required': ['X', 'y'],\n 'additionalProperties': False,\n 'properties': {\n 'X': {\n 'description': 'Features; the outer array is over samples.',\n 'type': 'array',\n 'items': {'type': 'array', 'items': {'type': 'number'}}},\n 'y': {\n 'description': 'Target class labels; the array is over samples.',\n 'anyOf': [\n {'type': 'array', 'items': {'type': 'number'}},\n {'type': 'array', 'items': {'type': 'string'}}]}}}\n\n_input_transform_schema = {\n '$schema': 'http://json-schema.org/draft-04/schema#',\n 'type': 'object',\n 'required': ['X', 'y'],\n 'additionalProperties': False,\n 'properties': {\n 'X': {\n 'description': 'Features; the outer array is over samples.',\n 'type': 'array',\n 'items': {'type': 'array', 'items': {'type': 'number'}}},\n 'y': {\n 'description': 'Target class labels; the array is over samples.',\n 'laleType': 'Any'\n}}}\n\n_output_transform_schema:Dict[str, Any] = {}\n\n_hyperparams_schema = {\n 'allOf': [\n { 'type': 'object',\n 'relevantToOptimizer': [],\n 'additionalProperties': False,\n 'properties': {\n 'sampling_strategy': {\n 'description': \"\"\"sampling_strategy : float, str, dict or callable, default='auto'. \nSampling information to resample the data set.\n\"\"\",\n 'anyOf': [\n { 'description':\"\"\"When ``float``, \nit corresponds to the desired ratio of the number of \nsamples in the minority class over the number of samples in the\nmajority class after resampling. Therefore, the ratio is expressed as\n:math:`\\\\alpha_{os} = N_{rm} / N_{M}` where :math:`N_{rm}` is the\nnumber of samples in the minority class after resampling and\n:math:`N_{M}` is the number of samples in the majority class.\n.. warning::\n ``float`` is only available for **binary** classification. An\n error is raised for multi-class classification.\"\"\",\n 'type': 'number'},\n { 'description':\"\"\"When ``str``, specify the class targeted by the resampling. \nThe number of samples in the different classes will be equalized.\nPossible choices are:\n``'minority'``: resample only the minority class;\n``'not minority'``: resample all classes but the minority class;\n``'not majority'``: resample all classes but the majority class;\n``'all'``: resample all classes;\n``'auto'``: equivalent to ``'not majority'``.\"\"\",\n 'enum': ['minority','not minority','not majority', 'all', 'auto']},\n { 'description':\"\"\"- When ``dict``, the keys correspond to the targeted classes. \nThe values correspond to the desired number of samples for each targeted\nclass.\"\"\",\n 'type': 'object'},\n { 'description':\"\"\"When callable, function taking ``y`` and returns a ``dict``. \nThe keys correspond to the targeted classes. The values correspond to the\ndesired number of samples for each class.\"\"\",\n 'laleType': 'Any'}],\n 'default': 'auto'},\n 'random_state': {\n 'description':\n 'Control the randomization of the algorithm.',\n 'anyOf': [\n { 'description': 'RandomState used by np.random',\n 'enum': [None]},\n { 'description': 'The seed used by the random number generator',\n 'type': 'integer'},\n { 'description': 'Random number generator instance.',\n 'laleType':'Any'}],\n 'default': None},\n 'k_neighbors':{\n 'description': \"\"\"If ``int``, number of nearest neighbours to used to construct synthetic samples. \nIf object, an estimator that inherits from\n:class:`sklearn.neighbors.base.KNeighborsMixin` that will be used to\nfind the k_neighbors.\"\"\",\n 'anyOf': [\n {'laleType':'Any'},\n {'type': 'integer'}],\n 'default': 5},\n 'n_jobs': {\n 'description': 'The number of threads to open if possible.',\n 'type': 'integer',\n 'default': 1}}}]}\n\n_combined_schemas = {\n '$schema': 'http://json-schema.org/draft-04/schema#',\n 'description': \"\"\" \"\"\",\n 'documentation_url': '',\n 'type': 'object',\n 'tags': {\n 'pre': ['~categoricals'],\n 'op': ['resampler'],\n 'post': []},\n 'properties': {\n 'hyperparams': _hyperparams_schema,\n 'input_fit': _input_fit_schema,\n 'input_transform': _input_transform_schema,\n 'output_transform': _output_transform_schema,\n}}\n\nSMOTE = lale.operators.make_operator(SMOTEImpl, _combined_schemas)", "repo_name": "somsirsa/lale", "sub_path": "lale/lib/imblearn/smote.py", "file_name": "smote.py", "file_ext": "py", "file_size_in_byte": 5841, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "github-code", "pt": "25", "api": [{"api_name": "imblearn.over_sampling.SMOTE", "line_number": 15, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 63, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 63, "usage_type": "name"}, {"api_name": "lale.helpers.operators.make_operator", "line_number": 146, "usage_type": "call"}, {"api_name": "lale.helpers.operators", "line_number": 146, "usage_type": "attribute"}, {"api_name": "lale.helpers", "line_number": 146, "usage_type": "name"}]}
+{"seq_id": "16635821260", "text": "#!/usr/bin/env python\n\nimport json\nimport argparse\nimport math\nimport os\n\ndef pretty_coord(x, y):\n\n def pretty(n):\n if n % 1 == 0.5:\n return f\"{math.floor(n)} ½\"\n else:\n return f\"{math.floor(n)}\"\n \n return f\"({pretty(x)}, {pretty(y)}) \"\n\ndef render_html(doc, output):\n '''Output an HTML description of the coordinates used.\n\n Being HTML, the styled text can then be copy-pasted into\n a Google Doc or similar.\n\n The implementation is deliberately very very simple (ugly, even).\n\n The document comes from a JSON document that has the following structure:\n\n {\n \"name\": \"Jerry, from Tom & Jerry\",\n \"author\": \"Conor Kerr\",\n \"world_coordinates\": {\n \"llx\": 0, \"lly\": 5, \"urx\": 20, \"ury\": 23\n },\n \"elements\": [\n {\n \"id\": \"A\",\n \"type\": \"filled_path\",\n \"coords\": [\n [3, 15], [1, 19], [1, 20], [1.5, 21], [2, 21.5],\n [3, 22], [4, 21.5], [5, 20], [5.5, 19], [5, 17.5],\n [4, 16], [3, 15]\n ],\n \"fill_color\": \"pink\" \n },\n {\n \"id\": \"N\",\n \"type\": \"path\",\n \"coords\": [\n [3, 22], [5, 21.5], [5.5, 20.5], [6, 19.5], [7, 18],\n [9, 19], [11, 19], [13, 18], [14, 19.5], [14.5, 20.5],\n [15, 21.5], [17, 22]\n ]\n }\n ]\n }\n\n At the root of the document, the c(name) and c(author) fields are\n currently not used.\n\n The c(world_coordinates) dictionary species the lower-left and\n upper-right x,y coordinates as used by turtle.set_worldcoordinates()\n\n The main element in the document is the c(elements) list. Each\n element of the list is drawn in the order given; so items earlier\n in the list will appear behind items that come later.\n\n Each element in the c(elements) list is a dictionary of that contains\n the following keys:\n\n * an c(id), which is just for identifying each element and can be\n useful in diagnostics.\n\n * a c(type), which currently must be one of \"path\" or \"filled_path\".\n The only difference currently is that \"filled_path\" may specify\n a fill colour and works on the assumption that the shape described\n is enclosed (the start and end coordinate is the same)\n\n * a c(coords) list of x,y coordinates. Each item is a list (not a\n tuple, but this is only due to the data coming from JSON)\n '''\n \n output.write('''\n \n \n \n {name} - by {author}\n \n \n \n
{name}
\n
by {author}
\n\n '''.format(\n name = doc.get('name', 'Unnamed masterpiece'),\n author = doc.get('author', 'unknown artist')))\n\n for element in doc.get('elements'):\n\n output.write('''
{id}: '''.format(\n id = element.get('id')\n ))\n\n for coord in element.get('coords'):\n output.write(pretty_coord(coord[0], coord[1]))\n\n if element.get('type') == 'filled_path':\n output.write(''' Colour {colour}