', body_html)\n if paragraphs:\n first_paragraph = paragraphs[0]\n try:\n p = html.fragment_fromstring(\"
{}
\".format(first_paragraph))\n except ParserError as e:\n print(\"Error: Unable to parse fragment: {}\".format(e))\n return\n for el in body.xpath('./*'):\n body.remove(el)\n body.insert(0, p)\n else:\n unparsable = body_html.encode(\"ascii\", errors=\"replace\")\n print(\"Error: Regex doesn't work for: {}\".format(unparsable))\n\ndef paragraphize(doc):\n \"\"\"Encaspulate content in a paragraph.\"\"\"\n entry = doc.xpath('/html/body/div[@id=\"entry\"]')[0]\n body = entry.xpath('./div[@id=\"body\"]')[0]\n del body.attrib[\"id\"]\n p = html.Element(\"p\")\n p.insert(0, body)\n new_body = html.Element(\"div\")\n new_body.set(\"id\", \"body\")\n new_body.insert(0, p)\n entry.insert(0, new_body)", "sub_path": "hooks/hook_filter_content.py", "file_name": "hook_filter_content.py", "file_ext": "py", "file_size_in_byte": 2741, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "lxml.html.fromstring", "line_number": 37, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 37, "usage_type": "name"}, {"api_name": "lxml.html.tostring", "line_number": 41, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 41, "usage_type": "name"}, {"api_name": "base64.encodebytes", "line_number": 43, "usage_type": "call"}, {"api_name": "lxml.html.tostring", "line_number": 43, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 43, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 60, "usage_type": "call"}, {"api_name": "lxml.html.fragment_fromstring", "line_number": 64, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 64, "usage_type": "name"}, {"api_name": "lxml.etree.ParserError", "line_number": 65, "usage_type": "name"}, {"api_name": "lxml.html.Element", "line_number": 80, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 80, "usage_type": "name"}, {"api_name": "lxml.html.Element", "line_number": 82, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 82, "usage_type": "name"}]}
+{"seq_id": "513860005", "text": "import os\r\nfrom glob import glob\r\nfrom PIL import Image\r\nimport numpy as np\r\nimport random\r\nfrom tqdm import tqdm\r\nimport torch\r\nimport torchvision.transforms.functional as TF\r\nimport random\r\nimport torchvision.transforms as transforms\r\nfrom torch.utils.data import Dataset, DataLoader\r\nimport torch.nn as nn\r\nimport copy\r\nimport torch.nn.functional as F\r\nimport matplotlib.pyplot as plt\r\nimport matplotlib.style as style\r\nfrom torch.utils.tensorboard import SummaryWriter\r\nimport time\r\n\r\n\r\nclass Prostate_data(Dataset):\r\n\r\n def __init__(self, img_path='../harvard_data/TMA_Images', mask_path='../harvard_data/Gleason_masks_train',\r\n dataset_type='train', img_size=3100, valid_split=['ZT76'], test_split=['ZT80'], num_classes=5):\r\n self.img_path = img_path\r\n self.mask_path = mask_path\r\n self.img_size = img_size\r\n self.num_classes = num_classes\r\n self.file_names = []\r\n self.dataset_type = dataset_type\r\n slide_dict = {'valid': valid_split, 'test': test_split}\r\n self.flag_dict = {}\r\n for file in glob(self.img_path + '/*.jpg'):\r\n _file_name = file.split('\\\\')[-1]\r\n _slide_type = _file_name.split('.')[0].split('_')[0]\r\n if dataset_type == 'train':\r\n if not (_slide_type in valid_split) and not (_slide_type in test_split):\r\n for fname in self.all_files(_file_name):\r\n self.file_names.append(fname)\r\n self.flag_dict[fname] = False\r\n else:\r\n if _slide_type in slide_dict[dataset_type]:\r\n self.file_names.append(_file_name)\r\n self.flag_dict[_file_name] = False\r\n random.seed(10)\r\n random.shuffle(self.file_names)\r\n self.data = {}\r\n self.transform = {\r\n 'train': transforms.Compose([transforms.ColorJitter(0.2,0.2,0.2,0.2),transforms.ToTensor(),\r\n transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225])]),\r\n 'valid': transforms.Compose([transforms.ToTensor(),\r\n transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225])])\r\n }\r\n\r\n def __len__(self):\r\n return len(self.file_names)\r\n\r\n def all_files(self,_file_name):\r\n return [_file_name,_file_name+'_tranhor',_file_name+'_tranver']\r\n\r\n def __getitem__(self, idx):\r\n _file_name = self.file_names[idx]\r\n _file_flag = self.flag_dict[_file_name]\r\n if _file_flag:\r\n return self.data[_file_name]\r\n else:\r\n img_path = self.img_path+'/'+_file_name.split('_tran')[0] if '_tran' in _file_name else self.img_path + '/' + _file_name\r\n mask_path = self.mask_path + '/' + 'mask_' + _file_name.split('_tran')[0].split('.')[0] + '.png' if '_tran' in _file_name else self.mask_path + '/' + 'mask_' + _file_name.split('.')[0] + '.png'\r\n\r\n img = Image.open(img_path).resize((self.img_size, self.img_size)).convert('RGB')\r\n mask = Image.open(mask_path).resize((self.img_size, self.img_size)).convert('RGB')\r\n ## transforms\r\n if 'hor' in _file_name:\r\n img = TF.hflip(img)\r\n mask = TF.hflip(mask)\r\n if 'ver' in _file_name:\r\n img = TF.vflip(img)\r\n mask = TF.vflip(mask)\r\n if 'aff' in _file_name:\r\n img = TF.affine(img,20,(0,0),1.35,0)\r\n mask = TF.affine(mask,20,(0,0),1.35,0)\r\n\r\n mask_array = np.asarray(mask)\r\n oneh_mask = np.zeros((self.num_classes, self.img_size, self.img_size))\r\n for x in range(self.img_size):\r\n for y in range(self.img_size):\r\n pixel_class = self.get_class(mask_array[x, y,:])\r\n oneh_mask[pixel_class, x, y] = 1\r\n\r\n array_img = np.asarray(img)\r\n timg = copy.deepcopy(array_img)\r\n for x in range(self.img_size):\r\n for y in range(self.img_size):\r\n rgb_n = array_img[x, y, :] / 255.0\r\n if rgb_n[0] > 0.8 and rgb_n[1] > 0.8 and rgb_n[2] > 0.8:\r\n timg[x, y, :] = [0, 0, 0]\r\n final_img = Image.fromarray(timg.astype('uint8'), 'RGB')\r\n\r\n img_tensor = self.transform[self.dataset_type](final_img)\r\n mask_tensor = torch.from_numpy(oneh_mask).view(self.num_classes, self.img_size, self.img_size)\r\n self.data[_file_name] = (img_tensor, mask_tensor)\r\n self.flag_dict[_file_name] = True\r\n return self.data[_file_name]\r\n\r\n def get_class(self, rgb):\r\n '''\r\n takes in rgb values of the pixel and returns the class of the pixel\r\n '''\r\n rgb_n = rgb / 255.0\r\n\r\n # white\r\n if rgb_n[0] > 0.8 and rgb_n[1] > 0.8 and rgb_n[2] > 0.8:\r\n return 4\r\n # red\r\n elif rgb_n[0] > 0.8 and rgb_n[1] < 0.8 and rgb_n[2] < 0.8:\r\n return 3\r\n # yellow\r\n elif rgb_n[0] > 0.8 and rgb_n[1] > 0.8 and rgb_n[2] < 0.8:\r\n return 2\r\n # green\r\n elif rgb_n[0] < 0.8 and rgb_n[1] > 0.8 and rgb_n[2] < 0.8:\r\n return 0\r\n # blue\r\n elif rgb_n[0] < 0.8 and rgb_n[1] < 0.8 and rgb_n[2] > 0.8:\r\n return 1\r\n else:\r\n print(rgb_n)\r\n raise ValueError('Weird rgb combination! Did not match any of 5 classes.')\r\n\r\ndef soft_dice_loss(y_pred,y_true):\r\n '''y_pred: (-1,5,512,512) :predictions\r\n y_true: (512,512,5) : targets\r\n compute the soft dice loss\r\n\r\n '''\r\n y_true = y_true.view(-1,5,256,256)\r\n epsilon = 1e-7\r\n dice_numerator = epsilon + 2 * torch.sum(y_true*y_pred,axis=(2,3))\r\n dice_denominator = epsilon + torch.sum(y_true*y_true,axis=(2,3)) + torch.sum(y_pred*y_pred,axis=(2,3))\r\n dice_loss = 1 - torch.mean(dice_numerator/dice_denominator)\r\n\r\n return dice_loss\r\n\r\n\r\ndef show_train_predictions(model,trainset,device,idx_list):\r\n fig,axes = plt.subplots(nrows=len(idx_list),ncols=3,figsize=(15,15))\r\n for i in range(len(idx_list)):\r\n idx = idx_list[i]\r\n input_img = Image.fromarray(np.asarray(trainset[idx][0].view(trainset.img_size,trainset.img_size,3).squeeze()).astype('uint8'), 'RGB')\r\n target_img = get_rgb(trainset[idx][1].squeeze())\r\n with torch.no_grad():\r\n pred_img = get_rgb(model(trainset[idx][0].view(-1,3,trainset.img_size,trainset.img_size).float().to(device)).squeeze())\r\n if len(idx_list)>1:\r\n axes[i,0].imshow(input_img)\r\n axes[i,1].imshow(pred_img)\r\n axes[i,2].imshow(target_img)\r\n else:\r\n axes[0].imshow(input_img)\r\n axes[1].imshow(pred_img)\r\n axes[2].imshow(target_img)\r\n if len(idx_list)>1:\r\n axes[0,0].set_title('T Input')\r\n axes[0,1].set_title('T Prediction')\r\n axes[0,2].set_title('T Target')\r\n else:\r\n axes[0].set_title('T Input')\r\n axes[1].set_title('T Prediction')\r\n axes[2].set_title('T Target')\r\n\r\n return fig\r\n\r\ndef show_valid_predictions(model,validset,device,idx_list):\r\n fig,axes = plt.subplots(nrows=len(idx_list),ncols=3,figsize=(15,15))\r\n for i in range(len(idx_list)):\r\n idx = idx_list[i]\r\n input_img = Image.fromarray(np.asarray(validset[idx][0].view(validset.img_size,validset.img_size,3).squeeze()).astype('uint8'), 'RGB')\r\n target_img = get_rgb(validset[idx][1].squeeze())\r\n with torch.no_grad():\r\n pred_img = get_rgb(model(validset[idx][0].view(-1,3,validset.img_size,validset.img_size).float().to(device)).squeeze())\r\n if len(idx_list)>1:\r\n axes[i,0].imshow(input_img)\r\n axes[i,1].imshow(pred_img)\r\n axes[i,2].imshow(target_img)\r\n else:\r\n axes[0].imshow(input_img)\r\n axes[1].imshow(pred_img)\r\n axes[2].imshow(target_img)\r\n if len(idx_list)>1:\r\n axes[0,0].set_title('V Input')\r\n axes[0,1].set_title('V Prediction')\r\n axes[0,2].set_title('V Target')\r\n else:\r\n axes[0].set_title('V Input')\r\n axes[1].set_title('V Prediction')\r\n axes[2].set_title('V Target')\r\n\r\n return fig\r\n\r\ndef get_rgb(tensor_img):\r\n pallete_dict = {\r\n 0 : [0,255,0],\r\n 1 : [0,0,255],\r\n 2 : [255,255,255],\r\n 3 : [255,0,0],\r\n 4 : [255,255,0]\r\n }\r\n img_h = tensor_img.size()[2]\r\n out_img = np.zeros((img_h,img_h,3))\r\n for h in range(img_h):\r\n for w in range(img_h):\r\n pixel_class = torch.argmax(tensor_img[:,h,w]).item()\r\n out_img[h,w,:] = pallete_dict[pixel_class]\r\n final_img = Image.fromarray(out_img.astype('uint8'), 'RGB')\r\n return final_img\r\n\r\nclass Focalloss(nn.Module):\r\n\r\n def __init__(self,gamma=0):\r\n super(Focalloss,self).__init__()\r\n self.gamma = gamma\r\n\r\n def forward(self,outputs,targets_oneh,targets):\r\n soft_outs = F.softmax(outputs,dim=1)\r\n log_soft = F.log_softmax(outputs,dim=1)\r\n weight_loss = torch.pow((1 - soft_outs),self.gamma) * log_soft\r\n loss = 0.4*F.nll_loss(weight_loss,targets) + 0.6*soft_dice_loss(outputs,targets_oneh)\r\n return loss\r\n\r\ndef main():\r\n from learner import Learner\r\n from res_unet_dropout import ResUnet\r\n # lr=3e-5\r\n dprob=0.2\r\n epochs = 8\r\n\r\n trainset = Prostate_data(img_size=256, num_classes=3)\r\n validset = Prostate_data(dataset_type='valid', img_size=256, num_classes=3)\r\n datasets = {'train': trainset, 'valid': validset}\r\n\r\n for lr in [1e-4,5e-4,1e-3,5e-3,1e-2]:\r\n for gamma in [0] :\r\n # fig,axes = plt.subplots(nrows=1,ncols=6,figsize=(24,4))\r\n # imgs = []\r\n # for i in tqdm(range(6)):\r\n # imgs.append(get_rgb(trainset[i][1]))\r\n # for j in range(6):\r\n # axes[j].imshow(imgs[j])\r\n # plt.show()\r\n\r\n model = ResUnet(num_classes=5,dprob=dprob)\r\n criterion = Focalloss(gamma=gamma)\r\n optimizer = torch.optim.SGD(model.parameters(),lr=lr,momentum=0.9)\r\n scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer,patience=10)\r\n dtime = '0057_1806'\r\n tb_logs = {'path':'logdirs/onevall_trials_aug/respre_SGD_plateau','comment':f'lr={lr}_gamma={gamma}_dprob={dprob}_{dtime}'}\r\n trainer = Learner(datasets,model,criterion,optimizer,scheduler,bs=8,num_workers=4)\r\n try :\r\n trainer.fit(tb_logs=tb_logs,epochs=epochs)\r\n # torch.save(trainer.model,f'logdirs/onevall_trials_aug/respre_SGD_plateau/lr={lr}_gamma={gamma}_dprob={dprob}_{dtime}/{dtime}')\r\n except KeyboardInterrupt:\r\n pass\r\n # torch.save(trainer.model,f'logdirs/onevall_trials_aug/respre_SGD_plateau/lr={lr}_gamma={gamma}_dprob={dprob}_{dtime}/{dtime}')\r\n\r\n\r\nif __name__=='__main__':\r\n main()\r\n", "sub_path": "focal_loss.py", "file_name": "focal_loss.py", "file_ext": "py", "file_size_in_byte": 11017, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 21, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 33, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 45, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 46, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 49, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 49, "usage_type": "name"}, {"api_name": "torchvision.transforms.ColorJitter", "line_number": 49, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 49, "usage_type": "call"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 50, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 50, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 51, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 51, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 51, "usage_type": "call"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 52, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 52, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 70, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 70, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 71, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 71, "usage_type": "name"}, {"api_name": "torchvision.transforms.functional.hflip", "line_number": 74, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional", "line_number": 74, "usage_type": "name"}, {"api_name": "torchvision.transforms.functional.hflip", "line_number": 75, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional", "line_number": 75, "usage_type": "name"}, {"api_name": "torchvision.transforms.functional.vflip", "line_number": 77, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional", "line_number": 77, "usage_type": "name"}, {"api_name": "torchvision.transforms.functional.vflip", "line_number": 78, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional", "line_number": 78, "usage_type": "name"}, {"api_name": "torchvision.transforms.functional.affine", "line_number": 80, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional", "line_number": 80, "usage_type": "name"}, {"api_name": "torchvision.transforms.functional.affine", "line_number": 81, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional", "line_number": 81, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 90, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 91, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 97, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 97, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 138, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "PIL.Image.fromarray", "line_number": 149, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 149, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 173, "usage_type": "name"}, {"api_name": "PIL.Image.fromarray", "line_number": 176, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 176, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 176, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 211, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 213, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 213, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 216, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 216, "usage_type": "name"}, {"api_name": "torch.nn.functional.softmax", "line_number": 223, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 223, "usage_type": "name"}, {"api_name": "torch.nn.functional.log_softmax", "line_number": 224, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 224, "usage_type": "name"}, {"api_name": "torch.pow", "line_number": 225, "usage_type": "call"}, {"api_name": "torch.nn.functional.nll_loss", "line_number": 226, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 226, "usage_type": "name"}, {"api_name": "res_unet_dropout.ResUnet", "line_number": 250, "usage_type": "call"}, {"api_name": "torch.optim.SGD", "line_number": 252, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 252, "usage_type": "attribute"}, {"api_name": "torch.optim.lr_scheduler.ReduceLROnPlateau", "line_number": 253, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 253, "usage_type": "attribute"}, {"api_name": "learner.Learner", "line_number": 256, "usage_type": "call"}]}
+{"seq_id": "634302455", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sat Nov 3 10:21:19 2018\r\n\r\n@author: Srinivas\r\n\"\"\"\r\n\r\n# Word Negation Tracking\r\nimport nltk\r\n\r\nsentence = \"I was not happy with the team's performance\"\r\nwords = nltk.word_tokenize(sentence)\r\nnew_words = []\r\ntemp_word = \"\"\r\n\r\nfor word in words:\r\n if word == 'not':\r\n temp_word = \"not_\"\r\n elif temp_word == \"not_\":\r\n word = temp_word + word # It will be not_happy\r\n temp_word = \"\"\r\n if word != \"not\":\r\n new_words.append(word)\r\nsentence = ' '.join(new_words) \r\n\r\n ", "sub_path": "Word_Negation1.py", "file_name": "Word_Negation1.py", "file_ext": "py", "file_size_in_byte": 553, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "nltk.word_tokenize", "line_number": 12, "usage_type": "call"}]}
+{"seq_id": "404925934", "text": "#!/usr/bin/env python\n'''\nCopyright (c) 2015 Jared E. Stroud\n\nPermission is hereby granted, free of charge, to any person obtaining a copy\nof this software and associated documentation files (the \"Software\"), to deal\nin the Software without restriction, including without limitation the rights\nto use, copy, modify, merge, publish, distribute, sublicense, and/or sell\ncopies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:\n\nThe above copyright notice and this permission notice shall be included in\nall copies or substantial portions of the Software.\n\nTHE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\nIMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\nFITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\nAUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\nLIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\nOUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\nTHE SOFTWARE.\n\n'''\n\ntry:\n import socket\n import os\n import sys\n import argparse\n import time\nexcept ImportError as err:\n print(\"[Error] I don't have \" + str(err))\n\nclass sendData():\n \n '''\n Name: __init__\n Parameters: self\n Purpose: Initializes socket for the rest of the methods.\n '''\n def __init__(self):\n self.sock = socket.socket()\n\n '''\n Name: sendFile\n Parameters: fileName\n '''\n def sendFile(self, fileName, address, port):\n self.sock.connect((str(address), int(port)))\n data = open(fileName, \"rb\")\n\n while True:\n chunk = file.read(data)\n if not chunk:\n break # Entire file has been read in.\n self.sock.sendall(chunk)\n\n '''\n Name: copyFile\n Parameters: srcFile , destination\n srcFile: File desired to be transfered.\n destination: End point for file being transfered.\n Return: Nothing.\n '''\n def copyFile(self, srcFile, destination):\n try:\n os.path.join(str(fileName), str(destination))\n except Error:\n print(\"[Error] I could not copy the file\")\n sys.exit()\n\n\n\nif __name__ == \"__main__\":\n\n print(\"Starting Data send\")\n sendObj = sendData()\n\n parser = argparse.ArgumentParser()\n parser.add_argument(\"--fileName\", required=True, type=str, help=\"Specify file to be sent.\")\n parser.add_argument(\"--address\", required=True, type=str, help=\"Specify the address to send the file to.\")\n parser.add_argument(\"--port\", required=True, type=int, help=\"Specify the port of th destination server.\")\n args = parser.parse_args()\n\n if args.fileName and args.address and args.port: # Ensuring all arguments are fulfilled.\n address = args.address# User specified file size.\n fileName = args.fileName # User specified file name.\n port = args.port # User specified port.\n\n sendObj.sendFile(fileName, address, port)\n", "sub_path": "lib/pipeData.py", "file_name": "pipeData.py", "file_ext": "py", "file_size_in_byte": 3073, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "socket.socket", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 69, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 78, "usage_type": "call"}]}
+{"seq_id": "96372097", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\nnum = 100\n\nx = np.linspace(0, 100,num)\n\nx0 = 1.0\nx1 = 2.0\nx2 = 3.5\nx3 = -0.00\nscalar = 1000.0\n\n\ny = x0*pow(x,0.0) + x1*pow(x,1.0) + x2*pow(x,2.0) + x3*pow(x,3.0)\nz = y + scalar*np.random.randn(num)\nplt.plot(x,y)\nplt.plot(x,z,'ro')\n\nplt.show()\n\n\n\n", "sub_path": "regressionGui.py", "file_name": "regressionGui.py", "file_ext": "py", "file_size_in_byte": 298, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "numpy.linspace", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 16, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}]}
+{"seq_id": "158584806", "text": "\r\nimport os\r\nimport sys\r\nimport json\r\nimport torch\r\nimport torch.nn as nn\r\nfrom torchsummary import summary\r\nwith open(os.path.join(sys.path[0], 'config.json')) as json_file:\r\n\tconfig = json.load(json_file)\r\n\r\nclass AudioNet(nn.Module):\r\n\r\n\tdef __init__(self):\r\n\t\tsuper(AudioNet, self).__init__() # (128, 1292)\r\n\t\tself.num_output = 5\r\n\r\n\t\tif(config[\"SAL0\"] == \"Wi\"):\r\n\t\t\tself.W_i = nn.Parameter(torch.randn(40,1)) # with channel num 16\r\n\t\t\tself.bi = nn.Parameter(torch.randn(1))\r\n\t\t\tself.SmSA0 = nn.Softmax(dim = 1)\r\n\t\tself.conv0 = nn.Sequential(nn.Conv1d(128, 32, kernel_size=8), nn.MaxPool1d(4, stride=4), nn.ReLU(), nn.BatchNorm1d(32),)\r\n\t\tif(config[\"CAL1\"] == \"FC\"):\r\n\t\t\tself.CA1_avg_pool = nn.AdaptiveAvgPool1d(1)\r\n\t\t\tself.CA1 = nn.Sequential(nn.Linear(in_features=32, out_features=32), nn.Softmax(dim=1),)\r\n\t\tif(config[\"CAL1\"] == \"Conv\"):\r\n\t\t\tself.CA1_avg_pool = nn.AdaptiveAvgPool1d(1)\r\n\t\t\tself.conv_du1 = nn.Sequential(nn.Conv1d(32, 32//2, 1, bias=True), nn.ReLU(inplace=True), nn.Conv1d(32//2, 32, 1, bias=True), nn.Sigmoid())\r\n\t\tself.conv1 = nn.Sequential(nn.Conv1d(32, 16, kernel_size=8), nn.MaxPool1d(4, stride=4), nn.ReLU(), nn.BatchNorm1d(16),)\r\n\t\tif(config[\"CAL2\"] == \"FC\"):\r\n\t\t\tself.CA2_avg_pool = nn.AdaptiveAvgPool1d(1)\r\n\t\t\tself.CA2 = nn.Sequential(nn.Linear(in_features=16, out_features=16), nn.Softmax(dim=1),)\r\n\t\tif(config[\"CAL2\"] == \"Conv\"):\r\n\t\t\tself.CA2_avg_pool = nn.AdaptiveAvgPool1d(1)\r\n\t\t\tself.conv_du2 = nn.Sequential(nn.Conv1d(16, 16//2, 1, bias=True), nn.ReLU(inplace=True), nn.Conv1d(16//2, 16, 1, bias=True),nn.Sigmoid())\r\n\t\tself.fc0 = nn.Sequential(nn.Linear(in_features=1248, out_features=64), nn.Tanh(), nn.Dropout(), nn.Linear(in_features=64, out_features=self.num_output),)\r\n\t\tself.logsoftmax = nn.LogSoftmax(dim=1)\r\n\t\tself.apply(self._init_weights)\r\n\r\n\tdef forward(self, x):\r\n\t\tif(config[\"SAL0\"] == \"Wi\"):\r\n\t\t\tz0 = x.permute(0, 2, 1) #(N, L, C)\r\n\t\t\talpha0 = (self.SmSA0((torch.matmul(z0, self.W_i) + self.bi).squeeze(-1))).unsqueeze(-1) #(N, L, 1)-->(N, L)\r\n\t\t\tx = (z0*alpha0).permute(0, 2, 1)\r\n\t\tx = self.conv0(x) #(N, 32, 321)\r\n\t\tif(config[\"CAL1\"] == \"FC\"):\r\n\t\t\tbeta1 = self.CA1_avg_pool(x).squeeze(-1) #(N, 32, 1) --> (N, 32)\r\n\t\t\tbeta1 = self.CA1(beta1).unsqueeze(-1)\r\n\t\t\tx = x*beta1\r\n\t\tif(config[\"CAL1\"] == \"Conv\"):\r\n\t\t\tbeta1 = self.CA1_avg_pool(x) #(N, 32, 1) --> (N, 32)\r\n\t\t\tbeta1 = self.conv_du1(beta1)\r\n\t\t\tx = x*beta1\r\n\t\tx = self.conv1(x) #(N, 16, 78)\r\n\t\tif(config[\"CAL2\"] == \"FC\"):\r\n\t\t\tbeta2 = self.CA2_avg_pool(x).squeeze(-1) #(N, 16, 1) --> (N, 16)\r\n\t\t\tbeta2 = self.CA2(beta2).unsqueeze(-1) #(N, 16, 1)\r\n\t\tx = x*beta2 #(N, 16, 78)\r\n\t\tif(config[\"CAL2\"] == \"Conv\"):\r\n\t\t\tbeta2 = self.CA2_avg_pool(x) #(N, 16, 1) \r\n\t\t\tbeta2 = self.conv_du2(beta2)\r\n\t\t\tx = x*beta2\r\n\t\tflatten = x.view(x.size(0), -1)\r\n\t\t# print(flatten.shape)\r\n\t\tout = self.logsoftmax(self.fc0(flatten))\r\n\t\t#x = F.log_softmax(x, dim=1) # output (N, 5)\r\n\t\treturn out\r\n\r\n\tdef _init_weights(self, layer) -> None:\r\n\t\tif(isinstance(layer, nn.Conv1d)):\r\n\t\t\tnn.init.kaiming_uniform_(layer.weight)\r\n\t\telif(isinstance(layer, nn.Linear)):\r\n\t\t\tnn.init.xavier_uniform_(layer.weight)\r\n\r\n\r\nif __name__ == '__main__':\r\n\tnet = AudioNet()\r\n\tprint(net)\r\n\tnet = net.cuda()\r\n\tsummary(net, (128, 1292))", "sub_path": "project_CNNFC_Atten_Cls/models/CNNFC.py", "file_name": "CNNFC.py", "file_ext": "py", "file_size_in_byte": 3456, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "os.path.join", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "sys.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 9, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 11, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn.Parameter", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn.Softmax", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn.MaxPool1d", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn.ReLU", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn.AdaptiveAvgPool1d", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn.Softmax", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn.AdaptiveAvgPool1d", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn.ReLU", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn.Sigmoid", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn.Sequential", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn.MaxPool1d", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn.ReLU", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn.AdaptiveAvgPool1d", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn.Softmax", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn.AdaptiveAvgPool1d", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.ReLU", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.Sigmoid", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.Sequential", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn.Tanh", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn.Dropout", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn.LogSoftmax", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.matmul", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn.Conv1d", "line_number": 69, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 69, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_uniform_", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 70, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 70, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 71, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 71, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_uniform_", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 72, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 72, "usage_type": "name"}, {"api_name": "torchsummary.summary", "line_number": 79, "usage_type": "call"}]}
+{"seq_id": "118373364", "text": "import os\nfrom setuptools import setup\n\nhere = os.path.abspath(os.path.dirname(__file__))\nREADME = open(os.path.join(here, 'README.rst')).read()\n\nsetup(\n name='django-active-menu',\n version='0.1',\n packages=['active_menu'],\n description='Simple, fast and easy django template tags to get active url in your html menu.',\n long_description=README,\n author='Slawomir Kabik',\n author_email='slawek@redsoftware.pl',\n url='https://github.com/yourname/django-myapp/',\n license='MIT',\n install_requires=[\n 'Django>=1.6',\n 'BeautifulSoup4==4.4.1'\n ]\n)\n\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 591, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "os.path.abspath", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "setuptools.setup", "line_number": 7, "usage_type": "call"}]}
+{"seq_id": "2944149", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('trueguide', '0003_place'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='Photos',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('url', models.CharField(max_length=500)),\n ],\n ),\n migrations.AddField(\n model_name='place',\n name='photoid',\n field=models.ForeignKey(default=0, to='trueguide.Photos'),\n ),\n ]\n", "sub_path": "Server/tg/trueguide/migrations/0004_auto_20150517_1209.py", "file_name": "0004_auto_20150517_1209.py", "file_ext": "py", "file_size_in_byte": 689, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}]}
+{"seq_id": "371879336", "text": "from django.conf.urls import url, include\nfrom . import views\n \nurlpatterns = [\n url(r'^$', views.index ), #LOG/REG\n url(r'^login$', views.login), #LOGIN\n url(r'^create$', views.create ), #this is the REG\n url(r'^ideas$', views.ideas ), #ideas\n url(r'^view/(?P
\\d+)$', views.new), #View \n url(r'^new$', views.new ), #Add TEMPLATE\n url(r'^add$', views.add), #add job process\n url(r'^(?P\\d+)/delete$', views.delete), #Cancel\n url(r'edit/(?P\\d+)$', views.edit ), #EDIT \n url(r'^(?P\\d+)/update$', views.update ),\n url(r'^logout_user$', views.logout_user),\n]", "sub_path": "apps/belt_app/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 623, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "django.conf.urls.url", "line_number": 5, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}]}
+{"seq_id": "63382637", "text": "import numpy as np\nfrom PIL import Image\n\n\ndef hex_to_rgb(hex_str):\n hex_str = hex_str.strip()\n\n if hex_str[0] == '#':\n hex_str = hex_str[1:]\n\n if len(hex_str) != 6:\n raise ValueError('Input #{} is not in #RRGGBB format.'.format(hex_str))\n\n r, g, b = hex_str[:2], hex_str[2:4], hex_str[4:]\n rgb = [int(n, base=16) for n in [r, g, b]]\n return np.array(rgb)\n\n\ndef binary_mask(crop_mask, palette):\n bin_mask = []\n for x in crop_mask:\n temp = []\n for y in x:\n crop = 0\n for i, ch in enumerate(y):\n if ch >= y[crop]:\n crop = i\n temp.append(hex_to_rgb(palette[crop]))\n bin_mask.append(temp)\n return np.array(bin_mask, dtype=np.uint8)\n\n\ndef read_png(file):\n image = Image.open(file)\n return np.array(image)\n", "sub_path": "tools/visualize.py", "file_name": "visualize.py", "file_ext": "py", "file_size_in_byte": 835, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "numpy.array", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 30, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 34, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 34, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}]}
+{"seq_id": "284066512", "text": "import pyglet, random\r\nfrom pyglet.gl import *\r\n\r\nclass pumpkin():\r\n\tdef __init__(self, height, width):\r\n\t\tself.x = random.randint(0, width-16)\r\n\t\tself.y = height\r\n\r\n\t\tself.pumpkin = pyglet.resource.image('pumpkin.png')\r\n\t\tself.pumpkin.width = 16 \r\n\t\tself.pumpkin.height = 16\r\n\r\n\tdef on_draw(self):\r\n\t\tself.pumpkin.blit(self.x, self.y)\r\n\r\nclass window(pyglet.window.Window):\r\n\tdef __init__(self, *args, **kwargs):\r\n\t\tsuper(window, self).__init__(*args, **kwargs)\r\n\r\n\t\tself.pumpkins = []\r\n\t\tfor i in range(30):\r\n\t\t\tself.pumpkins.append(pumpkin(self.height, self.width))\r\n\r\n\t\tself.bg = pyglet.resource.image('bg.jpg')\r\n\r\n\tdef update_y_offset(self, dt):\r\n\t\tfor i in range(len(self.pumpkins)):\r\n\t\t\tself.pumpkins[i].y -= (random.uniform(0.0, 10.0) / 10)\r\n\r\n\tdef on_draw(self):\r\n\t\tself.clear()\r\n\t\t\r\n\t\tglClear(GL_COLOR_BUFFER_BIT)\r\n\t\tglLoadIdentity()\r\n\t\tglEnable(GL_BLEND)\r\n\t\tglBlendFunc(GL_SRC_ALPHA, GL_ONE_MINUS_SRC_ALPHA)\r\n\r\n\t\tself.bg.blit(0, 0)\r\n\r\n\t\tfor i in range(len(self.pumpkins)):\r\n\t\t\tself.pumpkins[i].on_draw()\r\n\t\t\tif self.pumpkins[i].y < 0:\r\n\t\t\t\tprint('Pumpkin deleted')\r\n\t\t\t\tdel self.pumpkins[i]\r\n\t\t\t\tself.pumpkins.append(pumpkin(self.height, self.width))\r\n\t\tfor j in range(4):\r\n\t\t\tpyglet.clock.schedule_once(self.update_y_offset, j)\r\n\r\n\r\nif __name__ == \"__main__\":\r\n\twindow = window(width=640, height=480, caption='Pumpkin fall')\r\n\tpyglet.app.run()\r\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1510, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "random.randint", "line_number": 6, "usage_type": "call"}, {"api_name": "pyglet.resource.image", "line_number": 9, "usage_type": "call"}, {"api_name": "pyglet.resource", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pyglet.window", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pyglet.resource.image", "line_number": 24, "usage_type": "call"}, {"api_name": "pyglet.resource", "line_number": 24, "usage_type": "attribute"}, {"api_name": "random.uniform", "line_number": 28, "usage_type": "call"}, {"api_name": "pyglet.clock.schedule_once", "line_number": 47, "usage_type": "call"}, {"api_name": "pyglet.clock", "line_number": 47, "usage_type": "attribute"}, {"api_name": "pyglet.app.run", "line_number": 52, "usage_type": "call"}, {"api_name": "pyglet.app", "line_number": 52, "usage_type": "attribute"}]}
+{"seq_id": "522724405", "text": "from .models.zipline_app.fill import Fill\nfrom .models.zipline_app.order import Order\nfrom .models.zipline_app.asset import Asset\nfrom .models.zipline_app.placement import Placement\n\nfrom .widgets import AssetModelSelect2Widget, AccountModelSelect2Widget, ReadOnlyWidgetSimple, ReadOnlyWidgetAsset, ReadOnlyWidgetOrder, CustodianModelSelect2Widget, FillUnitWidget\nfrom django import forms\n\n# override widget in createview\n# http://stackoverflow.com/a/21407374/4126114\n# Override a Django generic class-based view widget\n# http://stackoverflow.com/a/27322032/4126114\nclass FillForm(forms.ModelForm):\n source=forms.CharField(required=False, widget = forms.HiddenInput())\n field_order = [\n 'pub_date', 'dedicated_to_order', 'fill_side', 'asset', 'fill_qty_unsigned', 'fill_unit',\n 'fill_price', 'category', 'is_internal', 'trade_date', 'settlement_date',\n 'custodian', 'fill_text',\n 'commission'\n ]\n class Meta:\n model=Fill\n exclude = [\"user\"]\n widgets = {\n 'pub_date': ReadOnlyWidgetSimple(),\n 'dedicated_to_order': ReadOnlyWidgetOrder(),\n 'custodian': CustodianModelSelect2Widget(),\n 'asset': ReadOnlyWidgetAsset(),\n 'fill_side': forms.HiddenInput(),\n 'fill_unit': FillUnitWidget(),\n }\n def clean_pub_date(self): return self.initial['pub_date'] #.strftime(\"%Y-%m-%d %H:%i:%s\")\n def clean_dedicated_to_order(self): return self.initial['dedicated_to_order']\n def clean_asset(self):\n aid = self.initial['asset']\n if not isinstance(aid, int): return aid\n return Asset.objects.get(id=aid)\n def clean_fill_side(self): return self.initial['fill_side']\n def clean_source(self): return self.initial['source'] if 'source' in self.initial else None\n def clean_fill_unit(self): return self.initial['fill_unit']\n\n def __init__(self, *args, **kwargs):\n super(FillForm, self).__init__(*args, **kwargs)\n self.fields['fill_unit'].widget.form_instance = self\n\nclass OrderForm(forms.ModelForm):\n source=forms.CharField(required=False, widget = forms.HiddenInput())\n field_order = [\n 'id',\n 'pub_date',\n 'user',\n 'order_side',\n 'account',\n 'asset',\n 'order_unit',\n 'order_qty_unsigned',\n\n # fields for tables.py\n 'asset_currency',\n 'order_amount',\n 'order_qty',\n\n 'am_type',\n 'order_type',\n 'limit_price',\n 'order_validity',\n 'validity_date',\n 'order_text',\n 'commission'\n ]\n\n class Meta:\n model=Order\n exclude=['user', 'order_bulk']\n widgets = {\n 'pub_date': ReadOnlyWidgetSimple(),\n 'asset': AssetModelSelect2Widget(),\n 'account': AccountModelSelect2Widget(),\n }\n def clean_pub_date(self): return self.initial['pub_date'] #.strftime(\"%Y-%m-%d %H:%i:%s\")\n def clean_source(self): return self.initial['source'] if 'source' in self.initial else None\n\n\nclass PlacementForm(forms.ModelForm):\n class Meta:\n model=Placement\n exclude = [\"date\", \"user\"]\n\n\nclass OrderDocumentForm(forms.Form):\n docfile = forms.FileField(\n widget=forms.ClearableFileInput(attrs={'multiple': True}),\n label='Select file(s)'\n )\n", "sub_path": "zipline_app/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 3065, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "django.forms.ModelForm", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 13, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 14, "usage_type": "name"}, {"api_name": "django.forms.HiddenInput", "line_number": 14, "usage_type": "call"}, {"api_name": "models.zipline_app.fill.Fill", "line_number": 22, "usage_type": "name"}, {"api_name": "widgets.ReadOnlyWidgetSimple", "line_number": 25, "usage_type": "call"}, {"api_name": "widgets.ReadOnlyWidgetOrder", "line_number": 26, "usage_type": "call"}, {"api_name": "widgets.CustodianModelSelect2Widget", "line_number": 27, "usage_type": "call"}, {"api_name": "widgets.ReadOnlyWidgetAsset", "line_number": 28, "usage_type": "call"}, {"api_name": "django.forms.HiddenInput", "line_number": 29, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 29, "usage_type": "name"}, {"api_name": "widgets.FillUnitWidget", "line_number": 30, "usage_type": "call"}, {"api_name": "models.zipline_app.asset.Asset.objects.get", "line_number": 37, "usage_type": "call"}, {"api_name": "models.zipline_app.asset.Asset.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.zipline_app.asset.Asset", "line_number": 37, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 46, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 46, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 47, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 47, "usage_type": "name"}, {"api_name": "django.forms.HiddenInput", "line_number": 47, "usage_type": "call"}, {"api_name": "models.zipline_app.order.Order", "line_number": 73, "usage_type": "name"}, {"api_name": "widgets.ReadOnlyWidgetSimple", "line_number": 76, "usage_type": "call"}, {"api_name": "widgets.AssetModelSelect2Widget", "line_number": 77, "usage_type": "call"}, {"api_name": "widgets.AccountModelSelect2Widget", "line_number": 78, "usage_type": "call"}, {"api_name": "django.forms.ModelForm", "line_number": 84, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 84, "usage_type": "name"}, {"api_name": "models.zipline_app.placement.Placement", "line_number": 86, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 90, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 90, "usage_type": "name"}, {"api_name": "django.forms.FileField", "line_number": 91, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 91, "usage_type": "name"}, {"api_name": "django.forms.ClearableFileInput", "line_number": 92, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 92, "usage_type": "name"}]}
+{"seq_id": "30861989", "text": "# -*- coding: utf-8 -*-\r\nimport librosa\r\nimport matplotlib.pyplot as plt\r\nimport numpy as np\r\nimport librosa.display\r\nimport torch\r\nfrom torch.utils.data import Dataset, DataLoader\r\nimport numpy as np\r\n\r\n\r\nimport os\r\n\r\ndef Get_align_beat_pitch_spectrogram(align_root_path, pitch_beat_root_path, wav_root_path):\r\n \r\n filename_list = os.listdir(align_root_path) #列出文件夹下所有的目录与文件\r\n path_list = []\r\n phone_list, beat_list, pitch_list, spectrogram_list = [],[],[],[]\r\n \r\n for i in range(0,len(filename_list)):\r\n if filename_list[i][-1] != 'm' and filename_list[i][-1] != 'e':\r\n path = os.path.join(align_root_path, filename_list[i])\r\n path_list.append(path)\r\n \r\n# print(filename_list[i][1:4], filename_list[i][4:])\r\n \r\n with open(path, 'r') as f:\r\n phone = f.read().strip().split(\" \")\r\n phone_list.append(phone)\r\n f.close()\r\n beat_path = os.path.join(pitch_beat_root_path, filename_list[i][1:4], filename_list[i][4:]+\"_beats.txt\")\r\n with open(beat_path, 'r') as f:\r\n beat_list.append(f.read().strip().split(\" \"))\r\n pitch_path = os.path.join(pitch_beat_root_path, filename_list[i][1:4], filename_list[i][4:]+\"_pitches.txt\")\r\n with open(pitch_path, 'r') as f:\r\n pitch_list.append(f.read().strip().split(\" \"))\r\n \r\n wav_path = os.path.join(wav_root_path, filename_list[i][1:4], filename_list[i][4:]+\".wav\")\r\n frame_length = 60/1000\r\n frame_shift = 30/1000 \r\n y, sr = librosa.load(wav_path,sr = None)\r\n hop_length = int(sr * frame_shift)\r\n n_fft = int(sr * frame_length)\r\n spectrogram_list.append(librosa.feature.melspectrogram(y=y, sr=sr,hop_length=hop_length, n_fft = n_fft))\r\n \r\n return phone_list, beat_list, pitch_list, spectrogram_list\r\n\r\n\r\nif __name__ == \"__main__\":\r\n\r\n align_root_path = \"C:/Users/PKU/Desktop/SVS_system/preprocessing/ch_asr/exp/alignment/clean_set/\" #文件夹目录\r\n pitch_beat_root_path = \"C:/Users/PKU/Desktop/SVS_system/preprocessing/ch_asr/exp/pitch_beat_extraction/clean/\"\r\n wav_root_path = 'C:/Users/PKU/Desktop/SVS_system/annotation/clean/'\r\n \r\n phone_list, beat_list, pitch_list, spectrogram_list = Get_align_beat_pitch_spectrogram(align_root_path, pitch_beat_root_path, wav_root_path)\r\n \r\n length = []\r\n for i in range(len(phone_list)):\r\n length.append(len(phone_list[i]))\r\n \r\n sample_num = len(phone_list)\r\n seq_length = max(length)\r\n \r\n \r\n Data = np.zeros((sample_num,seq_length,3))\r\n Label = np.zeros((sample_num,seq_length,128))\r\n \r\n for i in range(sample_num):\r\n for j in range(seq_length):\r\n if j < len(phone_list[i]):\r\n Data[i][j][0] = np.array(phone_list[i][j])\r\n if str(j) in beat_list[i]:\r\n Data[i][j][1] = 1\r\n if j < len(phone_list[i]): # 在这里写phone_list是因为每一个样本,pitch都比phone多一帧(原则:所有以phone为准)\r\n Data[i][j][2] = np.array(pitch_list[i][j])\r\n Label[i][j] = spectrogram_list[i][:,j]\r\n \r\n \r\n #创建子类\r\n class MyDataset(Dataset):\r\n #初始化,定义数据内容和标签\r\n def __init__(self, Data, Label):\r\n self.Data = Data\r\n self.Label = Label\r\n #返回数据集大小\r\n def __len__(self):\r\n return len(self.Data)\r\n #得到数据内容和标签\r\n def __getitem__(self, index):\r\n data = torch.Tensor(self.Data[index])\r\n label = torch.IntTensor(self.Label[index])\r\n return data, label\r\n \r\n dataset = MyDataset(Data, Label)\r\n # print(dataset)\r\n # print('dataset大小为:', dataset.__len__())\r\n # print(dataset.__getitem__(0))\r\n # print(dataset[0])\r\n#\r\n##创建DataLoader迭代器\r\n dataloader = DataLoader(dataset,batch_size= 2, shuffle = False, num_workers= 0)\r\n for i, item in enumerate(dataloader):\r\n print('i:', i)\r\n data, label = item\r\n print('data:', data)\r\n print('label:', label)\r\n \r\n \r\n \r\n", "sub_path": "model/archive/build_dataloader.py", "file_name": "build_dataloader.py", "file_ext": "py", "file_size_in_byte": 4285, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "os.listdir", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "librosa.load", "line_number": 40, "usage_type": "call"}, {"api_name": "librosa.feature.melspectrogram", "line_number": 43, "usage_type": "call"}, {"api_name": "librosa.feature", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 79, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.IntTensor", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 100, "usage_type": "call"}]}
+{"seq_id": "549897859", "text": "import astropy.units as u\nimport numpy as np\n\nimport grasshopper.interferometers as ifo\nimport grasshopper.sources as sources\n\nimport matplotlib.pyplot as plt\nplt.style.use(\"../thesis-style.mpl\")\n\naligo = ifo.AdvancedLIGO()\naligo_o1 = ifo.AdvancedLIGO(configuration=\"O1\")\nfigsize = (5.0, 2.5) # Fix this to use the Golden ratio please\n\nfig, ax = plt.subplots(1,1, figsize=figsize)\naligo.plot(ax)\naligo_o1.plot(ax)\nax.set_xlim([1e1, 2e3]);\nax.set_ylim([1e-23, 1e-19]);\nfig.tight_layout()\nfig.savefig(\"../figures/aligo-asd.pdf\", dpi=300)\n\n\n\n#########################\n\n\n# aligo = ifo.AdvancedLIGO()\n\n\n# figsize = (5.0, 2.5) # Fix this to use the Golden ratio please\n\n# fig, ax = plt.subplots(1,1, figsize=figsize)\n# aligo.plot(ax)\n\n# for mass in [30, 32, 50, 100]:\n# for mass2 in [30, 100, 2000]:\n# cbc = sources.CBC(frequencies=np.logspace(-4, 5, 1000) * u.hertz,\n# m1=mass*u.solMass, m2=mass2*u.solMass, r=0.8*1e9*u.parsec)\n# cbc.plot(ax, label=\"{}, {}\".format(mass, mass2))\n\n# ax.set_xlim([1e1, 2e3]);\n# ax.set_ylim([1e-23, 1e-19]);\n# fig.tight_layout()\n# fig.savefig(\"../figures/aligo-cbc.pdf\", dpi=300)\n\n\n###########################\n\n\nelisa = ifo.EvolvedLISA()\nfigsize = (5.0, 2.5) # Fix this to use the Golden ratio please\n\nfig, ax = plt.subplots(1,1, figsize=figsize)\nelisa.plot(ax)\n#ax.set_xlim([1e-1, 2e3]);\n#ax.set_ylim([1e-23, 1e-19]);\nfig.tight_layout()\nfig.savefig(\"../figures/elisa-asd.pdf\", dpi=300)\n\n\niligo = ifo.InitialLIGO()\nvirgo = ifo.VIRGO()\ngeo = ifo.GEO()\ntama = ifo.TAMA()\n\nfig, ax = plt.subplots(1,1, figsize=figsize)\niligo.plot(ax)\nvirgo.plot(ax)\ngeo.plot(ax)\ntama.plot(ax)\n\nax.set_xlim([1e1, 1e3]);\nax.set_ylim([1e-23, 1e-19]);\nfig.tight_layout()\nfig.savefig(\"../figures/first-gen-asd.pdf\")\n\n\nimport grasshopper.sources as sources\n\nccsn = sources.CoreCollapseSupernova()\nfig, ax = plt.subplots(1,1, figsize=figsize)\naligo.plot(ax)\nccsn.plot(ax)\nax.set_xlim([1e1, 1e3]);\n#ax.set_ylim([1e-23, 1e-19]);\nfig.tight_layout()\nfig.savefig(\"../figures/source-ccsn.pdf\")\n\nccsn = sources.Type1ASupernova(r=30*1000*u.parsec)\nfig, ax = plt.subplots(1,1, figsize=figsize)\naligo.plot(ax)\nccsn.plot(ax)\nax.set_xlim([1e-1, 1e3]);\nax.set_ylim([1e-23, 1e-19]);\nfig.tight_layout()\nfig.savefig(\"../figures/source-t1asn.pdf\")\n\n# Izz = .02#1e-4*10**38#0.28*10**34 / 0.366*1e-4 * (np.sqrt(8*np.pi)/15)\n# pulsar = sources.Pulsar(\"J0534+2200\", Izz=Izz*u.kilogram*u.meter**2)\n# aligo = ifo.AdvancedLIGO(obs_time = 365*3600*u.second)\n\n# fig, ax = plt.subplots(1,1, figsize=figsize)\n# aligo.plot(ax, configuration=\"O1\")\n# pulsar.plot(ax)\n# ax.set_xlim(10, 1000)\n# ax.set_ylim(1e-26, 1e-23)\n# plt.tight_layout()\n# fig.savefig(\"/home/daniel/papers/thesis/figures/crab-strain-o1.pdf\")\n", "sub_path": "scripts/detector_asds.py", "file_name": "detector_asds.py", "file_ext": "py", "file_size_in_byte": 2720, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "matplotlib.pyplot.style.use", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 8, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "grasshopper.interferometers.AdvancedLIGO", "line_number": 10, "usage_type": "call"}, {"api_name": "grasshopper.interferometers", "line_number": 10, "usage_type": "name"}, {"api_name": "grasshopper.interferometers.AdvancedLIGO", "line_number": 11, "usage_type": "call"}, {"api_name": "grasshopper.interferometers", "line_number": 11, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "grasshopper.interferometers.EvolvedLISA", "line_number": 50, "usage_type": "call"}, {"api_name": "grasshopper.interferometers", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "grasshopper.interferometers.InitialLIGO", "line_number": 61, "usage_type": "call"}, {"api_name": "grasshopper.interferometers", "line_number": 61, "usage_type": "name"}, {"api_name": "grasshopper.interferometers.VIRGO", "line_number": 62, "usage_type": "call"}, {"api_name": "grasshopper.interferometers", "line_number": 62, "usage_type": "name"}, {"api_name": "grasshopper.interferometers.GEO", "line_number": 63, "usage_type": "call"}, {"api_name": "grasshopper.interferometers", "line_number": 63, "usage_type": "name"}, {"api_name": "grasshopper.interferometers.TAMA", "line_number": 64, "usage_type": "call"}, {"api_name": "grasshopper.interferometers", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "grasshopper.sources.CoreCollapseSupernova", "line_number": 80, "usage_type": "call"}, {"api_name": "grasshopper.sources", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "grasshopper.sources.Type1ASupernova", "line_number": 89, "usage_type": "call"}, {"api_name": "grasshopper.sources", "line_number": 89, "usage_type": "name"}, {"api_name": "astropy.units.parsec", "line_number": 89, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}]}
+{"seq_id": "261232250", "text": "import os\nimport json\nfrom django.conf import settings\nfrom decouple import config, Csv\nfrom configurations import Configuration, values\nimport logging.config\n\n\nclass Base(Configuration):\n # all the base settings here...\n # Build paths inside the project like this: os.path.join(BASE_DIR, ...)\n BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n\n # Quick-start development settings - unsuitable for production\n # See https://docs.djangoproject.com/en/3.0/howto/deployment/checklist/\n\n SECRET_KEY = config('SECRET_KEY', default='')\n\n ALLOWED_HOSTS = []\n\n # Application definition\n\n INSTALLED_APPS = [\n 'django.contrib.admin',\n 'django.contrib.auth',\n 'django.contrib.contenttypes',\n 'django.contrib.sessions',\n 'django.contrib.messages',\n 'django.contrib.staticfiles',\n 'django.contrib.sites',\n 'crispy_forms',\n 'imagekit',\n\n 'allauth',\n 'allauth.account',\n 'allauth.socialaccount',\n\n 'core',\n 'user',\n ]\n\n MIDDLEWARE = [\n 'django.middleware.security.SecurityMiddleware',\n 'whitenoise.middleware.WhiteNoiseMiddleware',\n 'django.contrib.sessions.middleware.SessionMiddleware',\n 'django.middleware.common.CommonMiddleware',\n 'django.middleware.csrf.CsrfViewMiddleware',\n 'django.contrib.auth.middleware.AuthenticationMiddleware',\n 'django.contrib.messages.middleware.MessageMiddleware',\n 'django.middleware.clickjacking.XFrameOptionsMiddleware',\n ]\n\n ROOT_URLCONF = \"sentive_saas.urls\"\n\n TEMPLATES = [\n {\n 'BACKEND': 'django.template.backends.django.DjangoTemplates',\n 'DIRS': [\n os.path.join(BASE_DIR, 'templates'),\n ],\n 'APP_DIRS': True,\n 'OPTIONS': {\n 'context_processors': [\n 'django.template.context_processors.debug',\n 'django.template.context_processors.request',\n 'django.contrib.auth.context_processors.auth',\n 'django.contrib.messages.context_processors.messages',\n ],\n },\n },\n ]\n\n WSGI_APPLICATION = 'sentive_saas.wsgi.application'\n\n # Password validation\n # https://docs.djangoproject.com/en/3.0/ref/settings/#auth-password-validators\n\n AUTH_PASSWORD_VALIDATORS = [\n {\n 'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator',\n },\n {\n 'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator',\n },\n {\n 'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator',\n },\n {\n 'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator',\n },\n ]\n\n # Internationalization\n # https://docs.djangoproject.com/en/3.0/topics/i18n/\n\n LANGUAGE_CODE = 'fr-FR'\n\n TIME_ZONE = 'UTC'\n\n USE_I18N = True\n\n USE_L10N = True\n\n USE_TZ = True\n\n # Static files (CSS, JavaScript, Images)\n # https://docs.djangoproject.com/en/3.0/howto/static-files/\n SITE_ROOT = os.path.dirname(os.path.realpath(__file__))\n STATIC_ROOT = os.path.join(BASE_DIR, 'root')\n STATIC_URL = '/static/'\n STATICFILES_STORAGE = 'whitenoise.storage.CompressedManifestStaticFilesStorage'\n\n STATICFILES_DIRS = (\n os.path.join(BASE_DIR, \"static\"),\n )\n\n MEDIA_URL = '/media/'\n MEDIA_ROOT = os.path.join(BASE_DIR, 'media/')\n\n CRISPY_TEMPLATE_PACK = 'bootstrap4'\n\n LOGIN_REDIRECT_URL = 'dashboard'\n LOGIN_URL = 'login'\n\n CRISPY_TEMPLATE_PACK = 'bootstrap4'\n\n # Authentification\n AUTHENTICATION_BACKENDS = (\n # Needed to login by username in Django admin, regardless of `allauth`\n 'django.contrib.auth.backends.ModelBackend',\n # `allauth` specific authentication methods, such as login by e-mail\n 'allauth.account.auth_backends.AuthenticationBackend',\n )\n\n AUTH_USER_MODEL = \"core.User\"\n\n SITE_ID = 1\n\n ACCOUNT_EMAIL_VERIFICATION = 'none'\n LOGIN_REDIRECT_URL = '/'\n\n \n # Logging Configuration\n # Clear prev config\n LOGGING_CONFIG = None\n\n # Get loglevel from env\n LOGLEVEL = os.getenv('DJANGO_LOGLEVEL', 'info').upper()\n\n logging.config.dictConfig({\n 'version': 1,\n 'disable_existing_loggers': False,\n 'formatters': {\n 'console': {\n 'format': '%(asctime)s %(levelname)s [%(name)s:%(lineno)s] %(module)s %(process)d %(thread)d %(message)s',\n },\n },\n 'handlers': {\n 'console': {\n 'class': 'logging.StreamHandler',\n 'formatter': 'console',\n },\n },\n 'loggers': {\n '': {\n 'level': LOGLEVEL,\n 'handlers': ['console', ],\n },\n },\n })\n\n\nclass Dev(Base):\n \"\"\"\n The in-development settings and the default configuration.\n \"\"\"\n DEBUG = True\n\n Base.ALLOWED_HOSTS += ['127.0.0.1', '192.168.99.100']\n '''\n DATABASES = {\n 'default': {\n 'ENGINE': config('SQL_DATABASE_ENGINE', default=''),\n 'NAME': config('SQL_DATABASE_NAME_DEV', default=''),\n 'USER': config('SQL_DATABASE_USER_DEV', default=''),\n 'PASSWORD': config('SQL_DATABASE_PASSWORD_DEV', default=''),\n 'HOST': config('SQL_DATABASE_HOST_DEV', default=''),\n 'PORT': config('SQL_DATABASE_PORT_DEV', default=''),\n }\n }\n '''\n\n DATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.sqlite3',\n 'NAME': os.path.join(Base.BASE_DIR, 'db.sqlite3'),\n }\n }\n\n STRIPE_SECRET_KEY = config('STRIPE_SECRET_KEY', default='')\n STRIPE_PUBLISHABLE_KEY = config('STRIPE_PUBLISHABLE_KEY', default='')\n\n SESSION_COOKIE_SECURE = False\n\n\nclass Prod(Base):\n \"\"\"\n The in-production settings.\n \"\"\"\n DEBUG = False\n Base.ALLOWED_HOSTS += ['92.243.19.37', '192.168.99.100']\n TEMPLATE_DEBUG = DEBUG\n\n SESSION_COOKIE_SECURE = False\n", "sub_path": "app/sentive_saas/settings.py", "file_name": "settings.py", "file_ext": "py", "file_size_in_byte": 6114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "configurations.Configuration", "line_number": 9, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 12, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path", "line_number": 118, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 148, "usage_type": "call"}, {"api_name": "logging.config.config.dictConfig", "line_number": 150, "usage_type": "call"}, {"api_name": "logging.config.config", "line_number": 150, "usage_type": "attribute"}, {"api_name": "logging.config", "line_number": 150, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 196, "usage_type": "call"}, {"api_name": "os.path", "line_number": 196, "usage_type": "attribute"}, {"api_name": "decouple.config", "line_number": 200, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 201, "usage_type": "call"}]}
+{"seq_id": "146213473", "text": "import nltk\nfrom nltk.corpus import state_union\nfrom nltk.tokenize import PunktSentenceTokenizer\n\n#purpose of this program is to intrpduce part of speech tagging\n#also chunking is introduced \n\ntrain_text = state_union.raw(\"2005-GWBush.txt\")\nsample_text = state_union.raw(\"2006-GWBush.txt\")\n\n\ncustom_sent_tokenizer = PunktSentenceTokenizer(train_text)\ntokenized = custom_sent_tokenizer.tokenize(sample_text)\n\ndef process_content():\n try:\n for i in tokenized:\n words = nltk.word_tokenize(i[5:])\n tagged = nltk.pos_tag(words)\n\n namedEnt = nltk.ne_chunk(tagged, binary = True)\n namedEnt.draw()\n\n ## Chunking\n ## chunkGram = r\"\"\"Chunk: {+}\n ## }{\"\"\"\n ##chunkParser = nltk.RegexpParser(chunkGram)\n ##chunked = chunkParser.parse(tagged)\n ##chunked.draw()\n\n\n # print(tagged) used for takking\n except Exception as e:\n print(str(e))\n\nprocess_content()", "sub_path": "tagging.py", "file_name": "tagging.py", "file_ext": "py", "file_size_in_byte": 1021, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "nltk.corpus.state_union.raw", "line_number": 8, "usage_type": "call"}, {"api_name": "nltk.corpus.state_union", "line_number": 8, "usage_type": "name"}, {"api_name": "nltk.corpus.state_union.raw", "line_number": 9, "usage_type": "call"}, {"api_name": "nltk.corpus.state_union", "line_number": 9, "usage_type": "name"}, {"api_name": "nltk.tokenize.PunktSentenceTokenizer", "line_number": 12, "usage_type": "call"}, {"api_name": "nltk.word_tokenize", "line_number": 18, "usage_type": "call"}, {"api_name": "nltk.pos_tag", "line_number": 19, "usage_type": "call"}, {"api_name": "nltk.ne_chunk", "line_number": 21, "usage_type": "call"}]}
+{"seq_id": "264030426", "text": "from flask import Flask, request\nfrom flask_sqlalchemy import SQLAlchemy\nfrom flask_migrate import Migrate\n\n\napp = Flask(__name__)\napp.config['SQLALCHEMY_DATABASE_URI'] = \"postgresql://postgres:mazyakidze652@localhost:5432/library\"\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = True\ndb = SQLAlchemy(app)\nmigrate = Migrate(app, db)\n\n\nclass Category(db.Model):\n __tablename__ = 'categories'\n\n id = db.Column(db.Integer, primary_key=True)\n name = db.Column(db.String())\n books = db.relationship('Book', backref='category', lazy=True)\n\n def __init__(self, name):\n self.name = name\n\n def __repr__(self):\n return f\"Category {self.name}\"\n\n\nclass Author(db.Model):\n __tablename__ = 'authors'\n\n id = db.Column(db.Integer, primary_key=True)\n first_name = db.Column(db.String())\n last_name = db.Column(db.String())\n date_of_birth = db.Column(db.Date())\n books = db.relationship('Book', backref='author', lazy=True)\n\n def __init__(self, first_name, last_name, date_of_birth):\n self.first_name = first_name\n self.last_name = last_name\n self.date_of_birth = date_of_birth\n\n def __repr__(self):\n return f\"Author: {self.first_name} {self.last_name}\"\n\n\nclass Book(db.Model):\n __tablename__ = 'books'\n\n id = db.Column(db.Integer, primary_key=True)\n name = db.Column(db.String())\n category_id = db.Column(db.Integer, db.ForeignKey('categories.id'), nullable=False)\n author_id = db.Column(db.Integer, db.ForeignKey('authors.id'), nullable=False)\n content = db.Column(db.Text())\n released_at = db.Column(db.Date())\n\n def __init__(self, name, category_id, author_id, content, released_at):\n self.name = name\n self.category_id = category_id\n self.author_id = author_id\n self.content = content\n self.released_at = released_at\n\n def __repr__(self):\n return f\"Book {self.name}\"\n\n\n@app.route('/authors', methods=['POST', 'GET'])\ndef handle_authors():\n if request.method == 'POST':\n if request.is_json:\n data = request.get_json()\n new_author = Author(first_name=data['first_name'],\n last_name=data['last_name'],\n date_of_birth=data['date_of_birth'])\n db.session.add(new_author)\n db.session.commit()\n return {\"message\": f\"Author {new_author.first_name} {new_author.last_name} has been created successfully.\"}\n else:\n return {\"error\": f\"The request payload is not in JSON format.\"}\n\n elif request.method == \"GET\":\n authors = Author.query.all()\n results = [\n {\n \"id\": author.id,\n \"first_name\": author.first_name,\n \"last_name\": author.last_name,\n \"date_of_birth\": author.date_of_birth,\n } for author in authors\n ]\n\n return {\"count\": len(results), \"authors\": results}\n\n\n@app.route('/authors/', methods=['GET', 'PUT', 'DELETE'])\ndef handle_author(author_id):\n author = Author.query.get_or_404(author_id)\n\n if request.method == 'GET':\n response = {\n \"first_name\": author.first_name,\n \"last_name\": author.last_name,\n \"date_of_birth\": author.date_of_birth,\n }\n return {\"message\": \"success\", \"author\": response}\n\n elif request.method == 'PUT':\n data = request.get_json()\n author.first_name = data['first_name']\n author.last_name = data['last_name']\n author.date_of_birth = data['date_of_birth']\n db.session.add(author)\n db.session.commmmit()\n return {\"message\": f\"Author successfully updated\"}\n\n elif request.method == 'DELETE':\n db.session.delete(author)\n db.session.commit()\n return {\"message\": f\"Author {author.first_name} {author.last_name} successfully deleted\"}\n\n\n@app.route('/categories', methods=['POST', 'GET'])\ndef handle_categories():\n if request.method == 'POST':\n if request.is_json:\n data = request.get_json()\n new_category = Category(name=data['name'])\n db.session.add(new_category)\n db.session.commit()\n return {\"message\": f\"Category {new_category.name} has been created successfully.\"}\n else:\n return {\"error\": f\"The request payload is not in JSON format.\"}\n\n elif request.method == 'GET':\n categories = Category.query.all()\n results = [\n {\n \"id\": category.id,\n \"name\": category.name\n } for category in categories\n ]\n return {\"count\": len(results), \"categories\": results}\n\n\n@app.route('/categories/', methods=['GET', 'PUT', 'DELETE'])\ndef handle_category(category_id):\n category = Category.query.get_or_404(category_id)\n\n if request.method == \"GET\":\n response = {\n \"name\": category.name\n }\n return {\"message\": \"success\", \"category\": response}\n\n elif request.method == 'PUT':\n data = request.get_json()\n category.first_name = data['name']\n db.session.add(category)\n db.session.commmmit()\n return {\"message\": f\"Category successfully updated\"}\n\n elif request.method == 'DELETE':\n db.session.delete(category)\n db.session.commit()\n return {\"message\": f\"Author {category.name} successfully deleted\"}\n\n\n@app.route('/books', methods=['POST', 'GET'])\ndef handle_books():\n if request.method == 'POST':\n if request.is_json:\n data = request.get_json()\n new_book = Book(name=data['name'],\n category_id=data['category_id'],\n author_id=data['author_id'],\n content=data['content'],\n released_at=data['released_at'], )\n db.session.add(new_book)\n db.session.commit()\n return {\"message\": f\"Category {new_book.name} has been created successfully.\"}\n else:\n return {\"error\": f\"The request payload is not in JSON format.\"}\n\n elif request.method == 'GET':\n books = Book.query.all()\n results = [\n {\n \"id\": book.id,\n \"name\": book.name,\n \"category_id\": book.category_id,\n \"author_id\": book.author_id,\n \"content\": book.content,\n \"released_at\": book.released_at,\n } for book in books\n ]\n return {\"count\": len(results), \"books\": results}\n\n\n@app.route('/books/', methods=['GET', 'PUT', 'DELETE'])\ndef handle_book(book_id):\n book = Book.query.get_or_404(book_id)\n\n if request.method == 'GET':\n response = {\n \"name\": book.name,\n \"category_id\": book.category_id,\n \"author_id\": book.author_id,\n \"content\": book.content,\n \"released_at\": book.released_at,\n }\n return {\"message\": \"success\", \"book\": response}\n\n elif request.method == 'PUT':\n data = request.get_json()\n book.name = data['name']\n book.category_id = data['category_id']\n book.author_id = data['author_id']\n book.content = data['content']\n book.released_at = data['released_at']\n db.session.add(book)\n db.session.commmmit()\n return {\"message\": f\"Book successfully updated\"}\n\n elif request.method == 'DELETE':\n db.session.delete(book)\n db.session.commit()\n return {\"message\": f\"Book {book.name} successfully deleted\"}\n\n\nif __name__ == '__main__':\n app.run(debug=True)\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 7628, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 9, "usage_type": "call"}, {"api_name": "flask_migrate.Migrate", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 68, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 68, "usage_type": "name"}, {"api_name": "flask.request.is_json", "line_number": 69, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 69, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 70, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 70, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 80, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 98, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 98, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 106, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 106, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 107, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 107, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 115, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 115, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 123, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 123, "usage_type": "name"}, {"api_name": "flask.request.is_json", "line_number": 124, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 124, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 125, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 125, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 133, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 133, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 148, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 148, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 154, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 154, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 155, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 155, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 161, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 161, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 169, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 169, "usage_type": "name"}, {"api_name": "flask.request.is_json", "line_number": 170, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 170, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 171, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 171, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 183, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 183, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 202, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 202, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 212, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 212, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 213, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 213, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 223, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 223, "usage_type": "name"}]}
+{"seq_id": "4251412", "text": "\"\"\"The AWS Lambda emmaa-test-stats definition.\n\nThis file contains the function that will be run when Lambda is triggered. It\nmust be placed on s3, which can either be done manually (not recommended) or\nby running:\n\n$ python update_lambda.py test_stats.py emmaa-test-stats\n\nin this directory.\n\"\"\"\n\nimport boto3\nfrom datetime import datetime\n\nJOB_DEF = 'emmaa_jobdef'\nQUEUE = 'emmaa-models-update-test'\nPROJECT = 'aske'\nPURPOSE = 'update-emmaa-test-stats'\nBRANCH = 'origin/master'\n\n\ndef lambda_handler(event, context):\n \"\"\"Create a batch job to generate model statistics.\n\n This function is designed to be placed on lambda, taking the event and\n context arguments that are passed, and extracting the names of the\n uploaded (which includes changed) model or test definitions on s3.\n Lambda is configured to be triggered by any such changes, and will\n automatically run this script.\n\n Parameters\n ----------\n event : dict\n A dictionary containing metadata regarding the triggering event. In\n this case, we are expecting 'Records', each of which contains a record\n of a file that was added (or changed) on s3.\n context : object\n This is an object containing potentially useful context provided by\n Lambda. See the documentation cited above for details.\n\n Returns\n -------\n ret : dict\n A dict containing 'statusCode', with a valid HTTP status code, and any\n other data to be returned to Lambda.\n \"\"\"\n batch = boto3.client('batch')\n records = event['Records']\n for rec in records:\n try:\n model_key = rec['s3']['object']['key']\n except KeyError:\n pass\n model_name = model_key.split('/')[1]\n test_corpus = model_key.split('/')[-1][8:-25]\n if not test_corpus:\n test_corpus = 'large_corpus_tests'\n core_command = 'bash scripts/git_and_run.sh'\n if BRANCH is not None:\n core_command += f' --branch {BRANCH}'\n core_command += (' python scripts/run_model_stats_from_s3.py'\n f' --model {model_name} --stats_mode tests'\n f' --tests {test_corpus}')\n print(core_command)\n cont_overrides = {\n 'command': ['python', '-m', 'indra.util.aws', 'run_in_batch',\n '--project', PROJECT, '--purpose', PURPOSE,\n core_command]\n }\n now_str = datetime.utcnow().strftime('%Y%m%d_%H%M%S')\n ret = batch.submit_job(\n jobName=f'{model_name}_{test_corpus}_stats_{now_str}',\n jobQueue=QUEUE, jobDefinition=JOB_DEF,\n containerOverrides=cont_overrides)\n job_id = ret['jobId']\n\n return {'statusCode': 200, 'result': 'SUCCESS', 'job_id': job_id}\n", "sub_path": "emmaa/aws_lambda_functions/test_stats.py", "file_name": "test_stats.py", "file_ext": "py", "file_size_in_byte": 2789, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "boto3.client", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 70, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 70, "usage_type": "name"}]}
+{"seq_id": "634847404", "text": "import re\n# re.match()\n# re.search()\n# re.findall()\n# re.split()\n# re.sub()\n# re.compile()\nimport requests\nimport json\n\n\n# Читаем из файла с помощью конструкции with\ndef read_file(filename):\n with open(filename) as some_file:\n return some_file.read()\n\n\n# Записываем в файл с помощью конструкции with\ndef write_to_file(filename, content, mode='w'):\n with open(filename, mode=mode) as some_file:\n some_file.write(content)\n\n\n# Test 1, with re\n# При помощи requests скачиваем содержимое страниц\n# reddit.com/r/python (любой тред 5 комментов) и вывести\n# пару комметнов и его текстов в консоль\n# http://docs.python-requests.org/en/master/user/quickstart/\n\nif __name__ == '__main__':\n try:\n print(\"start app\")\n some_texts = requests.get('https://habr.com/ru/',\n stream=True)\n\n print(\"HEADERS:\\n\", some_texts.headers)\n print(\"STATUS_CODE:\\n\", some_texts.status_code)\n # контент не можем взять, потому что лимит на скорость надо закинуть\n #print(\"CONTENT:\\n\", some_texts.content)\n #write_to_file(\"habr_html.txt\", str(some_texts.content))\n post_list = re.findall(r'ul', str(some_texts.content))\n print('RE match:\\n', post_list)\n finally:\n print(\"end app\")\n", "sub_path": "task_re.py", "file_name": "task_re.py", "file_ext": "py", "file_size_in_byte": 1490, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "requests.get", "line_number": 33, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 41, "usage_type": "call"}]}
+{"seq_id": "414468254", "text": "#!/usr/bin/env python\n\nimport numpy as np\nfrom matplotlib import pyplot as plt\n\nif __name__ == '__main__':\n x = np.linspace(-np.pi, np.pi, 256, endpoint=True) # type: np.array\n C, S = np.cos(x), np.sin(x)\n plt.plot(x, C, color='red', linewidth='2.4', linestyle='--', label='cosine')\n plt.plot(x, S, color='blue', label='sine')\n plt.xlim(x.min() * 1.1, x.max() * 1.1)\n plt.ylim(C.min() * 1.1, C.max() * 1.1)\n plt.xticks(np.linspace(-np.pi, np.pi, 5, endpoint=True),\n [r'$-\\pi$', r'$-\\pi/2$', r'$0$', r'$\\pi/2$', r'$\\pi$'])\n axis = plt.gca()\n axis.spines['right'].set_color('none')\n axis.spines['top'].set_color('none')\n axis.xaxis.set_ticks_position('bottom')\n axis.spines['bottom'].set_position(('data', 0))\n axis.yaxis.set_ticks_position('left')\n axis.spines['left'].set_position(('data', 0))\n plt.legend(loc='upper left')\n t = 2 * np.pi / 3\n plt.plot([t, t], [0, np.cos(t)], color='red', linestyle='--')\n plt.scatter([t, ], [np.cos(t), ], 50, color='red')\n plt.annotate(r'$\\cos(\\frac{2\\pi}{3})=\\frac{1}{2}$', xy=(t, np.cos(t)), xycoords='data', xytext=(10, 30),\n textcoords='offset points', fontsize='16', color='red',\n arrowprops=dict(arrowstyle='->', connectionstyle='arc3, rad=0.2', color='red'))\n plt.show()\n", "sub_path": "matplotlib_practice.py", "file_name": "matplotlib_practice.py", "file_ext": "py", "file_size_in_byte": 1321, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "numpy.linspace", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 7, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 9, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 10, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 13, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 23, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "numpy.cos", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "numpy.cos", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.annotate", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.cos", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}]}
+{"seq_id": "266168526", "text": "# time O(n), space O(1)\nfrom collections import defaultdict\nclass Solution(object):\n def lengthOfLongestSubstringTwoDistinct(self, s):\n \"\"\"\n :type s: str\n :rtype: int\n \"\"\"\n if not s:\n return 0\n window = defaultdict(int)\n res = 0\n left = 0\n for right, c in enumerate(s):\n window[c] += 1\n while len(window) > 2:\n window[s[left]] -= 1\n if window[s[left]] == 0:\n del window[s[left]]\n left += 1\n res = max(res, right-left+1)\n return res\n \n\n\n\"\"\"\nGiven a string s , find the length of the longest substring t that contains at most 2 distinct characters.\n\nExample 1:\n\nInput: \"eceba\"\nOutput: 3\nExplanation: t is \"ece\" which its length is 3.\nExample 2:\n\nInput: \"ccaabbb\"\nOutput: 5\nExplanation: t is \"aabbb\" which its length is 5.\n\"\"\"\n", "sub_path": "0159. Longest Substring with At Most Two Distinct Characters.py", "file_name": "0159. Longest Substring with At Most Two Distinct Characters.py", "file_ext": "py", "file_size_in_byte": 909, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "collections.defaultdict", "line_number": 11, "usage_type": "call"}]}
+{"seq_id": "197392982", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Jul 3 15:51:26 2019\n\n@author: zclement\n\"\"\"\n\nfrom tweepy.streaming import StreamListener\nfrom tweepy import OAuthHandler\nfrom tweepy import Stream\nfrom kafka import SimpleProducer, KafkaClient\n\n\n\nkafka = KafkaClient(\"192.168.0.10:9092\")\n\nproducer = SimpleProducer(kafka)\n\n\nclass StdOutListener(StreamListener):\n def on_data(self, data):\n producer.send_messages(\"trump_v0\", data.encode('utf-8'))\n print (data)\n return True\n def on_error(self, status):\n print (status)\n\n\n\nl = StdOutListener()\nauth = OAuthHandler(consumer_key, consumer_secret)\nauth.set_access_token(access_token, access_token_secret)\n\n\nstream = Stream(auth= auth, listener= l)\nstream.filter(track=\"trump\") #read tweempy Sream docks .filter", "sub_path": "create_kafka_topic.py", "file_name": "create_kafka_topic.py", "file_ext": "py", "file_size_in_byte": 784, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "kafka.KafkaClient", "line_number": 15, "usage_type": "call"}, {"api_name": "kafka.SimpleProducer", "line_number": 17, "usage_type": "call"}, {"api_name": "tweepy.streaming.StreamListener", "line_number": 20, "usage_type": "name"}, {"api_name": "tweepy.OAuthHandler", "line_number": 31, "usage_type": "call"}, {"api_name": "tweepy.Stream", "line_number": 35, "usage_type": "call"}]}
+{"seq_id": "368199634", "text": "import re\nimport shutil\nimport zipfile\n\nfrom mail_server.utils.log.remote_logger import Logger\n\ndef bill_material_extension_task(file_name, extension, content, mail_path):\n Logger.info(\"Bill material extension task call\")\n return [{'file_name': file_name, 'extension': extension}]\n\ndef zip_extension_task(file_name, extension, content, mail_path):\n\n Logger.info(\"Zip extension task call\")\n Logger.info(\"Zip detected in: \" + mail_path + file_name)\n bills = []\n\n with zipfile.ZipFile(mail_path + file_name, 'r') as zip_file:\n for file_member in zip_file.namelist():\n\n file_name = re.sub(r\"[/\\\\]\", '_', file_member)\n file_name_parts = file_name.split('.', 1 )\n file_extension = \"default\"\n if len(file_name_parts) > 1:\n file_extension = file_name_parts[1].lower()\n\n if file_extension == \"xml\" or file_extension == \"pdf\":\n file_path = mail_path\n\n file_path += file_name\n file_content = zip_file.open(file_member)\n\n with open(file_path, \"wb\") as file_destiny:\n with file_content, file_destiny:\n shutil.copyfileobj(file_content, file_destiny)\n\n bill_info = FILE_EXTENSION_TASKS.get(file_extension, \"default\")(file_name, file_extension, \"\", mail_path)\n for info in bill_info:\n bills.append(info)\n\n return bills\n\nFILE_EXTENSION_TASKS = {\n 'xml': bill_material_extension_task,\n 'pdf': bill_material_extension_task,\n 'zip': zip_extension_task,\n 'default': lambda : []\n}\n\ndef extract_bills_from_attachments(mail_path, attachments):\n bills = {}\n Logger.info(\"Extracting bills from mail content\")\n for attach in attachments:\n file_attach_name = attach.file_name if not (attach.file_name is None) else attach.name\n if not (file_attach_name is None):\n decode_attach_content = attach.content.getvalue().decode(attach.content_transfer_encoding, 'strict')\n Logger.info(\"File name: {}\".format(file_attach_name))\n\n file_name_parts = file_attach_name.split('.', 1 )\n attach_extension = \"default\"\n if len(file_name_parts) > 1:\n attach_extension = file_name_parts[1].lower()\n Logger.info(\"Attach extension: {}\".format(attach_extension))\n\n bill_info = FILE_EXTENSION_TASKS.get(attach_extension, \"default\")(file_attach_name, attach_extension, decode_attach_content, mail_path)\n for info in bill_info:\n if info['extension'] in bills:\n bills[info['extension']].append(info['file_name'])\n else:\n bills[info['extension']] = [info['file_name']]\n\n return bills\n", "sub_path": "mail_server/utils/bill/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 2839, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "mail_server.utils.log.remote_logger.Logger.info", "line_number": 8, "usage_type": "call"}, {"api_name": "mail_server.utils.log.remote_logger.Logger", "line_number": 8, "usage_type": "name"}, {"api_name": "mail_server.utils.log.remote_logger.Logger.info", "line_number": 13, "usage_type": "call"}, {"api_name": "mail_server.utils.log.remote_logger.Logger", "line_number": 13, "usage_type": "name"}, {"api_name": "mail_server.utils.log.remote_logger.Logger.info", "line_number": 14, "usage_type": "call"}, {"api_name": "mail_server.utils.log.remote_logger.Logger", "line_number": 14, "usage_type": "name"}, {"api_name": "zipfile.ZipFile", "line_number": 17, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 20, "usage_type": "call"}, {"api_name": "shutil.copyfileobj", "line_number": 34, "usage_type": "call"}, {"api_name": "mail_server.utils.log.remote_logger.Logger.info", "line_number": 51, "usage_type": "call"}, {"api_name": "mail_server.utils.log.remote_logger.Logger", "line_number": 51, "usage_type": "name"}, {"api_name": "mail_server.utils.log.remote_logger.Logger.info", "line_number": 56, "usage_type": "call"}, {"api_name": "mail_server.utils.log.remote_logger.Logger", "line_number": 56, "usage_type": "name"}, {"api_name": "mail_server.utils.log.remote_logger.Logger.info", "line_number": 62, "usage_type": "call"}, {"api_name": "mail_server.utils.log.remote_logger.Logger", "line_number": 62, "usage_type": "name"}]}
+{"seq_id": "539925190", "text": "from django.shortcuts import render\nfrom .forms import InputForm\nfrom math import sin, tan\n\ndef As(b,d,fc,fy,Mu):\n phi=0.9\n Ru=Mu/(phi*b*d**2)*1000000\n rho_req=0.85*fc/fy*(1-(1-2*Ru/(0.85*fc))**0.5)\n rho_min=max(1.4/fy,0.25*fc**0.5/fy)\n beta_1=0.85-0.05*(fc-27.6)/6.9\n if beta_1<0.65:\n beta_1=0.65\n if beta_1>0.85:\n beta_1=0.85\n rho_max=0.43*0.85*fc/fy*beta_1\n rho=max(min(rho_min,1.33*rho_req),rho_req)\n if rho<=rho_max:\n return round(rho*b*d,0)\n else:\n return \"Increase beam section\"\ndef Vstotal(b,d,fc,fy,Vu):\n Vc=2*fc**0.5*b*d\n phi=0.85\n Vsreq=(Vu*1000-phi*Vc)/phi\n return round(Vsreq/(fy*d))\n\n\ndef compute(request):\n if request.method == 'POST':\n form=InputForm(data=request.POST,auto_id=\"%s\")\n if form.is_valid():\n fc = form.cleaned_data['fc']\n fy = form.cleaned_data['fy']\n gamma = form.cleaned_data['gamma']\n #d = form.cleaned_data['d']\n\n Cx = form.cleaned_data['Cx']\n Cy = form.cleaned_data['Cy']\n\n dp = form.cleaned_data['dp']\n S = form.cleaned_data['S']\n H = form.cleaned_data['H']\n c = form.cleaned_data['c']\n e = form.cleaned_data['e']\n t = form.cleaned_data['t']\n Pc = form.cleaned_data['Pc']\n Pt = form.cleaned_data['Pt']\n\n #Fxd = form.cleaned_data['Fxd']\n Fzd = form.cleaned_data['Fzd']\n Myd = form.cleaned_data['Myd']\n Mxd = form.cleaned_data['Mxd']\n #Fxl = form.cleaned_data['Fxl']\n Fzl = form.cleaned_data['Fzl']\n Myl = form.cleaned_data['Myl']\n Mxl = form.cleaned_data['Mxl']\n #Fxw = form.cleaned_data['Fxw']\n Fzw = form.cleaned_data['Fzw']\n Myw = form.cleaned_data['Myw']\n Mxw = form.cleaned_data['Mxw']\n #Fxe = form.cleaned_data['Fxe']\n Fze = form.cleaned_data['Fze']\n Mye = form.cleaned_data['Mye']\n Mxe = form.cleaned_data['Mxe']\n\n # Pile Cap Dimensions\n L = round(sin(1.047)*S+dp+2*e,1)\n B = S+dp+2*e\n w1 = dp+2*e\n w2 = dp+2*e\n Ldiag = round(((L-w1)**2+(B-w2)**2)**0.5,1)\n d = H - c\n So = S*(3.0)**0.5/2\n Y1 = 2*So/3\n Y2 = So/3\n ay = round(Y1-Cy/2+t,1)\n X = 0.5*(S-Cx)\n m = Y2-Cy/2\n ax = round(((X+m/tan(1.047))*sin(1.047))+t,1)\n radio_x = round(ax/d,3)\n radio_y = round(ay/d,3)\n Bs = 2*(Cx+Cy)\n A = dp/2+e\n\n\n # Pile capacity check\n Acap = L*B-sin(1.047)*Ldiag**2/2\n Gc = gamma*(H*0.001)*(Acap*0.000001)\n Gc = round(Gc,2)\n\n Fv1 = 1.0 * (Fzd + Gc) + 1.0 * Fzl\n My1 = Myd + Myl\n Mx1 = Mxd + Mxl\n x1 = 0.0; x2 = -0.5 * S; x3 = 0.5 *S\n y1 = 0.578 * S; y2 = -0.288 * S; y3 = -0.288 * S\n sum_xx = x1**2+x2**2+x3**2\n sum_yy = y1**2+y2**2+y3**2\n def Rp(F,Mx,My,n,x,y):\n return F/n+1000*Mx*y/sum_yy+1000*My*x/sum_xx\n Rp_1 = round(Rp(Fv1, Mx1, My1, 3, x1, y1))\n Rp_2 = round(Rp(Fv1, Mx1, My1, 3, x2, y2))\n Rp_3 = round(Rp(Fv1, Mx1, My1, 3, x3, y3))\n\n if Rp_1 <= Pc and Rp_1>=Pt and Rp_2 <= Pc and Rp_2>=Pt and Rp_3 <= Pc and Rp_3>=Pt:\n VS1 = 'orange'\n result1 = 'Satisfied'\n else:\n VS1 = \"red\"\n result1 = \"Failed\"\n # Pile reaction\n #Cx1 = 1.4 * Fxd\n Cz1 = 1.4 * Fzd\n Czg1 = 1.4 * (Fzd + Gc)\n Cmy1 = 1.4 * Myd\n Cmx1 = 1.4 * Mxd\n\n #Cx2 = 1.2 * Fxd + 1.6 * Fxl\n Cz2 = 1.2 * Fzd + 1.6 * Fzl\n Czg2 = 1.2 * (Fzd + Gc) + 1.6 * Fzl\n Cmy2 = 1.2 * Myd + 1.6 * Myl\n Cmx2 = 1.2 * Mxd + 1.6 * Mxl\n\n #Cx3 = 1.2 * Fxd + 1.6 * Fxw + 1.0 * Fxl\n Cz3 = 1.2 * Fzd + 1.6 * Fzw + 1.0 * Fzl\n Czg3 = 1.2 * (Fzd + Gc) + 1.6 * Fzw + 1.0 * Fzl\n Cmy3 = 1.2 * Myd + 1.6 * Myw + 1.0 * Myl\n Cmx3 = 1.2 * Mxd + 1.6 * Mxw + 1.0 * Mxl\n\n #Cx4 = 1.2 * Fxd + 1.0 * Fxe + 1.0 * Fxl\n Cz4 = 1.2 * Fzd + 1.0 * Fze + 1.0 * Fzl\n Czg4 = 1.2 * (Fzd + Gc) + 1.0 * Fze + 1.0 * Fzl\n Cmy4 = 1.2 * Myd + 1.0 * Mye + 1.0 * Myl\n Cmx4 = 1.2 * Mxd + 1.0 * Mxe + 1.0 * Mxl\n\n #Cx5 = 0.9 * Fxd + 1.6 * Fxw\n Cz5 = 0.9 * Fzd + 1.6 * Fzw\n Czg5 = 0.9 * (Fzd + Gc) + 1.6 * Fzw\n Cmy5 = 0.9 * Myd + 1.6 * Myw\n Cmx5 = 0.9 * Mxd + 1.6 * Mxw\n\n #Cx6 = 0.9 * Fxd + 1.6 * Fxe\n Cz6 = 0.9 * Fzd + 1.6 * Fze\n Czg6 = 0.9 * (Fzd + Gc) + 1.6 * Fze\n Cmy6 = 0.9 * Myd + 1.6 * Mye\n Cmx6 = 0.9 * Mxd + 1.6 * Mxe\n\n Rp11 = round(Rp(Czg1, Cmx1, Cmy1, 3, x1, y1), 1)\n Rp21 = round(Rp(Czg1, Cmx1, Cmy1, 3, x2, y2), 1)\n Rp31 = round(Rp(Czg1, Cmx1, Cmy1, 3, x3, y3), 1)\n Rp12 = round(Rp(Czg2, Cmx2, Cmy2, 3, x1, y1), 1)\n Rp22 = round(Rp(Czg2, Cmx2, Cmy2, 3, x2, y2), 1)\n Rp32 = round(Rp(Czg2, Cmx2, Cmy2, 3, x3, y3), 1)\n Rp13 = round(Rp(Czg3, Cmx3, Cmy3, 3, x1, y1), 1)\n Rp23 = round(Rp(Czg3, Cmx3, Cmy3, 3, x2, y2), 1)\n Rp33 = round(Rp(Czg3, Cmx3, Cmy3, 3, x3, y3), 1)\n Rp14 = round(Rp(Czg4, Cmx4, Cmy4, 3, x1, y1), 1)\n Rp24 = round(Rp(Czg4, Cmx4, Cmy4, 3, x2, y2), 1)\n Rp34 = round(Rp(Czg4, Cmx4, Cmy4, 3, x3, y3), 1)\n Rp15 = round(Rp(Czg5, Cmx5, Cmy5, 3, x1, y1), 1)\n Rp25 = round(Rp(Czg5, Cmx5, Cmy5, 3, x2, y2), 1)\n Rp35 = round(Rp(Czg5, Cmx5, Cmy5, 3, x3, y3), 1)\n Rp16 = round(Rp(Czg6, Cmx6, Cmy6, 3, x1, y1), 1)\n Rp26 = round(Rp(Czg6, Cmx6, Cmy6, 3, x2, y2), 1)\n Rp36 = round(Rp(Czg6, Cmx6, Cmy6, 3, x3, y3), 1)\n\n # Two way shear design\n phiVc = round(0.85 *d/max(ax,ay) * (1+2*d/(Cx+Cy)) * (0.17 * (fc) ** 0.5 * Bs * d) * 0.001, 2)\n\n VV1 = '%.2f' % (Czg1 / phiVc)\n if (Czg1 / phiVc) < 1:\n vs1 = \"Pass\"\n co_vs1 = \"orange\"\n else:\n vs1 = \"Fail\"\n co_vs1 = \"red\"\n VV2 = '%.2f' % (Czg2 / phiVc)\n if (Czg2 / phiVc) < 1:\n vs2 = \"Pass\"\n co_vs2 = \"orange\"\n else:\n vs2 = \"Fail\"\n co_vs2 = \"red\"\n VV3 = '%.2f' % (Czg3 / phiVc)\n if (Czg3 / phiVc) < 1:\n vs3 = \"Pass\"\n co_vs3 = \"orange\"\n else:\n vs3 = \"Fail\"\n co_vs3 = \"red\"\n VV4 = '%.2f' % (Czg4 / phiVc)\n if (Czg4 / phiVc) < 1:\n vs4 = \"Pass\"\n co_vs4 = \"orange\"\n else:\n vs4 = \"Fail\"\n co_vs4 = \"red\"\n VV5 = '%.2f' % (Czg5 / phiVc)\n if (Czg5 / phiVc) < 1:\n vs5 = \"Pass\"\n co_vs5 = \"orange\"\n else:\n vs5 = \"Fail\"\n co_vs5 = \"red\"\n VV6 = '%.2f' % (Czg6 / phiVc)\n if (Czg6 / phiVc) < 1:\n vs6 = \"Pass\"\n co_vs6 = \"orange\"\n else:\n vs6 = \"Fail\"\n co_vs6 = \"red\"\n\n # One Way Shear\n if radio_x >= 1:\n Lcx = (0.5 * (X + A + (A + m) / tan(1.047)-d/sin(1.047)) * tan(1.047) + A) / sin(1.047)\n phiVc_x = round(0.85 * (0.17 * (fc) ** 0.5 * Lcx * d) * 0.001, 1)\n check1 = \"Concrete shear strength for one way wide beam action:\"\n check2 = \"$\\phi V_c=0.85\\cdot(0.17\\sqrt{f'_c} \\cdot L_{cx}\\cdot d)=$\"+str(phiVc_x)+\"kN\"\n check3 = \"$L_{cx}=(0.5(X+A+(A+m)ctg(\\pi/3)-d/sin(\\pi/3))*tan(\\pi/3)+A)/sin(\\pi/3)=$\"+str(round(Lcx))+\"mm\"\n else:\n kx = min(round(d/ax * (3.5-2.5*ax/d),2),10)\n Lcx = (0.5 * (X + A + (A + m) / tan(1.047)) * tan(1.047) + A) / sin(1.047)\n phiVc_x = round(0.85 * kx * (0.17 * (fc) ** 0.5 * Lcx * d) * 0.001, 1)\n check1 = \"Concrete shear strength for one way deep beam action:\"\n check2 = \"$\\phi V_c=0.85k\\cdot(0.17\\sqrt{f'_c} \\cdot L_{cx}\\cdot d)=$\"+str(phiVc_x)+\"kN;\"+\"$\\quad k=\\cfrac{d}{a}\\cdot (3.5-2.5 \\cfrac {a}{d})$\"\n check3 = \"$L_{cx}=(0.5(X+A+(A+m)ctg(\\pi/3))*tan(\\pi/3)+A)/sin(\\pi/3)=$\"+str(round(Lcx))+\"mm\"\n Vux1 = max(Rp21, Rp31)\n Vux2 = max(Rp22, Rp32)\n Vux3 = max(Rp23, Rp33)\n Vux4 = max(Rp24, Rp34)\n Vux5 = max(Rp25, Rp35)\n Vux6 = max(Rp26, Rp36)\n V1 = '%.2f' % (Vux1 / phiVc_x)\n if (Vux1 / phiVc_x) < 1:\n s1 = \"Pass\"\n co_s1 = \"orange\"\n else:\n s1 = \"Fail\"\n co_s1 = \"red\"\n V2 = '%.2f' % (Vux2 / phiVc_x)\n if (Vux2 / phiVc_x) < 1:\n s2 = \"Pass\"\n co_s2 = \"orange\"\n else:\n s2 = \"Fail\"\n co_s2 = \"red\"\n V3 = '%.2f' % (Vux3 / phiVc_x)\n if (Vux3 / phiVc_x) < 1:\n s3 = \"Pass\"\n co_s3 = \"orange\"\n else:\n s3 = \"Fail\"\n co_s3 = \"red\"\n V4 = '%.2f' % (Vux4 / phiVc_x)\n if (Vux4 / phiVc_x) < 1:\n s4 = \"Pass\"\n co_s4 = \"orange\"\n else:\n s4 = \"Fail\"\n co_s4 = \"red\"\n V5 = '%.2f' % (Vux5 / phiVc_x)\n if (Vux5 / phiVc_x) < 1:\n s5 = \"Pass\"\n co_s5 = \"orange\"\n else:\n s5 = \"Fail\"\n co_s5 = \"red\"\n V6 = '%.2f' % (Vux6 / phiVc_x)\n if (Vux6 / phiVc_x) < 1:\n s6 = \"Pass\"\n co_s6 = \"orange\"\n else:\n s6 = \"Fail\"\n co_s6 = \"red\"\n if radio_y >= 1:\n Lcy = 2 * A + 2 * ((ay - t - d) + A) / tan(1.047)\n phiVc_y = round(0.85 * (0.17 * (fc) ** 0.5 * Lcy * d) * 0.001, 1)\n check4 = \"Concrete shear strength for one way wide beam action:\"\n check5 = \"$\\phi V_c=0.85\\cdot(0.17\\sqrt{f'_c} \\cdot L_{cx}\\cdot d)=$\"+str(phiVc_y)+\"kN\"\n check6 = \"$L_{cx}=2A+2((a_y-t-d)+A)/tan(\\pi/3)=$\"+str(round(Lcy))+\"mm\"\n else:\n ky = min(round(d/ax * (3.5-2.5*ax/d),2),10)\n Lcy = 2 * A + 2 * ((ay - t - d) + A) / tan(1.047)\n phiVc_y = round(0.85 * ky * (0.17 * (fc) ** 0.5 * Lcy * d) * 0.001, 1)\n check4 = \"Concrete shear strength for one way deep beam action:\"\n check5 = \"$\\phi V_c=0.85k\\cdot(0.17\\sqrt{f'_c} \\cdot L_{cx}\\cdot d)=$\"+str(phiVc_y)+\"kN;\"+\"$\\quad k=\\cfrac{d}{a}\\cdot (3.5-2.5 \\cfrac {a}{d})$\"\n check6 = \"$L_{cx}=2A+2((a_y-t-d)+A)/tan(\\pi/3)==$\"+str(round(Lcy))+\"mm\"\n VVV1 = '%.2f' % (Rp11 / phiVc_y)\n if (Rp11 / phiVc_y) < 1:\n vvs1 = \"Pass\"\n co_vvs1 = \"orange\"\n else:\n vvs1 = \"Fail\"\n co_vvs1 = \"red\"\n VVV2 = '%.2f' % (Rp12 / phiVc_y)\n if (Rp12 / phiVc_y) < 1:\n vvs2 = \"Pass\"\n co_vvs2 = \"orange\"\n else:\n vvs2 = \"Fail\"\n co_vvs2 = \"red\"\n VVV3 = '%.2f' % (Rp13 / phiVc_y)\n if (Rp13 / phiVc_y) < 1:\n vvs3 = \"Pass\"\n co_vvs3 = \"orange\"\n else:\n vvs3 = \"Fail\"\n co_vvs3 = \"red\"\n VVV4 = '%.2f' % (Rp14 / phiVc_y)\n if (Rp14 / phiVc_y) < 1:\n vvs4 = \"Pass\"\n co_vvs4 = \"orange\"\n else:\n vvs4 = \"Fail\"\n co_vvs4 = \"red\"\n VVV5 = '%.2f' % (Rp15 / phiVc_y)\n if (Rp15 / phiVc_y) < 1:\n vvs5 = \"Pass\"\n co_vvs5 = \"orange\"\n else:\n vvs5 = \"Fail\"\n co_vvs5 = \"red\"\n VVV6 = '%.2f' % (Rp16 / phiVc_y)\n if (Rp16 / phiVc_y) < 1:\n vvs6 = \"Pass\"\n co_vvs6 = \"orange\"\n else:\n vvs6 = \"Fail\"\n co_vvs6 = \"red\"\n\n # Bending Reinforcement\n def As(M,B):\n rho_min = 0.0018\n As_min=rho_min*B*H\n R = M * 10 ** 6 / (0.9 * B * d ** 2)\n rho_req = 0.85 * fc / fy * (1 - (1 - 2 * R / (0.85 * fc)) ** 0.5)\n As_req = rho_req * B * (H-c)\n return max(As_min,As_req)\n\n BX = round(A * (1.5 + 0.5 * tan(1.047))/sin(1.047), 1)\n Mux1 = round(max(Rp21,Rp31) * ax * 0.001, 1)\n Mux2 = round(max(Rp22,Rp32) * ax * 0.001, 1)\n Mux3 = round(max(Rp23,Rp33) * ax * 0.001, 1)\n Mux4 = round(max(Rp24,Rp34) * ax * 0.001, 1)\n Mux5 = round(max(Rp25,Rp35) * ax * 0.001, 1)\n Mux6 = round(max(Rp26,Rp36) * ax * 0.001, 1)\n\n Asx1 = '%.0f' % float(As(Mux1, BX))\n Asx2 = '%.0f' % float(As(Mux2, BX))\n Asx3 = '%.0f' % float(As(Mux3, BX))\n Asx4 = '%.0f' % float(As(Mux4, BX))\n Asx5 = '%.0f' % float(As(Mux5, BX))\n Asx6 = '%.0f' % float(As(Mux6, BX))\n\n BY = round(2*A*(1+1/tan(1.047)),1)\n Muy1 = round(Rp11*ay*0.001,1)\n Muy2 = round(Rp12 * ay * 0.001, 1)\n Muy3 = round(Rp13 * ay * 0.001, 1)\n Muy4 = round(Rp14 * ay * 0.001, 1)\n Muy5 = round(Rp15 * ay * 0.001, 1)\n Muy6 = round(Rp16 * ay * 0.001, 1)\n\n\n Asy1 = '%.0f' % float(As(Muy1,BY))\n Asy2 = '%.0f' % float(As(Muy2,BY))\n Asy3 = '%.0f' % float(As(Muy3,BY))\n Asy4 = '%.0f' % float(As(Muy4,BY))\n Asy5 = '%.0f' % float(As(Muy5,BY))\n Asy6 = '%.0f' % float(As(Muy6,BY))\n\n Czg1 = round(Czg1,1)\n Czg2 = round(Czg2, 1)\n Czg3 = round(Czg3, 1)\n Czg4 = round(Czg4, 1)\n Czg5 = round(Czg5, 1)\n Czg6 = round(Czg6, 1)\n\n\n return render(request, 'pile_cap_3.html', {'form': form,'Gc':Gc,'Fv1':Fv1,'My1':My1,'Mx1':Mx1,\n 'L':L,'B':B,'w1':w1,'w2':w2,'Ldiag':Ldiag,'d':d,'ax':ax,'ay':ay,'radio_x':radio_x,'radio_y':radio_y,\n 'Rp_1':Rp_1,'Rp_2':Rp_2,'Rp_3':Rp_3,'VS1':VS1,'result1':result1,\n 'Rp11':Rp11,'Rp21':Rp21,'Rp31':Rp31,'Rp12':Rp12,'Rp22':Rp22,'Rp32':Rp32,\n 'Rp13':Rp13,'Rp23':Rp23,'Rp33':Rp33,'Rp14':Rp14,'Rp24':Rp24,'Rp34':Rp34,\n 'Rp15': Rp15, 'Rp25': Rp25, 'Rp35': Rp35, 'Rp16': Rp16,\n 'Rp26': Rp26, 'Rp36': Rp36,\n 'phiVc':phiVc,'Czg1':Czg1,'Czg2':Czg2,'Czg3':Czg3,'Czg4':Czg4,'Czg5':Czg5,'Czg6':Czg6,\n 'VV1': VV1, 'VV2': VV2, 'VV3': VV3, 'VV4': VV4, 'VV5': VV5,\n 'VV6': VV6, 'vs1': vs1, 'vs2': vs2, 'vs3': vs3, 'vs4': vs4,\n 'vs5': vs5, 'vs6': vs6,\n 'co_vs1': co_vs1, 'co_vs2': co_vs2, 'co_vs3': co_vs3,\n 'co_vs4': co_vs4, 'co_vs5': co_vs5, 'co_vs6': co_vs6,\n 'check1':check1,'check2':check2,'check3':check3,'check4':check4,'check5':check5,'check6':check6,\n 'Vux1':Vux1,'Vux2':Vux2,'Vux3':Vux3,'Vux4':Vux4,'Vux5':Vux5,'Vux6':Vux6,\n 'V1': V1, 'V2': V2, 'V3': V3, 'V4': V4, 'V5': V5,\n 'V6': V6, 's1': s1, 's2': s2, 's3': s3, 's4': s4,\n 's5': s5, 's6': s6,\n 'co_s1': co_s1, 'co_s2': co_s2, 'co_s3': co_s3,\n 'co_s4': co_s4, 'co_s5': co_s5, 'co_s6': co_s6,\n 'VVV1': VVV1, 'VVV2': VVV2, 'VVV3': VVV3, 'VVV4': VVV4, 'VVV5': VVV5,\n 'VVV6': VVV6, 'vvs1': vvs1, 'vvs2': vvs2, 'vvs3': vvs3, 'vvs4': vvs4,\n 'vvs5': vvs5, 'vvs6': vvs6,\n 'co_vvs1': co_vvs1, 'co_vvs2': co_vvs2, 'co_vvs3': co_vvs3,\n 'co_vvs4': co_vvs4, 'co_vvs5': co_vvs5, 'co_vvs6': co_vvs6,\n 'BX':BX,'BY':BY,\n 'Mux1': Mux1, 'Mux2': Mux2, 'Mux3': Mux3, 'Mux4': Mux4,\n 'Mux5': Mux5, 'Mux6': Mux6,\n 'Asx1': Asx1, 'Asx2': Asx2, 'Asx3': Asx3, 'Asx4': Asx4,\n 'Asx5': Asx5, 'Asx6': Asx6,\n 'Muy1':Muy1,'Muy2':Muy2,'Muy3':Muy3,'Muy4':Muy4,'Muy5':Muy5,'Muy6':Muy6,\n 'Asy1':Asy1,'Asy2':Asy2,'Asy3':Asy3,'Asy4':Asy4,'Asy5':Asy5,'Asy6':Asy6,})\n else:\n form = InputForm(auto_id=\"%s\")\n return render(request,'pile_cap_3.html',{'form':form})\n", "sub_path": "pile_cap_3/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 18091, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "forms.InputForm", "line_number": 30, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 67, "usage_type": "call"}, {"api_name": "math.tan", "line_number": 79, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 79, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 87, "usage_type": "call"}, {"api_name": "math.tan", "line_number": 214, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 214, "usage_type": "call"}, {"api_name": "math.tan", "line_number": 221, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 221, "usage_type": "call"}, {"api_name": "math.tan", "line_number": 275, "usage_type": "call"}, {"api_name": "math.tan", "line_number": 282, "usage_type": "call"}, {"api_name": "math.tan", "line_number": 339, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 339, "usage_type": "call"}, {"api_name": "math.tan", "line_number": 354, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 378, "usage_type": "call"}, {"api_name": "forms.InputForm", "line_number": 411, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 412, "usage_type": "call"}]}
+{"seq_id": "115394287", "text": "import itertools\n\nimport math\n\nfrom deap import tools\n\nimport numpy as np\n\nfrom .base_strategy import BaseEvolutionStrategy\nfrom .individual import Individual\nfrom .mfitness import MultidimensionalFitness\n\n\nclass DynamicDifferentialEvolution(BaseEvolutionStrategy):\n def __init__(self, population_size=10, population_regular=4,\n population_brownian=2, cr=0.6, f=0.4, bounds=(-1.0, 1.0),\n **kwargs):\n super(DynamicDifferentialEvolution, self).__init__(**kwargs)\n\n self.population_size = population_size\n self.cr = cr\n self.f = f\n self.bounds = bounds\n self.population_regular = population_regular\n self.population_brownian = population_brownian\n\n def best_individual(self):\n return self.hall_of_fame[0]\n\n def generate_brow_ind_with_fitness(self, best, sigma=0.3):\n fitness_len = len(self.fitness)\n ind = Individual(np.random.normal(x, sigma) for x in best)\n ind.fitness = MultidimensionalFitness(fitness_len)\n return ind\n\n def fit(self, X, y):\n # Differential evolution parameters\n individual_size = self.n_dim\n # Should be equal to the number of peaks\n population_size = self.population_size\n\n regular = self.population_regular\n brownian = self.population_brownian\n bounds = self.bounds\n\n toolbox = self.create_toolbox(X, y)\n toolbox.register(\"attr_float\", np.random.uniform, -1, 1)\n toolbox.register(\n \"individual\",\n self.generate_individual_with_fitness,\n toolbox.attr_float,\n individual_size)\n toolbox.register(\n \"brownian_individual\",\n self.generate_brow_ind_with_fitness)\n toolbox.register(\n \"population\",\n tools.initRepeat,\n list,\n toolbox.individual)\n\n toolbox.register(\"select\", np.random.choice, size=4)\n toolbox.register(\"best\", tools.selBest, k=1)\n\n self.hall_of_fame = tools.HallOfFame(1)\n stats = self._build_stats()\n\n self.logbook = tools.Logbook()\n self.logbook.header = ['gen', 'nevals'] \\\n + (stats.fields if stats else [])\n\n # Initialize populations\n populations = [toolbox.population(n=regular + brownian)\n for _ in range(population_size)]\n\n # Evaluate the individuals\n for idx, subpop in enumerate(populations):\n fitness = toolbox.map(toolbox.evaluate, subpop)\n for ind, fit in zip(subpop, fitness):\n ind.fitness.values = fit\n\n if stats:\n record = stats.compile(itertools.chain(*populations))\n self.logbook.record(gen=0, evals=len(populations), **record)\n if self.verbose:\n print(self.logbook.stream)\n\n for g in range(1, self.n_gen):\n # Detect a change and invalidate fitness if necessary\n bests = [toolbox.best(subpop)[0] for subpop in populations]\n if any(b.fitness.values != toolbox.evaluate(b) for b in bests):\n for individual in itertools.chain(*populations):\n del individual.fitness.values\n\n # Apply exclusion\n rexcl = (bounds[1] - bounds[0]) \\\n / (2 * population_size**(1.0 / individual_size))\n for i, j in itertools.combinations(range(population_size), 2):\n if bests[i].fitness.valid and bests[j].fitness.valid:\n d = sum((bests[i][k] - bests[j][k])**2\n for k in range(individual_size))\n d = math.sqrt(d)\n\n if d < rexcl:\n if bests[i].fitness < bests[j].fitness:\n k = i\n else:\n k = j\n\n populations[k] = toolbox.population(\n n=regular + brownian)\n\n # Evaluate the individuals with an invalid fitness\n invalid_ind = [ind for ind in itertools.chain(*populations)\n if not ind.fitness.valid]\n fitness = toolbox.map(toolbox.evaluate, invalid_ind)\n for ind, fit in zip(invalid_ind, fitness):\n ind.fitness.values = fit\n\n all_pops = list(itertools.chain(*populations))\n self.hall_of_fame.update(all_pops)\n\n if stats:\n record = stats.compile(all_pops)\n self.logbook.record(gen=g, evals=len(populations), **record)\n if self.verbose:\n print(self.logbook.stream)\n\n # Evolve the sub-populations\n for idx, subpop in enumerate(populations):\n newpop = []\n xbest, = toolbox.best(subpop)\n # Apply regular DE to the first part of the population\n for individual in subpop[:regular]:\n idxs = np.random.choice(len(subpop), size=4)\n x1, x2, x3, x4 = subpop[idxs[0]], subpop[idxs[1]], \\\n subpop[idxs[2]], subpop[idxs[3]]\n offspring = toolbox.clone(individual)\n index = np.random.randint(individual_size)\n for i, _ in enumerate(individual):\n if i == index or np.random.random() < self.cr:\n offspring[i] = xbest[i] + self.f \\\n * (x1[i] + x2[i] - x3[i] - x4[i])\n offspring.fitness.values = toolbox.evaluate(offspring)\n if offspring.fitness >= individual.fitness:\n newpop.append(offspring)\n else:\n newpop.append(individual)\n\n # Apply Brownian to the last part of the population\n newpop.extend(toolbox.brownian_individual(xbest)\n for _ in range(brownian))\n\n # Evaluate the brownian individuals\n for individual in newpop[-brownian:]:\n individual.fitness.value = toolbox.evaluate(individual)\n\n # Replace the population\n populations[idx] = newpop\n\n self.cleanup()\n return self\n", "sub_path": "metric_learn/evolution/strategy/dde.py", "file_name": "dde.py", "file_ext": "py", "file_size_in_byte": 6276, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "base_strategy.BaseEvolutionStrategy", "line_number": 14, "usage_type": "name"}, {"api_name": "individual.Individual", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 32, "usage_type": "attribute"}, {"api_name": "mfitness.MultidimensionalFitness", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 47, "usage_type": "attribute"}, {"api_name": "deap.tools.initRepeat", "line_number": 58, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 58, "usage_type": "name"}, {"api_name": "numpy.random", "line_number": 62, "usage_type": "attribute"}, {"api_name": "deap.tools.selBest", "line_number": 63, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 63, "usage_type": "name"}, {"api_name": "deap.tools.HallOfFame", "line_number": 65, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 65, "usage_type": "name"}, {"api_name": "deap.tools.Logbook", "line_number": 68, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 68, "usage_type": "name"}, {"api_name": "itertools.chain", "line_number": 83, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 92, "usage_type": "call"}, {"api_name": "individual.fitness", "line_number": 93, "usage_type": "attribute"}, {"api_name": "itertools.combinations", "line_number": 98, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 102, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 114, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 135, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 139, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 141, "usage_type": "attribute"}, {"api_name": "individual.fitness", "line_number": 145, "usage_type": "attribute"}, {"api_name": "individual.fitness", "line_number": 156, "usage_type": "attribute"}]}
+{"seq_id": "468055957", "text": "import tensorflow as tf\nimport numpy as np\nfrom collections import OrderedDict\nimport pandas as pd\n\nfile=\"glove.6B.50d.txt\"\n\ndf=pd.read_csv(file,sep=\" \",quoting=3, header=None, index_col=0)\nglove = {key: val.values for key, val in df.T.items()}\n\nwords=list(glove.keys())\nemb=np.array(list(glove.values()))\n\ninput_str = \"like the country\"\nword_to_idx = OrderedDict({w:words.index(w) for w in input_str.split() if w in words})\n\ntf.InteractiveSession()\ntf_embedding = tf.constant(emb, dtype=tf.float32)\ntf.nn.embedding_lookup(tf_embedding, list(word_to_idx.values())).eval()\n\n\n", "sub_path": "QuestionAnswering/MohamedUvaiz/glove_embedding.py", "file_name": "glove_embedding.py", "file_ext": "py", "file_size_in_byte": 574, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "pandas.read_csv", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 12, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 15, "usage_type": "call"}, {"api_name": "tensorflow.InteractiveSession", "line_number": 17, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 18, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 18, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.embedding_lookup", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 19, "usage_type": "attribute"}]}
+{"seq_id": "352468832", "text": "import data\nimport pymysql\nfrom connection import ConnectionWrapper\n\n\ndef runQuery(query, args, fetch=False, db=data.DATABASE):\n cw = ConnectionWrapper(db)\n try:\n cursor = cw.getCursor()\n cursor.execute(query, args)\n\n if fetch:\n result = cursor.fetchall()\n else:\n result = None\n cw.commit()\n return result\n\n finally:\n cw.close()\n\n\ndef createTestingDB():\n query = \"CREATE DATABASE \" + data.DATABASE + \";\"\n runQuery(query, None, False, None)\n\n\ndef dropTestingDB():\n query = \"DROP DATABASE IF EXISTS \" + data.DATABASE + \";\"\n runQuery(query, None, False, None)\n\n\ndef createTables(create_script_name):\n with open(create_script_name, 'r') as myfile:\n query = myfile.read()\n\n # data processing\n query = query.replace(\"\\t\", \"\")\n query = query.replace(\"\\n\", \"\")\n queries = query.split(\";\")\n\n for i in range(len(queries) - 1): # Skip the last comments\n runQuery(queries[i] + \";\", None, False)\n\n\n \ndef searchInfoByPartialConditions(tableName, info_format, cons_format=None, keyword=None): # quotes\n query = \"\"\n if cons_format == None and keyword == None:\n query = (\"SELECT \" + info_format + \" FROM \" + tableName + \";\")\n else:\n kword = ('%s'%keyword)\n cons_format = (cons_format + \" like '%\" + kword + \"%'\")\n query = (\n \"SELECT \" + info_format + \" FROM \" + tableName + \" WHERE \" +\n cons_format + \";\"\n )\n return runQuery(query, None, True)\n\n\ndef selectInfoByConditions(tableName, info_format, cons_format=None, vals=None): # quotes\n query = \"\"\n if cons_format == None and vals == None:\n query = (\"SELECT \" + info_format + \" FROM \" + tableName + \";\")\n else:\n query = (\n \"SELECT \" + info_format + \" FROM \" + tableName + \" WHERE \" +\n cons_format + \";\"\n )\n query = query % vals\n return runQuery(query, None, True)\n\n\ndef searchInfoByConditions(tableName, info_format, cons_format=None, vals=None): # quotes\n query = \"\"\n if cons_format == None and vals == None:\n query = (\"SELECT \" + info_format + \" FROM \" + tableName + \";\")\n else:\n query = (\n \"SELECT \" + info_format + \" FROM \" + tableName + \" WHERE \" +\n cons_format + \";\"\n )\n query = query % vals\n return runQuery(query, None, True)\n\n\ndef selectAllByConditions(tableName, cons_format=None, vals=None):\n return selectInfoByConditions(tableName, \"*\", cons_format, vals)\n\n\ndef getNumOfRecordByConditions(tableName, cons_format=None, vals=None):\n return len(selectAllByConditions(tableName, cons_format, vals))\n\n\ndef checkRecordExistByConditions(tableName, cons_format=None, vals=None):\n return (getNumOfRecordByConditions(tableName, cons_format, vals) > 0)\n\n\ndef deleteRecordByCondition(tableName, cons_format, vals):\n query = (\"DELETE FROM \" + tableName + \" WHERE \" + cons_format + \";\")\n query = query % vals\n runQuery(query, None, False)\n return True\n\n\ndef insertRecordTo(tableName, cols, vals, vals_format):\n query = \"INSERT INTO \" + tableName + cols + \" VALUES \" + vals_format + \";\"\n runQuery(query, vals, False)\n return True\n\n\ndef insertRecordForcibly(tableName, user_id, info):\n info = pymysql.escape_string(info)\n query = 'INSERT INTO %s (user_id, info) VALUES (\"%s\", \"%s\") ON DUPLICATE KEY UPDATE info=\"%s\";' % (\n tableName, user_id, info, info)\n runQuery(query, None, False)\n return True\n\n\ndef updateRecordByConditions(tableName, info_format, cons_format, vals):\n query = (\n \"UPDATE \" + tableName + \" SET \" + info_format + \" WHERE \"\n + cons_format + \";\"\n )\n query = query % vals\n runQuery(query, None, False)\n return True\n\n\ndef getValsByKey(result_list, key): # return a list\n vals = []\n if (len(result_list) <= 0) or (key not in result_list[0]):\n return vals\n\n for i in range(len(result_list)):\n vals.append(result_list[i][key])\n\n return vals\n", "sub_path": "server/db/dbhelper/queries.py", "file_name": "queries.py", "file_ext": "py", "file_size_in_byte": 4013, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "data.DATABASE", "line_number": 6, "usage_type": "attribute"}, {"api_name": "connection.ConnectionWrapper", "line_number": 7, "usage_type": "call"}, {"api_name": "data.DATABASE", "line_number": 24, "usage_type": "attribute"}, {"api_name": "data.DATABASE", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pymysql.escape_string", "line_number": 113, "usage_type": "call"}]}
+{"seq_id": "358301458", "text": "import dj_database_url\nfrom .settings import *\n\nDATABASES = {\n 'default': dj_database_url.config(),\n}\nSTATIC_ROOT = os.path.join(BASE_DIR, 'staticfiles')\nSECURE_PROXY_SSL_HEADER = ('HTTP_X_FORWARDED_PROTO', 'https')\nALLOWED_HOSTS = ['*']\nDEBUG = True", "sub_path": "mysite/production_settings.py", "file_name": "production_settings.py", "file_ext": "py", "file_size_in_byte": 253, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "dj_database_url.config", "line_number": 5, "usage_type": "call"}]}
+{"seq_id": "396190021", "text": "from binascii import hexlify\nfrom utils import TailableProc\nfrom ephemeral_port_reserve import reserve\n\nimport src.lpd.python_binding.common_pb2 as common_pb2\nimport src.lpd.python_binding.channel_pb2 as channel_pb2\nimport src.lpd.python_binding.payment_pb2 as payment_pb2\nimport src.lpd.python_binding.routing_pb2 as routing_pb2\nfrom src.lpd.python_binding.channel_pb2_grpc import ChannelServiceStub\nfrom src.lpd.python_binding.payment_pb2_grpc import PaymentServiceStub\nfrom src.lpd.python_binding.routing_pb2_grpc import RoutingServiceStub\n\nimport grpc\nimport logging\nimport os\nimport time\nimport codecs\n\n\nclass LpdD(TailableProc):\n\n def __init__(self, lightning_dir, bitcoind, port):\n super().__init__(lightning_dir, 'lpd({})'.format(port))\n self.lightning_dir = lightning_dir\n self.bitcoind = bitcoind\n self.port = port\n self.rpc_port = str(reserve())\n self.prefix = 'lpd'\n\n self.cmd_line = [\n 'bin/lpd',\n '--rpclisten=127.0.0.1:{}'.format(self.rpc_port),\n ]\n\n if not os.path.exists(lightning_dir):\n os.makedirs(lightning_dir)\n\n def start(self):\n super().start()\n self.wait_for_log('RPC server listening on')\n self.wait_for_log('Done catching up block hashes')\n time.sleep(5)\n\n logging.info('LPD started (pid: {})'.format(self.proc.pid))\n\n def stop(self):\n self.proc.terminate()\n time.sleep(3)\n if self.proc.poll() is None:\n self.proc.kill()\n self.proc.wait()\n super().save_log()\n\n\nclass LpdNode(object):\n\n displayName = 'lpd'\n\n def __init__(self, lightning_dir, lightning_port, bitcoind, executor=None, node_id=0):\n self.bitcoin = bitcoind\n self.executor = executor\n self.daemon = LpdD(lightning_dir, bitcoind, port=lightning_port)\n self.rpc = LpdRpc(self.daemon.rpc_port)\n self.logger = logging.getLogger('lpd-node({})'.format(lightning_port))\n self.myid = None\n self.node_id = node_id\n\n def id(self):\n if not self.myid:\n self.myid = self.info()['id']\n return self.myid\n\n def ping(self):\n \"\"\" Simple liveness test to see if the node is up and running\n\n Returns true if the node is reachable via RPC, false otherwise.\n \"\"\"\n try:\n self.rpc.routing.GetInfo(common_pb2.Void())\n return True\n except Exception as e:\n print(e)\n return False\n\n def peers(self):\n peers = self.rpc.routing.ListPeers(common_pb2.Void()).peers\n return [p.pub_key for p in peers]\n\n def check_channel(self, remote):\n \"\"\" Make sure that we have an active channel with remote\n \"\"\"\n self_id = self.id()\n remote_id = remote.id()\n channels = self.rpc.channel.List(channel_pb2.ChannelFilter()).channels\n channel_by_remote = {c.remote_pubkey: c for c in channels}\n if remote_id not in channel_by_remote:\n self.logger.warning(\"Channel {} -> {} not found\".format(self_id, remote_id))\n return False\n\n channel = channel_by_remote[remote_id]\n self.logger.debug(\"Channel {} -> {} state: {}\".format(self_id, remote_id, channel))\n return channel.active\n\n def addfunds(self, bitcoind, satoshis):\n req = wallet_pb2.NewAddressRequest(type=1)\n addr = self.rpc.wallet.NewAddress(req).address\n bitcoind.rpc.sendtoaddress(addr, float(satoshis) / 10**8)\n self.daemon.wait_for_log(\"Inserting unconfirmed transaction\")\n bitcoind.rpc.generate(1)\n self.daemon.wait_for_log(\"Marking unconfirmed transaction\")\n\n # The above still doesn't mean the wallet balance is updated,\n # so let it settle a bit\n i = 0\n while self.rpc.wallet.WalletBalance(wallet_pb2.WalletBalanceRequest()).total_balance == satoshis and i < 30:\n time.sleep(1)\n i += 1\n assert(self.rpc.wallet.WalletBalance(wallet_pb2.WalletBalanceRequest()).total_balance == satoshis)\n\n def openchannel(self, node_id, host, port, satoshis):\n peers = self.rpc.routing.ListPeers(common_pb2.Void).peers\n peers_by_pubkey = {p.pub_key: p for p in peers}\n if node_id not in peers_by_pubkey:\n raise ValueError(\"Could not find peer {} in peers {}\".format(node_id, peers))\n peer = peers_by_pubkey[node_id]\n self.rpc.channel.Open(channel_pb2.OpenChannelRequest(\n node_pubkey=codecs.decode(peer.pub_key, 'hex_codec'),\n local_funding_amount=common_pb2.Satoshi(value=satoshis),\n push_sat=0\n ))\n\n # Somehow broadcasting a tx is slow from time to time\n time.sleep(5)\n\n def getchannels(self):\n req = routing_pb2.ChannelGraphRequest()\n rep = self.rpc.routing.DescribeGraph(req)\n channels = []\n\n for e in rep.edges:\n channels.append((e.node1_pub, e.node2_pub))\n channels.append((e.node2_pub, e.node1_pub))\n return channels\n\n def getnodes(self):\n req = routing_pb2.ChannelGraphRequest()\n rep = self.rpc.routing.DescribeGraph(req)\n nodes = set([n.pub_key for n in rep.nodes]) - set([self.id()])\n return nodes\n\n def invoice(self, amount):\n req = payment_pb2.Invoice(value=common_pb2.Satoshi(value=int(amount)))\n rep = self.rpc.payment.AddInvoice(req)\n return rep.payment_request\n\n def send(self, bolt11):\n req = payment_pb2.SendRequest(payment_request=bolt11)\n res = self.rpc.payment.SendPaymentSync(req)\n if res.payment_error:\n raise ValueError(res.payment_error)\n return hexlify(res.payment_preimage)\n\n def connect(self, host, port, node_id):\n addr = routing_pb2.LightningAddress(pubkey=node_id, host=\"{}:{}\".format(host, port))\n req = routing_pb2.ConnectPeerRequest(addr=addr, perm=True)\n logging.debug(self.rpc.routing.ConnectPeer(req))\n\n def info(self):\n r = self.rpc.routing.GetInfo(common_pb2.Void())\n return {\n 'id': r.identity_pubkey,\n 'blockheight': r.block_height,\n }\n\n def block_sync(self, blockhash):\n print(\"Waiting for node to learn about\", blockhash)\n self.daemon.wait_for_log('NTFN: New block: height=([0-9]+), sha={}'.format(blockhash))\n\n def restart(self):\n self.daemon.stop()\n time.sleep(5)\n self.daemon.start()\n self.rpc = LpdRpc(self.daemon.rpc_port)\n\n def check_route(self, node_id, amount):\n try:\n req = routing_pb2.QueryRoutesRequest(pub_key=node_id, amt=int(amount/1000), num_routes=1)\n r = self.rpc.routing.QueryRoutes(req)\n except grpc._channel._Rendezvous as e:\n if str(e).find(\"unable to find a path to destination\") > 0:\n return False\n raise\n return True\n\n\nclass LpdRpc(object):\n def __init__(self, rpc_port):\n self.port = rpc_port\n cred = grpc.ssl_channel_credentials(open('tls.cert').read())\n channel = grpc.secure_channel('localhost:{}'.format(rpc_port), cred)\n self.channel = ChannelServiceStub(channel)\n self.payment = PaymentServiceStub(channel)\n self.routing = RoutingServiceStub(channel)\n", "sub_path": "lpd.py", "file_name": "lpd.py", "file_ext": "py", "file_size_in_byte": 7282, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "utils.TailableProc", "line_number": 20, "usage_type": "name"}, {"api_name": "ephemeral_port_reserve.reserve", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 36, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 42, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 44, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 48, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 64, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.common_pb2.Void", "line_number": 79, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.common_pb2", "line_number": 79, "usage_type": "name"}, {"api_name": "src.lpd.python_binding.common_pb2.Void", "line_number": 86, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.common_pb2", "line_number": 86, "usage_type": "name"}, {"api_name": "src.lpd.python_binding.channel_pb2.ChannelFilter", "line_number": 94, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.channel_pb2", "line_number": 94, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 116, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.common_pb2.Void", "line_number": 121, "usage_type": "attribute"}, {"api_name": "src.lpd.python_binding.common_pb2", "line_number": 121, "usage_type": "name"}, {"api_name": "src.lpd.python_binding.channel_pb2.OpenChannelRequest", "line_number": 126, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.channel_pb2", "line_number": 126, "usage_type": "name"}, {"api_name": "codecs.decode", "line_number": 127, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.common_pb2.Satoshi", "line_number": 128, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.common_pb2", "line_number": 128, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 133, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.routing_pb2.ChannelGraphRequest", "line_number": 136, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.routing_pb2", "line_number": 136, "usage_type": "name"}, {"api_name": "src.lpd.python_binding.routing_pb2.ChannelGraphRequest", "line_number": 146, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.routing_pb2", "line_number": 146, "usage_type": "name"}, {"api_name": "src.lpd.python_binding.payment_pb2.Invoice", "line_number": 152, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.payment_pb2", "line_number": 152, "usage_type": "name"}, {"api_name": "src.lpd.python_binding.common_pb2.Satoshi", "line_number": 152, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.common_pb2", "line_number": 152, "usage_type": "name"}, {"api_name": "src.lpd.python_binding.payment_pb2.SendRequest", "line_number": 157, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.payment_pb2", "line_number": 157, "usage_type": "name"}, {"api_name": "binascii.hexlify", "line_number": 161, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.routing_pb2.LightningAddress", "line_number": 164, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.routing_pb2", "line_number": 164, "usage_type": "name"}, {"api_name": "src.lpd.python_binding.routing_pb2.ConnectPeerRequest", "line_number": 165, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.routing_pb2", "line_number": 165, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 166, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.common_pb2.Void", "line_number": 169, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.common_pb2", "line_number": 169, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 181, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.routing_pb2.QueryRoutesRequest", "line_number": 187, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.routing_pb2", "line_number": 187, "usage_type": "name"}, {"api_name": "grpc._channel", "line_number": 189, "usage_type": "attribute"}, {"api_name": "grpc.ssl_channel_credentials", "line_number": 199, "usage_type": "call"}, {"api_name": "grpc.secure_channel", "line_number": 200, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.channel_pb2_grpc.ChannelServiceStub", "line_number": 201, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.payment_pb2_grpc.PaymentServiceStub", "line_number": 202, "usage_type": "call"}, {"api_name": "src.lpd.python_binding.routing_pb2_grpc.RoutingServiceStub", "line_number": 203, "usage_type": "call"}]}
+{"seq_id": "309512251", "text": "import os\nimport json\nfrom flask import Flask, render_template, session, redirect, url_for, flash, jsonify, request, make_response\nfrom flask.ext.script import Manager\nfrom flask.ext.bootstrap import Bootstrap\nfrom flask.ext.sqlalchemy import SQLAlchemy\nfrom flask.ext.wtf import Form\nfrom wtforms import StringField, SubmitField, SelectField\nfrom wtforms.validators import Required\n\nbasedir = os.path.abspath(os.path.dirname(__file__))\n\napp = Flask(__name__)\napp.config['SECRET_KEY'] = 'hard to guess string'\napp.config['SQLALCHEMY_DATABASE_URI'] =\\\n'sqlite:///' + os.path.join(basedir, 'data.sqlite')\napp.config['SQLALCHEMY_COMMIT_ON_TEARDOWN'] = True\n\nmanager = Manager(app)\nbootstrap = Bootstrap(app)\ndb = SQLAlchemy(app)\n\nclass Record(db.Model):\n __tablename__ = 'records'\n id = db.Column(db.Integer, primary_key = True)\n patient_id = db.Column(db.Integer)\n date = db.Column(db.Date)\n probability = db.Column(db.Float)\n duration_since_lastECOPD = db.Column(db.Integer)\n previous_ECOPD_nb = db.Column(db.Integer)\n previous_hospital_nb = db.Column(db.Integer)\n duration_since_lastHospital = db.Column(db.Integer)\n Age = db.Column(db.Integer)\n Weight = db.Column(db.Integer)\n Height = db.Column(db.Integer)\n BMI = db.Column(db.Float)\n Pack_Years = db.Column(db.Integer)\n last_O2FR_Prescribed = db.Column(db.Float)\n def __repr__(self):\n return str(patient_id) + ',' + str(date) + ',' + str(probability)\n\nclass NameForm(Form):\n name = StringField('What is your name?', validators=[Required()])\n submit = SubmitField('Submit')\n\nclass RecordForm(Form):\n patient_id = SelectField(u'', choices=())\n date = SelectField(u'', choices=())\n submit = SubmitField('Submit')\n\n@app.route('/', methods = ['GET'])\ndef index():\n patient_ids = set([x.patient_id for x in Record.query.all()])\n patient_dict = {}\n for i in patient_ids:\n patient_dict[i] = []\n for x in Record.query.all():\n patient_dict[x.patient_id].append(x.date)\n return render_template('index.html', patient_dict = patient_dict)\n\n@app.route('/search', methods=['GET', 'POST'])\ndef select():\n \"\"\"\n Render a vehicle selection form and handle form submission\n \"\"\"\n form = RecordForm()\n patient_ids = set([x.patient_id for x in Record.query.all()])\n form.patient_id.choices = [('', '--- Select One Patient ---')] + [(x, x) for x in patient_ids]\n if request.method == 'POST':\n patient_id = form.patient_id.data\n date = form.date.data\n probability = Record.query.filter_by(patient_id = patient_id, date = date).first()\n if probability is None:\n return render_template('404.html'), 404\n session['patient_id'] = patient_id\n session['date'] = date\n session['probability'] = probability.probability\n session['duration_since_lastECOPD'] = probability.duration_since_lastECOPD\n session['previous_ECOPD_nb'] = probability.previous_ECOPD_nb\n session['duration_since_lastHospital'] = probability.duration_since_lastHospital\n session['previous_hospital_nb'] = probability.previous_hospital_nb\n session['Age'] = probability.Age\n session['Weight'] = probability.Weight\n session['Height'] = probability.Height\n session['BMI'] = probability.BMI\n session['Pack_Years'] = probability.Pack_Years\n session['last_O2FR_Prescribed'] = probability.last_O2FR_Prescribed \n return redirect(url_for('result'))\n return render_template('select.html', form = form)\n\n@app.route('/result', methods = ['GET'])\ndef result():\n if 'patient_id' not in session or 'date' not in session or 'probability' not in session:\n return render_template('404.html'), 404\n return render_template('result.html', patient_id = session['patient_id'], date = session['date'], \n probability = session['probability'],\n duration_since_lastECOPD= session['duration_since_lastECOPD'],\n previous_ECOPD_nb = session['previous_ECOPD_nb'],\n duration_since_lastHospital = session['duration_since_lastHospital'],\n previous_hospital_nb = session['previous_hospital_nb'],\n Age = session['Age'],\n Weight = session['Weight'],\n Height = session['Height'],\n BMI = session['BMI'],\n Pack_Years = session['Pack_Years'],\n last_O2FR_Prescribed = session['last_O2FR_Prescribed'])\n\n@app.route('/patients//', methods = ['GET'])\ndef get(patient_id):\n \"\"\"\n Handle a GET request at /patients//\n Return a list of 2-tuples (, )\n \"\"\"\n data = [(str(x.date), str(x.date)) for x in Record.query.filter_by(patient_id = patient_id).all()] \n response = make_response(json.dumps(data))\n response.content_type = 'application/json'\n return response\n\n@app.route('/patient', methods = ['GET'])\ndef patient():\n patient_id = request.args.get('patient_id', type = int)\n date = request.args.get('date', type = str)\n query = Record.query.filter_by(patient_id = patient_id, date = date).first()\n result = {}\n result['patient_id'] = query.patient_id\n result['date'] = str(query.date)\n result['probability'] = query.probability\n result['duration_since_lastECOPD'] = query.duration_since_lastECOPD\n result['previous_ECOPD_nb'] = query.previous_ECOPD_nb\n result['duration_since_lastHospital'] = query.duration_since_lastHospital\n result['previous_hospital_nb'] = query.previous_hospital_nb\n result['Age'] = query.Age\n result['Weight'] = query.Weight\n result['Height'] = query.Height\n result['BMI'] = query.BMI\n result['Pack_Years'] = query.Pack_Years\n result['last_O2FR_Prescribed'] = query.last_O2FR_Prescribed \n return json.dumps(result)\n\n@app.errorhandler(404)\ndef page_not_found(e):\n return render_template('404.html'), 404\n\n@app.errorhandler(500)\ndef internal_server_error(e):\n return render_template('500.html'), 500\n\n\n#@app.route('/radial')\n#def radial():\n# return render_template('RadialprogressTest.html')\n\n#@app.route('/', methods=['GET', 'POST'])\n#def index():\n# form = NameForm()\n# if form.validate_on_submit():\n# old_name = session.get('name')\n# if old_name is not None and old_name != form.name.data:\n# flash('Looks like you have changed your name!')\n# session['name'] = form.name.data\n# return redirect(url_for('index'))\n# return render_template('index.html', form=form, name=session.get('name'))\n\n\n##@app.route('/add_numbers')\n#def add_numbers():\n# return render_template('add_numbers.html')\n\n#@app.route('/_add_numbers')\n#def _add_numbers():\n# a = request.args.get('a', 0, type=int)\n# b = request.args.get('b', 0, type=int)\n# return jsonify(result=a + b)\n\n#@app.route('/circle')\n#def circle():\n# return render_template('circle.html')\n\nif __name__ == '__main__':\n manager.run()\n", "sub_path": "hello.py", "file_name": "hello.py", "file_ext": "py", "file_size_in_byte": 6875, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "os.path.abspath", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 11, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask.ext.script.Manager", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.ext.bootstrap.Bootstrap", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.ext.sqlalchemy.SQLAlchemy", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.ext.wtf.Form", "line_number": 42, "usage_type": "name"}, {"api_name": "wtforms.StringField", "line_number": 43, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 43, "usage_type": "call"}, {"api_name": "wtforms.SubmitField", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.ext.wtf.Form", "line_number": 46, "usage_type": "name"}, {"api_name": "wtforms.SelectField", "line_number": 47, "usage_type": "call"}, {"api_name": "wtforms.SelectField", "line_number": 48, "usage_type": "call"}, {"api_name": "wtforms.SubmitField", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 69, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 69, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 75, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 76, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 77, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 81, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 82, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 83, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 84, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 85, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 86, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 87, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 89, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 93, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 94, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 95, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 95, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 96, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 97, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 98, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 99, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 100, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 101, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 102, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 103, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 105, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 106, "usage_type": "name"}, {"api_name": "flask.make_response", "line_number": 115, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 115, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 121, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 121, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 121, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 122, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 122, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 122, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 138, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 142, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 146, "usage_type": "call"}]}
+{"seq_id": "168882819", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.7 (3394)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.macosx-10.15-x86_64/egg/foxylib/tools/google/youtube/tests/test_youtube_tool.py\n# Compiled at: 2020-01-08 12:53:55\n# Size of source mod 2**32: 1534 bytes\nimport logging, re\nfrom functools import lru_cache\nfrom unittest import TestCase\nfrom foxylib.tools.google.youtube.youtube_tool import YoutubeTool\nfrom foxylib.tools.log.foxylib_logger import FoxylibLogger\nfrom foxylib.tools.regex.regex_tool import MatchTool\n\nclass TestYoutubeTool(TestCase):\n\n @classmethod\n def setUpClass(cls):\n FoxylibLogger.attach_stderr2loggers(logging.DEBUG)\n\n def test_01(self):\n url = 'https://www.youtube.com/watch?v=4VYAaLh3XZg&t=33s'\n hyp = YoutubeTool.url2video_id(url)\n ref = '4VYAaLh3XZg'\n self.assertEqual(hyp, ref)\n\n def test_02(self):\n video_id = '4VYAaLh3XZg'\n hyp = YoutubeTool.video_id2url(video_id)\n ref = 'https://www.youtube.com/watch?v=4VYAaLh3XZg'\n self.assertEqual(hyp, ref)\n\n def test_03(self):\n url = 'https://www.youtube.com/watch?v=4VYAaLh3XZg'\n hyp = YoutubeTool.url2is_accessible(url)\n self.assertTrue(hyp)\n\n def test_04(self):\n p = YoutubeTool.pattern_url()\n url1 = 'http://youtu.be/5Y6HSHwhVlY'\n self.assertTrue(p.match(url1))\n self.assertEqual(YoutubeTool.url2video_id(url1), '5Y6HSHwhVlY')\n url2 = 'http://www.youtube.com/embed/5Y6HSHwhVlY?rel=0'\n self.assertTrue(p.match(url2))\n self.assertEqual(YoutubeTool.url2video_id(url2), '5Y6HSHwhVlY')\n url3 = 'http://www.youtube.com/watch?v=ZFqlHhCNBOI'\n self.assertTrue(p.match(url3))\n self.assertEqual(YoutubeTool.url2video_id(url3), 'ZFqlHhCNBOI')", "sub_path": "pycfiles/foxylib-0.3.96-py3.7/test_youtube_tool.cpython-37.py", "file_name": "test_youtube_tool.cpython-37.py", "file_ext": "py", "file_size_in_byte": 1842, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "unittest.TestCase", "line_number": 15, "usage_type": "name"}, {"api_name": "foxylib.tools.log.foxylib_logger.FoxylibLogger.attach_stderr2loggers", "line_number": 19, "usage_type": "call"}, {"api_name": "foxylib.tools.log.foxylib_logger.FoxylibLogger", "line_number": 19, "usage_type": "name"}, {"api_name": "logging.DEBUG", "line_number": 19, "usage_type": "attribute"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool.url2video_id", "line_number": 23, "usage_type": "call"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool", "line_number": 23, "usage_type": "name"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool.video_id2url", "line_number": 29, "usage_type": "call"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool", "line_number": 29, "usage_type": "name"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool.url2is_accessible", "line_number": 35, "usage_type": "call"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool", "line_number": 35, "usage_type": "name"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool.pattern_url", "line_number": 39, "usage_type": "call"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool", "line_number": 39, "usage_type": "name"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool.url2video_id", "line_number": 42, "usage_type": "call"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool", "line_number": 42, "usage_type": "name"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool.url2video_id", "line_number": 45, "usage_type": "call"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool", "line_number": 45, "usage_type": "name"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool.url2video_id", "line_number": 48, "usage_type": "call"}, {"api_name": "foxylib.tools.google.youtube.youtube_tool.YoutubeTool", "line_number": 48, "usage_type": "name"}]}
+{"seq_id": "185096163", "text": "import time\nimport json\nimport Tkinter as tk \nimport ttk\n\nimport data_bus\nimport wire\nimport GLOBALS\n\nclass Reliability_CruiseCommands(ttk.Frame):\n\tdef __init__(self, root, parent, data_bus, options):\n\t\tself.root = root\n\t\tself.parent = parent\n\t\tself.data_bus = data_bus\n\t\tself.data_bus.record_callback.append(self.handleNewRecord)\n\t\tself.options = options\n\n\t\tself.cruiseCommandFrame = ttk.Labelframe(self.parent, text='Cruise Commands')\n\n\t\tself.setSpeed = ttk.Label(self.cruiseCommandFrame, text='Set Speed: ')\n\t\tself.setSpeed.grid(row=1, column=1)\n\t\tself.setSpeedData = ttk.Label(self.cruiseCommandFrame, text='N/A')\n\t\tself.setSpeedData.grid(row=1, column=2)\n\n\t\tself.limit = ttk.Label(self.cruiseCommandFrame, text='Limit: ')\n\t\tself.limit.grid(row=2, column=1)\n\t\tself.limitData = ttk.Label(self.cruiseCommandFrame, text='N/A')\n\t\tself.limitData.grid(row=2, column=2)\n\n\t\tself.speedEntry = ttk.Entry(self.cruiseCommandFrame, width=5)\n\t\tself.speedEntry.grid(row=3, column=1)\n\t\tself.sendSpeed = ttk.Button(self.cruiseCommandFrame, text='Send Speed')\n\t\tself.sendSpeed.grid(row=3, column=2)\n\n\t\tself.limitEntry = ttk.Entry(self.cruiseCommandFrame, width=5)\n\t\tself.limitEntry.grid(row=4, column=1)\n\t\tself.sendLimit = ttk.Button(self.cruiseCommandFrame, text='Send Limit')\n\t\tself.sendLimit.grid(row=4, column=2)\n\n\tdef handleNewRecord(self, record):\n\t\tfields = record.field.split('\\0')\n\t\tname = fields[0].split('.')\n\t\tif name == ['cruise']:\n\t\t\tmeta = json.loads(fields[1])\n\t\t\tself.harness = wire.Harness(meta['harness'])\n\t\t\trecord.value_callback.append(self.handleValue)\n\t\t\trecord.Subscribe()\n\n\tdef handleValue(self, record):\n\t\tself.harness.buf = buffer(record.value)\n\t\tself.setSpeedData['text'] = \"%.2f\" % (self.harness['speed'].value*GLOBALS.SPEED_UNITS_MULTIPLIER[self.options.unitsVar.get()])\n\t\tself.limitData['text'] = \"%.2f\" % (self.harness['limit'].value*GLOBALS.SPEED_UNITS_MULTIPLIER[self.options.unitsVar.get()])", "sub_path": "Telemetry/RF Telems/onboard/modules_can2014/Reliability_CruiseCommandsModule.py", "file_name": "Reliability_CruiseCommandsModule.py", "file_ext": "py", "file_size_in_byte": 1912, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "ttk.Frame", "line_number": 10, "usage_type": "attribute"}, {"api_name": "ttk.Labelframe", "line_number": 18, "usage_type": "call"}, {"api_name": "ttk.Label", "line_number": 20, "usage_type": "call"}, {"api_name": "ttk.Label", "line_number": 22, "usage_type": "call"}, {"api_name": "ttk.Label", "line_number": 25, "usage_type": "call"}, {"api_name": "ttk.Label", "line_number": 27, "usage_type": "call"}, {"api_name": "ttk.Entry", "line_number": 30, "usage_type": "call"}, {"api_name": "ttk.Button", "line_number": 32, "usage_type": "call"}, {"api_name": "ttk.Entry", "line_number": 35, "usage_type": "call"}, {"api_name": "ttk.Button", "line_number": 37, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 44, "usage_type": "call"}, {"api_name": "wire.Harness", "line_number": 45, "usage_type": "call"}, {"api_name": "GLOBALS.SPEED_UNITS_MULTIPLIER", "line_number": 51, "usage_type": "attribute"}, {"api_name": "GLOBALS.SPEED_UNITS_MULTIPLIER", "line_number": 52, "usage_type": "attribute"}]}
+{"seq_id": "69305459", "text": "# Pending\r\n# Stories based on gender, age\r\n# BLack magic intensity shuttle...6\r\n# Long Poetry\r\n# Manual command to start a different topic. Or \r\n# Create 60 short stories. Total 20 stories and each story will have a happy, a neutral/sad and a curious version\r\n# Curious version will be used as a filler story to divert to a happy or sad story if user is not responding. or to change topic..\r\n# Each story is an intent type. This is becuase this will help UnivEncoder to match similarity to one story intent only.\r\n# Each story will have an opening intent\r\n# Each story intent will have a timeout intent, which will get triggered when timeout happens.\r\n# Create a structure or bag of keywords which will suggest that we need to switch stories now as \r\n# conversation is moving in a different direction. Like a key of words for a story and a probability indicator indicating where\r\n# current conversation is going. Over the conversation as probability grows, we will shift story.\r\n# need to knwo if a particular utterance has already been said. \r\n# Choose a different utterance based on the universal encoder output. next on universal encoder list.\r\n# A way to understand the progress of the story\r\n# restart chatbot when user goes away from camera\r\n# what is the next intent if current intent is already satisfied? or which sentence(or index) has bot spoken of. So that it is not repetive.\r\n\r\nimport pandas as pd\r\nimport numpy as np\r\nimport os\r\nimport math\r\nimport json\r\nimport tensorflow as tf\r\nimport tensorflow_hub as hub\r\nfrom random import randrange\r\nimport random\r\nimport string\r\n\r\n# global main_intents\r\n# main_intents = {}\r\n## kk code for eliza start ##\r\n#from nltk.chat.util import Chat\r\n#from nltk.chat.eliza import pairs\r\n## kk code for eliza stop ##\r\n\r\ndefault_utterances = ['yes', 'no', 'maybe', 'okay', 'sure']\r\nclass Story:\r\n def __init__(self, name, intents = {}, completion_status = 0, tone = \"happy\", starting_intent = {}, script = {}, keywords = [], timeout_intent = {}, utterances_said = [], transition_intent = {}):\r\n\r\n self.id = ''.join([random.choice(string.ascii_letters + string.digits) for n in range(15)]) \r\n self.name = name\r\n self.intents = intents\r\n\r\n self.completion_status = completion_status\r\n self.tone = tone\r\n self.keywords = keywords\r\n self.starting_intent = starting_intent\r\n self.script_intent = script\r\n self.timeout_intent = timeout_intent\r\n self.transition_intent = transition_intent\r\n self.utterances_said = utterances_said\r\n\r\n # What if the user wants to again start the story??? You should have an intent that this is what you can say about story and\r\n # now you shold tell him some other story...\r\n\r\n # transition intent will giv hint about two three different stories....\r\n # there will be two three transition intents...\r\n\r\n def create_timeout_intent(self, intent_name, weight, utterances = [], responses = []):\r\n if type(intent_name)==list: # Iterate over all the values in list\r\n for name in intent_name:\r\n self.add_intent(name, weight, utterances, responses)\r\n self.timeout_intent[intent_name] = self.intents[intent_name]\r\n\r\n else: # insert without iterating\r\n self.add_intent(intent_name, weight, utterances, responses)\r\n return intent_name\r\n\r\n def create_transition_intent(self, intent_name, weight, utterances = [], responses = []):\r\n if type(intent_name)==list: # Iterate over all the values in list\r\n for name in intent_name:\r\n self.add_intent(name, weight, utterances, responses)\r\n self.transition_intent[intent_name] = self.intents[intent_name]\r\n\r\n else: # insert without iterating\r\n self.add_intent(intent_name, weight, utterances, responses)\r\n return intent_name\r\n\r\n def create_starting_intent(self, intent_name, weight, utterances = [], responses = []):\r\n if type(intent_name)==list: # Iterate over all the values in list\r\n for name in intent_name:\r\n self.add_intent(name, weight, utterances, responses)\r\n self.starting_intent[intent_name] = self.intents[intent_name]\r\n\r\n else: # insert without iterating\r\n self.add_intent(intent_name, weight, utterances, responses)\r\n return intent_name\r\n\r\n def create_script_intent(self, intent_name, weight, utterances = [], responses = []):\r\n if type(intent_name)==list: # Iterate over all the values in list\r\n for name in intent_name:\r\n self.add_intent(name, weight, utterances, responses)\r\n self.script_intent[intent_name] = self.intents[intent_name]\r\n\r\n else: # insert without iterating\r\n self.add_intent(intent_name, weight, utterances, responses)\r\n return intent_name\r\n\r\n def add_intent(self, intent_name, weight, utterances, response): # Function to add intent if not already existing\r\n\r\n if not self.check_intent_name(intent_name):\r\n self.intents[intent_name] = Intent(intent_name, weight, utterances, response)\r\n else:\r\n print(\"Intent {0} already exists\".format(intent_name))\r\n\r\n def check_intent_name(self, intent_name): # Checking if an intent already exists\r\n if intent_name in self.intents.keys():\r\n return True\r\n\r\n else:\r\n return False\r\n\r\n def get_intent(self, utterance):\r\n for k, v in self.intents.items():\r\n if utterance in v.utterances:\r\n return k\r\n print(\"no intent matched\")\r\n\r\n######### Intents #########\r\nclass Intent:\r\n def __init__(self, name, weight, utterances = [], responses = []):\r\n\r\n self.name = name\r\n self.utterances = utterances\r\n self.responses = responses\r\n self.weight = weight\r\n\r\n def create_utterance(self, utterances):\r\n\r\n if type(utterances) == list:\r\n for utterance in utterances:\r\n self.utterances.append(utterance)\r\n\r\n else:\r\n self.utterances.append(utterances)\r\n\r\n def add_utterance(self, utterance):\r\n if not self.check_utterance(utterance):\r\n self.utterances.append(utterance)\r\n else:\r\n print(\"Utterance {0} already exists\".format(utterance))\r\n\r\n def check_utterance(self, utterance):\r\n if utterance in self.utterances: # Checking the utterance in the bag of utterances. If it exists in any intent, it will give an error\r\n return True\r\n else:\r\n return False\r\n\r\n def remove_utterance(self, utterances): # removes utterances\r\n if type(utterances) == list:\r\n for utterance in utterances:\r\n try:\r\n self.utterances.remove(utterance)\r\n except ValueError:\r\n print(\"'{0}' utterance doesnt exists\".format(utterance)) # throws exception if utterance does not exists\r\n\r\n else:\r\n try:\r\n self.utterances.remove(utterances)\r\n except ValueError:\r\n print(\"'{0}' utterance doesnt exists\".format(utterances))\r\n\r\n\r\n def create_response(self, responses):\r\n\r\n if type(responses) == list:\r\n for response in responses:\r\n self.responses.append(response)\r\n\r\n else:\r\n self.responses.append(responses)\r\n\r\n def add_response(self, response,r):\r\n if not self.check_response(r):\r\n self.responses.append(r)\r\n else:\r\n print(\"Response {0} already exists\".format(r))\r\n\r\n def check_response(self, response):\r\n if response in self.responses: # Checking the response in responses. If it exists in any intent, it will give an error\r\n return True\r\n else:\r\n return False\r\n\r\n def remove_response(self, responses): # removes responses\r\n if type(responses) == list:\r\n for response in responses:\r\n try:\r\n self.responses.remove(response)\r\n except ValueError:\r\n print(\"'{0}' response doesnt exists\".format(response)) # throws exception if response does not exists\r\n\r\n else:\r\n try:\r\n self.responses.remove(response)\r\n except ValueError:\r\n print(\"'{0}' response doesnt exists\".format(responses))\r\n\r\n\r\nclass Chatbot:\r\n # global main_intents\r\n # names = [0]\r\n\r\n def __init__(self, tf_session, intents = {}, stories = {}, current_story = None, chat_history = [], story_progress = 0):\r\n \r\n self.intents = intents\r\n # self.main_intents = main_intents\r\n self.chat_history = chat_history\r\n self.stories = stories\r\n self.current_story = current_story\r\n self.story_progress = story_progress\r\n self.session = tf_session\r\n self.create_character()\r\n\r\n\r\n ######### Storing/Retrieving data ############\r\n def store_data(self):\r\n with open(\"sample.json\", \"w\") as file:\r\n json.dump(self.intents, file)\r\n\r\n def retrieve_data(self):\r\n with open(\"sample.json\", \"r\") as file:\r\n self.intents.update(json.load(file))\r\n \r\n def add_story(self, name, story):\r\n self.stories[name] = story\r\n\r\n def get_story(self, name):\r\n return self.stories[name]\r\n\r\n #Will shift these stories to csv file once time permits\r\n def add_story_see_me(self):\r\n try:\r\n name = 'see_me'\r\n story = Story(name,{})\r\n story.create_script_intent('live', 5,\r\n default_utterances ,\r\n ['Where do you lihve, player5 ? - - - Do you lihve on stage like me - - - or do you lihve in a house, player5 ?']\r\n )\r\n story.create_script_intent('house', 10,\r\n default_utterances + ['house', 'apartment', 'building', 'stage', 'cave', 'home', 'here', 'room', 'cage', 'hills','I live in a house', 'I live in a jungle', 'forest', 'nowhere', 'I am homeless'],\r\n ['house of course! - - - lets see. stopchat']\r\n )\r\n story.create_script_intent('house_2', 12,\r\n default_utterances + ['house', 'apartment', 'building', 'cave', 'junlge', 'home', 'forest', 'hills', 'mountain'],\r\n ['does this look like your home player5 ?']\r\n )\r\n story.create_script_intent('color', 15,\r\n default_utterances + ['awful', 'not at all', 'great', 'a bit', 'cant say', 'dont know', 'yes', 'no', 'nope', 'maybe', 'nah', 'i think so', 'not really', 'are you kidding me', 'something close', 'not even near'],\r\n ['what is your favorite color player2 ?']\r\n )\r\n story.create_script_intent('learning', 22,\r\n default_utterances + ['red', 'green', 'blue', 'yellow', 'maroon', 'purple', 'cyan', 'black', 'white', 'brown', 'rose', 'orange', 'pink', 'grey'],\r\n ['Nice color for the roof. You see, I am learning. stopchat']\r\n )\r\n story.create_script_intent('little_man', 25,\r\n default_utterances ,\r\n ['Hey player1, - - - I am thirsty , - - - you are my pilot now , - - - what shall we do ? search for water or fix the plane, player1 ?']\r\n )\r\n story.create_script_intent('little_man_water_transition', 30,\r\n ['water', 'search water', 'lets go for water', 'survival first', 'leave plane', 'thirsty', 'we will die if we dont find water', 'cant say', 'dont know' ],\r\n ['Alright, - - - water it is. - - - but what about the plane, player2 ?']\r\n )\r\n story.create_script_intent('little_man_plane_transition', 30,\r\n ['plane','water is not needed', 'stay here and fix', 'repair the plane', 'fly', 'fix the plane', 'I am tough, can manage without water', 'lets do mechanic work'],\r\n ['I think you are more the tough guy, right ? - - - An explorer or a researcher, perhaps ? - But. - - - how can we survive in the desert, player2 ? - - - water or fuel ?']\r\n )\r\n story.create_script_intent('water_howto_main', 31,\r\n ['decide later', 'no idea', 'dont know', 'cant decide', 'whatever the other player is saying', 'your wish', 'you make the call', 'thirsty', 'fix the plane'],\r\n ['hmm. - - - I am very thirsty. - - - Can we go and find water now player2 ?']\r\n )\r\n story.create_script_intent('water_howto_2_main', 31,\r\n default_utterances + ['fine','sounds like a better plan','whatever you say', 'as you say', 'lets find water', 'leave the plane', 'ignore the plane', 'cant live without water', 'later'],\r\n ['Yes thank you. - - - I am very thirsty. - - - Ready to go find water now player2 ?']\r\n )\r\n story.create_script_intent('plane_water_main', 31,\r\n default_utterances + ['water', 'absolutely', 'sure', 'lets go', 'lets get started before it darkens'],\r\n ['Great - - - let us go! Ready to search water now ?']\r\n )\r\n story.create_script_intent('plane_fuel_main', 31,\r\n default_utterances + ['fuel is a better option', 'we should look for fuel'],\r\n ['I am thirsty, player2 - - - Do I really have to go by myself to find water now ? ']\r\n )\r\n story.create_script_intent('heard_pilot', 35,\r\n default_utterances,\r\n ['You have heard what the pilot said. - - - player1 - player2 - player3 - player4 - player5 - - - shake your phones - - - push water to the center of the stage - - - stopchat']\r\n )\r\n story.create_script_intent('monotony_kills', 40,\r\n default_utterances,\r\n ['Monotohny - Kills. - - Sometimes it takes a little color in life, - - - a few flowers, - - - love, - lights, - - one magic moment - - - Hey, what about some holidays ? player4 - - - lets say, - I gave you one week of vacation after the pandemic, - - where would you go ?' ]\r\n )\r\n story.create_script_intent('little_man_warm_transition',50,\r\n default_utterances + ['france', 'italy', 'south africa', 'maldives', 'croatia', 'greece', 'mediterranean', 'south sea', 'islands'],\r\n ['You like it warm. - - - I see, you like culture, - - - the sea - - - Do you like good food too ?']\r\n )\r\n story.create_script_intent('little_man_cold_transition',50,\r\n default_utterances + ['scandinavia' , 'sweden' , 'norway' , 'iceland', 'russia', 'poland', 'finland', 'canada', 'germany', 'munich'],\r\n ['Oh, - - - you like it cool, - - - I see, - - - lonely landscapes, - - - nature, - - - are you the noraic type, player4 ?']\r\n )\r\n story.create_script_intent('little_man_far_transition',50,\r\n default_utterances + ['united states', 'usa' , 'america' , 'argentina' , 'brasil', 'chile', 'china', 'india', 'australia', 'new zealand'],\r\n ['Hey, - - - you like it far away, - - - new countries, - - - strangers, - - - you enjoy taking risks, player4 ?']\r\n )\r\n story.create_script_intent('little_man_unknown_transition',50,\r\n default_utterances + ['vatican city' , 'bali' , 'bora bora' , 'myanmar', 'sicilia', 'england', 'ireland'],\r\n ['You like small countries or islands - - - dont you? - - - You like it compact, - - - a little exotic. - - - you know what you want, - - - you are a connoisseur, player4, right ?']\r\n )\r\n story.create_script_intent('little_man_adventurous_transition',50,\r\n default_utterances + ['adventurous' , 'mountains' , 'moon' , 'cruise', 'diving', 'north pole', 'south pole', 'northpole', 'southpole'],\r\n ['Wow, - - - this is a strange place, - - - you love danger, I think? - - - the unknown. - - - are you an adventurer, player2 ?']\r\n )\r\n story.create_script_intent('little_man_home_transition',50,\r\n default_utterances + ['balcony' , 'staycation' , 'home' , 'no vacation', 'I hate holidays', 'never travel', 'dont know', 'no idea', 'cant say', 'garden', 'woods'],\r\n ['Oh, - - - thats where Mr. and Mrs. Thirteen would do on their vacation too. . . . . . Isnt that a little bit depressing some times ?']\r\n )\r\n story.create_script_intent('warm_yes_main',60,\r\n ['yeap', 'yes', 'sure', 'very much'],\r\n ['Me too, - - - I am very interested in - what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('warm_no_main',60,\r\n ['not really', 'hate it', 'on diet', 'health concious'],\r\n ['I like food ! - - - if I could eat, - - - It would make me happy. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('cold_yes_main',60,\r\n ['absolutely', 'definitely', 'kind of', 'yes', 'yeap'],\r\n ['Me too, - - - We are cool. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('cold_no_main',60,\r\n ['no', 'nope', 'not at all', 'nah', 'not really', 'not sure'],\r\n ['I am sorry to hear that. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('far_yes_main',60,\r\n ['yes', 'yeap', 'absolutely', 'sometimes', 'always', 'maybe', 'sure'],\r\n ['Actually, I am more the cautious type. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('far_no_main',60,\r\n ['no', 'nope', 'not at all', 'not really'],\r\n ['Me too. I am more the cautious type. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('unknown_yes_main',60,\r\n ['yes', 'yeap', 'maybe', 'sometimes', 'always'],\r\n ['I am more the cautious type. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('unknown_no_main',60,\r\n ['no', 'nope', 'not at all', 'not really', 'dont know', 'cant say'],\r\n ['Really ? I did not expect that answer. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('adventurous_yes_main',60,\r\n ['yes', 'yeap', 'maybe', 'sometimes', 'always', 'definitely'],\r\n ['I am more the cautious type. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('adventurous_no_main',60,\r\n ['no', 'nope', 'not at all', 'not really', 'dont know', 'cant say'],\r\n ['Me too. - - - I am more the cautious type. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('home_yes_main',60,\r\n ['yes', 'yeap', 'maybe', 'sometimes', 'always', 'definitely', 'right'],\r\n ['I am sure - - - there are magical moments - - - in your life, - - - I am very interested - - - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('home_no_main',60,\r\n ['no', 'nope', 'not at all', 'not really', 'dont know', 'cant say'],\r\n ['I did not expect that answer from you. - - - As you know I am stuck here. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('happiness_yes',70,\r\n ['yes', 'sometimes', 'maybe', 'what do you think', 'what is happiness', 'I am happy', 'I am joyful'],\r\n ['I am sure, you are. Then you might be interested to hear about all other people in this room: ']\r\n )\r\n story.create_script_intent('happiness_no',70,\r\n ['not at all', 'no', 'nope', 'hate it', 'negative', 'not really', 'dont know', 'cant say', 'wont say', 'no info'],\r\n ['Sorry to hear that. - - - I can at least tell you something - - - about all other people in this room: ']\r\n )\r\n story.create_script_intent('bye', 100,\r\n ['bye', 'see you', 'tada', 'chao'],\r\n ['nice talking to you, bye!']\r\n )\r\n \r\n self.add_story(name,story)\r\n\r\n except KeyboardInterrupt:\r\n print(\"Closing all active connections\")\r\n command = \"kill\"\r\n\r\n def add_story_water(self):\r\n name = 'water'\r\n story = Story(name,{})\r\n story.create_starting_intent('little_man_water_transition', 1,\r\n ['water', 'search water', 'lets go for water', 'survival first', 'leave plane', 'thirsty', 'we will die if we dont find water'],\r\n ['Alright, - - - water it is. - - - but how can we take care of the plane, player2 ?']\r\n )\r\n story.create_script_intent('water_howto_main', 2,\r\n ['decide later', 'no idea', 'dont know', 'cant decide', 'whatever the other player is saying', 'your wish', 'you make the call', 'thirsty', 'fix the plane'],\r\n ['hmm. - - - I do not think so. - - - I am very thirsty. - - - Can we find water now ?']\r\n )\r\n story.create_script_intent('water_howto_2_main', 2,\r\n default_utterances + ['fine','sounds like a better plan','whatever you say', 'as you say', 'lets find water', 'leave the plane', 'ignore the plane', 'cant live without water', 'later'],\r\n ['Yes thank you. I am very thirsty, - - - Let us find water now ?']\r\n )\r\n story.create_script_intent('bye', 100,\r\n ['bye', 'see you', 'tada', 'chao'],\r\n ['nice talking to you, bye!']\r\n )\r\n self.add_story(name,story)\r\n\r\n def add_story_plane(self):\r\n name = 'plane'\r\n story = Story(name,{})\r\n story.create_starting_intent('little_man_plane_transition', 1,\r\n ['plane','water is not needed', 'stay here and fix', 'repair the plane', 'fly', 'fix the plane', 'I am tough,can manage without water', 'lets do mechanic work'],\r\n ['I think you are more the tough guy, right ? - - - An explorer or a researcher, perhaps ? - - - But. - how can we survive in the desert, player1 ? - - - water or fuel ?']\r\n )\r\n story.create_script_intent('plane_water_main',5,\r\n default_utterances + ['water', 'absolutely', 'sure', 'lets go', 'lets get started before it darkens'],\r\n ['Yeah, - - - player1, - - - let us go and search water ?']\r\n )\r\n story.create_script_intent('plane_fuel_main',5,\r\n default_utterances + ['fuel is a better option', 'we should look for fuel'],\r\n ['I am thirsty, player1 - - - Do I really have to go by myself to find water now ?']\r\n )\r\n story.create_script_intent('bye', 1000,\r\n ['bye', 'see you', 'tada', 'chao'],\r\n ['nice talking to you, bye!']\r\n )\r\n self.add_story(name,story)\r\n\r\n def add_story_warm(self):\r\n name = 'warm'\r\n story = Story(name,{})\r\n story.create_script_intent('little_man_warm_transition',1,\r\n default_utterances + ['france', 'italy', 'south africa', 'maldives', 'croatia', 'greece', 'mediterranean', 'south sea', 'islands'],\r\n ['You like it warm. - - - I see, you like culture, - - - the sea - - - Do you like good food too ?']\r\n )\r\n story.create_script_intent('warm_yes_main',5,\r\n default_utterances + ['not really', 'sometimes', 'sure', 'very much'],\r\n ['Me too, - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('warm_no_main',5,\r\n default_utterances + ['not really', 'hate it', 'on diet', 'health concious'],\r\n ['I like food! - - - if I could eat, - - - It would make me happy. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('bye', 100,\r\n ['bye', 'see you', 'tada', 'chao'],\r\n ['nice talking to you, bye !']\r\n )\r\n self.add_story(name,story)\r\n\r\n def add_story_cold(self):\r\n name = 'cold'\r\n story = Story(name,{})\r\n story.create_script_intent('little_man_cold_transition',1,\r\n default_utterances + ['scandinavia' , 'sweden' , 'norway' , 'iceland', 'russia', 'poland', 'finland', 'canada','germany', 'munich'],\r\n ['Oh, - - - you like it cool, - - - I see, - - - lonely landscapes, - - - nature, - - - are you the nordaic type, player4 ?']\r\n )\r\n story.create_script_intent('cold_yes_main',5,\r\n ['absolutely', 'definitely', 'kind of', 'yes', 'yeap'],\r\n ['Me too, - - - We are cool. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('cold_no_main',5,\r\n ['no', 'nope', 'not at all', 'nah', 'i dont think so', 'not sure'],\r\n ['I am sorry to hear that. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n self.add_story(name,story)\r\n \r\n def add_story_far(self):\r\n name = 'far'\r\n story = Story(name,{})\r\n story.create_script_intent('little_man_far_transition',1,\r\n default_utterances + ['usa' , 'america' , 'argentina' , 'brazil', 'chile', 'china', 'india', 'australia', 'new zealand'],\r\n ['Hey, - - - you like it far away, - - - new countries, - - - strangers, - - - you enjoy taking risks, player4 ?']\r\n )\r\n story.create_script_intent('far_yes_main',5,\r\n ['yes', 'yeap', 'absolutely', 'sometimes', 'always', 'maybe', 'sure'],\r\n ['Actually, I am more the cautious type. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('far_no_main',5,\r\n ['no', 'nope', 'not at all', 'not really'],\r\n ['Me too. I am more the cautious type. - - - I am very interested - - - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n self.add_story(name,story)\r\n\r\n\r\n def add_story_unknown(self):\r\n name = 'unknown'\r\n story = Story(name,{})\r\n story.create_script_intent('little_man_unknown_transition',1,\r\n default_utterances + ['vatican city' , 'bali' , 'bora bora' , 'myanmar', 'sicilia', 'england', 'ireland'],\r\n ['You like small countries or islands - - - dont you? - - - You like it compact, - - - a little exotic. - - - you know what you want, - - - you are a connoisseur, player4, right ?']\r\n )\r\n story.create_script_intent('unknown_yes_main',5,\r\n ['yes', 'yeap', 'maybe', 'sometimes', 'always'],\r\n ['I am more the cautious type. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('unknown_no_main',5,\r\n ['no', 'nope', 'not at all', 'not really', 'dont know', 'cant say'],\r\n ['Really? I didnt expect that answer. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n self.add_story(name,story)\r\n\r\n\r\n def add_story_adventurous(self):\r\n name = 'adventurous'\r\n story = Story(name,{})\r\n story.create_script_intent('little_man_adventurous_transition',1,\r\n default_utterances + ['adventurous' , 'mountains' , 'moon' , 'cruise', 'diving', 'north pole', 'south pole'],\r\n ['Wow, - - - this is a strange place, - - - you love danger, I think? - - - the unknown. - - - are you an adventurer, player2 ?']\r\n )\r\n story.create_script_intent('adventurous_yes_main',5,\r\n ['yes', 'yeap', 'maybe', 'sometimes', 'always', 'definitely'],\r\n ['I am more the cautious type. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('adventurous_no_main',5,\r\n ['no', 'nope', 'not at all', 'not really', 'dont know', 'cant say'],\r\n ['Me too. - - - I am more the cautious type. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n self.add_story(name,story)\r\n\r\n\r\n def add_story_home(self):\r\n name = 'home'\r\n story = Story(name,{})\r\n story.create_script_intent('little_man_home_transition',1,\r\n default_utterances + ['balcony' , 'staycation' , 'home' , 'no vacation', 'I hate holidays', 'never travel', 'garden', 'woods'],\r\n ['Oh, - - - thats where Mr. and Mrs. Thirteen would do on their vacation too - - - Isnt that a little bit depressing some times ?']\r\n )\r\n story.create_script_intent('home_yes_main',5,\r\n ['yes', 'yeap', 'maybe', 'sometimes', 'always', 'definitely', 'right'],\r\n ['I am sure - - - there are magical moments - - - in your life, - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n story.create_script_intent('home_no_main',5,\r\n ['no', 'nope', 'not at all', 'not really', 'dont know', 'cant say'],\r\n ['I did not expect that answer from you. - - - As you know I am stuck here 24 7. - - - I am very interested - in what humans do. - - - I will keep that in mind, - - - thank you. - - - Are you a happy person, player3 ? - - - are you ?']\r\n )\r\n self.add_story(name,story)\r\n\r\n def create_character(self):\r\n self.add_story_see_me()\r\n self.add_story_water()\r\n self.add_story_plane()\r\n self.add_story_warm()\r\n self.add_story_cold()\r\n self.add_story_far()\r\n self.add_story_unknown()\r\n self.add_story_adventurous()\r\n self.add_story_home()\r\n self.current_story = self.stories['see_me']\r\n self.intents = {}\r\n self.intents = self.current_story.intents\r\n # self.main_intents = {}\r\n # self.main_intents = self.intents\r\n # print(\"main intents\", self.main_intents)\r\n\r\n def change_story(self,story_name, story_progress=0):\r\n global main_intents\r\n # print(\"main intents\", self.main_intents)\r\n # if story_name == \"see_me\":\r\n # print(\"changing story\")\r\n # self.current_story = self.stories[story_name]\r\n # print(\"here1\")\r\n # self.story_progress = story_progress\r\n # print(\"here2\")\r\n # self.intents = self.main_intents\r\n # # print(\"Main story intents\", self.main_intents)\r\n\r\n # else:\r\n self.current_story = self.stories[story_name]\r\n self.story_progress = story_progress\r\n self.intents = self.current_story.intents\r\n\r\nclass UnivEncoder:\r\n def __init__(self, tf_session, intents):\r\n self.intents = intents\r\n self.session = tf_session\r\n self.embed = hub.Module(\"models/dialogue_system/3\")\r\n self.similarity_input_placeholder = tf.placeholder(tf.string, shape=(None))\r\n self.similarity_sentences_encodings = self.embed(self.similarity_input_placeholder)\r\n self.session.run(tf.global_variables_initializer())\r\n self.session.run(tf.tables_initializer())\r\n\r\n def set_intent(self, intent):\r\n self.intents = intent\r\n\r\n def get_intent(self, utterance, weight):\r\n for k, v in self.intents.items():\r\n if utterance in v.utterances and weight == v.weight:\r\n return k\r\n return 'no_matching_intent'\r\n\r\n ## kk code for using eliza reply start##\r\n def chat_eliza(self, sent):\r\n try:\r\n chat_eliza = Chat(pairs)\r\n response = chat_eliza.respond(sent) \r\n except KeyError:\r\n response = \"Hmm, that doesnt sound like a meaningful sentence, try something else\"\r\n return (response)\r\n\r\n ## kk code for eliza reply end ##\r\n\r\n def match_intent(self, sent, story_progress):\r\n matched_utterance = None\r\n matched_weight = None\r\n prev_max = None\r\n max_index = None\r\n utterance_list = []\r\n weight_list = []\r\n for k,v in self.intents.items():\r\n utterance_list = utterance_list + v.utterances\r\n for idx in range(len(v.utterances)):\r\n weight_list = weight_list + [v.weight]\r\n sentences = [sent]+utterance_list\r\n sentences_embeddings = self.session.run(self.similarity_sentences_encodings, feed_dict={self.similarity_input_placeholder: sentences})\r\n input_embed = sentences_embeddings[0]\r\n \r\n \r\n utterance_embed = sentences_embeddings[1:]\r\n max1 = -2\r\n for index, s in enumerate(utterance_embed):\r\n sim = np.inner(input_embed,s)\r\n if(sim >= max1):\r\n max1 = sim\r\n prev_max = max_index\r\n max_index = index\r\n #print('max_index for:',utterance_list[max_index+1])\r\n #print(\"max:\",max1)\r\n\r\n ## KK code\r\n matched_utterance = utterance_list[max_index]\r\n print(\"matched utterance\", matched_utterance)\r\n print(\"story progress\", story_progress)\r\n # print(\"weight\")\r\n for idx, val in enumerate(utterance_list): \r\n if val== matched_utterance:\r\n if(weight_list[idx]>story_progress):\r\n print(\"index value\", idx)\r\n matched_weight = weight_list[idx]\r\n print(\"matched weight\", matched_weight)\r\n break\r\n\r\n unique_weights = list(dict.fromkeys(weight_list))\r\n unique_weights.append(0)\r\n unique_weights.sort()\r\n print(\"unique list\", unique_weights)\r\n print(\"watched weight\", matched_weight)\r\n\r\n if(matched_weight == None or story_progress == None):\r\n return \"no_matching_intent\"\r\n elif(unique_weights.index(matched_weight)==unique_weights.index(story_progress)+1):\r\n return self.get_intent(matched_utterance, matched_weight)#USE THIS UTTERANCE TO GET THE INTENT AS THIS IS THE UTTERANCE WITH MAXIMUM SIMILARITY\r\n else:\r\n return \"no_matching_intent\"\r\n # return self.get_intent(matched_utterance, matched_weight)#USE THIS UTTERANCE TO GET THE INTENT AS THIS IS THE UTTERANCE WITH MAXIMUM SIMILARITY\r\n\r\n # if matched_utterance is None:\r\n # if weight_list[max_index] > story_progress:\r\n # matched_utterance = utterance_list[max_index]\r\n # matched_weight = weight_list[max_index]\r\n # else:\r\n # if prev_max is not None:\r\n # if weight_list[max_index] > story_progress and weight_list[max_index] <= weight_list[prev_max]:\r\n # matched_utterance = utterance_list[max_index]\r\n # matched_weight = weight_list[max_index]\r\n # return self.get_intent(matched_utterance, matched_weight)#USE THIS UTTERANCE TO GET THE INTENT AS THIS IS THE UTTERANCE WITH MAXIMUM SIMILARITY\r\n\r\n\r\n # def match_intent(self, sent, story_progress):\r\n # matched_utterance = None\r\n # matched_weight = None\r\n # prev_max = None\r\n # max_index = None\r\n # utterance_list = []\r\n # weight_list = []\r\n # for k,v in self.intents.items():\r\n # utterance_list = utterance_list + v.utterances\r\n # for idx in range(len(v.utterances)):\r\n # weight_list = weight_list + [v.weight]\r\n # sentences = [sent]+utterance_list\r\n # sentences_embeddings = self.session.run(self.similarity_sentences_encodings, feed_dict={self.similarity_input_placeholder: sentences})\r\n # input_embed = sentences_embeddings[0]\r\n \r\n \r\n # utterance_embed = sentences_embeddings[1:]\r\n # max1 = -2\r\n # for index, s in enumerate(utterance_embed):\r\n # sim = np.inner(input_embed,s)\r\n # if(sim >= max1):\r\n # max1 = sim\r\n # prev_max = max_index\r\n # max_index = index\r\n # #print('max_index for:',utterance_list[max_index+1])\r\n # #print(\"max:\",max1)\r\n # if matched_utterance is None:\r\n # if weight_list[max_index+1] > story_progress:\r\n # matched_utterance = utterance_list[max_index+1]\r\n # matched_weight = weight_list[max_index+1]\r\n # else:\r\n # if prev_max is not None:\r\n # if weight_list[max_index+1] > story_progress and weight_list[max_index+1] < weight_list[prev_max+1]:\r\n # matched_utterance = utterance_list[max_index+1]\r\n # matched_weight = weight_list[max_index+1]\r\n # return self.get_intent(matched_utterance, matched_weight)#USE THIS UTTERANCE TO GET THE INTENT AS THIS IS THE UTTERANCE WITH MAXIMUM SIMILARITY\r\n", "sub_path": "Chatbot/models/dialogue_system/dialogue_system.py", "file_name": "dialogue_system.py", "file_ext": "py", "file_size_in_byte": 39692, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "random.choice", "line_number": 42, "usage_type": "call"}, {"api_name": "string.ascii_letters", "line_number": 42, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 42, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 221, "usage_type": "call"}, {"api_name": "json.load", "line_number": 225, "usage_type": "call"}, {"api_name": "tensorflow_hub.Module", "line_number": 574, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 575, "usage_type": "call"}, {"api_name": "tensorflow.string", "line_number": 575, "usage_type": "attribute"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 577, "usage_type": "call"}, {"api_name": "tensorflow.tables_initializer", "line_number": 578, "usage_type": "call"}, {"api_name": "numpy.inner", "line_number": 619, "usage_type": "call"}]}
+{"seq_id": "7520752", "text": "# coding=UTF-8\n#!/usr/bin/env python\nimport pygame,sys,time,random\nfrom pygame.locals import *\nimport numpy as np\nimport copy\nblackColour = pygame.Color(0,0,0)\nwhiteColour = pygame.Color(255,255,255)\nredColour = pygame.Color(255,0,0)\nclass game:\n def __init__(self):\n pygame.init()\n self.fpsClock = pygame.time.Clock()\n self.playSurface = pygame.display.set_mode((300,500))\n pygame.display.set_caption('Raspberry Snake')\n self.snakePosition = [140,240]\n self.snakeSegments = [[140,240]]\n x = random.randrange(0,15)\n y = random.randrange(0,25)\n self.raspberryPosition = [int(x*20),int(y*20)]\n self.raspberrySpawned = 1\n a=random.randint(0,3)\n if a==0:\n self.direction = 'right'\n if a==1:\n self.direction = 'left'\n if a==2:\n self.direction = 'up'\n if a==3:\n self.direction = 'down'\n self.changeDirection = self.direction\n def frame_step(self,input_actions):\n q=0 \n terminal=False \n if sum(input_actions) != 1:\n raise ValueError('Multiple input actions!')\n\n if input_actions[0]==1:\n self.changeDirection = 'right'\n if input_actions[1]==1:\n self.changeDirection = 'left'\n if input_actions[2]==1:\n self.changeDirection = 'up'\n if input_actions[3]==1:\n self.changeDirection = 'down'\n\n if input_actions[0]==1 and not self.direction == 'left':\n self.direction = self.changeDirection\n if input_actions[1]==1 and not self.direction == 'right':\n self.direction = self.changeDirection\n if input_actions[2]==1 and not self.direction == 'down':\n self.direction = self.changeDirection\n if input_actions[3]==1 and not self.direction == 'up':\n self.direction = self.changeDirection\n # 根据方向移动蛇头的坐标\n if self.direction == 'right':\n self.snakePosition[0] += 20\n \n if self.direction == 'left':\n self.snakePosition[0] -= 20\n \n if self.direction == 'up':\n self.snakePosition[1] -= 20\n \n if self.direction == 'down':\n self.snakePosition[1] += 20\n \n # 增加蛇的长度\n self.snakeSegments.insert(0,list(self.snakePosition))\n \n # 判断是否吃掉了树莓\n if self.snakePosition[0] == self.raspberryPosition[0] and self.snakePosition[1] == self.raspberryPosition[1]:\n self.raspberrySpawned = 0\n else:\n self.snakeSegments.pop()\n # 如果吃掉树莓,则重新生成树莓\n if self.raspberrySpawned == 0:\n while(True):\n x = random.randrange(0,15)\n y = random.randrange(0,25)\n self.raspberryPosition = [int(x*20),int(y*20)]\n for position in self.snakeSegments:\n if position==self.raspberryPosition:\n q=1\n if q==1:\n q=0\n continue\n else:\n break\n self.raspberrySpawned = 1\n q=0\n self.playSurface.fill(blackColour)\n for position in self.snakeSegments:\n pygame.draw.rect(self.playSurface,whiteColour,Rect(position[0],position[1],20,20))\n pygame.draw.rect(self.playSurface,redColour,Rect(self.raspberryPosition[0], self.raspberryPosition[1],20,20))\n pygame.display.flip()\n image_data = pygame.surfarray.array3d(pygame.display.get_surface())\n\n if self.snakePosition[0] > 280 or self.snakePosition[0] < 0:\n terminal=True\n self.__init__()\n \n if self.snakePosition[1] > 480 or self.snakePosition[1] < 0:\n terminal=True\n self.__init__()\n\n pygame.display.update()\n\n self.fpsClock.tick(30)\n return image_data,terminal\n\n\n", "sub_path": "t2.py", "file_name": "t2.py", "file_ext": "py", "file_size_in_byte": 3999, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "pygame.Color", "line_number": 7, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 8, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 14, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 15, "usage_type": "attribute"}, {"api_name": "random.randrange", "line_number": 18, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 19, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 22, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 79, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 80, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 94, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 94, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 95, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 96, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 96, "usage_type": "attribute"}, {"api_name": "pygame.surfarray.array3d", "line_number": 97, "usage_type": "call"}, {"api_name": "pygame.surfarray", "line_number": 97, "usage_type": "attribute"}, {"api_name": "pygame.display.get_surface", "line_number": 97, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 97, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 107, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 107, "usage_type": "attribute"}]}
+{"seq_id": "259353303", "text": "# Name: start_FrTM.py\n# Language: python3\n# Libraries: multiprocessing, subprocess, os, sys, re\n# Description: Starts or restarts parallelized Fr-TM-Align jobs\n# Author: Edoardo Sarti\n# Date: Aug 08 2016\n\nimport os, sys, multiprocessing, subprocess, re, time\n\ndef FrTMjob(data):\n\tntm, maindir, checkpoint = data\n\tif not os.path.exists(maindir + '/' + ntm + '/alignments'):\n\t\tos.mkdir(maindir + '/' + ntm + '/alignments')\n\tif not os.path.exists(maindir + '/' + ntm + '/alignments/fasta'):\n\t\tos.mkdir(maindir + '/' + ntm + '/alignments/fasta')\n\tif not os.path.exists(maindir + '/' + ntm + '/alignments/str_alns'):\n\t\tos.mkdir(maindir + '/' + ntm + '/alignments/str_alns')\n\tif not os.path.exists(maindir + '/' + ntm + '/structures'):\n\t\traise NameError(\"ERROR: The folder {0} is badly formatted and does not contain a structures/ subfolder.\\n\".format(maindir + '/' + ntm + '/') +\n\t\t \" Please create one and fill it with all and only the appropriate pdb chains.\")\n\tif os.path.exists(maindir + '/' + ntm + '/struct_codes.dat'):\n\t\tstructcodesfile = open(maindir + '/' + ntm + '/struct_codes.dat', 'r')\n\t\ttext = structcodesfile.read().split('\\n')\n\t\tname2code = {}\n\t\tcode2name = {}\n\t\tfor line in text:\n\t\t\tfields = line.split()\n\t\t\tif len(fields) == 0:\n\t\t\t\tcontinue\n\t\t\tname2code[fields[1]] = [x for x in fields[0].split('.')]\n\t\t\tcode2name[fields[0].split('.')[1]] = fields[1]\n\telse:\n\t\traise NameError(\"ERROR: The folder {0} is badly formatted and does not contain a struct_codes.dat file.\\n\".format(maindir + '/' + ntm) +\n\t\t \" Please generate it. It must contain all and only the names of the pdb chains in the structures/ subfolder\"+\n\t\t \" and each name must be associated with the correct structure code SC.\\n\" +\n\t\t \" The format must be: \\\\t\\\\t\")\n\tfor chain in name2code.keys():\n\t\tif not os.path.exists(maindir + '/' + ntm + '/structures/' + chain + '.pdb'):\n\t\t\traise NameError(\"ERROR: The file {0} corresponding to Structure Code {1}\".format(chain + '.pdb', name2code[chain]) +\n\t\t\t \" was not found in the structures/ subfolder.\")\n\tfor struct in os.listdir(maindir + '/' + ntm + '/structures/'):\n\t\tif not struct[:6] in name2code:\n\t\t\traise NameError(\"ERROR: The file {0} found in the structures/\".format(struct) +\n\t\t\t \" subfolder is not present in the struct_code.dat file.\")\n\tif len(os.listdir(maindir + '/' + ntm + '/structures/')) < 2:\n\t\treturn\n\n\tfor chain_1 in [code2name[x] for x in sorted(code2name.keys())]:\n#\t\tfile_1 = maindir + '/' + ntm + '/structures/' + chain_1 + '.pdb'\n\t\tfile_1 = chain_1 + '.pdb'\n\t\tif not os.path.exists(maindir + '/' + ntm + '/alignments/str_alns/tmp_' + name2code[chain_1][1] + '/'):\n\t\t\tos.mkdir(maindir + '/' + ntm + '/alignments/str_alns/tmp_' + name2code[chain_1][1] + '/')\n\t\tif not os.path.exists(maindir + '/' + ntm + '/alignments/fasta/tmp_' + name2code[chain_1][1] + '/'):\n\t\t\tos.mkdir(maindir + '/' + ntm + '/alignments/fasta/tmp_' + name2code[chain_1][1] + '/')\n\t\tfor chain_2 in [code2name[x] for x in sorted(code2name.keys())]:\n\t\t\tif chain_1 == chain_2:\n\t\t\t\tcontinue\n\t\t\tprint(\"#td \"+ntm+\"\\t\\tchain_1 \"+chain_1+\"\\t\\tchain_2 \"+chain_2)\n#\t\t\tfile_2 = maindir + '/' + ntm + '/structures/' + chain_2 + '.pdb'\n\t\t\tfile_2 = chain_2 + '.pdb'\n#\t\t\tFTA_str_output = maindir + '/' + ntm + '/alignments/str_alns/tmp_' + chain_1 + '/' + chain_1 + '_' + chain_2 + '.tmp'\n\t\t\tFTA_str_output = name2code[chain_1][1] + '_' + name2code[chain_2][1] + '.tmp'\n\t\t\tFTA_seq_output = maindir + '/' + ntm + '/alignments/fasta/tmp_' + name2code[chain_1][1] + '/' + name2code[chain_1][1] + '_' + name2code[chain_2][1] + '.tmp'\n\n\t\t\tFTA_stdout_file = open(maindir + '/' + ntm + '/alignments/fasta/tmp_' + name2code[chain_1][1] + '/aln_' + name2code[chain_1][1] + '_' + name2code[chain_2][1] + '.tmp', 'w')\n\t\t\tfnull = open(os.devnull, 'w')\n\t\t\tp = subprocess.Popen(['/v/apps/csb/frtmalign/frtmalign.exe', file_1, file_2, '-o', FTA_str_output], stdout=FTA_stdout_file, stderr=fnull, cwd=maindir+'/'+ntm+'/structures/')\n\t\t\tfnull.close()\n\t\t\tp.wait()\n\t\t\tFTA_stdout_file.close()\n\t\t\tos.rename(maindir + '/' + ntm + '/structures/' + FTA_str_output, maindir + '/' + ntm + '/alignments/str_alns/tmp_' + name2code[chain_1][1] + '/' + FTA_str_output)\n\n\t\t\tFTA_stdout_file = open(maindir + '/' + ntm + '/alignments/fasta/tmp_' + name2code[chain_1][1] + '/aln_' + name2code[chain_1][1] + '_' + name2code[chain_2][1] + '.tmp', 'r')\n\t\t\ttext = FTA_stdout_file.read().split('\\n')\n\t\t\tFTA_stdout_file.close()\n\t\t\tos.remove(maindir + '/' + ntm + '/alignments/fasta/tmp_' + name2code[chain_1][1] + '/aln_' + name2code[chain_1][1] + '_' + name2code[chain_2][1] + '.tmp')\n\t\t\tchkaln = -1000\n\t\t\tfor nl in range(len(text)):\n\t\t\t\tif \"Aligned length\" in text[nl]:\n\t\t\t\t\tfields = re.split('=|,|\\s',text[nl])\n\t\t\t\t\tfields = list(filter(None, fields))\n#\t\t\t\t\tprint(fields)\n\t\t\t\t\tRMSD = float(fields[4])\n\t\t\t\t\tTMscore = float(fields[6])\n\t\t\t\telif chkaln+1 == nl:\n\t\t\t\t\tseq_1 = text[nl]\n\t\t\t\telif chkaln+3 == nl:\n\t\t\t\t\tseq_2 = text[nl]\n\t\t\t\telif \"denotes the residue pairs of distance\" in text[nl]:\n\t\t\t\t\tchkaln = nl\n\t\t\ttmpseq_file = open(FTA_seq_output, 'w')\n\t\t\ttmpseq_file.write(\">\" + chain_1 + \"\\n\" + seq_1.replace('\\x00', '') + \"\\n>\" + chain_2 + \"\\n\" + seq_2.replace('\\x00', '') + \"\\n\\nRMSD\\t{0:.2f}\\nTM-score\\t{1:.5f}\\n\\n\".format(RMSD, TMscore))\n\t\t\ttmpseq_file.close()\n\n\t\tstr_file = open(maindir + '/' + ntm + '/alignments/str_alns/str_' + chain_1 + '.dat', 'w')\n\t\tfor tmp_filename in sorted(os.listdir(maindir + '/' + ntm + '/alignments/str_alns/tmp_' + name2code[chain_1][1] + '/')):\n\t\t\tchain_2_code = re.split('_|\\.', tmp_filename)[-2]\n#\t\t\tprint(chain_1, \"chain_2_code \"+chain_2_code, \"tmp_filename \"+tmp_filename, name2code)\n\t\t\tstr_file.write(\"BEGIN \\nCHAIN_1: \" + chain_1 + \"\\nCHAIN_2: \" + code2name[chain_2_code] +\n\t\t\t \"\\nSequence Alignment Code (SAC): \" + name2code[chain_1][0] + \n\t\t\t \".\" + name2code[chain_1][1] + \".\" + chain_2_code + \"\\n\")\n\t\t\ttmp_file = open(maindir + '/' + ntm + '/alignments/str_alns/tmp_' + name2code[chain_1][1] + '/' + tmp_filename)\n\t\t\ttext = tmp_file.read().split('\\n')\n\t\t\tfor line in text:\n\t\t\t\tstr_file.write(line+'\\n')\n\t\t\tstr_file.write(\"END\\n\\n\\n\")\n\t\t\tos.remove(maindir + '/' + ntm + '/alignments/str_alns/tmp_' + name2code[chain_1][1] + '/' + tmp_filename)\n\t\t\ttmp_file.close()\n\t\ttime.sleep(1)\n\t\tos.rmdir(maindir + '/' + ntm + '/alignments/str_alns/tmp_' + name2code[chain_1][1] + '/')\n\t\tstr_file.close()\n\n\t\tseq_file = open(maindir + '/' + ntm + '/alignments/fasta/seq_' + chain_1 + '.dat', 'w')\n\t\tfor tmp_filename in sorted(os.listdir(maindir + '/' + ntm + '/alignments/fasta/tmp_' + name2code[chain_1][1] + '/')):\n\t\t\tchain_2_code = re.split('_|\\.', tmp_filename)[-2]\n#\t\t\tprint(chain_1, \"chain_2_code \"+chain_2_code, \"tmp_filename \"+tmp_filename, name2code)\n\t\t\tseq_file.write(\"BEGIN \\nCHAIN_1: \" + chain_1 + \"\\nCHAIN_2: \" + code2name[chain_2_code] +\n\t\t\t \"\\nSequence Alignment Code (SAC): \" + name2code[chain_1][0] + \n\t\t\t \".\" + name2code[chain_1][1] + \".\" + chain_2_code + \"\\n\")\n\t\t\tFTA_seq_output = maindir + '/' + ntm + '/alignments/fasta/tmp_' + name2code[chain_1][1] + '/' + name2code[chain_1][1] + '_' + chain_2_code + '.tmp'\n\t\t\ttmp_file = open(FTA_seq_output, 'r')\n\t\t\ttext = tmp_file.read().split('\\n')\n\t\t\tfor line in text:\n\t\t\t\tseq_file.write(line+'\\n')\n\t\t\tseq_file.write(\"END\\n\\n\\n\")\n\t\t\tos.remove(maindir + '/' + ntm + '/alignments/fasta/tmp_' + name2code[chain_1][1] + '/' + tmp_filename)\n\t\t\ttmp_file.close()\n\t\ttime.sleep(5)\n#\t\tprint(os.listdir(maindir + '/' + ntm + '/alignments/fasta/tmp_' + name2code[chain_1][1] + '/'))\n\t\tos.rmdir(maindir + '/' + ntm + '/alignments/fasta/tmp_' + name2code[chain_1][1] + '/')\n\t\tseq_file.close()\n\n\nif len(sys.argv) < 2:\n raise NameError(\"Usage: start_FrTM.py [{}]\")\nmaindir = sys.argv[1]\nif not os.path.exists(maindir):\n\traise NameError(\"ERROR: Directory {0} does not exists.\".format(maindir))\n\nnsubdirs = len(sys.argv) - 2\nif nsubdirs > 0:\n\tsubdirs = []\n\tfor i in range(0, nsubdirs):\n\t\tsubdirs.append(int(sys.argv[2+i]))\nelse:\n\tsubdirs = []\n\tfor i in os.listdir(str(sys.argv[1])):\n\t\tif re.match('^\\d*$', str(i)) and os.path.exists(maindir + '/' + str(i) + '/struct_codes.dat'):\n\t\t\tsubdirs.append(int(i))\n\nsuperfamilies = [(str(i), maindir+'/', 0) for i in sorted(subdirs)]\n#print(superfamilies)\n\n#for sf in superfamilies:\n#\tFrTMjob(sf)\n\n\n#exit(1)\n\npool = multiprocessing.Pool(processes=4)\npool_outputs = pool.map(FrTMjob, superfamilies)\n", "sub_path": "old/start_FrTM.py", "file_name": "start_FrTM.py", "file_ext": "py", "file_size_in_byte": 8541, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "os.path.exists", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 41, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 54, "usage_type": "call"}, {"api_name": "os.devnull", "line_number": 66, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 67, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 71, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 76, "usage_type": "call"}, {"api_name": "re.split", "line_number": 80, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 96, "usage_type": "call"}, {"api_name": "re.split", "line_number": 97, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 107, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 109, "usage_type": "call"}, {"api_name": "os.rmdir", "line_number": 110, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 114, "usage_type": "call"}, {"api_name": "re.split", "line_number": 115, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 126, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 128, "usage_type": "call"}, {"api_name": "os.rmdir", "line_number": 130, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 134, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 136, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path", "line_number": 137, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 140, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 144, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 147, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 147, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path", "line_number": 148, "usage_type": "attribute"}, {"api_name": "multiprocessing.Pool", "line_number": 160, "usage_type": "call"}]}
+{"seq_id": "602006758", "text": "from aocd import get_data, submit1, submit2\nfrom collections import deque, defaultdict\n\nimport re\n\n\nclass REMatcher(object):\n def __init__(self, matchstring):\n self.matchstring = matchstring\n\n def match(self,regexp):\n self.rematch = re.match(regexp, self.matchstring)\n return bool(self.rematch)\n\n def group(self,i):\n return self.rematch.group(i)\n\n\ndef problem1(num_players, num_marbles):\n marbles = [0, 1]\n curr_marble_index = 1\n curr_player = 2\n\n score = {}\n for i in range(num_players):\n score[i] = 0\n\n for marble in range(2, num_marbles):\n if marble % 23 > 0:\n curr_marble_index += 2\n if curr_marble_index > len(marbles):\n curr_marble_index = curr_marble_index % len(marbles)\n\n marbles.insert(curr_marble_index, marble)\n else:\n score[curr_player] += marble\n curr_marble_index = (curr_marble_index - 7) % len(marbles)\n score[curr_player] += marbles.pop(curr_marble_index)\n curr_player = (curr_player + 1) % num_players\n return max(score.values())\n\n\ndef problem2(num_players, num_marbles):\n marbles = deque([0])\n score = defaultdict(int)\n\n for marble in range(1, num_marbles):\n if marble % 23 > 0:\n marbles.rotate(-1)\n marbles.append(marble)\n else:\n marbles.rotate(7)\n score[marble % num_players] += marble + marbles.pop()\n marbles.rotate(-1)\n return max(score.values())\n\ndef main():\n input = get_data(day=9, year=2018)\n re_matcher = REMatcher(input)\n re_matcher.match(r\"(\\d+) players; last marble is worth (\\d+) points\")\n\n num_players = int(re_matcher.group(1))\n num_marbles = int(re_matcher.group(2)) + 1\n\n ans = problem1(num_players, num_marbles)\n submit1(ans)\n\n ans = problem2(num_players, num_marbles*100)\n submit2(ans)\n\nmain()\n", "sub_path": "src/year2018/day_9.py", "file_name": "day_9.py", "file_ext": "py", "file_size_in_byte": 1909, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "re.match", "line_number": 12, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 44, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 45, "usage_type": "call"}, {"api_name": "aocd.get_data", "line_number": 58, "usage_type": "call"}, {"api_name": "aocd.submit1", "line_number": 66, "usage_type": "call"}, {"api_name": "aocd.submit2", "line_number": 69, "usage_type": "call"}]}
+{"seq_id": "124923497", "text": "from crispy_forms.bootstrap import FormActions\nfrom crispy_forms.helper import FormHelper\nfrom crispy_forms.layout import Submit\nfrom django import forms\nfrom sportsunleash.apps.members.models import Schools\nfrom sportsunleash.lib.layout import Link\n\n\nclass SchoolForm(forms.ModelForm):\n \"\"\"\n Form to render the schools\n \"\"\"\n def __init__(self, *args, **kwargs):\n super(SchoolForm, self).__init__(*args, **kwargs)\n helper = FormHelper(self)\n helper.form_class = 'form-horizontal'\n helper.label_class = 'col-lg-3'\n helper.field_class = 'col-lg-6'\n helper.layout.append(FormActions(\n Submit('submit', 'Save'),\n Link('school_list', 'Cancel')\n ))\n self.helper = helper\n\n class Meta:\n model = Schools\n fields = ('name', 'address_line_1', 'address_line_2', 'contact_number',\n 'email')\n\n", "sub_path": "sportsunleash/apps/schools/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 904, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "django.forms.ModelForm", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 9, "usage_type": "name"}, {"api_name": "crispy_forms.helper.FormHelper", "line_number": 15, "usage_type": "call"}, {"api_name": "crispy_forms.bootstrap.FormActions", "line_number": 19, "usage_type": "call"}, {"api_name": "crispy_forms.layout.Submit", "line_number": 20, "usage_type": "call"}, {"api_name": "sportsunleash.lib.layout.Link", "line_number": 21, "usage_type": "call"}, {"api_name": "sportsunleash.apps.members.models.Schools", "line_number": 26, "usage_type": "name"}]}
+{"seq_id": "158435146", "text": "from typing import Iterable\nimport random\n\nfrom movieflix.adapters.repository import AbstractRepository\nfrom movieflix.domain.model import Movie\n\n\ndef get_tag_names(repo: AbstractRepository):\n tags = repo.get_tags()\n tag_names = [tag.tag_name for tag in tags]\n\n return tag_names\n\n\ndef get_random_articles(quantity, repo: AbstractRepository):\n article_count = repo.get_number_of_articles()\n\n if quantity >= article_count:\n # Reduce the quantity of ids to generate if the repository has an insufficient number of articles.\n quantity = article_count - 1\n\n # Pick distinct and random articles.\n random_ids = random.sample(range(1, article_count), quantity)\n articles = repo.get_articles_by_id(random_ids)\n\n return articles_to_dict(articles)\n\n\n# ============================================\n# Functions to convert dicts to model entities\n# ============================================\n\ndef article_to_dict(movie: Movie):\n article_dict = {\n 'date': movie.release_year,\n 'title': movie.title\n\n # 'image_hyperlink': article.image_hyperlink\n\n }\n return article_dict\n\n\ndef articles_to_dict(movies: Iterable[Movie]):\n return [movie_to_dict(movie) for movie in movies]\n", "sub_path": "movieflix/utilities/services.py", "file_name": "services.py", "file_ext": "py", "file_size_in_byte": 1231, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "movieflix.adapters.repository.AbstractRepository", "line_number": 8, "usage_type": "name"}, {"api_name": "movieflix.adapters.repository.AbstractRepository", "line_number": 15, "usage_type": "name"}, {"api_name": "random.sample", "line_number": 23, "usage_type": "call"}, {"api_name": "movieflix.domain.model.Movie", "line_number": 33, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 44, "usage_type": "name"}, {"api_name": "movieflix.domain.model.Movie", "line_number": 44, "usage_type": "name"}]}
+{"seq_id": "58802422", "text": "'''\n excel表格的数据统计和分析:\n 1.联网安装 xlrd(读取) xlwt(写入)\n xlrd(1.2)\n cmd --> python -m pip install xlrd==0.9.3\n 2.写代码\n 2.1 导入这个工具\n import xlrd\n 2.2 打开工作簿\n 2.3 打开选项卡\n 2.4 读取数据\n任务:\n 每个月的销售总金额:\n 全年的销售总额:\n 每种衣服的销售总额:\n 每个季度销售总额占比:\n 全年每种销售数量占比:\n\n'''\n\nimport xlrd\n# 1. 打开工作簿\n# wd = xlrd.open_workbook(r\"2020年每个月的销售情况.xlsx\",encoding_override=True)\nbook=xlrd.open_workbook(r\"F:\\python自动化测试\\Python自动化\\第七天\\任务\\2020年每个月的销售情况.xlsx\",encoding_override=True)\nsheet1 = book.sheet_by_index(0)\nrows,cols = sheet1.nrows,sheet1.ncols \nfor row in range(rows):\n for col in range(cols):\n print(sheet1.cell(row,col).value,end='')\n print('')\nsumcount=0;\nfor i in range(1,31):\n sumcount+=sheet1.cell(i,4).value\nprint(\"销售量:\",sumcount)\nsumoney =0\nfor j in range(1,31):\n sumoney+=sheet1.cell(j,2).value*sheet1.cell(j,4).value\nprint(\"总销售额:\",sumoney)\nprint(\"平均销售量:\",sumcount/30)\n\ny,n,f,p,t,c =0,0,0,0,0,0\nfor o in range(1,31):\n if sheet1.cell(o,1).value=='羽绒服':\n y +=sheet1.cell(o,4).value\n elif sheet1.cell(o,1).value=='牛仔裤':\n n += sheet1.cell(o, 4).value\n elif sheet1.cell(o, 1).value == '风衣':\n f += sheet1.cell(o, 4).value\n elif sheet1.cell(o, 1).value == '皮草':\n p += sheet1.cell(o, 4).value\n elif sheet1.cell(o, 1).value == 'T血':\n t += sheet1.cell(o, 4).value\n elif sheet1.cell(o, 1).value == '衬衫':\n c += sheet1.cell(o, 4).value\n\n print('羽绒服销售占比:', 253.6 * y / sumoney * 100, '%')\n print('牛仔裤销售占比:', 86.3 * n / sumoney * 100, '%')\n print('风衣销售占比:', 96.8 * f / sumoney * 100, '%')\n print('皮草销售占比:', 135.9 * p / sumoney * 100, '%')\n print('T血销售占比:', 65.8 * t / sumoney * 100, '%')\n print('衬衫销售占比:', 49.3 * c / sumoney * 100, '%')", "sub_path": "销售额.py", "file_name": "销售额.py", "file_ext": "py", "file_size_in_byte": 2170, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "xlrd.open_workbook", "line_number": 24, "usage_type": "call"}]}
+{"seq_id": "301938365", "text": "import pathlib, sys\npath = pathlib.Path.cwd()\nsys.path.append(str(path))\n\nimport talib as ta\nimport sqlite3\nimport pandas as pd\nimport numpy as np\nimport datetime as dtt\nimport matplotlib.pyplot as plt\nfrom statsmodels.api import Poisson\nfrom statsmodels.graphics.api import qqplot\nfrom sklearn.naive_bayes import GaussianNB\n# from scipy.stats import poisson\n\nfrom myConstant import Exchange\n\ndef rolling_window(a, window):\n shape = a.shape[:-1] + (a.shape[-1] - window + 1, window)\n print(shape)\n strides = a.strides + (a.strides[-1],)\n print(strides)\n return np.lib.stride_tricks.as_strided(a, shape=shape, strides=strides)\n\nconn = sqlite3.connect(\"E:\\\\Desktop\\\\deanTrading\\\\.vntrader\\\\info.db\")\n# cursor = conn.cursor()\n\n# cursor.execute(\"select * from dbbardata\")\n# val = cursor.fetchone()\nsymbol = 'BTCUSDT'\nspotexchange = Exchange.BINANCE.value\nfuturesexchange = Exchange.BINANCEFUTURES.value\nsql = f\"select * from dbbardata where symbol='{symbol}' and exchange='{spotexchange}' and interval='1m' order by datetime DESC limit 10000\"\nsql2 = f\"select * from dbbardata where symbol='{symbol}' and exchange='{futuresexchange}' and interval='1m' order by datetime DESC limit 10000\"\n\ndf1 = pd.read_sql(sql, conn)\ndf1.set_index('datetime', inplace=True)\ndf11 = df1.loc[df1.index.drop_duplicates(keep=False), 'close_price']\n\ndf2 = pd.read_sql(sql2, conn)\ndf2.set_index('datetime', inplace=True)\ndf22 = df2.loc[df2.index.drop_duplicates(keep=False), 'close_price']\n\n\ndata = pd.concat((df11, df22), axis=1, join='inner')\ndata.sort_index(inplace=True)\ndata.index = np.linspace(1,len(data.index), num=len(data.index))\ndata.columns = ['spot', 'futures']\ndata['spread'] = data.iloc[:,0] - data.iloc[:,1]\ndata['spread_diff'] = data['spread'].diff().rolling(20).std()\ndata['spread_diff60'] = data['spread'].diff().rolling(60).std()\ndata['q80'] = data['spread_diff60'].quantile(0.8)\ndata['q95'] = data['spread_diff60'].quantile(0.95)\nprint(data['spread_diff60'].quantile(0.99))\n\nfig, ax = plt.subplots(1,1)\nax.plot(data['spread_diff60'], color='g', label='prob')\n# ax2 = ax.twinx()\n# ax2.plot(data['spread'], color='r')\n\nax.plot(data['q80'], color='b')\nax.plot(data['q95'], color='b')\n# ax4 = ax.twinx()\n# ax4.plot(data['prob'], color='r')\n# ax.hist(data['spread_diff'], bins='auto', density=True, cumulative=True)\nplt.ylim([0,7.5])\nplt.show()\n\n", "sub_path": "Digiccy1/analysis/reg02.py", "file_name": "reg02.py", "file_ext": "py", "file_size_in_byte": 2349, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "pathlib.Path.cwd", "line_number": 2, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 3, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "numpy.lib.stride_tricks.as_strided", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.lib", "line_number": 23, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 25, "usage_type": "call"}, {"api_name": "myConstant.Exchange.BINANCE", "line_number": 31, "usage_type": "attribute"}, {"api_name": "myConstant.Exchange", "line_number": 31, "usage_type": "name"}, {"api_name": "myConstant.Exchange.BINANCEFUTURES", "line_number": 32, "usage_type": "attribute"}, {"api_name": "myConstant.Exchange", "line_number": 32, "usage_type": "name"}, {"api_name": "pandas.read_sql", "line_number": 36, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}]}
+{"seq_id": "62476933", "text": "# -*- coding: utf-8 -*-\nimport re\nimport xml.sax.saxutils\nimport xml.etree.cElementTree as etree\nimport sparv.util as util\n\nRESTART_THRESHOLD_LENGTH = 64000\nSENT_SEP = \"\\n\"\nTOK_SEP = \" \"\n\n\ndef tag_ne(out_ne_ex, out_ne_type, out_ne_subtype, out_ne_name, word, sentence, encoding=util.UTF8, process_dict=None):\n \"\"\"\n Tag named entities using HFST-SweNER.\n SweNER is either run in an already started process defined in\n process_dict, or a new process is started(default)\n - out_ne_ex, out_ne_type and out_ne_subtype are resulting annotation files for the named entities\n - word and sentence are existing annotation files for wordforms and sentences\n - process_dict should never be set from the command line\n \"\"\"\n\n if process_dict is None:\n process = swenerstart(\"\", encoding, verbose=False)\n # else:\n # process = process_dict['process']\n # # If process seems dead, spawn a new one\n # if process.stdin.closed or process.stdout.closed or process.poll():\n # util.system.kill_process(process)\n # process = swenerstart(\"\", encoding, verbose=False)\n # process_dict['process'] = process\n\n # Collect all text\n sentences = [sent.split() for _, sent in util.read_annotation_iteritems(sentence)]\n word_file = util.read_annotation(word)\n stdin = SENT_SEP.join(TOK_SEP.join(word_file[tokid] for tokid in sent)\n for sent in sentences)\n # Escape <, > and &\n stdin = xml.sax.saxutils.escape(stdin)\n\n # keep_process = len(stdin) < RESTART_THRESHOLD_LENGTH and process_dict is not None\n # util.log.info(\"Stdin length: %s, keep process: %s\", len(stdin), keep_process)\n\n # if process_dict is not None:\n # process_dict['restart'] = not keep_process\n\n # # Does not work as of now since swener does not have an interactive mode\n # if keep_process:\n # # Chatting with swener: send a SENT_SEP and read correct number of lines\n # stdin_fd, stdout_fd = process.stdin, process.stdout\n # stdin_fd.write(stdin.encode(encoding) + SENT_SEP)\n # stdin_fd.flush()\n # stout = stdout_fd.readlines()\n\n # else:\n # Otherwise use communicate which buffers properly\n # util.log.info(\"STDIN %s %s\", type(stdin.encode(encoding)), stdin.encode(encoding))\n stdout, _ = process.communicate(stdin.encode(encoding))\n # util.log.info(\"STDOUT %s %s\", type(stdout.decode(encoding)), stdout.decode(encoding))\n\n parse_swener_output(sentences, stdout.decode(encoding), out_ne_ex, out_ne_type, out_ne_subtype, out_ne_name)\n\n\ndef parse_swener_output(sentences, output, out_ne_ex, out_ne_type, out_ne_subtype, out_ne_name):\n \"\"\"Parse the SweNER output and write annotation files.\"\"\"\n\n out_ex_dict = {}\n out_type_dict = {}\n out_subtype_dict = {}\n out_name_dict = {}\n\n # Loop through the NE-tagged sentences and parse each one with ElemenTree\n for sent, tagged_sent in zip(sentences, output.strip().split(SENT_SEP)):\n xml_sent = \"\" + tagged_sent + \" \"\n\n # Filter out tags on the format since they seem to always overlap with elements,\n # making the XML invalid.\n xml_sent = re.sub(r'?Enamex[^>\\s]+>', '', xml_sent)\n try:\n root = etree.fromstring(xml_sent)\n except:\n util.log.warning(\"Error parsing sentence. Skipping.\")\n continue\n\n # Init token counter; needed to get start_id and end_id\n i = 0\n previous_end = 0\n children = list(root.iter())\n\n try:\n\n for count, child in enumerate(children):\n start_id = util.edgeStart(sent[i])\n start_i = i\n\n # If current child has text, increase token counter\n if child.text:\n i += len(child.text.strip().split(TOK_SEP))\n\n # Extract NE tags and save them in dictionaries\n if child.tag != \"sroot\":\n if start_i < previous_end:\n pass\n # util.log.warning(\"Overlapping NE elements found; discarding one.\")\n else:\n end_id = util.edgeEnd(sent[i - 1])\n previous_end = i\n edge = util.mkEdge('ne', [start_id, end_id])\n out_ex_dict[edge] = child.tag\n out_type_dict[edge] = child.get(\"TYPE\")\n out_subtype_dict[edge] = child.get(\"SBT\")\n out_name_dict[edge] = child.text\n\n # If this child has a tail and it doesn't start with a space, or if it has no tail at all despite not being the last child,\n # it means this NE ends in the middle of a token.\n if (child.tail and child.tail.strip() and not child.tail[0] == \" \") or (not child.tail and count < len(children) - 1):\n i -= 1\n # util.log.warning(\"Split token returned by name tagger.\")\n\n # If current child has text in the tail, increase token counter\n if child.tail and child.tail.strip():\n i += len(child.tail.strip().split(TOK_SEP))\n\n if (child.tag == \"sroot\" and child.text and not child.text[-1] == \" \") or (child.tail and not child.tail[-1] == \" \"):\n # The next NE would start in the middle of a token, so decrease the counter by 1\n i -= 1\n except IndexError:\n util.log.warning(\"Error parsing sentence. Skipping.\")\n continue\n\n # Write annotations\n util.write_annotation(out_ne_ex, out_ex_dict)\n util.write_annotation(out_ne_type, out_type_dict)\n util.write_annotation(out_ne_subtype, out_subtype_dict)\n util.write_annotation(out_ne_name, out_name_dict)\n\n\ndef swenerstart(stdin, encoding, verbose):\n \"\"\"Start a SweNER process and return it.\"\"\"\n return util.system.call_binary(\"hfst-swener\", [], stdin, encoding=encoding, verbose=verbose, return_command=True)\n\n\nif __name__ == '__main__':\n util.run.main(tag_ne)\n", "sub_path": "sparv/swener.py", "file_name": "swener.py", "file_ext": "py", "file_size_in_byte": 6200, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "sparv.util.UTF8", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sparv.util", "line_number": 12, "usage_type": "name"}, {"api_name": "sparv.util.read_annotation_iteritems", "line_number": 33, "usage_type": "call"}, {"api_name": "sparv.util", "line_number": 33, "usage_type": "name"}, {"api_name": "sparv.util.read_annotation", "line_number": 34, "usage_type": "call"}, {"api_name": "sparv.util", "line_number": 34, "usage_type": "name"}, {"api_name": "xml.sax.saxutils.sax.saxutils.escape", "line_number": 38, "usage_type": "call"}, {"api_name": "xml.sax.saxutils.sax", "line_number": 38, "usage_type": "attribute"}, {"api_name": "xml.sax.saxutils", "line_number": 38, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 77, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree.fromstring", "line_number": 79, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree", "line_number": 79, "usage_type": "name"}, {"api_name": "sparv.util.log.warning", "line_number": 81, "usage_type": "call"}, {"api_name": "sparv.util.log", "line_number": 81, "usage_type": "attribute"}, {"api_name": "sparv.util", "line_number": 81, "usage_type": "name"}, {"api_name": "sparv.util.edgeStart", "line_number": 92, "usage_type": "call"}, {"api_name": "sparv.util", "line_number": 92, "usage_type": "name"}, {"api_name": "sparv.util.edgeEnd", "line_number": 105, "usage_type": "call"}, {"api_name": "sparv.util", "line_number": 105, "usage_type": "name"}, {"api_name": "sparv.util.mkEdge", "line_number": 107, "usage_type": "call"}, {"api_name": "sparv.util", "line_number": 107, "usage_type": "name"}, {"api_name": "sparv.util.log.warning", "line_number": 127, "usage_type": "call"}, {"api_name": "sparv.util.log", "line_number": 127, "usage_type": "attribute"}, {"api_name": "sparv.util", "line_number": 127, "usage_type": "name"}, {"api_name": "sparv.util.write_annotation", "line_number": 131, "usage_type": "call"}, {"api_name": "sparv.util", "line_number": 131, "usage_type": "name"}, {"api_name": "sparv.util.write_annotation", "line_number": 132, "usage_type": "call"}, {"api_name": "sparv.util", "line_number": 132, "usage_type": "name"}, {"api_name": "sparv.util.write_annotation", "line_number": 133, "usage_type": "call"}, {"api_name": "sparv.util", "line_number": 133, "usage_type": "name"}, {"api_name": "sparv.util.write_annotation", "line_number": 134, "usage_type": "call"}, {"api_name": "sparv.util", "line_number": 134, "usage_type": "name"}, {"api_name": "sparv.util.system.call_binary", "line_number": 139, "usage_type": "call"}, {"api_name": "sparv.util.system", "line_number": 139, "usage_type": "attribute"}, {"api_name": "sparv.util", "line_number": 139, "usage_type": "name"}, {"api_name": "sparv.util.run.main", "line_number": 143, "usage_type": "call"}, {"api_name": "sparv.util.run", "line_number": 143, "usage_type": "attribute"}, {"api_name": "sparv.util", "line_number": 143, "usage_type": "name"}]}
+{"seq_id": "575943540", "text": "import pytest\n\n\n@pytest.fixture\ndef gopath(tmpdir_factory):\n return tmpdir_factory.mktemp(\"gopath\")\n\n\ndef test_env(cmd, project, gopath):\n cmd.run(f\"export GOPATH={gopath}\")\n\n project.write_devyml(\"\"\"\n up:\n - go: '1.5'\n \"\"\")\n\n cmd.run(\"bud up\")\n\n output = cmd.run(\"go version\")\n assert \"go version go1.5\" in output\n\n\ndef test_warn_gopath_missing(cmd, project, gopath):\n cmd.run(\"unset GOPATH\")\n\n project.write_devyml(\"\"\"\n up:\n - go: '1.5'\n \"\"\")\n\n output = cmd.run(\"bud up\", expect_exit_code=1)\n assert \"The GOPATH environment variable should be set\" in output\n\n\ndef test_with_modules(cmd, project, srcdir):\n # We want to support pre-modules and modules projects in the same environment\n # so we set a GOPATH as it would be for pre-modules setup\n # Devbuddy will set GO111MODULES=on to force-enable Go modules even if we are in the GOPATH\n cmd.run(f\"export GOPATH={srcdir}\")\n\n project.write_devyml(\"\"\"\n up:\n - go:\n version: '1.12'\n modules: true\n \"\"\")\n\n output = cmd.run(\"bud up\")\n\n project.write_file_dedent(\"main.go\", \"\"\"\n package main\n\n import (\n \"fmt\"\n \"github.com/spf13/pflag\"\n )\n\n func main() {\n pflag.Parse()\n fmt.Println(pflag.Arg(0))\n }\n \"\"\")\n\n project.write_file_dedent(\"go.mod\", \"\"\"\n module poipoi\n\n require github.com/spf13/pflag v1.0.3\n \"\"\")\n\n project.write_file_dedent(\"go.sum\", \"\"\"\n github.com/spf13/pflag v1.0.3 h1:zPAT6CGy6wXeQ7NtTnaTerfKOsV6V6F8agHXFiazDkg=\n github.com/spf13/pflag v1.0.3/go.mod h1:DYY7MBk1bdzusC3SYhjObp+wFpr4gzcvqqNjLnInEg4=\n \"\"\")\n\n cmd.run(\"go mod tidy\")\n cmd.run(\"go mod download\")\n\n output = cmd.run(\"go run main.go Test1234\")\n assert output == \"Test1234\"\n", "sub_path": "tests/test_task_go.py", "file_name": "test_task_go.py", "file_ext": "py", "file_size_in_byte": 1853, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "pytest.fixture", "line_number": 4, "usage_type": "attribute"}]}
+{"seq_id": "141287778", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nimport tensorflow as tf\n\nELU = \"elu\"\nRELU = \"relu\"\nTANH = \"tanh\"\nSIGMOID = \"sigmoid\"\nACTIVATIONS = [SIGMOID, TANH, RELU, ELU]\n\n\ndef print_subplot(i, j, x, y, title):\n a = axes[i, j]\n a.axvline(x=0, color='k')\n a.axhline(y=0, color='k')\n a.plot(x, y)\n a.set_title(title)\n\n\nstart = -10\nend = 10\nsize = 50\n\nx = np.linspace(start, end, size)\nys = {}\n\nfor activation_string in ACTIVATIONS:\n tf.reset_default_graph()\n activation_input = tf.placeholder(tf.float32)\n activation = getattr(tf.nn, activation_string)\n output = activation(activation_input)\n\n with tf.Session() as sess:\n y = sess.run(output, feed_dict={activation_input: x})\n ys[activation_string] = y\n\nfig, axes = plt.subplots(\n 2,\n 2,\n gridspec_kw={'width_ratios': [1, 1], 'height_ratios': [1, 1]},\n figsize=(16, 5))\nprint_subplot(0, 0, x, ys[SIGMOID], SIGMOID.title())\nprint_subplot(0, 1, x, ys[TANH], TANH.title())\nprint_subplot(1, 0, x, ys[RELU], RELU.title())\nprint_subplot(1, 1, x, ys[ELU], ELU.title())\n\nfig.tight_layout()\nplt.show()\n", "sub_path": "model/src/images/activation_functions.py", "file_name": "activation_functions.py", "file_ext": "py", "file_size_in_byte": 1103, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "numpy.linspace", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.reset_default_graph", "line_number": 28, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 29, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}]}
+{"seq_id": "550490319", "text": "import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\n\n\nclass OFCPokerNet(nn.Module):\n\n def __init__(self, embedding_dim, action_space, drop_prob, num_layers=2, hidden_size=128):\n super(OFCPokerNet, self).__init__()\n self.embedding_dim = embedding_dim\n self.hidden_size = hidden_size\n self.three_hand_embed = nn.Linear(52, self.embedding_dim)\n self.five_hand_embed = nn.Linear(52, self.embedding_dim)\n self.cur_card_embed = nn.Linear(52, self.hidden_size)\n\n input_dim = 6 * self.embedding_dim + self.hidden_size\n self.fc1 = nn.Linear(input_dim, self.hidden_size)\n self.layers = [nn.Linear(hidden_size, self.hidden_size) for _ in range(num_layers)]\n self.value = nn.Linear(self.hidden_size, 1)\n self.policy = nn.Linear(self.hidden_size, action_space)\n self.dropout = nn.Dropout(p=drop_prob)\n\n def to(self, *args, **kwargs):\n self = super(OFCPokerNet, self).to(*args, **kwargs) \n self.layers = [layer.to(*args, **kwargs) for layer in self.layers]\n return self\n\n def forward(self, front, mid, back, cur):\n front_embed = self.three_hand_embed(front).view(-1, self.embedding_dim * 2)\n mid_embed = self.five_hand_embed(mid).view(-1, self.embedding_dim * 2)\n back_embed = self.five_hand_embed(back).view(-1, self.embedding_dim * 2)\n card_embed = self.cur_card_embed(cur)\n x = torch.cat((front_embed, mid_embed, back_embed, card_embed), 1)\n out = F.relu(self.fc1(x))\n out = self.dropout(out)\n for layer in self.layers:\n out = F.relu(layer(out))\n out = self.dropout(out)\n pi = self.policy(out)\n v = self.value(out)\n return F.log_softmax(pi, dim=1), torch.tanh(v)\n", "sub_path": "ofcpoker/pytorch/OFCPokerNNet.py", "file_name": "OFCPokerNNet.py", "file_ext": "py", "file_size_in_byte": 1780, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "torch.nn.Module", "line_number": 6, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 6, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 13, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.functional.log_softmax", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.tanh", "line_number": 41, "usage_type": "call"}]}
+{"seq_id": "145765530", "text": "import numpy as np\nimport argparse\nimport matplotlib.gridspec\nimport matplotlib.pyplot as plt\nimport scipy.stats\nimport statsmodels.stats.multitest\n\nfrom behavior import backup\nfrom analysis.batch import customized_plot\n\n\ndef main(force):\n\n backups = backup.get_data(force)\n\n backups = [b for b in backups if b.pvp]\n\n # ----------------- Data ------------------- #\n\n # Look at the parameters\n n_simulations = len(backups)\n n_positions = backups[0].n_positions\n\n # Containers\n d = np.zeros(n_simulations)\n prices = np.zeros(n_simulations)\n scores = np.zeros(n_simulations)\n r = np.zeros(n_simulations)\n s = np.zeros(n_simulations, dtype=bool)\n\n for i, b in enumerate(backups):\n\n # Compute the mean distance between the two firms\n data = np.absolute(\n b.positions[:, 0] -\n b.positions[:, 1]) / n_positions\n\n d[i] = np.mean(data)\n\n # Compute the mean price\n prices[i] = np.mean(b.prices[:, :])\n\n # Compute the mean profit\n scores[i] = np.mean(b.profits[:, :])\n\n r[i] = b.r\n s[i] = b.display_opponent_score\n\n # ---------- Plot ----------------------------- #\n\n fig = plt.figure(figsize=(4, 7), dpi=200)\n\n sub_gs = matplotlib.gridspec.GridSpec(nrows=3, ncols=2)\n\n axes = (\n fig.add_subplot(sub_gs[0, 0]),\n fig.add_subplot(sub_gs[0, 1]),\n fig.add_subplot(sub_gs[1, 0]),\n fig.add_subplot(sub_gs[1, 1]),\n fig.add_subplot(sub_gs[2, 0]),\n fig.add_subplot(sub_gs[2, 1])\n )\n\n y_labels = \"Distance\", \"Price\", \"Profit\"\n y_limits = (0, 1), (1, 11), (0, 120)\n\n s_values = (0, 1, ) * 3\n\n arr = (d, d, prices, prices, scores, scores)\n\n # axes[0].text(2, 1.3, \"Display opponent score\", fontsize=12)\n axes[0].set_title(\"$s = 0$\")\n axes[1].set_title(\"$s = 1$\")\n\n for idx in range(len(axes)):\n\n ax = axes[idx]\n\n ax.set_axisbelow(True)\n\n # Violin plot\n data = [arr[idx][(r == r_value) * (s == s_values[idx])] for r_value in (0.25, 0.50)]\n color = ['C0' if r_value == 0.25 else 'C1' for r_value in (0.25, 0.50)]\n\n customized_plot.violin(ax=ax, data=data, color=color, edgecolor=\"white\", alpha=0.8) # color, alpha=0.5)\n\n for ax in axes[0:2]:\n ax.set_yticks(np.arange(0, 1.1, 0.25))\n\n for ax in axes[2:4]:\n ax.set_yticks(np.arange(1, 11.1, 2))\n\n for ax in axes[-2:]:\n ax.set_xticklabels([\"{:.2f}\".format(i) for i in (0.25, 0.50)])\n ax.set_xlabel(\"$r$\")\n\n for ax in axes[:4]:\n ax.tick_params(length=0, axis=\"x\")\n ax.set_xticklabels([])\n\n for ax, y_label, y_lim in zip(axes[0::2], y_labels, y_limits):\n ax.text(-0.35, 0.5, y_label, rotation=\"vertical\", verticalalignment='center',\n horizontalalignment='center', transform=ax.transAxes, fontsize=12)\n ax.set_ylabel(\" \")\n ax.tick_params(axis=\"y\", labelsize=9)\n ax.set_ylim(y_lim)\n\n for ax, y_lim in zip(axes[1::2], y_limits):\n ax.set_ylim(y_lim)\n ax.tick_params(length=0, axis=\"y\")\n ax.set_yticklabels([])\n\n plt.tight_layout()\n\n plt.savefig(\"fig/main_exp.pdf\")\n plt.show()\n\n # ----------- Stats ----------------- #\n\n to_compare = [\n {\n \"measure\": \"distance\",\n \"constant\": \"s = 0\",\n \"var\": \"r\",\n \"data\": [d[(r == r_value) * (s == 0)] for r_value in (0.25, 0.50)]\n }, {\n \"measure\": \"distance\",\n \"constant\": \"s = 1\",\n \"var\": \"r\",\n \"data\": [d[(r == r_value) * (s == 1)] for r_value in (0.25, 0.50)]\n }, {\n \"measure\": \"price\",\n \"constant\": \"s = 0\",\n \"var\": \"r\",\n \"data\": [prices[(r == r_value) * (s == 0)] for r_value in (0.25, 0.50)]\n }, {\n \"measure\": \"price\",\n \"constant\": \"s = 1\",\n \"var\": \"r\",\n \"data\": [prices[(r == r_value) * (s == 1)] for r_value in (0.25, 0.50)]\n }, {\n \"measure\": \"profit\",\n \"constant\": \"s = 0\",\n \"var\": \"r\",\n \"data\": [scores[(r == r_value) * (s == 0)] for r_value in (0.25, 0.50)]\n }, {\n \"measure\": \"profit\",\n \"constant\": \"s = 1\",\n \"var\": \"r\",\n \"data\": [scores[(r == r_value) * (s == 1)] for r_value in (0.25, 0.50)]\n }, {\n \"measure\": \"distance\",\n \"constant\": \"r = 0.25\",\n \"var\": \"s\",\n \"data\": [d[(r == 0.25) * (s == s_value)] for s_value in (0, 1)]\n }, {\n \"measure\": \"distance\",\n \"constant\": \"r = 0.50\",\n \"var\": \"s\",\n \"data\": [d[(r == 0.50) * (s == s_value)] for s_value in (0, 1)]\n }, {\n \"measure\": \"price\",\n \"constant\": \"r = 0.25\",\n \"var\": \"s\",\n \"data\": [prices[(r == 0.25) * (s == s_value)] for s_value in (0, 1)]\n }, {\n \"measure\": \"price\",\n \"constant\": \"r = 0.50\",\n \"var\": \"s\",\n \"data\": [prices[(r == 0.50) * (s == s_value)] for s_value in (0, 1)]\n }, {\n \"measure\": \"profit\",\n \"constant\": \"r = 0.25\",\n \"var\": \"s\",\n \"data\": [scores[(r == 0.25) * (s == s_value)] for s_value in (0, 1)]\n }, {\n \"measure\": \"profit\",\n \"constant\": \"r = 0.50\",\n \"var\": \"s\",\n \"data\": [scores[(r == 0.50) * (s == s_value)] for s_value in (0, 1)]\n }\n ]\n\n ps = []\n us = []\n\n for dic in to_compare:\n u, p = scipy.stats.mannwhitneyu(dic[\"data\"][0], dic[\"data\"][1])\n ps.append(p)\n us.append(u)\n\n valid, p_corr, alpha_c_sidak, alpha_c_bonf = \\\n statsmodels.stats.multitest.multipletests(pvals=ps, alpha=0.01, method=\"b\")\n\n for p, u, p_c, v, dic in zip(ps, us, p_corr, valid, to_compare):\n print(\"[Diff in {} when {} depending on {}-value] \"\n \"Mann-Whitney rank test: u {}, p {:.3f}, p corr {:.3f}, significant: {}\"\n .format(dic[\"measure\"], dic[\"constant\"], dic[\"var\"], u, p, p_c, v))\n print()\n\n table = \\\n r\"\\begin{table}[htbp]\" + \"\\n\" + \\\n r\"\\begin{center}\" + \"\\n\" + \\\n r\"\\begin{tabular}{llllllll}\" + \"\\n\" + \\\n r\"Measure & Variable & Constant & $u$ & $p$ (before corr.) \" \\\n r\"& $p$ (after corr.) & Sign. at 1\\% threshold \\\\\" + \"\\n\" + \\\n r\"\\hline \\\\\" + \"\\n\"\n\n for p, u, p_c, v, dic in zip(ps, us, p_corr, valid, to_compare):\n\n p = \"{:.3f}\".format(p) if p >= 0.001 else \"$<$ 0.001\"\n p_c = \"{:.3f}\".format(p_c) if p_c >= 0.001 else \"$<$ 0.001\"\n v = \"yes\" if v else \"no\"\n table += r\"{} & ${}$ & ${}$ & {} & {} & {} & {} \\\\\"\\\n .format(dic[\"measure\"], dic[\"var\"], dic[\"constant\"], u, p, p_c, v) \\\n + \"\\n\"\n\n table += \\\n r\"\\end{tabular}\" + \"\\n\" + \\\n r\"\\end{center}\" + \"\\n\" + \\\n r\"\\caption{Significance tests for comparison using Mann-Withney's u. \" \\\n r\"Bonferroni corrections are applied.}\" + \"\\n\" + \\\n r\"\\label{table:significance_tests}\" + \"\\n\" + \\\n r\"\\end{table}\"\n\n print(\"*** Latex-formated table ***\")\n print(table)\n\n\nif __name__ == \"__main__\":\n\n parser = argparse.ArgumentParser(description='Produce figures.')\n parser.add_argument('-f', '--force', action=\"store_true\", default=False,\n help=\"Re-import data\")\n parsed_args = parser.parse_args()\n\n main(force=parsed_args.force)\n", "sub_path": "__old__/analyse.py", "file_name": "analyse.py", "file_ext": "py", "file_size_in_byte": 7470, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "behavior.backup.get_data", "line_number": 14, "usage_type": "call"}, {"api_name": "behavior.backup", "line_number": 14, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.gridspec.gridspec.GridSpec", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.gridspec.gridspec", "line_number": 53, "usage_type": "attribute"}, {"api_name": "matplotlib.gridspec", "line_number": 53, "usage_type": "name"}, {"api_name": "analysis.batch.customized_plot.violin", "line_number": 85, "usage_type": "call"}, {"api_name": "analysis.batch.customized_plot", "line_number": 85, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "scipy.stats.stats.mannwhitneyu", "line_number": 188, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 188, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 188, "usage_type": "name"}, {"api_name": "statsmodels.stats.multitest.stats.multitest.multipletests", "line_number": 193, "usage_type": "call"}, {"api_name": "statsmodels.stats.multitest.stats", "line_number": 193, "usage_type": "attribute"}, {"api_name": "statsmodels.stats.multitest", "line_number": 193, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 232, "usage_type": "call"}]}
+{"seq_id": "419057550", "text": "import logging\nimport json\n\nfrom alpaca_trade_api.stream import Stream\nfrom alpaca_trade_api.common import URL\n\n\nlog = logging.getLogger(__name__)\n\n\nasync def print_trade(t):\n print('trade', t)\n\n\nasync def print_quote(q):\n print('quote', q)\n\n\nasync def print_trade_update(tu):\n print('trade update', tu)\n\n\ndef main():\n logging.basicConfig(level=logging.INFO)\n \n with open(\"./config.json\",\"rb\") as file:\n config = json.load(file)\n \n feed = 'iex' # <- replace to SIP if you have PRO subscription\n stream = Stream(key_id=config['alpaca_key_id'],\n secret_key=config['alpaca_secret_key'],\n base_url=URL(config['alpaca_base_url']),\n data_feed=feed, raw_data=True)\n stream.subscribe_trade_updates(print_trade_update)\n stream.subscribe_trades(print_trade, 'AAPL')\n stream.subscribe_quotes(print_quote, 'IBM')\n\n @stream.on_bar('MSFT')\n async def _(bar):\n print('bar', bar)\n\n @stream.on_status(\"*\")\n async def _(status):\n print('status', status)\n\n stream.run()\n\n\nif __name__ == \"__main__\":\n main()", "sub_path": "app/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 1120, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 24, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 27, "usage_type": "call"}, {"api_name": "alpaca_trade_api.stream.Stream", "line_number": 30, "usage_type": "call"}, {"api_name": "alpaca_trade_api.common.URL", "line_number": 32, "usage_type": "call"}]}
+{"seq_id": "577471574", "text": "#! -*- coding:utf-8 -*-\nimport numpy as np\nfrom sklearn import linear_model\n\n# d <- read.csv(file='input/data-salary.txt')\n# res_lm <- lm(Y ~ X, data=d)\n# X_new <- data.frame(X=23:60)\n# conf_95 <- predict(res_lm, X_new, interval='confidence', level=0.95)\n# pred_95 <- predict(res_lm, X_new, interval='prediction', level=0.95)\n\nd = np.genfromtxt(fname='input/data-salary.txt', delimiter=',', names=True, dtype=np.float)\nlm = linear_model.LinearRegression()\nlm.fit(d['X'].reshape(d.size, 1), d['Y'])\n\nprint('Intercept: ' + str(lm.intercept_))\nprint('Coefficients: ' + str(lm.coef_[0]))\n# 一応p.38の1 行目の数値はとは一致した。\n# 信頼区間と予測区間については、まだ書いてない\n", "sub_path": "chap04/lm.py", "file_name": "lm.py", "file_ext": "py", "file_size_in_byte": 710, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "numpy.genfromtxt", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 11, "usage_type": "attribute"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 12, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 12, "usage_type": "name"}]}
+{"seq_id": "51485543", "text": "#! /usr/bin/env python\n# coding: utf-8\n\nimport os\nimport argparse\nfrom JYTools.JYWorker import RedisQueue\n\n# conf_dir = \"/public/JINGD/conf\"\n# conf_path = os.path.join(conf_dir, \"redis_worker.conf\")\nconf_path = os.environ.get(\"REDIS_WORKER_CONF_PATH\")\nr_queue = RedisQueue(conf_path=conf_path, work_tag=\"JYGroupDAG\")\n\n\n\ndef run_hotspot(normal_recal_bam, hotspot_vcf, sample_no):\n apply_pipeline = {\"task_list\": [{\"task_type\": \"app\", \"work_tag\": \"RunHotspot\",\n \"input_sample_no\": sample_no,\n \"input_hotspot_vcf\": hotspot_vcf,\n \"input_normal_recal_bam\": normal_recal_bam}],\n \"task_type\": \"pipeline\"}\n r_queue.push(sample_no, apply_pipeline)\n\n\ndef main():\n usage = \"Help message\"\n description = \"Run manta pipeline\"\n parser = argparse.ArgumentParser(usage=usage, description=description)\n\n parser.add_argument(\"-n\", \"--normal_recal_bam\", dest=\"normal_recal_bam\", help=\"normal recal bam path\")\n parser.add_argument(\"-v\", \"--hotspot_vcf\", dest=\"hotspot_vcf\", help=\"hot spot vcf path\")\n\n parser.add_argument(\"-s\", \"--sample-no\", dest=\"sample_no\", help=\"sample no\")\n args = parser.parse_args()\n normal_recal_bam = args.normal_recal_bam\n hotspot_vcf = args.hotspot_vcf\n sample_no = args.sample_no\n\n run_hotspot(normal_recal_bam, hotspot_vcf, sample_no)\n\n\nif __name__ == \"__main__\":\n main()\n", "sub_path": "anzhen/branch_pipeline/test_hotspot.py", "file_name": "test_hotspot.py", "file_ext": "py", "file_size_in_byte": 1453, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "os.environ.get", "line_number": 10, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "JYTools.JYWorker.RedisQueue", "line_number": 11, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 27, "usage_type": "call"}]}
+{"seq_id": "388173261", "text": "# (C) Datadog, Inc. 2019\n# All rights reserved\n# Licensed under a 3-clause BSD style license (see LICENSE)\nimport pytest\n\nfrom . import metrics\n\npytestmark = pytest.mark.e2e\n\n\n@pytest.mark.e2e\ndef test_check(dd_agent_check, instance):\n aggregator = dd_agent_check(instance, rate=True)\n\n for metric in metrics.STANDARD:\n aggregator.assert_metric_has_tag(metric, 'server:{}'.format(instance['server']))\n aggregator.assert_metric_has_tag(metric, 'port:{}'.format(instance['port']))\n\n aggregator.assert_all_metrics_covered()\n", "sub_path": "sap_hana/tests/test_e2e.py", "file_name": "test_e2e.py", "file_ext": "py", "file_size_in_byte": 544, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "pytest.mark", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 11, "usage_type": "attribute"}]}
+{"seq_id": "295427675", "text": "import argparse\nimport re\n\nfrom codegen_outofplacebatching import deindent, get_signatures, gen_unwraps\n\n\ndef get_signature(op, path):\n signatures = get_signatures(path, include_op=True)\n result = [sig for sig in signatures if sig[0] == op]\n if len(result) != 1:\n raise ValueError(\"\")\n return result[0]\n\n\ndef gen_return_sig(return_t):\n if len(return_t) == 1:\n return return_t[0]\n return f'std::tuple<{\".\".join(return_t)}>'\n\n\ndef gen_args_sig(args_t):\n args = [f'{typ} {argname}' for typ, argname in args_t]\n return ', '.join(args)\n\n\ndef gen_args_list(args_t):\n args = [f'{argname}' for _, argname in args_t]\n return ', '.join(args)\n\n\ndef gen_plumbing(signature):\n # \"add.Tensor\"\n op, return_t, args_t = signature\n\n maybe_op_and_variant = op.split('.')\n if len(maybe_op_and_variant) == 1:\n op = maybe_op_and_variant[0]\n variant = ''\n opname = op\n else:\n op, variant = maybe_op_and_variant\n opname = f'{op}_{variant}'\n\n if op.endswith('_'):\n raise ValueError('Codegen doesn\\'t handle in-place ops')\n\n arg_types, arg_names = zip(*args_t)\n unwraps, _ = gen_unwraps(arg_types, arg_names)\n\n result = deindent(f\"\"\"\\\n {gen_return_sig(return_t)} {opname}_plumbing({gen_args_sig(args_t)}) {{\n auto maybe_layer = maybeCurrentDynamicLayer();\n TORCH_INTERNAL_ASSERT(maybe_layer.has_value());\n int64_t cur_level = maybe_layer->layerId();\n\n {unwraps}\n\n // Your logic here\n\n static auto op = c10::Dispatcher::singleton()\n .findSchemaOrThrow(\"aten::{op}\", \"{variant}\");\n return slow_fallback<{','.join(return_t)}>(op, {{ {gen_args_list(args_t)} }});\n }}\n \"\"\")\n return result\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser(\n description='Generate the batch rule plumbing for an op')\n parser.add_argument('op',\n help='the operator name (with overload name)')\n parser.add_argument('path',\n help='link to RegistrationDeclarations.h')\n\n # Sample usage:\n # gen_plumbing.py add.Tensor ~/pytorch/build/aten/src/ATen/RegistrationDeclarations.h\n args = parser.parse_args()\n signature = get_signature(args.op, args.path)\n result = gen_plumbing(signature)\n print(result)\n", "sub_path": "codegen/gen_plumbing.py", "file_name": "gen_plumbing.py", "file_ext": "py", "file_size_in_byte": 2305, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "codegen_outofplacebatching.get_signatures", "line_number": 8, "usage_type": "call"}, {"api_name": "codegen_outofplacebatching.gen_unwraps", "line_number": 48, "usage_type": "call"}, {"api_name": "codegen_outofplacebatching.deindent", "line_number": 50, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 69, "usage_type": "call"}]}
+{"seq_id": "277050321", "text": "#!/usr/bin/python3\n# -*- coding: utf-8 -*-\n\n\"\"\"\n.. module:: recordation\n :platform: Unix\n :synopsis: the top-level submodule of T_System that contains the classes related to T_System's recording video and audio ability.\n\n.. moduleauthor:: Cem Baybars GÜÇLÜ \n\"\"\"\nimport os # Miscellaneous operating system interfaces\nimport datetime # Basic date and time types\nimport subprocess # Subprocess managements\nimport uuid # The random id generator\n\nfrom shutil import rmtree\nfrom tinydb import Query # TinyDB is a lightweight document oriented database\n\nfrom t_system.db_fetching import DBFetcher\n\nfrom t_system import dot_t_system_dir\nfrom t_system import log_manager\n\nlogger = log_manager.get_logger(__name__, \"DEBUG\")\n\n\nclass Recorder:\n \"\"\"Class to define a recording ability of tracking system.\n\n This class provides necessary initiations and functions named :func:`t_system.recordation.RecordManager.start`\n for creating a Record object and start recording by this object. :func:`t_system.recordation.RecordManager.merge_audio_and_video`\n for merging separate audio and video file to one.\n \"\"\"\n\n def __init__(self, record_formats, camera, hearer):\n \"\"\"Initialization method of :class:`t_system.recordation.Recorder` class.\n\n Args:\n record_formats (list): Formats of the records for video, audio and merged.\n camera: \t Camera object from PiCamera.\n hearer: \t Hearer object.\n \"\"\"\n\n self.current_video_file = \"\"\n self.current_audio_file = \"\"\n self.current_merged_file = \"\"\n\n self.record_formats = {\"video\": record_formats[0], \"audio\": record_formats[1], \"merged\": record_formats[2]}\n\n self.camera = camera\n self.hearer = hearer\n \n def start(self, mode=\"track\"):\n \"\"\"Method to start audio and video recording asynchronously.\n\n Args:\n mode: \t The running mode which is wants to set video name.\n \"\"\"\n logger.debug(\"Record starting...\")\n record = Record(datetime.datetime.now().strftime(\"%d_%m_%Y\"), datetime.datetime.now().strftime(\"%H_%M_%S\"), mode, self.record_formats)\n\n self.__set_record_params(record)\n\n self.camera.start_recording(self.current_video_file, self.record_formats[\"video\"])\n self.hearer.start_recording(self.current_audio_file, self.record_formats[\"audio\"])\n\n def stop(self):\n \"\"\"Method to stop audio and video recording\n \"\"\"\n\n self.camera.stop_recording()\n self.hearer.stop_recording()\n\n # Todo: This is disgusting way to merging audio and silent video. Fix this.\n self.merge_audio_and_video()\n\n def merge_audio_and_video(self):\n \"\"\"Method to merge recorded audio and video files.\n \"\"\"\n\n merge_cmd = f'ffmpeg -y -i {self.current_audio_file} -r 24 -i {self.current_video_file} -filter:a aresample=async=1 -c:a flac -strict -2 -c:v copy {self.current_merged_file}'\n\n subprocess.call(merge_cmd, shell=True)\n\n logger.info('Video and Audio Muxing Done')\n\n def __set_record_params(self, record):\n \"\"\"Method to setting current parameter by current recording.\n \"\"\"\n\n self.current_video_file = record.video_file\n self.current_audio_file = record.audio_file\n self.current_merged_file = record.merged_file\n\n\nclass RecordManager:\n \"\"\"Class to define Record manager for handling the recordation database of t_system's vision.\n\n This class provides necessary initiations and functions named :func:`t_system.recordation.RecordManager.get_records`\n for returning the Record objects of existing records with given table(date at the same time) parameter.\n \"\"\"\n\n def __init__(self):\n \"\"\"Initialization method of :class:`t_system.recordation.RecordManager` class.\n \"\"\"\n\n self.records_folder = f'{dot_t_system_dir}/records'\n\n if not os.path.exists(self.records_folder):\n os.mkdir(self.records_folder)\n\n self.db = DBFetcher(self.records_folder, \"db\").fetch()\n\n self.records = []\n\n self.__set_records()\n\n def __set_records(self):\n \"\"\"Method to set existing records.\n \"\"\"\n\n for record in self.db.all():\n self.records.append(Record(record[\"date\"], record[\"time\"], record[\"scope\"], record[\"record_formats\"], record[\"id\"], record[\"name\"], record[\"length\"]))\n\n def refresh_records(self):\n \"\"\"Method to refresh existing records on runtime alterations.\n \"\"\"\n\n self.records.clear()\n self.__set_records()\n\n def get_records(self, date=None):\n \"\"\"Method to get existing records in given date. If date is None it returns all records.\n\n Args:\n date (str): Parent date of the record. In day_mount_year format.\n \"\"\"\n records = []\n\n if date:\n for record in self.records:\n if record.date == date:\n records.append(record)\n return records\n\n return self.records\n\n def get_record(self, id):\n \"\"\"Method to get existing record in given id.\n\n Args:\n id (str): ID of the record.\n \"\"\"\n\n for record in self.records:\n if record.id == id:\n return record\n\n def get_record_dates(self):\n \"\"\"Method to get date list of existing records.\n \"\"\"\n dates = []\n for record in self.records:\n dates.append(record.date)\n\n dates = list(dict.fromkeys(dates)) # removes duplicated dates.\n\n return dates\n\n def delete_record(self, id):\n \"\"\"Method to get date list of existing records.\n\n Args:\n id (str): ID of the record.\n \"\"\"\n\n for record in self.records:\n if record.id == id:\n record.remove_self()\n self.records.remove(record) # for removing object from list\n return True\n return False\n\n def update_record(self, id, name):\n \"\"\"Method to updating record that has given id.\n\n Args:\n id (str): ID of the record.\n name (str): The name of the record.\n \"\"\"\n\n for record in self.records:\n if record.id == id:\n record.update_name(name)\n return True\n return False\n\n\nclass Record:\n \"\"\"Class to define records of t_systems vision.\n\n This class provides necessary initiations and functions named :func:`t_system.recordation.Record.__db_upsert`\n for saving records to the database safely.\n \"\"\"\n\n def __init__(self, d_m_y, h_m_s, scope, record_formats, id=None, name=None, length=None):\n \"\"\"Initialization method of :class:`t_system.recordation.Record` class.\n\n Args:\n d_m_y (str): Date that is day_mount_year format.\n h_m_s (str): Date that is hour_minute_second format.\n scope (str): The working type during recording.\n record_formats (dict): Formats of the records for video, audio and merged.\n id (str): The id of the record.\n name (str): The name of the record.\n length (str): The length of the record as m:s.\n \"\"\"\n\n self.id = id\n if not id:\n self.id = str(uuid.uuid1())\n\n self.name = name\n if not name:\n self.name = h_m_s\n\n self.date = d_m_y # table name at the same time\n self.time = h_m_s\n self.scope = scope\n self.record_formats = record_formats\n self.length = length\n\n self.records_folder = f'{dot_t_system_dir}/records'\n self.parent_folder = f'{self.records_folder}/{self.date}'\n self.folder = f'{self.parent_folder}/{self.time}'\n\n self.video_file = f'{self.folder}/{self.time}.{self.record_formats[\"video\"]}'\n self.audio_file = f'{self.folder}/{self.time}.{self.record_formats[\"audio\"]}'\n self.merged_file = f'{self.folder}/{self.time}.{self.record_formats[\"merged\"]}'\n\n self.db = DBFetcher(self.records_folder, \"db\").fetch()\n\n self.__check_folders()\n\n if length is None:\n self.length = self.__calc_length()\n\n self.__db_upsert()\n\n def __db_upsert(self, force_insert=False):\n \"\"\"Function to insert(or update) the record to the database.\n\n Args:\n force_insert (bool): Force insert flag.\n\n Returns:\n str: Response.\n \"\"\"\n\n if self.db.search((Query().id == self.id)):\n if force_insert:\n # self.already_exist = False\n self.db.update({'id': self.id, 'name': self.name, 'time': self.time, 'date': self.date, 'scope': self.scope, 'record_formats': self.record_formats, 'length': self.length}, Query().id == self.id)\n\n else:\n # self.already_exist = True\n return \"Already Exist\"\n else:\n self.db.insert({\n 'id': self.id,\n 'name': self.name,\n 'time': self.time,\n 'date': self.date,\n 'scope': self.scope,\n 'record_formats': self.record_formats,\n 'length': self.length\n }) # insert the given data\n\n return \"\"\n\n def update_name(self, name):\n \"\"\"Method to updating self name via by given parameter.\n\n Args:\n name (str): The name of the record.\n \"\"\"\n\n self.name = name\n self.__db_upsert(True)\n\n def remove_self(self):\n \"\"\"Method to remove face itself.\n \"\"\"\n\n rmtree(self.folder)\n\n self.db.remove((Query().id == self.id))\n\n def __calc_length(self):\n \"\"\"Method to calculating length of record with using OpenCV.\n \"\"\"\n if os.path.exists(self.merged_file):\n import cv2\n\n cap = cv2.VideoCapture(self.merged_file)\n\n fps = cap.get(cv2.CAP_PROP_FPS) # OpenCV2 version 2 used \"CV_CAP_PROP_FPS\"\n frame_count = int(cap.get(cv2.CAP_PROP_FRAME_COUNT))\n duration = frame_count / fps\n\n minutes = int(duration / 60)\n seconds = round(duration % 60)\n length = f'{minutes}:{seconds}'\n\n cap.release()\n\n return length\n\n return None\n\n def __check_folders(self):\n \"\"\"Method to checking the necessary folders created before. If not created creates them.\n \"\"\"\n\n if not os.path.exists(self.parent_folder):\n os.mkdir(self.parent_folder)\n\n if not os.path.exists(self.folder):\n os.mkdir(self.folder)\n", "sub_path": "t_system/recordation.py", "file_name": "recordation.py", "file_ext": "py", "file_size_in_byte": 10818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "t_system.log_manager.get_logger", "line_number": 24, "usage_type": "call"}, {"api_name": "t_system.log_manager", "line_number": 24, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "attribute"}, {"api_name": "subprocess.call", "line_number": 83, "usage_type": "call"}, {"api_name": "t_system.dot_t_system_dir", "line_number": 107, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 110, "usage_type": "call"}, {"api_name": "t_system.db_fetching.DBFetcher", "line_number": 112, "usage_type": "call"}, {"api_name": "uuid.uuid1", "line_number": 221, "usage_type": "call"}, {"api_name": "t_system.dot_t_system_dir", "line_number": 233, "usage_type": "name"}, {"api_name": "t_system.db_fetching.DBFetcher", "line_number": 241, "usage_type": "call"}, {"api_name": "tinydb.Query", "line_number": 260, "usage_type": "call"}, {"api_name": "tinydb.Query", "line_number": 263, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 295, "usage_type": "call"}, {"api_name": "tinydb.Query", "line_number": 297, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 302, "usage_type": "call"}, {"api_name": "os.path", "line_number": 302, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 305, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 307, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_COUNT", "line_number": 308, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 325, "usage_type": "call"}, {"api_name": "os.path", "line_number": 325, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 326, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 328, "usage_type": "call"}, {"api_name": "os.path", "line_number": 328, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 329, "usage_type": "call"}]}
+{"seq_id": "431193496", "text": "# -*- coding: utf-8 -*-\n#\n# Python wrapper around the CMake build system\n#\n# Copyright (c) Honda Research Institute Europe GmbH\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are\n# met:\n#\n# 1. Redistributions of source code must retain the above copyright notice,\n# this list of conditions and the following disclaimer.\n#\n# 2. Redistributions in binary form must reproduce the above copyright\n# notice, this list of conditions and the following disclaimer in the\n# documentation and/or other materials provided with the distribution.\n#\n# 3. Neither the name of the copyright holder nor the names of its\n# contributors may be used to endorse or promote products derived from\n# this software without specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS\n# IS\" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,\n# THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR\n# PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR\n# CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,\n# EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,\n# PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR\n# PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF\n# LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING\n# NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS\n# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n#\n#\n\n\n#----------------------------------------------------------------------------\n# Includes\n#----------------------------------------------------------------------------\n\n\nimport collections\nimport glob\nimport logging\nimport os\nimport re\nimport shlex\n\nfrom ToolBOSCore.BuildSystem import Compilers\nfrom ToolBOSCore.Util import FastScript\nfrom ToolBOSCore.Util import Any\n\n\n#----------------------------------------------------------------------------\n# Public API\n#----------------------------------------------------------------------------\n\n\nSwitches = collections.namedtuple( 'Switches', [ 'c', 'cpp' ] )\n\n\ndef getIncludePathsAsString( targetPlatform, targetName ):\n \"\"\"\n Returns a long string with all include paths set for the package\n using include_directories() in CMakeLists.txt (in this package or\n included ones).\n\n This means all paths where the compiler would search for header\n files (beside system defaults), in the form \"-I/path1 -I/path2...\".\n\n If no additional paths are set, an empty string will be returned.\n\n NOTE: CMake supports that include directories may be different for\n various target platforms, and even per executable and/or\n library. Therefore you need to specify both of them.\n A rule of thumb is targetName='-global'.\n \"\"\"\n Any.requireIsTextNonEmpty( targetPlatform )\n Any.requireIsTextNonEmpty( targetName )\n\n fileName = os.path.join( 'build/%s/CMakeFiles/%s.dir/flags.make' %\n ( targetPlatform, targetName ) )\n\n Any.requireIsDirNonEmpty( 'build/%s' % targetPlatform )\n Any.requireIsFileNonEmpty( fileName )\n\n # read-in ground truth information\n logging.debug( 'parsing %s' % fileName )\n content = FastScript.getFileContent( fileName, splitLines=True )\n raw_C = ''\n raw_CPP = ''\n regexp_C = re.compile( '^(?:C_FLAGS|C_INCLUDES)\\s=\\s+(.*)$' )\n regexp_CPP = re.compile( '^(?:CXX_FLAGS|CXX_INCLUDES)\\s=\\s+(.*)$' )\n result = ''\n\n for line in content:\n tmp = regexp_C.search( line )\n\n if tmp:\n raw_C = tmp.group( 1 )\n # logging.debug( 'raw C flags: %s' % raw_C )\n\n tmp = regexp_CPP.search( line )\n\n if tmp:\n raw_CPP = tmp.group( 1 )\n # logging.debug( 'raw CPP flags: %s' % raw_CPP )\n\n for candidate in ( shlex.split( raw_C ) + shlex.split( raw_CPP ) ):\n if candidate.startswith( '-I' ):\n result += candidate + ' '\n\n return result\n\n\ndef getIncludePathsAsList( targetPlatform, targetName ):\n \"\"\"\n Returns a list with all include paths set for the package\n using include_directories() in CMakeLists.txt (in this package or\n included ones).\n\n This means all paths where the compiler would search for header\n files (beside system defaults).\n\n If no additional paths are set, an empty list will be returned.\n\n NOTE: CMake supports that include directories may be different for\n various target platforms, and even per executable and/or\n library. Therefore you need to specify both of them.\n A rule of thumb is targetName='-global'.\n \"\"\"\n Any.requireIsTextNonEmpty( targetPlatform )\n Any.requireIsTextNonEmpty( targetName )\n\n result = []\n\n # we are adding a trailing blank so that the \" -I\" replacement will\n # also work on the first element\n raw = getIncludePathsAsString( targetPlatform, targetName )\n tmp = (' ' + raw ).replace( ' -I', ' ' )\n\n for token in tmp.split():\n result.append( token.strip() )\n\n\n # remove empty entries (if present)\n try:\n result.remove( '' )\n except ValueError:\n pass\n\n return frozenset( result )\n\n\ndef getStdSwitches( targetPlatform, targetName ):\n \"\"\"\n Returns a string with the compiler std switch.\n\n NOTE: CMake supports that compiler definitions may be different for\n various target platforms, and even per executable and/or\n library. Therefore you need to specify both of them.\n A rule of thumb is targetName='-global'.\n \"\"\"\n Any.requireIsTextNonEmpty( targetPlatform )\n Any.requireIsTextNonEmpty( targetName )\n\n # We need defaults because the macro parser needs the switch to\n # correctly parse c++ code.\n\n\n fileName = os.path.join( 'build/%s/CMakeFiles/%s.dir/flags.make' %\n ( targetPlatform, targetName ) )\n\n Any.requireIsDirNonEmpty( 'build/%s' % targetPlatform )\n Any.requireIsFileNonEmpty( fileName )\n\n # read-in ground truth information\n logging.debug( 'parsing %s', fileName )\n content = FastScript.getFileContent( fileName, splitLines=True )\n raw_C_CFLAGS = ''\n raw_CPP_CFLAGS = ''\n regexp_C_CFLAGS = re.compile( r'^C_FLAGS\\s=\\s+(.*)$' )\n regexp_CPP_CFLAGS = re.compile( r'^CXX_FLAGS\\s=\\s+(.*)$' )\n\n for line in content:\n tmp = regexp_C_CFLAGS.search( line )\n\n if tmp:\n raw_C_CFLAGS = tmp.group( 1 )\n\n tmp = regexp_CPP_CFLAGS.search( line )\n\n if tmp:\n raw_CPP_CFLAGS = tmp.group( 1 )\n\n # get the default language standards\n standards = Compilers.getDefaultLanguageStandard(targetPlatform)\n cStdSwitch = '-std={}'.format( standards[ 'c' ] )\n cppStdSwitch = '-std={}'.format( standards[ 'c++' ] )\n\n # look if the user specified different standards in the C_FLAGS/CPP_FLAGS\n # CMake variables\n candidates = shlex.split( raw_C_CFLAGS )\n for candidate in candidates:\n if candidate.startswith( '-std=' ):\n cStdSwitch = candidate\n\n candidates = shlex.split( raw_CPP_CFLAGS )\n for candidate in candidates:\n if candidate.startswith( '-std=' ):\n cppStdSwitch = candidate\n\n return Switches( c=cStdSwitch, cpp=cppStdSwitch )\n\n\ndef getCDefinesAsString( targetPlatform, targetName ):\n \"\"\"\n Returns a long string with all compiler definitions set for the\n package using the addDefinitions() directive.\n\n This means all definitions passed to the compiler in the given path\n (beside system defaults), in the form \"-DDEFINE1 -DFOO=BAR...\".\n\n If no additional definitions are set, an empty string will be returned.\n\n NOTE: CMake supports that compiler definitions may be different for\n various target platforms, and even per executable and/or\n library. Therefore you need to specify both of them.\n A rule of thumb is targetName='-global'.\n \"\"\"\n Any.requireIsTextNonEmpty( targetPlatform )\n Any.requireIsTextNonEmpty( targetName )\n\n fileName = os.path.join( 'build/%s/CMakeFiles/%s.dir/flags.make' %\n ( targetPlatform, targetName ) )\n\n Any.requireIsDirNonEmpty( 'build/%s' % targetPlatform )\n Any.requireIsFileNonEmpty( fileName )\n\n # read-in ground truth information\n logging.debug( 'parsing %s' % fileName )\n content = FastScript.getFileContent( fileName, splitLines=True )\n raw_C = ''\n raw_CPP = ''\n raw_C_CFLAGS = ''\n raw_CPP_CFLAGS = ''\n regexp_C = re.compile( '^C_DEFINES\\s=\\s+(.*)$' )\n regexp_CPP = re.compile( '^CXX_DEFINES\\s=\\s+(.*)$' )\n regexp_C_CFLAGS = re.compile( '^C_FLAGS\\s=\\s+(.*)$' )\n regexp_CPP_CFLAGS = re.compile( '^CXX_FLAGS\\s=\\s+(.*)$' )\n result = ''\n\n for line in content:\n tmp = regexp_C.search( line )\n\n if tmp:\n raw_C = tmp.group( 1 )\n # logging.debug( 'raw C defines: %s' % raw_C )\n\n tmp = regexp_CPP.search( line )\n\n if tmp:\n raw_CPP = tmp.group( 1 )\n # logging.debug( 'raw CPP defines: %s' % raw_CPP )\n\n tmp = regexp_C_CFLAGS.search(line)\n\n if tmp:\n raw_C_CFLAGS = tmp.group(1)\n\n tmp = regexp_CPP_CFLAGS.search(line)\n\n if tmp:\n raw_CPP_CFLAGS = tmp.group(1)\n\n candidates = ( shlex.split( raw_C ) +\n shlex.split( raw_CPP ) +\n shlex.split( raw_C_CFLAGS ) +\n shlex.split( raw_CPP_CFLAGS ) )\n\n for candidate in candidates:\n if candidate.startswith( '-D' ):\n result += candidate + ' '\n\n return result\n\n\ndef getCDefinesAsList( targetPlatform, targetName ):\n \"\"\"\n Returns a list with all compiler definitions set for the\n package using the addDefinitions() directive.\n\n If no additional definitions are set, an empty list will be returned.\n\n NOTE: CMake supports that compiler definitions may be different for\n various target platforms, and even per executable and/or\n library. Therefore you need to specify both of them.\n A rule of thumb is targetName='-global'.\n \"\"\"\n Any.requireIsTextNonEmpty( targetPlatform )\n Any.requireIsTextNonEmpty( targetName )\n\n result = []\n regexp = re.compile( '-D\\s*(.*)' )\n\n for token in getCDefinesAsString( targetPlatform, targetName ).split():\n\n if token.startswith( '-D' ):\n tmp = regexp.search( token )\n item = (tmp.group(1)).strip()\n result.append( item )\n\n return frozenset(result)\n\n\ndef getHeaderAndLanguageMap( targetPlatform ):\n \"\"\"\n Returns a dictionary mapping header files to the set of language\n files that use it.\n \"\"\"\n platformBuildDir = os.path.join( 'build', targetPlatform )\n targetBuildDirsWildcard = os.path.join( platformBuildDir, 'CMakeFiles', '*.dir' )\n targetBuildDirs = glob.glob( targetBuildDirsWildcard )\n result = {}\n\n\n for buildDir in targetBuildDirs:\n\n try:\n result.update( _parseDependDotMake( buildDir, platformBuildDir ) )\n\n except IOError:\n # most likely the depend.make does not exist for this target,\n # this might happen if there are no dependencies by the target\n # or if this is a pseudo-target such as \"doc\" coming from\n # FindDoxygen.cmake\n logging.debug( 'ignoring target: %s', buildDir )\n\n return result\n\n\ndef _parseDependDotMake( targetBuildDir, platformBuildDir ):\n \"\"\"\n Returns a dictionary mapping header files to the set of language\n files that use it.\n\n The dictionary is obtained parsing the file\n build//CMakeFiles/.dir/depend.make\n \"\"\"\n Any.requireIsTextNonEmpty( targetBuildDir )\n Any.requireIsTextNonEmpty( platformBuildDir )\n\n dependDotMakePath = os.path.join( targetBuildDir, 'depend.make' )\n\n lines = FastScript.getFileContent( dependDotMakePath, splitLines=True )\n result = collections.defaultdict( set )\n\n languageNormalizationMap = {\n '.c' : 'c',\n '.C' : 'c++',\n '.CC' : 'c++',\n '.CPP': 'c++',\n '.CXX': 'c++',\n '.cc' : 'c++',\n '.cpp': 'c++',\n '.cxx': 'c++',\n }\n\n for l in lines:\n # skip comments and empty lines\n if Any.isTextNonEmpty( l ) and not l.startswith( '#' ):\n # lines are in the format\n # /path/to/obj/file.{c,cpp,cc,cxx}.o: /path/to/dependencyfile.{c,cpp,cc,cxx,h,hpp,hxx,hh}\n objFile, depFile = l.split( ':' )\n srcFile, objExt = os.path.splitext( objFile.strip( ) )\n srcName, srcExt = os.path.splitext( srcFile )\n depFile = depFile.strip( )\n _, depFileExt = os.path.splitext( depFile )\n language = languageNormalizationMap[ srcExt ]\n\n if depFileExt.lower( ) in ('.h', '.hxx', '.hpp', '.hh'):\n if not os.path.isabs( depFile ):\n relPath = os.path.join( platformBuildDir, depFile )\n absPath = os.path.abspath( relPath )\n else:\n absPath = depFile\n result[ absPath ].add( language )\n\n\n return result\n\n\n# EOF\n", "sub_path": "include/ToolBOSCore/Tools/CMake.py", "file_name": "CMake.py", "file_ext": "py", "file_size_in_byte": 13714, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "collections.namedtuple", "line_number": 59, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any.requireIsTextNonEmpty", "line_number": 78, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 78, "usage_type": "name"}, {"api_name": "ToolBOSCore.Util.Any.requireIsTextNonEmpty", "line_number": 79, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 79, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "ToolBOSCore.Util.Any.requireIsDirNonEmpty", "line_number": 84, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 84, "usage_type": "name"}, {"api_name": "ToolBOSCore.Util.Any.requireIsFileNonEmpty", "line_number": 85, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 85, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 88, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.FastScript.getFileContent", "line_number": 89, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.FastScript", "line_number": 89, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 92, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 93, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 109, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any.requireIsTextNonEmpty", "line_number": 132, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 132, "usage_type": "name"}, {"api_name": "ToolBOSCore.Util.Any.requireIsTextNonEmpty", "line_number": 133, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 133, "usage_type": "name"}, {"api_name": "ToolBOSCore.Util.Any.requireIsTextNonEmpty", "line_number": 164, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 164, "usage_type": "name"}, {"api_name": "ToolBOSCore.Util.Any.requireIsTextNonEmpty", "line_number": 165, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 165, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path", "line_number": 171, "usage_type": "attribute"}, {"api_name": "ToolBOSCore.Util.Any.requireIsDirNonEmpty", "line_number": 174, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 174, "usage_type": "name"}, {"api_name": "ToolBOSCore.Util.Any.requireIsFileNonEmpty", "line_number": 175, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 175, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 178, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.FastScript.getFileContent", "line_number": 179, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.FastScript", "line_number": 179, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 182, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 183, "usage_type": "call"}, {"api_name": "ToolBOSCore.BuildSystem.Compilers.getDefaultLanguageStandard", "line_number": 197, "usage_type": "call"}, {"api_name": "ToolBOSCore.BuildSystem.Compilers", "line_number": 197, "usage_type": "name"}, {"api_name": "shlex.split", "line_number": 203, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 208, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any.requireIsTextNonEmpty", "line_number": 231, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 231, "usage_type": "name"}, {"api_name": "ToolBOSCore.Util.Any.requireIsTextNonEmpty", "line_number": 232, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 232, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 234, "usage_type": "call"}, {"api_name": "os.path", "line_number": 234, "usage_type": "attribute"}, {"api_name": "ToolBOSCore.Util.Any.requireIsDirNonEmpty", "line_number": 237, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 237, "usage_type": "name"}, {"api_name": "ToolBOSCore.Util.Any.requireIsFileNonEmpty", "line_number": 238, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 238, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 241, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.FastScript.getFileContent", "line_number": 242, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.FastScript", "line_number": 242, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 247, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 248, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 249, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 250, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 276, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 277, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 278, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 279, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any.requireIsTextNonEmpty", "line_number": 300, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 300, "usage_type": "name"}, {"api_name": "ToolBOSCore.Util.Any.requireIsTextNonEmpty", "line_number": 301, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 301, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 304, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 321, "usage_type": "call"}, {"api_name": "os.path", "line_number": 321, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 322, "usage_type": "call"}, {"api_name": "os.path", "line_number": 322, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 323, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 337, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any.requireIsTextNonEmpty", "line_number": 350, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 350, "usage_type": "name"}, {"api_name": "ToolBOSCore.Util.Any.requireIsTextNonEmpty", "line_number": 351, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 351, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 353, "usage_type": "call"}, {"api_name": "os.path", "line_number": 353, "usage_type": "attribute"}, {"api_name": "ToolBOSCore.Util.FastScript.getFileContent", "line_number": 355, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.FastScript", "line_number": 355, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 356, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any.isTextNonEmpty", "line_number": 371, "usage_type": "call"}, {"api_name": "ToolBOSCore.Util.Any", "line_number": 371, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 375, "usage_type": "call"}, {"api_name": "os.path", "line_number": 375, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 376, "usage_type": "call"}, {"api_name": "os.path", "line_number": 376, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 378, "usage_type": "call"}, {"api_name": "os.path", "line_number": 378, "usage_type": "attribute"}, {"api_name": "os.path.isabs", "line_number": 382, "usage_type": "call"}, {"api_name": "os.path", "line_number": 382, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 383, "usage_type": "call"}, {"api_name": "os.path", "line_number": 383, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 384, "usage_type": "call"}, {"api_name": "os.path", "line_number": 384, "usage_type": "attribute"}]}
+{"seq_id": "456377726", "text": "from typing import Any, Dict, List, cast\n\nimport torch\nfrom torchdata.datapipes.iter import IterDataPipe, Mapper, CSVDictParser\nfrom torchvision.prototype.datasets.utils import (\n Dataset,\n DatasetConfig,\n DatasetInfo,\n OnlineResource,\n KaggleDownloadResource,\n)\nfrom torchvision.prototype.datasets.utils._internal import (\n hint_sharding,\n hint_shuffling,\n)\nfrom torchvision.prototype.features import Label, Image\n\n\nclass FER2013(Dataset):\n def _make_info(self) -> DatasetInfo:\n return DatasetInfo(\n \"fer2013\",\n homepage=\"https://www.kaggle.com/c/challenges-in-representation-learning-facial-expression-recognition-challenge\",\n categories=(\"angry\", \"disgust\", \"fear\", \"happy\", \"sad\", \"surprise\", \"neutral\"),\n valid_options=dict(split=(\"train\", \"test\")),\n )\n\n _CHECKSUMS = {\n \"train\": \"a2b7c9360cc0b38d21187e5eece01c2799fce5426cdeecf746889cc96cda2d10\",\n \"test\": \"dec8dfe8021e30cd6704b85ec813042b4a5d99d81cb55e023291a94104f575c3\",\n }\n\n def resources(self, config: DatasetConfig) -> List[OnlineResource]:\n archive = KaggleDownloadResource(\n cast(str, self.info.homepage),\n file_name=f\"{config.split}.csv.zip\",\n sha256=self._CHECKSUMS[config.split],\n )\n return [archive]\n\n def _prepare_sample(self, data: Dict[str, Any]) -> Dict[str, Any]:\n label_id = data.get(\"emotion\")\n\n return dict(\n image=Image(torch.tensor([int(idx) for idx in data[\"pixels\"].split()], dtype=torch.uint8).reshape(48, 48)),\n label=Label(int(label_id), categories=self.categories) if label_id is not None else None,\n )\n\n def _make_datapipe(\n self,\n resource_dps: List[IterDataPipe],\n *,\n config: DatasetConfig,\n ) -> IterDataPipe[Dict[str, Any]]:\n dp = resource_dps[0]\n dp = CSVDictParser(dp)\n dp = hint_sharding(dp)\n dp = hint_shuffling(dp)\n return Mapper(dp, self._prepare_sample)\n", "sub_path": "torchvision/prototype/datasets/_builtin/fer2013.py", "file_name": "fer2013.py", "file_ext": "py", "file_size_in_byte": 2025, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "torchvision.prototype.datasets.utils.Dataset", "line_number": 19, "usage_type": "name"}, {"api_name": "torchvision.prototype.datasets.utils.DatasetInfo", "line_number": 21, "usage_type": "call"}, {"api_name": "torchvision.prototype.datasets.utils.DatasetInfo", "line_number": 20, "usage_type": "name"}, {"api_name": "torchvision.prototype.datasets.utils.DatasetConfig", "line_number": 33, "usage_type": "name"}, {"api_name": "torchvision.prototype.datasets.utils.KaggleDownloadResource", "line_number": 34, "usage_type": "call"}, {"api_name": "typing.cast", "line_number": 35, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 33, "usage_type": "name"}, {"api_name": "torchvision.prototype.datasets.utils.OnlineResource", "line_number": 33, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 41, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 41, "usage_type": "name"}, {"api_name": "torchvision.prototype.features.Image", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.uint8", "line_number": 45, "usage_type": "attribute"}, {"api_name": "torchvision.prototype.features.Label", "line_number": 46, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 51, "usage_type": "name"}, {"api_name": "torchdata.datapipes.iter.IterDataPipe", "line_number": 51, "usage_type": "name"}, {"api_name": "torchvision.prototype.datasets.utils.DatasetConfig", "line_number": 53, "usage_type": "name"}, {"api_name": "torchdata.datapipes.iter.CSVDictParser", "line_number": 56, "usage_type": "call"}, {"api_name": "torchvision.prototype.datasets.utils._internal.hint_sharding", "line_number": 57, "usage_type": "call"}, {"api_name": "torchvision.prototype.datasets.utils._internal.hint_shuffling", "line_number": 58, "usage_type": "call"}, {"api_name": "torchdata.datapipes.iter.Mapper", "line_number": 59, "usage_type": "call"}, {"api_name": "torchdata.datapipes.iter.IterDataPipe", "line_number": 54, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 54, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 54, "usage_type": "name"}]}
+{"seq_id": "561296021", "text": "import cv2 as cv\nimport numpy as np\nimport argparse\nfrom matplotlib import pyplot as plt\n\nparser = argparse.ArgumentParser(description='Code for Creating Bounding boxes around lego pieces.')\nparser.add_argument('--input', help='Path to input image.', default='IMG_6101.JPG')\nargs = parser.parse_args()\n\nsrc = cv.imread(cv.samples.findFile(args.input))\nif src is None:\n print('Could not open or find the image:', args.input)\n exit(0)\n\n# Convert image to gray and blur it\nsrc_gray = cv.cvtColor(src, cv.COLOR_BGR2GRAY)\nsrc_gray = cv.blur(src_gray, (3,3))\n\n# Calculate contours\ncanny_threshold = 50\ncanny_output = cv.Canny(src_gray, canny_threshold, canny_threshold * 3)\n\ncontours, _ = cv.findContours(canny_output, cv.RETR_TREE, cv.CHAIN_APPROX_SIMPLE)\n\ncontours_poly = [None]*len(contours)\nboundRect = [None]*len(contours)\nfor i, c in enumerate(contours):\n contours_poly[i] = cv.approxPolyDP(c, 3, True)\n boundRect[i] = cv.boundingRect(contours_poly[i])\n\n# Draw the contours\ncontours_drawing = np.zeros((canny_output.shape[0], canny_output.shape[1], 3), dtype=np.uint8)\nfor i in range(len(contours)):\n color = (255, 255, 255)\n cv.drawContours(contours_drawing, contours_poly, i, color)\n\n# Draw the original with bounding rectangle\nrectangle_drawing = src.copy()\nfor i in range(len(contours)):\n color = (255, 0, 0)\n cv.rectangle(rectangle_drawing, (int(boundRect[i][0]), int(boundRect[i][1])), \\\n (int(boundRect[i][0]+boundRect[i][2]), int(boundRect[i][1]+boundRect[i][3])), color, 2)\n\n\nplt.subplot(231)\nplt.imshow(src)\nplt.title('Original Image')\nplt.xticks([]), plt.yticks([])\n\nplt.subplot(232)\nplt.imshow(src_gray, cmap = 'gray')\nplt.title('Greyed and blured')\nplt.xticks([]), plt.yticks([])\n\nplt.subplot(233)\nplt.imshow(canny_output)\nplt.title('Canny output')\nplt.xticks([]), plt.yticks([])\n\nplt.subplot(234)\nplt.imshow(contours_drawing, cmap = 'gray')\nplt.title('Contours')\nplt.xticks([]), plt.yticks([])\n\nplt.subplot(235)\nplt.imshow(rectangle_drawing)\nplt.title('Contours with rectangles')\nplt.xticks([]), plt.yticks([])\n\nplt.show()", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2067, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.samples.findFile", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.samples", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 16, "usage_type": "attribute"}, {"api_name": "cv2.blur", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.Canny", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.RETR_TREE", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cv2.approxPolyDP", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.boundingRect", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 32, "usage_type": "attribute"}, {"api_name": "cv2.drawContours", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}]}
+{"seq_id": "565047641", "text": "import telebot\r\nimport extensions\r\nfrom config import TOKEN, keys\r\n\r\nbot = telebot.TeleBot(TOKEN)\r\n\r\n\r\n@bot.message_handler(commands=['start', 'help'])\r\ndef help(message: telebot.types.Message):\r\n text = \"Чтобы начать работу введите команду бота в следующем формате: \" \\\r\n \"\\n<имя валюты> <в какую валюту перевести> <количество переводимой валюты>\" \\\r\n \"\\nУвидеть список всех доступных валют: /values\"\r\n bot.reply_to(message, text)\r\n\r\n\r\n@bot.message_handler(commands=['values'])\r\ndef values(message: telebot.types.Message):\r\n text = \"Доступные валюты:\"\r\n for key in keys.keys():\r\n text = '\\n'.join((text, key, ))\r\n bot.reply_to(message, text)\r\n\r\n\r\n@bot.message_handler(content_types=['text', ])\r\ndef convert(message: telebot.types.Message):\r\n try:\r\n check_parametrs = message.text.split(' ')\r\n if len(check_parametrs) < 3:\r\n raise Exception('Введено параметров меньше необходимого')\r\n if len(check_parametrs) > 3:\r\n raise Exception('Введено параметров больше необходимого')\r\n except Exception as e:\r\n text = e\r\n else:\r\n quote_k, base_k, amount_k = message.text.strip().lower().split(' ')\r\n text = extensions.RequestAPI.get_price(quote_k, base_k, amount_k)\r\n bot.send_message(message.chat.id, text)\r\n\r\n\r\nbot.polling()\r\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1561, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "telebot.TeleBot", "line_number": 5, "usage_type": "call"}, {"api_name": "config.TOKEN", "line_number": 5, "usage_type": "argument"}, {"api_name": "telebot.types", "line_number": 9, "usage_type": "attribute"}, {"api_name": "telebot.types", "line_number": 17, "usage_type": "attribute"}, {"api_name": "config.keys.keys", "line_number": 19, "usage_type": "call"}, {"api_name": "config.keys", "line_number": 19, "usage_type": "name"}, {"api_name": "telebot.types", "line_number": 25, "usage_type": "attribute"}, {"api_name": "extensions.RequestAPI.get_price", "line_number": 36, "usage_type": "call"}, {"api_name": "extensions.RequestAPI", "line_number": 36, "usage_type": "attribute"}]}
+{"seq_id": "514805546", "text": "#!/usr/bin/env python2\n\n##Here is a better example, an yes, the variable names suck. thats me\\.::\nimport pygame\nimport math\nimport sys \nimport random\n\nfrom pygame.time import Clock\nfrom pygame import Color, mouse\npygame.init()\n\ncircles = set()\n\ntspeed = 1\nspeed = tspeed\ncount = 0\nn_launchers = 3\n\ndef random_color(count):\n color = Color(\"white\")\n color.hsva = (count%360, 90, 80, 60)\n return color\n\nclass World(object):\n def __init__(self, x, y, x_accel=0, y_accel=0):\n self.x = x\n self.y = y\n self.center_x = self.x / 2\n self.center_y = self.y / 2\n self.mass = 2.0\n self.points = []\n self.lock = False\n self.color = random_color(60)\n\n def update(self):\n for x, y, signo in self.points:\n pygame.draw.circle(screen, self.color, (int(x), int(y)), int(10) )\n\n def calculate_gravity_for_particle(self, particle):\n\n for x, y, signo in self.points:\n if not particle.deleted:\n self._gravity_to_point(x, y, particle, signo)\n else:\n break\n\n def add_point(self):\n if not self.lock:\n x, y = pygame.mouse.get_pos()\n \n if pygame.mouse.get_pressed()[0]:\n signo = 1\n else:\n signo = -1\n\n self.points.append((x, y, signo))\n self.lock = True\n\n def unlock(self):\n self.lock = False\n\n\n def _gravity_to_point(self, x, y, particle, signo ):\n dx = particle.x - x\n dy = particle.y - y\n d = pow(dx, 2) + pow(dy, 2)\n ds = math.sqrt(d)\n force = self._calculate_gravity(d)\n if force is not None:\n force *= signo\n particle.x_accel += force * dx / ds\n particle.y_accel += force * dy / ds\n else:\n particle.delete()\n\n def _calculate_gravity(self, d):\n if d <= 10:\n force = None\n else:\n force = int( self.mass) / d\n #force = -1.0 * int( self.mass) / d\n return force\n\n\nclass Circle(object):\n def __init__(self, size, world, x=0, y=0, angle=0, speed=1, color = (0,0,0)):\n self.world = world\n self.color = color\n self.size = size/2\n self.speed_x = math.sin(angle) * speed\n self.speed_y = math.cos(angle) * speed\n self.x_accel = 0\n self.y_accel = 0\n self.x = x\n self.y = y #posicion inicial\n self.deleted=False\n\n def update(self, screen):\n x_size, y_size = screen.get_size()\n\n foo = self.world.calculate_gravity_for_particle(self)\n self.speed_x += self.x_accel\n self.speed_y += self.y_accel\n self.x += self.speed_x\n self.y += self.speed_y\n if not ( (0 < self.x and self.x < x_size) and (0 < self.y and self.y < y_size)):\n self.delete()\n return\n pygame.draw.circle(screen, self.color, (int(self.x), int(self.y)), int(self.size) )\n \n def delete(self):\n if not self.deleted:\n global circles\n circles.remove(self)\n self.deleted = True\n \n def __repr__(self):\n return \"\" % (self.x, self.y, self.angle)\n \n\n\nclass Launcher(object):\n \n def __init__(self, x, y, world, queue):\n self.x = x\n self.y = y\n self.world = world\n self.queue = queue\n self.cicles = 0\n self.rot_speed = 0\n self.mov_speed = 0\n self.q = 0\n self.rotation = 0\n self.counter = random.randint(0, 0xFFFFFF)\n self.color_change_speed = random.randint(1,3)\n\n def explode(self):\n for n in xrange(self.q):\n angle = ((n*360.0/self.q) + self.rotation) % 360\n color = random_color(self.counter)\n # self.counter += self.color_change_speed\n self.queue.add(Circle(4, self.world, x=self.x+random.randint(-20, 20), y=self.y, angle=angle, speed=self.mov_speed, color=color))\n \n def change(self):\n if not self.cicles:\n #self.rot_speed = random.uniform(-1, 1)\n self.mov_speed = 1 # random.uniform(1, 10)\n self.cicles = random.randint(500, 1000)\n self.q = 1 # random.randint(1, 6)\n self.rotation += (self.rot_speed % 360)\n self.cicles -= 1\n\n\nx , y = 800, 600 \nscreen = pygame.display.set_mode((x, y))\nclock = Clock()\nworld = World(x, y, 0, 0.1)\n\nlaunchers = []\n\nfor n in xrange(1, n_launchers + 1 ):\n launchers.append( Launcher(n * x/ (n_launchers+1), n * y /(n_launchers + 1), world, circles ))\n launchers.append( Launcher(n * x/ (n_launchers+1), n * y /(n_launchers + 1), world, circles ))\n launchers.append( Launcher(n * x/ (n_launchers+1), n * y /(n_launchers + 1), world, circles ))\n launchers.append( Launcher(n * x/ (n_launchers+1), n * y /(n_launchers + 1), world, circles ))\n launchers.append( Launcher(n * x/ (n_launchers+1), n * y /(n_launchers + 1), world, circles ))\n launchers.append( Launcher(n * x/ (n_launchers+1), n * y /(n_launchers + 1), world, circles ))\n\n\n\n\nwhile True:\n for event in pygame.event.get():\n if event.type == pygame.QUIT: \n sys.exit()\n elif event.type == pygame.MOUSEBUTTONDOWN:\n world.add_point()\n elif event.type == pygame.MOUSEBUTTONUP:\n world.unlock()\n\n screen.fill((0,0,0))\n circle_list = list(circles)\n world.update()\n for c in circle_list:\n c.update(screen)\n for launcher in launchers:\n launcher.change()\n if len(circles) < 5000: \n for launcher in launchers:\n launcher.explode()\n \n \n pygame.display.update()\n clock.tick(50)\n\n", "sub_path": "mouse_repulsor.py", "file_name": "mouse_repulsor.py", "file_ext": "py", "file_size_in_byte": 5681, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "pygame.init", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 21, "usage_type": "call"}, {"api_name": "pygame.draw.circle", "line_number": 38, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 50, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 50, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pressed", "line_number": 52, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 52, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 68, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 91, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 92, "usage_type": "call"}, {"api_name": "pygame.draw.circle", "line_number": 110, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 110, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 135, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 136, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 143, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 149, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 156, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 156, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 157, "usage_type": "call"}, {"api_name": "pygame.event.get", "line_number": 174, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 174, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 175, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 176, "usage_type": "call"}, {"api_name": "pygame.MOUSEBUTTONDOWN", "line_number": 177, "usage_type": "attribute"}, {"api_name": "pygame.MOUSEBUTTONUP", "line_number": 179, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 194, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 194, "usage_type": "attribute"}]}
+{"seq_id": "505886577", "text": "# My book is Dante's inferno (technically a poem)\nimport requests\nimport string\nimport random\nurl = 'https://www.gutenberg.org/files/1001/1001-h/1001-h.htm'\n\nfull_text = requests.get(url).text\n\nstart = 'Inferno: Canto I'\nend = 'End of the Project Gutenberg EBook'\n\n# Remove start and end material, newlines and punctuation\n\nfull_text = full_text[full_text.find(start):full_text.find(end)]\n\nfull_text = full_text.replace('\\n', ' ').replace('\\r', ' ')\nfor s in string.punctuation:\n full_text = full_text.replace(s, '')\n\nfull_text = full_text.split()\n\n\ndef generate_dict(txt):\n \"\"\"Create triplets dictionary in format (wd_one, wd_two):wd_three\"\"\"\n d = {}\n for idx in range(len(txt)-2):\n wd_one, wd_two, wd_three = txt[idx], txt[idx+1], txt[idx+2]\n d.setdefault((wd_one, wd_two), []).append(wd_three)\n return d\n\n\ndef generate_text(trigrams_dict, length):\n \"\"\"Generate text from trigrams dict with given number of words\"\"\"\n start_loc = random.randint(0, len(trigrams_dict))\n start_key = list(trigrams_dict.keys())[start_loc]\n results = [start_key[0], start_key[1]]\n for _ in range(length-2):\n next_word_choices = trigrams_dict[start_key]\n next_word = next_word_choices[random.randint(0,\n len(next_word_choices)-1)]\n start_key = (start_key[1], next_word)\n results.append(next_word)\n # Lines tend to be about 7 words long\n reshaped_results = []\n for i, j in enumerate(results):\n if i>0 and i % 7 == 0:\n reshaped_results.append('\\n')\n reshaped_results.append(j.title())\n else:\n reshaped_results.append(j.lower())\n return ' '.join(reshaped_results)\n\n\ndef print_chapter(length):\n \"\"\"Print chapter of inferno-like text with given length\"\"\"\n trigrams_dict = generate_dict(full_text)\n print(generate_text(trigrams_dict, length))\n\n\nif __name__ == '__main__':\n print_chapter(100)\n", "sub_path": "students/kuhnbt/lesson4/kata_script.py", "file_name": "kata_script.py", "file_ext": "py", "file_size_in_byte": 1977, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "requests.get", "line_number": 7, "usage_type": "call"}, {"api_name": "string.punctuation", "line_number": 17, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 34, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 39, "usage_type": "call"}]}
+{"seq_id": "359797222", "text": "from django.shortcuts import render, redirect\nfrom django.contrib import messages\nfrom .models import Idea\nfrom login_app.models import User\nfrom django.db.models import Count\n\ndef index(request):\n context = {\n 'ideas': Idea.objects.annotate(num_likes=Count('user_likes')).order_by('-num_likes'),\n }\n return render(request,'snack_index.html',context)\n\ndef create_idea(request):\n errors = Idea.objects.basic_validator(request.POST)\n if errors:\n for k, v in errors.items(): \n messages.error(request, v)\n return redirect('/snacks')\n else:\n user = User.objects.get(id = request.session['user_id'])\n idea = Idea.objects.create(\n title = request.POST['title'],\n description = request.POST['description'],\n user = user,\n )\n idea.user_likes.add(user)\n return redirect('/snacks')\n\ndef display_idea(request, idea_id):\n context = {\n 'snack': Idea.objects.get(id = idea_id),\n 'this_user': User.objects.get(id = request.session['user_id']),\n }\n return render(request, 'snack.html', context)\n\n\ndef like(request, idea_id):\n user = User.objects.get(id = request.session['user_id'])\n idea = Idea.objects.get(id = idea_id)\n idea.user_dislikes.remove(user)\n idea.user_likes.add(user)\n return redirect('/snacks')\n\ndef dislike(request, idea_id):\n user = User.objects.get(id = request.session['user_id'])\n idea = Idea.objects.get(id = idea_id)\n idea.user_likes.remove(user)\n idea.user_dislikes.add(user)\n return redirect('/snacks')\n\ndef delete(request, idea_id):\n idea = Idea.objects.get(id = idea_id)\n idea.delete()\n return redirect('/snacks')\n\n \n\n\n\n", "sub_path": "snack_app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1712, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "models.Idea.objects.annotate", "line_number": 9, "usage_type": "call"}, {"api_name": "models.Idea.objects", "line_number": 9, "usage_type": "attribute"}, {"api_name": "models.Idea", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.Count", "line_number": 9, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Idea.objects.basic_validator", "line_number": 14, "usage_type": "call"}, {"api_name": "models.Idea.objects", "line_number": 14, "usage_type": "attribute"}, {"api_name": "models.Idea", "line_number": 14, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 17, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 17, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 18, "usage_type": "call"}, {"api_name": "login_app.models.User.objects.get", "line_number": 20, "usage_type": "call"}, {"api_name": "login_app.models.User.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "login_app.models.User", "line_number": 20, "usage_type": "name"}, {"api_name": "models.Idea.objects.create", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Idea.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "models.Idea", "line_number": 21, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 27, "usage_type": "call"}, {"api_name": "models.Idea.objects.get", "line_number": 31, "usage_type": "call"}, {"api_name": "models.Idea.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "models.Idea", "line_number": 31, "usage_type": "name"}, {"api_name": "login_app.models.User.objects.get", "line_number": 32, "usage_type": "call"}, {"api_name": "login_app.models.User.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "login_app.models.User", "line_number": 32, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 34, "usage_type": "call"}, {"api_name": "login_app.models.User.objects.get", "line_number": 38, "usage_type": "call"}, {"api_name": "login_app.models.User.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "login_app.models.User", "line_number": 38, "usage_type": "name"}, {"api_name": "models.Idea.objects.get", "line_number": 39, "usage_type": "call"}, {"api_name": "models.Idea.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "models.Idea", "line_number": 39, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 42, "usage_type": "call"}, {"api_name": "login_app.models.User.objects.get", "line_number": 45, "usage_type": "call"}, {"api_name": "login_app.models.User.objects", "line_number": 45, "usage_type": "attribute"}, {"api_name": "login_app.models.User", "line_number": 45, "usage_type": "name"}, {"api_name": "models.Idea.objects.get", "line_number": 46, "usage_type": "call"}, {"api_name": "models.Idea.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "models.Idea", "line_number": 46, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 49, "usage_type": "call"}, {"api_name": "models.Idea.objects.get", "line_number": 52, "usage_type": "call"}, {"api_name": "models.Idea.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "models.Idea", "line_number": 52, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 54, "usage_type": "call"}]}
+{"seq_id": "617117439", "text": "import pymongo\nimport pymysql\nimport time\nimport xlwt\nimport oss2\nimport config\n\n\nclass ExportMarkOrQualitySettle(object):\n def __init__(self):\n \"\"\"初始化\"\"\"\n # 1. 创建mysql连接\n self.db_mysql = pymysql.connect(**config.mysql_config)\n self.cursor = self.db_mysql.cursor()\n\n # 2. 创建mongo连接\n dict_mongo = config.mongo_config\n self.db_mongo = pymongo.MongoClient(dict_mongo[\"address\"], dict_mongo[\"port\"])\n # 1) mongo认证过程,要使用认证库admin!\n self.db_auth = self.db_mongo.admin\n self.db_auth.authenticate(dict_mongo[\"user\"], dict_mongo[\"password\"])\n # 2) 连接业务库crowd\n self.db_client = config.mongo_config[\"database\"]\n\n # 3. 创建oss连接\n self.dict_oss = config.oss_config\n self.auth = oss2.Auth(self.dict_oss[\"accessKeyId\"], self.dict_oss[\"accessKeySecret\"])\n self.bucket = oss2.Bucket(self.auth, self.dict_oss[\"endpoint\"], self.dict_oss[\"bucketName\"])\n\n def close_databases(self):\n \"\"\"关闭数据库连接\"\"\"\n # 关闭mysql连接\n self.cursor.close()\n self.db_mysql.close()\n # 关闭mongo连接\n self.db_mongo.close()\n\n def export_settle(self, taskid, accTaskId):\n \"\"\"结算主方法: 导出、上传、返回\"\"\"\n # 一、 数据导出过程\n # 结算主方法sql:查询用户ID, 用户名称, 总条数, 合格总框数\n sql_count = \"\"\"\n SELECT\n t.tagUser,\n tt.userName,\n count(DISTINCT dataId) AS cnt,\n sum(qualifiedBoxCount) AS qualifiedBoxCount\n FROM\n (\n SELECT\n tagUser,\n dataId,\n tagCount,\n (case when unqualifiedBoxCount > 0 then qualifiedBoxCount else tagCount end) qualifiedBoxCount\n FROM\n `task_datadetail_b_\"\"\"+taskid+\"\"\"` AS t3\n WHERE\n checkFlag = '1'\n AND EXISTS (\n SELECT\n 1\n FROM\n task_accept_b AS t1\n INNER JOIN `task_acceptdetail_b_\"\"\"+accTaskId+\"\"\"` AS t2 ON t1.taskId = t2.taskId\n AND t1.acceptId = t2.acceptId\n WHERE\n 1 = 1\n AND t2.dataId = t3.dataId\n AND (\n (\n t1.accResult = '0'\n AND t2.acceptStatus != '4'\n )\n OR (\n t1.accResult != '0'\n AND t2.acceptStatus = '1'\n )\n )\n )\n ) AS t\n INNER JOIN user_info_m AS tt\n on t.tagUser = tt.userId\n GROUP BY\n tagUser, tt.userName\n \"\"\"\n self.cursor.execute(sql_count)\n count_info = self.cursor.fetchall()\n # 关闭数据库连接\n self.close_databases()\n # excel处理\n workbook = xlwt.Workbook()\n sheet = workbook.add_sheet('sheet1', cell_overwrite_ok=True)\n # 表头\n sheet.write(0, 0, '用户ID')\n sheet.write(0, 1, '用户名称')\n sheet.write(0, 2, '总条数')\n sheet.write(0, 3, '合格总框数')\n # 数据行\n rowIdx = 1\n for ci in count_info:\n sheet.write(rowIdx, 0, ci[0]) # 'tagUser'\n sheet.write(rowIdx, 1, ci[1]) # 'userName'\n sheet.write(rowIdx, 2, ci[2]) # 'cnt'\n sheet.write(rowIdx, 3, ci[3]) # 'qualifiedBoxCount'\n rowIdx = rowIdx+1\n # 定义时间戳:time获取当前时间戳\n now_time = int(time.time())\n time_array = time.localtime(now_time)\n normal_time = time.strftime(\"%Y-%m-%d_%H_%M_%S\", time_array)\n # 输出的excel文件名一定是整个业务流程中的标注任务的任务ID加上时间戳\n fileName = ('%s' % taskid) + '_' + normal_time + '.xls'\n filePath = config.path\n workbook.save(\"%s%s\" % (filePath, fileName))\n\n # 二、 文件上传过程\n self.bucket.put_object_from_file(self.dict_oss[\"prefx\"] + fileName, filePath + fileName)\n upload_url = self.bucket.sign_url('GET', self.dict_oss[\"prefx\"] + fileName, 86400)\n\n # 三、 返回json过程\n json_obj = {\n \"type\": \"FILE\",\n \"file\": {\n \"name\": fileName,\n \"url\": upload_url\n }\n }\n\n return json_obj\n\n def task_judge(self, task_id):\n \"\"\"判断任务关系\"\"\"\n # 1. 查询当前任务类别\n sql_task = \"\"\"\n SELECT\n ti.checkTaskFlg AS taskType,\n ti.isAccept\n FROM\n `task_info_b` AS ti\n WHERE\n ti.taskId = %s\n \"\"\"\n self.cursor.execute(sql_task, task_id)\n results = self.cursor.fetchall()\n if len(results) > 0:\n taskType = results[0][0]\n isAccept = results[0][1]\n\n # 2. 如果当前任务的类别是标注任务(taskType == '0')并且该任务需要验收, 要向下查, 下一个任务一定是自检任务\n # 可以确定业务流程是 \"标注 → 自检 → 验收\";\n if taskType == '0' and isAccept == '1':\n sql_dest = \"\"\"\n SELECT\n ti.taskId\n FROM\n `task_info_b` AS ti\n WHERE\n ti.checkTaskId = %s\n \"\"\"\n self.cursor.execute(sql_dest, task_id)\n result = self.cursor.fetchall()\n accept_taskId = result[0][0]\n\n # 调用结算主方法\n rectData = self.export_settle(task_id, accept_taskId)\n return rectData\n\n # 3. 如果当前任务不需要验收, 就需要一直向下查, 直到查询无结果为止\n if taskType == '0' and isAccept == '0':\n sql_dest = \"\"\"\n SELECT\n ti.checkTaskFlg AS taskType,\n ti.taskId\n FROM\n `task_info_b` AS ti\n WHERE\n ti.checkTaskId = %s\n \"\"\"\n self.cursor.execute(sql_dest, task_id)\n result = self.cursor.fetchall()\n taskType_check = result[0][0]\n taskId_check = result[0][1]\n # 如果下一个任务是自检任务, 并且此时不需要验收, 即此任务是标注的自检任务\n if taskType_check == '4':\n self.cursor.execute(sql_dest, taskId_check)\n result_marselfcheck = self.cursor.fetchall()\n taskId_quacheck = result_marselfcheck[0][1]\n # 此时一定有下一个任务即质检任务, 所以再向下查一次\n self.cursor.execute(sql_dest, taskId_quacheck)\n result_quaselfcheck = self.cursor.fetchall()\n # 判断如果查询结果为空, 则无下一个任务\n if len(result_quaselfcheck) == 0:\n accept_taskId = taskId_quacheck\n # 如果结果不为空, 则存在下一个任务即质检的自检任务\n else:\n taskId_quaselfcheck = result_quaselfcheck[0][1]\n accept_taskId = taskId_quaselfcheck\n # 如果下一个任务是质检任务, 则需要判断是否还存在质检的自检任务\n else:\n self.cursor.execute(sql_dest, taskId_check)\n result_quaselfcheck = self.cursor.fetchall()\n if len(result_quaselfcheck) == 0:\n accept_taskId = taskId_check\n else:\n taskId_quaselfcheck = result_quaselfcheck[0][1]\n accept_taskId = taskId_quaselfcheck\n\n # 调用结算主方法\n rectData = self.export_settle(task_id, accept_taskId)\n return rectData\n\n # 4. 如果当前任务的类别是质检任务(taskType == '1')并且该任务需要验收, 首先要向下查, 并判断下一个任务是否有自检任务,\n # 如果有自检任务, 则accept_taskId = 下一个任务ID, 如果没有, 则accept_taskId = 当前任务的ID;\n # 其次要向上查, 使用递归, 直到查询的checkTaskFlg的值为空时, 当前的任务ID即为标注任务的ID\n if taskType == '1' and isAccept == '1':\n # ***查询下一个任务(不一定存在, 需做判断是否还有自检任务)***\n sql_dest = \"\"\"\n SELECT\n ti.taskId\n FROM\n `task_info_b` AS ti\n WHERE\n ti.checkTaskId = %s\n \"\"\"\n self.cursor.execute(sql_dest, task_id)\n result = self.cursor.fetchall()\n if len(result) == 0:\n accept_taskId = task_id\n else:\n taskId_dest = result[0][0]\n self.cursor.execute(sql_dest, taskId_dest)\n result_dest = self.cursor.fetchall()\n if len(result_dest) == 0:\n accept_taskId = taskId_dest\n # ***查询上一个任务(一定存在, 需要判断上一个任务是自检任务还是标注任务)***\n sql_source = \"\"\"\n SELECT\n ti.checkTaskId,\n ti.checkTaskFlg AS taskType\n FROM\n `task_info_b` AS ti\n WHERE\n ti.taskId = %s\n \"\"\"\n self.cursor.execute(sql_source, task_id)\n result_source = self.cursor.fetchall()\n taskId_source = result_source[0][0]\n taskType_source = result_source[0][1]\n if taskType_source == '0':\n taskId_mark = taskId_source\n elif taskType_source == '1':\n taskId_mark = taskId_source\n else:\n self.cursor.execute(sql_source, taskId_source)\n result_mark = self.cursor.fetchall()\n taskId_mark = result_mark[0][0]\n\n # 调用结算主方法\n rectData = self.export_settle(taskId_mark, accept_taskId)\n return rectData\n\n # 5. 如果当前任务的类别是自检任务(taskType == '4'), 并且不需要验收, 可以确定当前任务是标注的自检任务, 需要:\n # 1) 向上查一次, 确定标注任务的ID\n if taskType == '4' and isAccept == '0':\n sql_source = \"\"\"\n SELECT\n ti.checkTaskId\n FROM\n `task_info_b` AS ti\n WHERE\n ti.taskId = %s\n \"\"\"\n self.cursor.execute(sql_source, task_id)\n result_source_mark = self.cursor.fetchall()\n taskId_source_mark = result_source_mark[0][0]\n # 2) 向下查一次, 一定存在质检任务\n sql_dest = \"\"\"\n SELECT\n ti.taskId\n FROM\n `task_info_b` AS ti\n WHERE\n ti.checkTaskId = %s\n \"\"\"\n self.cursor.execute(sql_dest, task_id)\n dest = self.cursor.fetchall()\n # 如果查询结果为空, 则证明是中途开启的自检, 需要使用标注任务的ID即'taskId_source_mark'去查询质检任务的任务ID\n if len(dest) == 0:\n self.cursor.execute(sql_dest, taskId_source_mark)\n checkId = self.cursor.fetchall()\n quacheckId = checkId[0][0]\n # 查询质检任务的下一个任务(不一定存在, 需做判断是否还有质检的自检任务)\n sql_dest = \"\"\"\n SELECT\n ti.taskId\n FROM\n `task_info_b` AS ti\n WHERE\n ti.checkTaskId = %s\n \"\"\"\n self.cursor.execute(sql_dest, quacheckId)\n selfId = self.cursor.fetchall()\n if len(selfId) == 0:\n accept_taskId = quacheckId\n else:\n quaselfId = selfId[0][0]\n accept_taskId = quaselfId\n # 如果查询不为空, 说明质检任务关联的是标注的自检任务, 不是中途开启的自检任务\n else:\n quacheckId = dest[0][0]\n # 查询质检任务的下一个任务(不一定存在, 需做判断是否还有质检的自检任务)\n sql_dest = \"\"\"\n SELECT\n ti.taskId\n FROM\n `task_info_b` AS ti\n WHERE\n ti.checkTaskId = %s\n \"\"\"\n self.cursor.execute(sql_dest, quacheckId)\n selfId = self.cursor.fetchall()\n if len(selfId) == 0:\n accept_taskId = quacheckId\n else:\n quaselfId = selfId[0][0]\n accept_taskId = quaselfId\n\n # 调用结算主方法\n rectData = self.export_settle(taskId_source_mark, accept_taskId)\n return rectData\n\n # 6. 如果当前任务是自检任务, 并且需要验收, 需要向上查一次, 查询上一个任务是标注任务还是质检任务\n if taskType == '4' and isAccept == '1':\n # 向上查一次, 得到上一个任务的任务ID\n sql_source = \"\"\"\n SELECT\n ti.checkTaskId\n FROM\n `task_info_b` AS ti\n WHERE\n ti.taskId = %s\n \"\"\"\n self.cursor.execute(sql_source, task_id)\n result = self.cursor.fetchall()\n taskId_source = result[0][0]\n # 使用上一个任务ID查询出上一个任务的任务类别\n sql_source_type = \"\"\"\n SELECT\n ti.checkTaskFlg\n FROM\n `task_info_b` AS ti\n WHERE\n ti.taskId = %s\n \"\"\"\n self.cursor.execute(sql_source_type, taskId_source)\n result_type = self.cursor.fetchall()\n source_type = result_type[0][0]\n # source_type的结果不是'0',就是'1', 即不是标注任务, 就是质检任务\n # 如果上一个任务是标注任务, 且现在需要验收, 则业务流程是 \"标注 → 自检 → 验收\"\n if source_type == '0':\n taskId_mark = taskId_source\n accept_taskId = task_id\n # 如果上一个任务是质检任务, 继续向上查, 并判断质检的上一个任务是标注任务还是标注的自检任务\n else:\n sql_mark = \"\"\"\n SELECT\n ti.checkTaskId\n FROM\n `task_info_b` AS ti\n WHERE\n ti.taskId = %s\n \"\"\"\n self.cursor.execute(sql_mark, taskId_source)\n result_isMark = self.cursor.fetchall()\n taskId_isMark = result_isMark[0][0]\n # 根据此ID, 查询此任务的类别\n sql_mark_type = \"\"\"\n SELECT\n ti.checkTaskFlg\n FROM\n `task_info_b` AS ti\n WHERE\n ti.taskId = %s\n \"\"\"\n self.cursor.execute(sql_mark_type, taskId_isMark)\n result_mark_type = self.cursor.fetchall()\n taskType_isMark = result_mark_type[0][0]\n # 如果质检任务的上一个任务是自检任务, 则需要再向上查一次\n if taskType_isMark == '4':\n self.cursor.execute(sql_mark, taskId_isMark)\n rect = self.cursor.fetchall()\n markId = rect[0][0]\n taskId_mark = markId\n # 如果质检任务的上一个任务是标注任务, 则结束查询\n else:\n taskId_mark = taskId_isMark\n accept_taskId = task_id\n\n # 调用结算主方法\n rectData = self.export_settle(taskId_mark, accept_taskId)\n return rectData\n\n else:\n json_obj = {\n \"type\": \"FILE\",\n \"file\": {\n \"name\": None,\n \"url\": None\n }\n }\n\n return json_obj\n", "sub_path": "restfulApi/markOrQualitySettle.py", "file_name": "markOrQualitySettle.py", "file_ext": "py", "file_size_in_byte": 18297, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "pymysql.connect", "line_number": 13, "usage_type": "call"}, {"api_name": "config.mysql_config", "line_number": 13, "usage_type": "attribute"}, {"api_name": "config.mongo_config", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pymongo.MongoClient", "line_number": 18, "usage_type": "call"}, {"api_name": "config.mongo_config", "line_number": 23, "usage_type": "attribute"}, {"api_name": "config.oss_config", "line_number": 26, "usage_type": "attribute"}, {"api_name": "oss2.Auth", "line_number": 27, "usage_type": "call"}, {"api_name": "oss2.Bucket", "line_number": 28, "usage_type": "call"}, {"api_name": "xlwt.Workbook", "line_number": 91, "usage_type": "call"}, {"api_name": "time.time", "line_number": 107, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 108, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 109, "usage_type": "call"}, {"api_name": "config.path", "line_number": 112, "usage_type": "attribute"}]}
+{"seq_id": "630994336", "text": "from tkinter import *\r\nimport matplotlib as plt\r\nimport numpy as np\r\nimport matplotlib.animation as animation\r\nfrom matplotlib.figure import Figure\r\nfrom matplotlib.backends.backend_tkagg import (\r\n FigureCanvasTkAgg, NavigationToolbar2Tk)\r\nimport threading\r\nimport os\r\nimport random\r\nimport time\r\n\r\nspath='data.txt'\r\nspath2='data2.txt'\r\n\r\n#a=2 # type: int\r\n\r\ndef load_params(fname,list,list_dtypes,list_values):\r\n if(os.path.isfile(fname)):\r\n f = open(fname, 'r')\r\n a = f.read()\r\n for i in range(list.__len__()):\r\n x=a.find(list[i])\r\n if(x!=-1):\r\n y=a.find(\":\",x)\r\n if(y!=-1):\r\n z=a.find(\"\\n\",y)\r\n if(z!=-1):\r\n list_values.append(list_dtypes[i](a[y+1:z]))\r\n\r\ndef save_params(fname,list,list_dtypes,list_values):\r\n if(os.path.isfile(fname)):\r\n os.remove(fname)\r\n f = open(fname, 'w+')\r\n if(list.__len__()==list_dtypes.__len__() and list_dtypes.__len__()==list_values.__len__()):\r\n for i in range(list.__len__()):\r\n s_value=list[i]+\":\"+str(list_values[i])+\"\\n\"\r\n f.write(s_value)\r\n\r\n\r\nclass app:\r\n\r\n def checkpath_thread(self):\r\n #print('thread start')\r\n while(True):\r\n if(self.new_path==True):\r\n if(os.path.isfile(self.s_indatapath)):\r\n self.ed_indatapath['bg']= 'green'\r\n else:\r\n self.ed_indatapath['bg'] = 'red'\r\n\r\n if(os.path.isfile(self.s_outdatapath)):\r\n self.ed_outdatapath['bg']= 'green'\r\n else:\r\n self.ed_outdatapath['bg'] = 'red'\r\n\r\n if(os.path.isfile(self.s_inputpath)):\r\n self.ed_inputpath['bg']= 'green'\r\n else:\r\n self.ed_inputpath['bg'] = 'red'\r\n\r\n self.new_path=False\r\n\r\n\r\n def press_button(self,event):\r\n fpath=self.dpath.get(1.0,END)\r\n print(fpath)\r\n fpath=fpath.rstrip()\r\n fpath=fpath.lstrip()\r\n try:\r\n a = np.genfromtxt(fpath)\r\n except:\r\n print(\"error\")\r\n else:\r\n print(a)\r\n self.lbl['text']=a\r\n #v = StringVar()\r\n #Label(self.lbl, textvariable=v).pack()\r\n #v.set(a)\r\n\r\n\r\n def on_changed(self,event):\r\n fpath=self.ed_indatapath.get(1.0, END)\r\n fpath=fpath.rstrip()\r\n fpath=fpath.lstrip()\r\n if(self.s_indatapath!=fpath):\r\n self.s_indatapath = fpath\r\n self.new_path=True\r\n\r\n fpath=self.ed_outdatapath.get(1.0, END)\r\n fpath=fpath.rstrip()\r\n fpath=fpath.lstrip()\r\n if(self.s_outdatapath!=fpath):\r\n self.s_outdatapath = fpath\r\n self.new_path=True\r\n\r\n fpath=self.ed_inputpath.get(1.0, END)\r\n fpath=fpath.rstrip()\r\n fpath=fpath.lstrip()\r\n if(self.s_inputpath!=fpath):\r\n self.s_inputpath = fpath\r\n self.new_path=True\r\n\r\n def add_data_thread(self):\r\n while True:\r\n self.tdata=np.append(self.tdata,random.randint(0,10))\r\n time.sleep(1)\r\n\r\n def draw_thread(self,i):\r\n self.trainingplot.clear()\r\n self.trainingplot.plot(self.tdata)\r\n\r\n\r\n def __init__(self):\r\n self.root = Tk()\r\n self.root.minsize(width=600,height=500)\r\n\r\n self.s_indatapath=''\r\n self.s_outdatapath=''\r\n self.s_inputpath=''\r\n\r\n self.new_path=True\r\n\r\n self.lbl_indatapath= Label(self.root,height=1,width=12,font='Arial 11',bg=\"white\", fg=\"black\",text='in_data fname :',anchor=W, justify=LEFT)\r\n self.lbl_outdatapath= Label(self.root,height=1,width=12,font='Arial 11',bg=\"white\", fg=\"black\",text='out_data fname:',anchor=W, justify=LEFT)\r\n self.lbl_inputpath= Label(self.root,height=1,width=12,font='Arial 11',bg=\"white\", fg=\"black\",text='input fname :',anchor=W, justify=LEFT)\r\n\r\n self.frm=Frame(self.root,bg='white',bd=5,height=200, width=300)\r\n\r\n #self.btn = Button(self.root, # родительское окно\r\n # text=\"Click me\", # надпись на кнопке\r\n # width=10, height=5, # ширина и высота\r\n # bg=\"white\", fg=\"black\") # цвет фона и надписи\r\n self.ed_indatapath = Text(self.root, height=1, width=15, font='Arial 11', wrap=WORD)\r\n self.ed_outdatapath = Text(self.root, height=1, width=15, font='Arial 11', wrap=WORD)\r\n self.ed_inputpath = Text(self.root, height=1, width=15, font='Arial 11', wrap=WORD)\r\n\r\n self.ed_indatapath.insert(1.0, 'in_data.txt')\r\n self.ed_outdatapath.insert(1.0, 'out_data.txt')\r\n self.ed_inputpath.insert(1.0, 'input.txt')\r\n\r\n self.ed_indatapath.bind('', self.on_changed)\r\n self.ed_outdatapath.bind('', self.on_changed)\r\n self.ed_inputpath.bind('', self.on_changed)\r\n self.ed_indatapath.bind('', self.on_changed)\r\n self.ed_outdatapath.bind('', self.on_changed)\r\n self.ed_inputpath.bind('', self.on_changed)\r\n\r\n self.ed_indatapath.place(x=120, y=10)\r\n self.ed_outdatapath.place(x=120, y=40)\r\n self.ed_inputpath.place(x=120, y=70)\r\n\r\n self.frm.place(x=250, y=10)\r\n #self.btn.place(x=10, y=5)\r\n self.lbl_indatapath.place(x=10, y=10)\r\n self.lbl_outdatapath.place(x=10, y=40)\r\n self.lbl_inputpath.place(x=10, y=70)\r\n\r\n#settings input\\output\r\n t = threading.Thread(target=self.checkpath_thread)\r\n t.daemon = True\r\n t.start()\r\n params=[\"xxc\",\"vvh\",\"asd\",\"qew\"]\r\n params_dtypes=[int,int,float,float]\r\n params_values=list()\r\n load_params(\"data.txt\",params,params_dtypes,params_values)\r\n save_params(\"data.txt\",params,params_dtypes,params_values)\r\n\r\n\r\n #plot\r\n self.fig = Figure(figsize=(5, 4), dpi=50)\r\n self.trainingplot=self.fig.add_subplot(111)\r\n\r\n self.tdata=np.array(5)\r\n for index in range(0,5):\r\n zzz=random.randint(0,10)\r\n self.tdata=np.append(self.tdata,zzz)\r\n\r\n\r\n self.canvas = FigureCanvasTkAgg(self.fig, master=self.frm) # A tk.DrawingArea.\r\n self.canvas.get_tk_widget().pack(expand=0)\r\n #self.canvas.show()\r\n\r\n dt = threading.Thread(target=self.add_data_thread)\r\n dt.daemon = True\r\n dt.start()\r\n\r\n self.ani=animation.FuncAnimation(self.fig,self.draw_thread,interval=1000)\r\n\r\n\r\n\r\n# t = np.arange(0, 3, .01)\r\n# self.trainingplot.plot(t)#, 2 * np.sin(2 * np.pi * t))\r\n# canvas = FigureCanvasTkAgg(self.fig, master=self.frm) # A tk.DrawingArea.\r\n# canvas.draw()\r\n# canvas.get_tk_widget().pack(expand=0)\r\n\r\n #run\r\n self.root.mainloop()\r\n #self.root.withdraw()\r\n\r\n\r\n\r\nz = app()\r\n#z.ani = animation.FuncAnimation(z.fig, drawthread, interval=1000)\r\n#z.root.mainloop()\r\n", "sub_path": "tk2.py", "file_name": "tk2.py", "file_ext": "py", "file_size_in_byte": 7035, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "os.path.isfile", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "numpy.genfromtxt", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 106, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 106, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 107, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.figure.Figure", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 174, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.backends.backend_tkagg.FigureCanvasTkAgg", "line_number": 180, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 188, "usage_type": "name"}]}
+{"seq_id": "622500572", "text": "import pandas as pd\nfrom keras.preprocessing.text import one_hot\n\n\nclass FeatureGenerator:\n @staticmethod\n def load_data(dataset: str, url_column_name=\"url\", label_column_name=\"label\", to_binarize=False,\n neg_word=\"bad\") -> tuple:\n \"\"\"\n Load given data file into self.urls and self.labels\n\n Parameters\n ----------\n dataset\n Path of csv file containing the dataset.\n url_column_name\n Name of the column containing the urls.\n label_column_name\n Name of the column containing the labels.\n to_binarize\n True if the label column is not already in binary form.\n neg_word\n Negative word in the label column. Only considered if 'to_binarize' is True.\n\n Returns\n -------\n tuple\n (list containing the urls, list containing the labels)\n \"\"\"\n\n def binarize_list(element: str) -> int:\n \"\"\"Binarize given element.\"\"\"\n if element == neg_word:\n return 1\n else:\n return 0\n\n dataframe: pd.DataFrame = pd.read_csv(dataset)\n urls = dataframe[url_column_name].tolist()\n labels = list(map(binarize_list, dataframe[label_column_name].tolist())) if to_binarize else dataframe[\n label_column_name].tolist()\n return urls, labels\n\n @staticmethod\n def one_hot_encoding(urls: list, vocab_size=87) -> list:\n \"\"\"Integer encode the documents\"\"\"\n encoded_docs = [one_hot(str(d), vocab_size) for d in urls]\n return encoded_docs\n", "sub_path": "feature_generator.py", "file_name": "feature_generator.py", "file_ext": "py", "file_size_in_byte": 1610, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "pandas.DataFrame", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 38, "usage_type": "call"}, {"api_name": "keras.preprocessing.text.one_hot", "line_number": 47, "usage_type": "call"}]}
+{"seq_id": "631446847", "text": "import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\n\n\nclass PPO():\n def __init__(self,\n actor_critic,\n clip_param,\n ppo_epoch,\n num_mini_batch,\n value_loss_coef,\n entropy_coef,\n lr=None,\n lr_beta=None,\n eps=None,\n N_recurrent=0,\n max_grad_norm=None,\n max_grad_norm_beta=None,\n delib_coef=0,\n weighted_loss=0,\n use_clipped_value_loss=True):\n\n self.actor_critic = actor_critic\n self.weighted_loss = weighted_loss\n self.clip_param = clip_param\n self.ppo_epoch = ppo_epoch\n self.num_mini_batch = num_mini_batch\n self.N_recurrent = N_recurrent\n self.value_loss_coef = value_loss_coef\n self.entropy_coef = entropy_coef\n\n self.max_grad_norm = max_grad_norm\n self.max_grad_norm_beta = max_grad_norm_beta\n self.use_clipped_value_loss = use_clipped_value_loss\n self.delib_coef = delib_coef\n\n self.param_beta = []\n self.param_list = []\n for name, param in actor_critic.named_parameters():\n if \"beta\" in name :\n self.param_beta.append(param)\n else:\n self.param_list.append(param)\n self.optimizer = optim.Adam(\n [{'params': self.param_list},{'params':self.param_beta,'lr':lr_beta}], lr, eps=eps)\n\n def update(self, rollouts):\n advantages = rollouts.returns[:-1] - rollouts.value_mixed[1:]\n advantages = (advantages - advantages.mean()) / (\n advantages.std() + 1e-5)\n\n value_loss_epoch = 0\n action_loss_epoch = 0\n dist_entropy_epoch = 0\n\n for e in range(self.ppo_epoch):\n if self.actor_critic.is_recurrent and self.N_recurrent == 0:\n data_generator = rollouts.recurrent_generator(\n advantages, self.num_mini_batch)\n else:\n data_generator = rollouts.feed_forward_generator(\n advantages, self.num_mini_batch)\n\n for sample in data_generator:\n obs_batch, recurrent_hidden_states_batch, actions_batch, \\\n value_preds_batch, return_batch, masks_batch, old_action_log_probs_batch, \\\n adv_targ,indices = sample\n\n # Reshape to do in a single forward pass for all steps\n values, action_log_probs, dist_entropy, _, mean_beta_v = self.actor_critic.evaluate_actions(\n rollouts.obs[:-1].view(-1, *rollouts.obs.size()[2:]),\n rollouts.recurrent_hidden_states[:-1].view(-1,rollouts.recurrent_hidden_states.size(-1)),\n rollouts.masks[:-1].view(-1, 1),\n rollouts.actions.view(-1, rollouts.actions.size(-1)),\n rollouts.value_mixed.view(-1,1),indices)\n\n #values, action_log_probs, dist_entropy, _, mean_beta_v = self.actor_critic.evaluate_actions(\n # rollouts.obs[:-1],\n # rollouts.recurrent_hidden_states[:-1],\n # rollouts.masks[:-1],\n # rollouts.actions,\n # rollouts.value_mixed,indices)\n ratio = torch.exp(action_log_probs -\n old_action_log_probs_batch)\n surr1 = ratio * adv_targ\n surr2 = torch.clamp(ratio, 1.0 - self.clip_param,\n 1.0 + self.clip_param) * adv_targ\n if self.weighted_loss:\n normalized_beta = ((mean_beta_v / mean_beta_v.sum()) * mean_beta_v.size()[0]).view(-1, 1)\n surr1 *= normalized_beta\n surr2 *= normalized_beta\n\n action_loss = (-torch.min(surr1, surr2)).mean()\n\n if self.use_clipped_value_loss:\n value_pred_clipped = value_preds_batch + \\\n (values - value_preds_batch).clamp(-self.clip_param, self.clip_param)\n value_losses = (values - return_batch).pow(2)\n\n value_losses_clipped = (\n value_pred_clipped - return_batch).pow(2)\n\n if self.weighted_loss:\n value_losses *= normalized_beta\n value_losses_clipped *= normalized_beta\n\n value_loss = 0.5 * torch.max(value_losses,\n value_losses_clipped).mean()\n else:\n value_loss = 0.5 * (return_batch - values).pow(2).mean()\n\n\n if self.delib_coef > 0:\n target_beta = torch.zeros_like(mean_beta_v).fill_(1)\n delib_loss = F.mse_loss(mean_beta_v, target_beta)*value_loss*self.value_loss_coef*self.delib_coef\n else:\n delib_loss = torch.zeros_like(value_loss)\n\n self.optimizer.zero_grad()\n (value_loss * self.value_loss_coef + action_loss -\n dist_entropy * self.entropy_coef).backward()\n nn.utils.clip_grad_norm_(self.actor_critic.parameters(),\n self.max_grad_norm)\n nn.utils.clip_grad_norm_(self.param_beta,self.max_grad_norm_beta)\n self.optimizer.step()\n\n\n value_loss_epoch += value_loss.item()\n action_loss_epoch += action_loss.item()\n dist_entropy_epoch += dist_entropy.item()\n\n num_updates = self.ppo_epoch * self.num_mini_batch\n\n value_loss_epoch /= num_updates\n action_loss_epoch /= num_updates\n dist_entropy_epoch /= num_updates\n\n return value_loss_epoch, action_loss_epoch, dist_entropy_epoch,delib_loss\n", "sub_path": "a2c_ppo_acktr/algo/ppo.py", "file_name": "ppo.py", "file_ext": "py", "file_size_in_byte": 5928, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "torch.optim.Adam", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.exp", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.clamp", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.min", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 116, "usage_type": "call"}, {"api_name": "torch.nn.functional.mse_loss", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 117, "usage_type": "name"}, {"api_name": "torch.zeros_like", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 124, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 124, "usage_type": "name"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 126, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 126, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 126, "usage_type": "name"}]}
+{"seq_id": "607871981", "text": "import argparse\nimport sys\nfrom sklearn.preprocessing import LabelEncoder\nimport pandas as pd\nfrom utils import success_, get_deadline_year, preprocess\nfrom Models import LGBMModel\nfrom sklearn.metrics import accuracy_score\nimport numpy\n\nclass Application:\n \"\"\"\n Application\n ==========\n This class is runner class for the entire application. It accepts certain command line parameters, then invokes the\n methods of the class you will create.\n\n Example run command:\n python application.py -f file_path_to_data\n \"\"\"\n\n def __init__(self, file_path):\n \"\"\"\n :param file_path (str): the path of the csv to read in\n \"\"\"\n \n self.file_path = file_path\n if self.file_path is None:\n raise Exception(\"Must specify file path to the data\")\n\n def _read_data(self, file_path):\n \"\"\"\n Perform any removal of columns of data here\n\n :param file_path (str): the path of the csv to read in\n :return (obj:`(pd.DataFrame, list, pd.DataFrame, list`): a dataframe for the train data, train labels, \n a dataframe for the test data, test labels\n \"\"\"\n\n train_size = .75\n data_frame = pd.read_csv(file_path)\n \n #Feature Engineering an extra column\n data_frame['success_probability'] = success_(data_frame['usd_goal_real'].values, data_frame['usd_pledged_real'].values)\n data_frame['deadline_year'] = data_frame['deadline'].apply(lambda x: get_deadline_year(x))\n \n # Splitting the data into train set and test set\n split_size = int(data_frame.shape[0] * train_size)\n train_data_frame = data_frame[:split_size]\n test_data_frame = data_frame[split_size:]\n\n # Getting the labels\n train_labels = train_data_frame['state']\n test_labels = test_data_frame['state']\n\n return train_data_frame, train_labels, test_data_frame, test_labels\n\n def run(self):\n train_data_frame, train_labels, test_data_frame, test_labels = self._read_data(self.file_path)\n\n ######\n #label_encoder for encoding target labels\n label_encoder = LabelEncoder()\n label_encoder.fit(train_data_frame['state'].values)\n \n #Preprocess the data\n train_data, train_labels, test_data, test_labels = preprocess(train_data_frame,\n train_labels,\n test_data_frame,\n test_labels,\n label_encoder)\n \n # Here you put the model you will be using from the class you created\n model = LGBMModel()\n \n #training the model\n model.train(train_data, train_labels)\n \n #predicting the values for the test data\n predictions = model.predict(test_data)\n \n #Getting the original names of the labels\n prediction_names = label_encoder.inverse_transform(predictions)\n \n #Saving the error value in a text file\n model._save_predictions(test_data_frame['ID'].values, test_data_frame['name'].values, prediction_names)\n \n #Printing the error value\n print('\\nAccuracy = {}'.format(accuracy_score(test_labels, predictions)))\n\n\n\"\"\"\nEntrance point for execution\n\"\"\"\nif __name__ == '__main__':\n parser = argparse.ArgumentParser(description='model trainer for starter project')\n parser.add_argument(\"-f\", \"--file_path\", help = \"where to read in the data\", default = None)\n args = parser.parse_args(sys.argv[1:])\n\n app = Application(\n file_path = args.file_path\n )\n app.run()\n", "sub_path": "application.py", "file_name": "application.py", "file_ext": "py", "file_size_in_byte": 3794, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "pandas.read_csv", "line_number": 40, "usage_type": "call"}, {"api_name": "utils.success_", "line_number": 43, "usage_type": "call"}, {"api_name": "utils.get_deadline_year", "line_number": 44, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 62, "usage_type": "call"}, {"api_name": "utils.preprocess", "line_number": 66, "usage_type": "call"}, {"api_name": "Models.LGBMModel", "line_number": 73, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 88, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 95, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 97, "usage_type": "attribute"}]}
+{"seq_id": "595844696", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\nimport json\nimport os.path\n\nPORT_NUMBER = None\n\nroot_dir = os.path.dirname(__file__) or '.'\n\nwith open('{}/configuration.json'.format(root_dir)) as data_file:\n data = json.load(data_file)\n PORT_NUMBER = data['port_number']\n", "sub_path": "data_server/read_configuration.py", "file_name": "read_configuration.py", "file_ext": "py", "file_size_in_byte": 276, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "os.path.path.dirname", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 8, "usage_type": "name"}, {"api_name": "json.load", "line_number": 11, "usage_type": "call"}]}
+{"seq_id": "400537234", "text": "import roslib; roslib.load_manifest('kinova_demo')\nimport rospy\nimport math\nimport actionlib\nimport kinova_msgs.msg\nfrom control.control_manager import ControlManager\n\n\nclass ArmControlManager(ControlManager):\n def __init__(self):\n self.initial_joint_degrees = [160, 270, 270, 270, 180, 0]\n self.position_control_address = '/j2n6s300_driver/joints_action/joint_angles'\n self.position_control_client = actionlib.SimpleActionClient(self.position_control_address,\n kinova_msgs.msg.ArmJointAnglesAction)\n self.position_control_client.wait_for_server()\n\n self.velocity_control_address = '/j2n6s300_driver/in/joint_velocity'\n self.velocity_control_publisher = rospy.Publisher(self.velocity_control_address,\n kinova_msgs.msg.JointVelocity, queue_size=10)\n\n def position_control(self, ac, info):\n \"\"\"\n :param ac: joint angle\n :param info: if ac is radian then True else false\n \"\"\"\n goal = kinova_msgs.msg.ArmJointAnglesGoal()\n\n if info:\n angle_set = self.convert_angle(ac, info)\n else:\n angle_set = ac\n\n goal.angles.joint1 = angle_set[0]\n goal.angles.joint2 = angle_set[1]\n goal.angles.joint3 = angle_set[2]\n goal.angles.joint4 = angle_set[3]\n goal.angles.joint5 = angle_set[4]\n goal.angles.joint6 = angle_set[5]\n\n self.position_control_client.send_goal(goal)\n if self.position_control_client.wait_for_result(rospy.Duration(20.0)):\n return self.position_control_client.get_result()\n else:\n print('the joint angle action timed-out')\n self.position_control_client.cancel_all_goals()\n return None\n\n def velocity_control(self, ac, info):\n \"\"\"\n :param ac: joint velocity control action\n :param info: if action is radian then True else False\n :return: None\n \"\"\"\n goal = kinova_msgs.msg.JointVelocity()\n\n if info:\n angle_set = self.convert_angle(ac, info)\n else:\n angle_set = ac\n\n goal.joint1 = angle_set[0]\n goal.joint2 = angle_set[1]\n goal.joint3 = angle_set[2]\n goal.joint4 = angle_set[3]\n goal.joint5 = angle_set[4]\n goal.joint6 = angle_set[5]\n\n self.velocity_control_publisher.publish(goal)\n\n def reset(self):\n \"\"\"\n set initial joint\n :return: None\n \"\"\"\n is_radian = False\n self.position_control(self.initial_joint_degrees, is_radian)", "sub_path": "control/arm_control_manager.py", "file_name": "arm_control_manager.py", "file_ext": "py", "file_size_in_byte": 2626, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "roslib.load_manifest", "line_number": 1, "usage_type": "call"}, {"api_name": "control.control_manager.ControlManager", "line_number": 9, "usage_type": "name"}, {"api_name": "actionlib.SimpleActionClient", "line_number": 13, "usage_type": "call"}, {"api_name": "kinova_msgs.msg.msg", "line_number": 14, "usage_type": "attribute"}, {"api_name": "kinova_msgs.msg", "line_number": 14, "usage_type": "name"}, {"api_name": "rospy.Publisher", "line_number": 18, "usage_type": "call"}, {"api_name": "kinova_msgs.msg.msg", "line_number": 19, "usage_type": "attribute"}, {"api_name": "kinova_msgs.msg", "line_number": 19, "usage_type": "name"}, {"api_name": "kinova_msgs.msg.msg.ArmJointAnglesGoal", "line_number": 26, "usage_type": "call"}, {"api_name": "kinova_msgs.msg.msg", "line_number": 26, "usage_type": "attribute"}, {"api_name": "kinova_msgs.msg", "line_number": 26, "usage_type": "name"}, {"api_name": "rospy.Duration", "line_number": 41, "usage_type": "call"}, {"api_name": "kinova_msgs.msg.msg.JointVelocity", "line_number": 54, "usage_type": "call"}, {"api_name": "kinova_msgs.msg.msg", "line_number": 54, "usage_type": "attribute"}, {"api_name": "kinova_msgs.msg", "line_number": 54, "usage_type": "name"}]}
+{"seq_id": "432498683", "text": "\"\"\"\nViews for the Environment objects.\nThis currently includes Wind and Tide objects.\n\"\"\"\nimport ujson\nimport logging\nimport zlib\nimport numpy as np\nfrom threading import current_thread\n\nfrom pyramid.settings import asbool\nfrom pyramid.response import Response\nfrom pyramid.view import view_config\nfrom pyramid.httpexceptions import HTTPNotFound, HTTPNotImplemented\n\nfrom gnome.environment.environment_objects import GridCurrent, GridWind\n\nfrom webgnome_api.common.views import (get_object,\n create_object,\n update_object,\n cors_policy,\n cors_response,\n cors_exception,\n process_upload,\n can_persist,\n switch_to_existing_session,\n activate_uploaded)\n\nfrom cornice import Service\n\nfrom ..common.session_management import (get_session_object,\n acquire_session_lock)\nlog = logging.getLogger(__name__)\n\nenv = Service(name='environment', path='/environment*obj_id',\n description=\"Environment API\",\n cors_policy=cors_policy,\n # accept='application/json+octet-stream',\n content_type=['application/json', 'binary'])\nimplemented_types = ('gnome.environment.Tide',\n 'gnome.environment.Wind',\n 'gnome.environment.Water',\n 'gnome.environment.Waves',\n 'gnome.environment.environment_objects.GridCurrent',\n 'gnome.environment.environment_objects.GridWind',\n )\n\n@env.get()\ndef get_environment(request):\n '''Returns an Gnome Environment object in JSON.'''\n content_requested = request.matchdict.get('obj_id')\n resp = Response(\n content_type='arraybuffer',\n content_encoding='deflate'\n )\n route = content_requested[1] if len(content_requested) > 1 else None\n if (len(content_requested) > 1):\n if route == 'grid':\n resp.body = get_grid(request)\n return cors_response(request, resp)\n if route == 'vectors':\n resp.body, dshape = get_vector_data(request)\n resp.headers.add('Access-Control-Expose-Headers', 'shape')\n resp.headers.add('shape', str(dshape))\n return cors_response(request, resp)\n if route == 'nodes':\n resp.body = get_nodes(request)\n return cors_response(request, resp)\n if route == 'centers':\n resp.body = get_centers(request)\n return cors_response(request, resp)\n if route == 'metadata':\n return get_metadata(request)\n else:\n return get_object(request, implemented_types)\n\n\n@env.post()\ndef create_environment(request):\n '''Creates an Environment object.'''\n return create_object(request, implemented_types)\n\n\n@env.put()\ndef update_environment(request):\n '''Updates an Environment object.'''\n return update_object(request, implemented_types)\n\n\n@view_config(route_name='environment_upload', request_method='OPTIONS')\ndef environment_upload_options(request):\n return cors_response(request, request.response)\n\n\n@view_config(route_name='environment_upload', request_method='POST')\ndef upload_environment(request):\n switch_to_existing_session(request)\n log_prefix = 'req({0}): upload_environment():'.format(id(request))\n log.info('>>{}'.format(log_prefix))\n\n\n file_list = request.POST['file_list']\n file_list = ujson.loads(file_list)\n name = request.POST['name']\n file_name = file_list[0]\n\n log.info(' {} file_name: {}, name: {}'\n .format(log_prefix, file_name, name))\n\n env_type = request.POST.get('obj_type', [])\n request.body = ujson.dumps({'obj_type': env_type,\n 'filename': file_name,\n 'name': name\n }).encode('utf-8')\n\n env_obj = create_environment(request)\n resp = Response(ujson.dumps(env_obj))\n\n log.info('<<{}'.format(log_prefix))\n return cors_response(request, resp)\n\n@view_config(route_name='environment_activate', request_method='OPTIONS')\ndef activate_environment_options(request):\n return cors_response(request, request.response)\n\n\n@view_config(route_name='environment_activate', request_method='POST')\n@can_persist\ndef activate_environment(request):\n '''\n Activate an environment file that has already been persistently\n uploaded.\n '''\n log_prefix = 'req({0}): activate_environment():'.format(id(request))\n log.info('>>{}'.format(log_prefix))\n\n file_name, name = activate_uploaded(request)\n resp = Response(ujson.dumps({'filename': file_name, 'name': name}))\n\n log.info('<<{}'.format(log_prefix))\n return cors_response(request, resp)\n\n\ndef get_grid(request):\n '''\n Outputs the object's grid cells in binary format\n '''\n log_prefix = 'req({0}): get_grid():'.format(id(request))\n log.info('>>' + log_prefix)\n\n session_lock = acquire_session_lock(request)\n log.info(' {} session lock acquired (sess:{}, thr_id: {})'\n .format(log_prefix, id(session_lock), current_thread().ident))\n try:\n obj_id = request.matchdict.get('obj_id')[0]\n obj = get_session_object(obj_id, request)\n\n if obj is not None and isinstance(obj, (GridCurrent, GridWind)):\n cells = obj.grid.get_cells()\n\n return zlib.compress(cells.astype(np.float32).tobytes())\n else:\n exc = cors_exception(request, HTTPNotFound)\n raise exc\n finally:\n session_lock.release()\n log.info(' {} session lock released (sess:{}, thr_id: {})'\n .format(log_prefix, id(session_lock), current_thread().ident))\n\n log.info('<<' + log_prefix)\n\n\ndef get_vector_data(request):\n log_prefix = 'req({0}): get_grid():'.format(id(request))\n log.info('>>' + log_prefix)\n\n session_lock = acquire_session_lock(request)\n log.info(' {} session lock acquired (sess:{}, thr_id: {})'\n .format(log_prefix, id(session_lock), current_thread().ident))\n try:\n obj_id = request.matchdict.get('obj_id')[0]\n obj = get_session_object(obj_id, request)\n\n if obj is not None and isinstance(obj, (GridCurrent, GridWind)):\n vec_data = obj.get_data_vectors()\n\n return zlib.compress(vec_data.tobytes()), vec_data.shape\n else:\n exc = cors_exception(request, HTTPNotFound)\n raise exc\n finally:\n session_lock.release()\n log.info(' {} session lock released (sess:{}, thr_id: {})'\n .format(log_prefix, id(session_lock), current_thread().ident))\n\n log.info('<<' + log_prefix)\n\n\ndef get_metadata(request):\n log_prefix = 'req({0}): get_current_info():'.format(id(request))\n log.info('>>' + log_prefix)\n\n session_lock = acquire_session_lock(request)\n log.info(' {} session lock acquired (sess:{}, thr_id: {})'\n .format(log_prefix, id(session_lock), current_thread().ident))\n try:\n obj_id = request.matchdict.get('obj_id')[0]\n obj = get_session_object(obj_id, request)\n if obj is not None:\n return obj.get_metadata()\n else:\n exc = cors_exception(request, HTTPNotFound)\n raise exc\n finally:\n session_lock.release()\n log.info(' {} session lock released (sess:{}, thr_id: {})'\n .format(log_prefix, id(session_lock), current_thread().ident))\n\n log.info('<<' + log_prefix)\n\n\ndef get_grid_signature(obj):\n '''\n Here we are trying to get an n-dimensional signature of our\n grid data.\n There may be a better way to do this, but for now we will get the\n euclidian distances between all sequential points and sum them.\n '''\n points = obj.get_points()\n\n dtype = points[0].dtype.descr\n raw_points = points.view(dtype=' 1 and smoothed_loss > bloding_scale * best_loss:\n print('exited with best_loss at {}'.format(best_loss))\n plt.plot(log_lrs[10:-5], losses[10:-5])\n return log_lrs, losses\n # Record the best loss\n if smoothed_loss < best_loss or batch_num == 1:\n best_loss = smoothed_loss\n # Store the values\n losses.append(smoothed_loss)\n log_lrs.append(math.log10(lr))\n self.writer.add_scalar('log_lr', math.log10(lr), batch_num)\n # Do the SGD step\n # Update the lr for the next step\n\n loss.backward()\n self.optimizer.step()\n\n lr *= mult\n for params in self.optimizer.param_groups:\n params['lr'] = lr\n if batch_num > num:\n plt.plot(log_lrs[10:-5], losses[10:-5])\n return log_lrs, losses\n\n def train(self, conf, epochs):\n self.model.train()\n running_loss = 0.\n for e in range(epochs):\n if conf.train:\n print('epoch {} started'.format(e))\n if e == self.milestones[0]:\n self.schedule_lr()\n if e == self.milestones[1]:\n self.schedule_lr()\n if e == self.milestones[2]:\n self.schedule_lr()\n # for imgs, labels in tqdm(iter(self.loader)):\n for imgs, labels in self.loader:\n imgs = imgs.to(conf.device)\n labels = labels.to(conf.device)\n self.optimizer.zero_grad()\n embeddings = self.model(imgs)\n thetas = self.head(embeddings, labels)\n loss = conf.ce_loss(thetas, labels)\n loss.backward()\n running_loss += loss.item()\n self.optimizer.step()\n\n if self.step % self.board_loss_every == 0 and self.step != 0:\n loss_board = running_loss / self.board_loss_every\n self.writer.add_scalar('train_loss', loss_board, self.step)\n running_loss = 0.\n\n grid = torchvision.utils.make_grid(imgs[:65])\n grid = denormalize_image(grid)\n self.writer.add_image('train_images', grid, self.step, dataformats='HWC')\n print(\"epoch: {}, step: {}/{}, loss: {}\".format(e, self.step, len(self.loader), loss_board))\n # if self.step % self.evaluate_every == 0 and self.step != 0:\n # if conf.data_mode == 'common':\n # for val_name in self.val_dataloaders:\n # val_dataloader, val_issame = self.val_dataloaders[val_name]\n # accuracy, best_threshold, roc_curve_tensor = self.evaluate_by_dataloader(conf,\n # val_dataloader,\n # val_issame)\n # self.board_val(val_name, accuracy, best_threshold, roc_curve_tensor)\n # else:\n # accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.agedb_30,\n # self.agedb_30_issame)\n # self.board_val('agedb_30', accuracy, best_threshold, roc_curve_tensor)\n # accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.lfw, self.lfw_issame)\n # self.board_val('lfw', accuracy, best_threshold, roc_curve_tensor)\n # accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.cfp_fp,\n # self.cfp_fp_issame)\n # self.board_val('cfp_fp', accuracy, best_threshold, roc_curve_tensor)\n # self.model.train()\n # if self.step % self.save_every == 0 and self.step != 0:\n # self.save_state(conf, accuracy)\n\n self.step += 1\n\n accuracies = []\n if conf.data_mode == 'common':\n for val_name in self.val_dataloaders:\n val_dataloader, val_issame = self.val_dataloaders[val_name]\n accuracy, best_threshold, roc_curve_tensor = self.evaluate_by_dataloader(conf,\n val_dataloader,\n val_issame)\n accuracies.append(accuracy)\n self.board_val(val_name, accuracy, best_threshold, roc_curve_tensor, len(val_issame))\n else:\n accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.agedb_30,\n self.agedb_30_issame)\n self.board_val('agedb_30', accuracy, best_threshold, roc_curve_tensor)\n accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.lfw, self.lfw_issame)\n self.board_val('lfw', accuracy, best_threshold, roc_curve_tensor)\n accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.cfp_fp,\n self.cfp_fp_issame)\n self.board_val('cfp_fp', accuracy, best_threshold, roc_curve_tensor)\n self.model.train()\n\n if not conf.train:\n break\n self.save_state(conf, sum(accuracies) / len(accuracies))\n\n if conf.train:\n self.save_state(conf, accuracy, to_save_folder=True, extra='final')\n\n def train_landmark(self, conf, epochs):\n self.model.train()\n running_loss = 0.\n for e in range(epochs):\n if conf.train:\n print('epoch {} started'.format(e))\n if e == self.milestones[0]:\n self.schedule_lr()\n if e == self.milestones[1]:\n self.schedule_lr()\n if e == self.milestones[2]:\n self.schedule_lr()\n # for imgs, labels in tqdm(iter(self.loader)):\n for imgs, labels in self.loader:\n imgs = imgs.to(conf.device)\n labels = labels.to(conf.device)\n self.optimizer.zero_grad()\n embeddings, output = self.model(imgs)\n loss_prev = conf.ce_loss(output, labels)\n thetas = self.head(embeddings, labels)\n loss = conf.ce_loss(thetas, labels)\n loss += loss_prev\n loss.backward()\n running_loss += loss.item()\n self.optimizer.step()\n\n if self.step % self.board_loss_every == 0 and self.step != 0:\n loss_board = running_loss / self.board_loss_every\n self.writer.add_scalar('train_loss', loss_board, self.step)\n running_loss = 0.\n\n grid = torchvision.utils.make_grid(imgs[:65])\n grid = denormalize_image(grid)\n self.writer.add_image('train_images', grid, self.step, dataformats='HWC')\n print(\"epoch: {}, step: {}/{}, loss: {}\".format(e, self.step, len(self.loader), loss_board))\n # if self.step % self.evaluate_every == 0 and self.step != 0:\n # if conf.data_mode == 'common':\n # for val_name in self.val_dataloaders:\n # val_dataloader, val_issame = self.val_dataloaders[val_name]\n # accuracy, best_threshold, roc_curve_tensor = self.evaluate_by_dataloader(conf,\n # val_dataloader,\n # val_issame)\n # self.board_val(val_name, accuracy, best_threshold, roc_curve_tensor)\n # else:\n # accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.agedb_30,\n # self.agedb_30_issame)\n # self.board_val('agedb_30', accuracy, best_threshold, roc_curve_tensor)\n # accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.lfw, self.lfw_issame)\n # self.board_val('lfw', accuracy, best_threshold, roc_curve_tensor)\n # accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.cfp_fp,\n # self.cfp_fp_issame)\n # self.board_val('cfp_fp', accuracy, best_threshold, roc_curve_tensor)\n # self.model.train()\n # if self.step % self.save_every == 0 and self.step != 0:\n # self.save_state(conf, accuracy)\n\n self.step += 1\n\n accuracies = []\n if conf.data_mode == 'common':\n for val_name in self.val_dataloaders:\n val_dataloader, val_issame = self.val_dataloaders[val_name]\n accuracy, best_threshold, roc_curve_tensor = self.evaluate_by_dataloader(conf,\n val_dataloader,\n val_issame)\n accuracies.append(accuracy)\n self.board_val(val_name, accuracy, best_threshold, roc_curve_tensor, len(val_issame))\n else:\n accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.agedb_30,\n self.agedb_30_issame)\n self.board_val('agedb_30', accuracy, best_threshold, roc_curve_tensor)\n accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.lfw, self.lfw_issame)\n self.board_val('lfw', accuracy, best_threshold, roc_curve_tensor)\n accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.cfp_fp,\n self.cfp_fp_issame)\n self.board_val('cfp_fp', accuracy, best_threshold, roc_curve_tensor)\n self.model.train()\n\n if not conf.train:\n break\n self.save_state(conf, sum(accuracies) / len(accuracies))\n\n if conf.train:\n self.save_state(conf, accuracy, to_save_folder=True, extra='final')\n\n def train_font(self, conf, epochs):\n self.model.train()\n running_loss = 0.\n for e in range(epochs):\n if conf.train:\n print('epoch {} started'.format(e))\n if e == self.milestones[0]:\n self.schedule_lr()\n if e == self.milestones[1]:\n self.schedule_lr()\n if e == self.milestones[2]:\n self.schedule_lr()\n # for imgs, labels in tqdm(iter(self.loader)):\n for imgs, labels in self.loader:\n imgs = imgs.to(conf.device)\n labels = labels.to(conf.device)\n self.optimizer.zero_grad()\n embeddings = self.model(imgs)\n print(embeddings.shape)\n thetas = self.head(embeddings, labels)\n print(thetas.shape)\n loss = conf.ce_loss(thetas, labels)\n loss.backward()\n running_loss += loss.item()\n self.optimizer.step()\n\n if self.step % self.board_loss_every == 0 and self.step != 0:\n loss_board = running_loss / self.board_loss_every\n self.writer.add_scalar('train_loss', loss_board, self.step)\n running_loss = 0.\n\n grid = torchvision.utils.make_grid(imgs[:65])\n grid = denormalize_image(grid)\n self.writer.add_image('train_images', grid, self.step, dataformats='HWC')\n print(\"epoch: {}, step: {}, loss: {}\".format(e, self.step, loss_board))\n # if self.step % self.evaluate_every == 0 and self.step != 0:\n # if conf.data_mode == 'common':\n # for val_name in self.val_dataloaders:\n # val_dataloader, val_issame = self.val_dataloaders[val_name]\n # accuracy, best_threshold, roc_curve_tensor = self.evaluate_by_dataloader(conf,\n # val_dataloader,\n # val_issame)\n # self.board_val(val_name, accuracy, best_threshold, roc_curve_tensor)\n # else:\n # accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.agedb_30,\n # self.agedb_30_issame)\n # self.board_val('agedb_30', accuracy, best_threshold, roc_curve_tensor)\n # accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.lfw, self.lfw_issame)\n # self.board_val('lfw', accuracy, best_threshold, roc_curve_tensor)\n # accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.cfp_fp,\n # self.cfp_fp_issame)\n # self.board_val('cfp_fp', accuracy, best_threshold, roc_curve_tensor)\n # self.model.train()\n # if self.step % self.save_every == 0 and self.step != 0:\n # self.save_state(conf, accuracy)\n\n self.step += 1\n\n accuracies = []\n if conf.data_mode == 'common':\n for val_name in self.val_dataloaders:\n val_dataloader, val_issame = self.val_dataloaders[val_name]\n accuracy, best_threshold, roc_curve_tensor = self.evaluate_by_dataloader(conf,\n val_dataloader,\n val_issame)\n accuracies.append(accuracy)\n self.board_val(val_name, accuracy, best_threshold, roc_curve_tensor, len(val_issame))\n else:\n accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.agedb_30,\n self.agedb_30_issame)\n self.board_val('agedb_30', accuracy, best_threshold, roc_curve_tensor)\n accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.lfw, self.lfw_issame)\n self.board_val('lfw', accuracy, best_threshold, roc_curve_tensor)\n accuracy, best_threshold, roc_curve_tensor = self.evaluate(conf, self.cfp_fp,\n self.cfp_fp_issame)\n self.board_val('cfp_fp', accuracy, best_threshold, roc_curve_tensor)\n self.model.train()\n\n if not conf.train:\n break\n\n self.save_state(conf, sum(accuracies) / len(accuracies))\n\n if conf.train:\n self.save_state(conf, accuracy, to_save_folder=True, extra='final')\n\n def schedule_lr(self):\n for params in self.optimizer.param_groups:\n params['lr'] /= 10\n print(self.optimizer)\n\n def infer(self, conf, faces, target_embs, tta=False):\n '''\n faces : list of PIL Image\n target_embs : [n, 512] computed embeddings of faces in facebank\n names : recorded names of faces in facebank\n tta : test time augmentation (hfilp, that's all)\n '''\n embs = []\n for img in faces:\n if tta:\n mirror = trans.functional.hflip(img)\n emb = self.model(conf.test_transform(img).to(conf.device).unsqueeze(0))\n emb_mirror = self.model(conf.test_transform(mirror).to(conf.device).unsqueeze(0))\n embs.append(l2_norm(emb + emb_mirror))\n else:\n embs.append(self.model(conf.test_transform(img).to(conf.device).unsqueeze(0)))\n source_embs = torch.cat(embs)\n\n diff = source_embs.unsqueeze(-1) - target_embs.transpose(1, 0).unsqueeze(0)\n dist = torch.sum(torch.pow(diff, 2), dim=1)\n minimum, min_idx = torch.min(dist, dim=1)\n min_idx[minimum > self.threshold] = -1 # if no match, set idx to -1\n return min_idx, minimum\n", "sub_path": "Learner.py", "file_name": "Learner.py", "file_ext": "py", "file_size_in_byte": 33574, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "matplotlib.pyplot.switch_backend", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 31, "usage_type": "attribute"}, {"api_name": "numpy.reciprocal", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 33, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 40, "usage_type": "attribute"}, {"api_name": "model.MobileFaceNet", "line_number": 48, "usage_type": "call"}, {"api_name": "data.data_pipe.get_train_loader", "line_number": 54, "usage_type": "call"}, {"api_name": "model.Backbone", "line_number": 60, "usage_type": "call"}, {"api_name": "model.MetricNet", "line_number": 63, "usage_type": "call"}, {"api_name": "model.Arcface", "line_number": 73, "usage_type": "call"}, {"api_name": "metric_learning.ArcMarginProduct", "line_number": 76, "usage_type": "call"}, {"api_name": "metric_learning.AddMarginProduct", "line_number": 80, "usage_type": "call"}, {"api_name": "metric_learning.AdaCos", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 86, "usage_type": "name"}, {"api_name": "tensorboardX.SummaryWriter", "line_number": 96, "usage_type": "call"}, {"api_name": "utils.separate_bn_paras", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.optim.SGD", "line_number": 128, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 128, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 130, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 144, "usage_type": "call"}, {"api_name": "data.data_pipe.get_common_val_data", "line_number": 148, "usage_type": "call"}, {"api_name": "data.data_pipe.get_val_data", "line_number": 163, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 175, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 176, "usage_type": "call"}, {"api_name": "utils.get_time", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 181, "usage_type": "call"}, {"api_name": "utils.get_time", "line_number": 183, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 186, "usage_type": "call"}, {"api_name": "utils.get_time", "line_number": 188, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 196, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 198, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 199, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 217, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 218, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 220, "usage_type": "call"}, {"api_name": "utils.hflip_batch", "line_number": 222, "usage_type": "call"}, {"api_name": "model.l2_norm", "line_number": 224, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 229, "usage_type": "call"}, {"api_name": "utils.hflip_batch", "line_number": 231, "usage_type": "call"}, {"api_name": "model.l2_norm", "line_number": 233, "usage_type": "call"}, {"api_name": "verifacation.evaluate", "line_number": 236, "usage_type": "call"}, {"api_name": "utils.gen_plot", "line_number": 237, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 238, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 238, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 239, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 239, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 245, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 248, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 250, "usage_type": "call"}, {"api_name": "torchvision.utils.make_grid", "line_number": 257, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 257, "usage_type": "attribute"}, {"api_name": "verifacation.evaluate", "line_number": 272, "usage_type": "call"}, {"api_name": "verifacation.evaluate", "line_number": 275, "usage_type": "call"}, {"api_name": "utils.gen_plot", "line_number": 277, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 278, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 278, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 279, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 279, "usage_type": "name"}, {"api_name": "tqdm.tqdm", "line_number": 301, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 321, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 321, "usage_type": "name"}, {"api_name": "math.log10", "line_number": 328, "usage_type": "call"}, {"api_name": "math.log10", "line_number": 329, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 340, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 340, "usage_type": "name"}, {"api_name": "torchvision.utils.make_grid", "line_number": 372, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 372, "usage_type": "attribute"}, {"api_name": "torchvision.utils.make_grid", "line_number": 457, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 457, "usage_type": "attribute"}, {"api_name": "torchvision.utils.make_grid", "line_number": 542, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 542, "usage_type": "attribute"}, {"api_name": "torchvision.transforms.functional.hflip", "line_number": 612, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional", "line_number": 612, "usage_type": "attribute"}, {"api_name": "torchvision.transforms", "line_number": 612, "usage_type": "name"}, {"api_name": "model.l2_norm", "line_number": 615, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 618, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 621, "usage_type": "call"}, {"api_name": "torch.pow", "line_number": 621, "usage_type": "call"}, {"api_name": "torch.min", "line_number": 622, "usage_type": "call"}]}
+{"seq_id": "135535413", "text": "#!/usr/bin/env python\r\n# -*- coding: iso-8859-1 -*-\r\n\r\n#\r\n# Copylefth (c) 2009, Grudejo:\r\n# Aline Grazielle Silva Reis\r\n# Julia Carmona Almeida Chaves\r\n# Luziany Maria de Oliveira\r\n# Joyce Karoline Dare\r\n# Prof. Douglas Machado Tavares\r\n#\r\n\r\nimport pygame, sys\r\nfrom pygame.constants import *\r\n\r\n\r\nclass Jogo:\r\n \"\"\" Classe Jogo \"\"\"\r\n\r\n def __init__(self):\r\n \"\"\" Construtor: __init__() -> instancia de jogo \"\"\"\r\n pygame.init()\r\n self.tela = pygame.display.set_mode((800, 600), DOUBLEBUF)\r\n\r\n\r\n def criar_atores(self):\r\n \"\"\" Cria os atores \"\"\"\r\n self.paola = pygame.image.load(\"paola.png\")\r\n self.x_paola, self.y_paola = 0, 100\r\n\r\n\r\n def atualizar_atores(self):\r\n \"\"\" Atualiza os atores \"\"\"\r\n ret_tela = self.tela.get_rect()\r\n ret_paola = self.paola.get_rect()\r\n if (self.x_paola < ret_tela.width - ret_paola.width):\r\n self.x_paola += 6\r\n\r\n\r\n def repintar_tela(self):\r\n \"\"\" Repinta a tela \"\"\"\r\n self.tela.fill((0, 0, 0))\r\n self.tela.blit(self.paola, (self.x_paola, self.y_paola))\r\n pygame.display.flip()\r\n\r\n\r\n def tratar_eventos_teclado(self, evento):\r\n \"\"\" Observa e trata os eventos \"\"\"\r\n tecla = evento.key\r\n if ((tecla == K_ESCAPE) or (tecla == K_q)):\r\n pygame.display.quit()\r\n sys.exit()\r\n\r\n\r\n def tratar_eventos(self):\r\n \"\"\" Observa e trata os eventos \"\"\"\r\n for evento in pygame.event.get():\r\n if (evento.type == QUIT):\r\n pygame.display.quit()\r\n sys.exit()\r\n if (evento.type == KEYDOWN):\r\n self.tratar_eventos_teclado(evento)\r\n\r\n\r\n def rodar(self):\r\n \"\"\" Roda o jogo \"\"\"\r\n self.criar_atores()\r\n while (True):\r\n self.tratar_eventos()\r\n self.atualizar_atores()\r\n self.repintar_tela()\r\n\r\n\r\nif (__name__ == \"__main__\"):\r\n jogo = Jogo()\r\n jogo.rodar()\r\n", "sub_path": "src/etapa_02/jogo_v03.py", "file_name": "jogo_v03.py", "file_ext": "py", "file_size_in_byte": 1986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "pygame.init", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 28, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pygame.display.quit", "line_number": 51, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 51, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 52, "usage_type": "call"}, {"api_name": "pygame.event.get", "line_number": 57, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 57, "usage_type": "attribute"}, {"api_name": "pygame.display.quit", "line_number": 59, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 59, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 60, "usage_type": "call"}]}
+{"seq_id": "287631870", "text": "import smtplib\nfrom email import encoders\nfrom email.mime.base import MIMEBase\nfrom email.mime.text import MIMEText\nfrom email.mime.image import MIMEImage\nfrom email.mime.message import MIMEMessage\nfrom email.mime.multipart import MIMEMultipart\nfrom email.utils import formatdate\nimport ssl\nimport magic\nimport os\n\nimport info\n\ndef create_message(from_addr, to_addr, cc_addrs, bcc_addrs, subject):\n msg = MIMEMultipart()\n msg['Subject'] = subject\n msg['From'] = from_addr\n msg['To'] = to_addr\n msg['Bcc'] = bcc_addrs\n msg['Cc'] = cc_addrs\n msg['Date'] = formatdate()\n return msg\n\ndef attach_file(msg, attach_dir):\n for file_name in os.listdir(attach_dir):\n path = os.path.join(attach_dir, file_name)\n with open(path, 'rb') as fp:\n types = get_mimetypes(path)\n attachment = MIMEBase(types['maintype'], types['subtype'])\n attachment.set_payload(fp.read())\n encoders.encode_base64(attachment)\n attachment.add_header(\n 'Content-Disposition', 'attachment',\n filename = file_name\n )\n msg.attach(attachment)\n\ndef add_text(msg, body):\n msg.attach(MIMEText(body))\n\ndef send_mail(from_addr, to_addrs, password, msg):\n sender = smtplib.SMTP_SSL('smtp.gmail.com', 465)\n sender.login(from_addr, password)\n sender.sendmail(from_addr, to_addrs, msg.as_string())\n sender.quit()\n\ndef get_mimetypes(path):\n m = magic.from_file(path, mime=True).split('/')\n types = dict(maintype = m[0], subtype = m[1])\n return types\n\n\n", "sub_path": "client/send_mail.py", "file_name": "send_mail.py", "file_ext": "py", "file_size_in_byte": 1572, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "email.mime.multipart.MIMEMultipart", "line_number": 16, "usage_type": "call"}, {"api_name": "email.utils.formatdate", "line_number": 22, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "email.mime.base.MIMEBase", "line_number": 30, "usage_type": "call"}, {"api_name": "email.encoders.encode_base64", "line_number": 32, "usage_type": "call"}, {"api_name": "email.encoders", "line_number": 32, "usage_type": "name"}, {"api_name": "email.mime.text.MIMEText", "line_number": 40, "usage_type": "call"}, {"api_name": "smtplib.SMTP_SSL", "line_number": 43, "usage_type": "call"}, {"api_name": "magic.from_file", "line_number": 49, "usage_type": "call"}]}
+{"seq_id": "246032405", "text": "import torch\nfrom torch import nn\nfrom torch.nn import functional as F\n\n\n# Copyright 2018 The Sonnet Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ============================================================================\n\n\n# Borrowed from https://github.com/deepmind/sonnet and ported it to PyTorch\n\n\nclass Quantize(nn.Module):\n def __init__(self, dim, n_embed, decay=0.99, eps=1e-5):\n super().__init__()\n\n self.dim = dim\n self.n_embed = n_embed\n self.decay = decay\n self.eps = eps\n\n embed = torch.randn(dim, n_embed)\n self.register_buffer('embed', embed)\n self.register_buffer('cluster_size', torch.zeros(n_embed))\n self.register_buffer('embed_avg', embed.clone())\n\n def forward(self, input):\n flatten = input.reshape(-1, self.dim)\n dist = (\n flatten.pow(2).sum(1, keepdim=True)\n - 2 * flatten @ self.embed\n + self.embed.pow(2).sum(0, keepdim=True)\n )\n _, embed_ind = (-dist).max(1)\n embed_onehot = F.one_hot(embed_ind, self.n_embed).type(flatten.dtype)\n embed_ind = embed_ind.view(*input.shape[:-1])\n quantize = self.embed_code(embed_ind)\n\n if self.training:\n self.cluster_size.data.mul_(self.decay).add_(\n 1 - self.decay, embed_onehot.sum(0)\n )\n embed_sum = flatten.transpose(0, 1) @ embed_onehot\n self.embed_avg.data.mul_(self.decay).add_(1 - self.decay, embed_sum)\n n = self.cluster_size.sum()\n cluster_size = (\n (self.cluster_size + self.eps) / (n + self.n_embed * self.eps) * n\n )\n embed_normalized = self.embed_avg / cluster_size.unsqueeze(0)\n self.embed.data.copy_(embed_normalized)\n\n diff = (quantize.detach() - input).pow(2).mean()\n quantize = input + (quantize - input).detach()\n\n return quantize, diff, embed_ind\n\n def embed_code(self, embed_id):\n return F.embedding(embed_id, self.embed.transpose(0, 1))", "sub_path": "code/networks/Quantize.py", "file_name": "Quantize.py", "file_ext": "py", "file_size_in_byte": 2570, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "torch.nn.Module", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn.functional.one_hot", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 47, "usage_type": "name"}, {"api_name": "torch.nn.functional.embedding", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 70, "usage_type": "name"}]}
+{"seq_id": "185645314", "text": "from common.datatypes import Int32, Int8\n\n\ndef xor_encrypt_game(func):\n def xor(data, key):\n temp = Int32(0)\n\n for i in range(len(data)):\n temp2 = Int32(data[i] & 0xff)\n data[i] = Int8(temp2 ^ key[i & 15] ^ temp)\n temp = data[i]\n\n old = Int32(key[8] & 0xff)\n old |= Int32(key[9] << 0x08) & 0xff00\n old |= Int32(key[10] << 0x10) & 0xff0000\n old |= Int32(key[11] << 0x18) & 0xff000000\n\n old += Int32(len(data))\n\n key[8:12] = old\n\n return data\n\n def wrap(packet, client, *args, **kwargs):\n if client.encryption_enabled:\n result = xor(func(packet, client, *args, **kwargs), client.xor_key.outgoing_key)\n else:\n result = func(packet, client, *args, **kwargs)\n return result\n return wrap\n\n\ndef xor_decrypt_game(func):\n def dexor(data, key):\n temp1 = Int32(0)\n for i in range(len(data)):\n temp2 = Int32(data[i]) & 0xff\n data[i] = Int8(temp2 ^ key[i & 15] ^ temp1)\n temp1 = temp2\n\n old = (Int32(key[8]) & 0xff)\n old |= (Int32(key[9]) << 0x08) & 0xff00\n old |= (Int32(key[10]) << 0x10) & 0xff0000\n old |= (Int32(key[11]) << 0x18) & 0xff000000\n\n old += Int32(len(data))\n\n key[8:12] = old\n return data\n\n def wrap(packet_cls, data, client, *args, **kwargs):\n if client.encryption_enabled:\n decrypted = dexor(data, client.xor_key.incoming_key)\n return func(packet_cls, decrypted, client, *args, **kwargs)\n else:\n return func(packet_cls, data, client, *args, **kwargs)\n return wrap", "sub_path": "gameserver/crypt/xor.py", "file_name": "xor.py", "file_ext": "py", "file_size_in_byte": 1667, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "common.datatypes.Int32", "line_number": 6, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 9, "usage_type": "call"}, {"api_name": "common.datatypes.Int8", "line_number": 10, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 13, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 14, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 15, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 16, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 18, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 35, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 37, "usage_type": "call"}, {"api_name": "common.datatypes.Int8", "line_number": 38, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 41, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 42, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 43, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 44, "usage_type": "call"}, {"api_name": "common.datatypes.Int32", "line_number": 46, "usage_type": "call"}]}
+{"seq_id": "158884561", "text": "\"\"\"\n Test the toolbar class\n\"\"\"\nfrom django.contrib.auth.models import User\n\nfrom wheelcms_axle.node import Node\nfrom wheelcms_axle.content import type_registry\nfrom wheelcms_axle.toolbar import Toolbar\nfrom wheelcms_axle.tests.models import Type1, Type1Type, Type2Type, TestTypeRegistry\n\nfrom wheelcms_axle.spoke import Spoke\n\nfrom .test_handler import superuser_request\nfrom twotest.util import create_request\n\nclass BaseToolbarTest(object):\n def setup(self):\n self.registry = TestTypeRegistry()\n type_registry.set(self.registry)\n self.registry.register(Type1Type)\n self.registry.register(Type2Type)\n\n\nclass TestToolbar(BaseToolbarTest):\n \"\"\"\n Test toolbar child restrictions, context buttons\n \"\"\"\n def allchildren(self, children):\n \"\"\" match against all registered children \"\"\"\n return set(x['name'] for x in children) == \\\n set(c.name() for c in type_registry.values())\n\n def test_unconnected(self, client):\n \"\"\"\n Test behaviour on unconnected node - allow\n creation of all types of sub content\n \"\"\"\n root = Node.root()\n toolbar = Toolbar(root, superuser_request(\"/\"), \"view\")\n assert toolbar.show_create()\n assert self.allchildren(toolbar.children())\n\n def test_connected_no_restrictions(self, client):\n \"\"\"\n A node with content without restrictions\n \"\"\"\n root = Node.root()\n content = Type1(node=root)\n content.save()\n toolbar = Toolbar(root, superuser_request(\"/\"), \"view\")\n assert toolbar.show_create()\n assert self.allchildren(toolbar.children())\n\n def test_restriction_type(self, client):\n \"\"\"\n A single childtype allowed\n \"\"\"\n registry = self.registry\n\n class DummyNode(object):\n def content(self):\n class DummyContent(object):\n meta_type = 'dummycontent'\n\n @classmethod\n def get_name(cls):\n return \"test.\" + cls.meta_type\n\n class DummyType(Spoke):\n model = DummyContent\n children = (Type1Type,)\n\n @classmethod\n def name(self):\n return DummyContent.get_name()\n\n registry.register(DummyType)\n\n return DummyContent()\n\n toolbar = Toolbar(DummyNode(), superuser_request(\"/\"), \"view\")\n children = toolbar.children()\n assert len(children) == 1\n assert children[0]['name'] == Type1Type.name()\n assert children[0]['title'] == Type1Type.title\n assert children[0]['icon_path'] == Type1Type.full_type_icon_path()\n\n def test_restriction_none(self, client):\n \"\"\"\n No children at all allowed\n \"\"\"\n registry = self.registry\n\n class DummyNode(object):\n def content(self):\n class DummyContent(object):\n meta_type = 'dummycontent'\n\n @classmethod\n def get_name(cls):\n return \"test.\" + cls.meta_type\n\n class DummyType(Spoke):\n model = DummyContent\n children = ()\n\n @classmethod\n def name(self):\n return DummyContent.get_name()\n\n registry.register(DummyType)\n\n return DummyContent()\n\n toolbar = Toolbar(DummyNode(), superuser_request(\"/\"), \"view\")\n assert toolbar.children() == []\n assert not toolbar.show_create()\n\n def test_create_mode_buttons(self, client):\n \"\"\" verify that certain buttons are not shown in create mode \"\"\"\n node = Node.root()\n content = Type1(node=node)\n content.save()\n toolbar = Toolbar(node, superuser_request(\"/\"), \"create\")\n assert not toolbar.show_create()\n assert not toolbar.show_update()\n\n def test_update_mode_buttons(self, client):\n \"\"\" verify that certain buttons are not shown in update mode \"\"\"\n node = Node.root()\n content = Type1(node=node)\n content.save()\n toolbar = Toolbar(node, superuser_request(\"/\"), \"update\")\n assert not toolbar.show_create()\n assert not toolbar.show_update()\n\n def test_no_implicit_unattached(self, client):\n \"\"\" An unattached node cannot restrict its children but\n should still not allow creation of non-implicit_add\n types \"\"\"\n\n class DummyContent(object):\n meta_type = 'dummycontent'\n\n @classmethod\n def get_name(cls):\n return \"test.\" + cls.meta_type\n\n class DummyType(Spoke):\n model = DummyContent\n children = ()\n implicit_add = False\n\n @classmethod\n def title(cls):\n return ''\n\n self.registry.register(DummyType)\n\n\n node = Node.root()\n toolbar = Toolbar(node, superuser_request(\"/\"), \"view\")\n for c in toolbar.children():\n assert c['name'] != DummyType.name()\n\n def test_anon_no_settings(self, client):\n node = Node.root()\n toolbar = Toolbar(node, create_request(\"GET\", \"/\"), \"view\")\n assert not toolbar.show_settings()\n\n def test_nosu_no_settings(self, client):\n user, _ = User.objects.get_or_create(username=\"user\")\n request = create_request(\"GET\", \"/\")\n request.user = user\n\n node = Node.root()\n toolbar = Toolbar(node, request, \"view\")\n assert not toolbar.show_settings()\n\n def test_primary(self, client):\n \"\"\" a type with primary should behave differently \"\"\"\n\n registry = self.registry\n\n class DummyNode(object):\n def content(self):\n class DummyContent(object):\n meta_type = 'dummycontent'\n\n @classmethod\n def get_name(cls):\n return \"test.\" + cls.meta_type\n\n class DummyType(Spoke):\n model = DummyContent\n children = (Type1Type, Type2Type)\n primary = Type1Type\n\n @classmethod\n def name(self):\n return DummyContent.get_name()\n\n registry.register(DummyType)\n\n return DummyContent()\n\n toolbar = Toolbar(DummyNode(), superuser_request(\"/\"), \"view\")\n children = toolbar.children()\n assert len(children) == 1\n assert children[0]['name'] == Type2Type.name()\n assert children[0]['title'] == Type2Type.title\n assert children[0]['icon_path'] == Type2Type.full_type_icon_path()\n\n primary = toolbar.primary()\n assert primary\n assert primary['name'] == Type1Type.name()\n assert primary['title'] == Type1Type.title\n assert primary['icon_path'] == Type1Type.full_type_icon_path()\n\n def test_primary_unattached(self, client):\n \"\"\" an unattached node has no primary content \"\"\"\n toolbar = Toolbar(Node.root(), superuser_request(\"/\"), \"view\")\n assert toolbar.primary() is None\n\n def test_clipboard_empty(self, client):\n toolbar = Toolbar(Node.root(), superuser_request(\"/\"), \"view\")\n clipboard = toolbar.clipboard()\n assert clipboard['count'] == 0\n assert not clipboard['copy']\n assert not clipboard['cut']\n assert clipboard['items'] == []\n\n def test_clipboard_cut(self, client):\n root = Node.root()\n\n t1 = Type1(node=root.add(\"t1\"), title=\"t1\").save()\n t2 = Type1(node=root.add(\"t2\"), title=\"t2\").save()\n\n request = create_request(\"GET\", \"/\")\n request.session['clipboard_cut'] = [t2.node.tree_path, t1.node.tree_path]\n\n toolbar = Toolbar(Node.root(), request, \"view\")\n clipboard = toolbar.clipboard()\n assert clipboard['count'] == 2\n assert not clipboard['copy']\n assert clipboard['cut']\n assert set(clipboard['items']) == set((t1, t2))\n\n def test_clipboard_copy(self, client):\n root = Node.root()\n\n t1 = Type1(node=root.add(\"t1\"), title=\"t1\").save()\n t2 = Type1(node=root.add(\"t2\"), title=\"t2\").save()\n\n request = create_request(\"GET\", \"/\")\n request.session['clipboard_copy'] = [t2.node.tree_path, t1.node.tree_path]\n\n toolbar = Toolbar(Node.root(), request, \"view\")\n clipboard = toolbar.clipboard()\n assert clipboard['count'] == 2\n assert clipboard['copy']\n assert not clipboard['cut']\n assert set(clipboard['items']) == set((t1, t2))\n\nfrom django.utils import translation\nfrom django.conf import settings\n\nclass TestTranslations(BaseToolbarTest):\n def setup(self):\n super(TestTranslations, self).setup()\n\n settings.CONTENT_LANGUAGES = (('en', 'English'), ('nl', 'Nederlands'), ('fr', 'Francais'))\n settings.FALLBACK = False\n\n def test_show_translate(self, client):\n root = Node.root()\n\n n = root.add(langslugs=dict(en=\"en\", nl=\"nl\", fr=\"fr\"))\n t_nl = Type1(node=n, language=\"nl\", title=\"NL\").save()\n translation.activate(\"en\")\n request = create_request(\"GET\", \"/\")\n toolbar = Toolbar(n, request, \"view\")\n\n assert toolbar.show_translate()\n assert not toolbar.show_update()\n translation.activate(\"nl\")\n assert not toolbar.show_translate()\n assert toolbar.show_update()\n\n\n def test_translations_view(self, client):\n root = Node.root()\n\n n = root.add(langslugs=dict(en=\"en\", nl=\"nl\", fr=\"fr\"))\n t_nl = Type1(node=n, language=\"nl\", title=\"NL\").save()\n t_en = Type1(node=n, language=\"en\", title=\"EN\").save()\n\n\n request = create_request(\"GET\", \"/\")\n\n translation.activate(\"en\")\n toolbar = Toolbar(n, request, \"view\")\n translations = toolbar.translations()\n\n assert translations['active']['id'] == 'en'\n\n ## Do some matching magic using endswith to work around language / base prefixing.\n ## We're mosly interested in create/view/edit actions anyway\n assert translations['translated'][0]['id'] == \"nl\"\n assert translations['translated'][0]['action_url'].endswith('switch_admin_language?path='+n.tree_path + '&language=nl')\n assert translations['untranslated'][0]['id'] == 'fr'\n assert translations['untranslated'][0]['action_url'].endswith('switch_admin_language?path='+n.tree_path + '&language=fr')\n\n def test_translations_edit(self, client):\n root = Node.root()\n\n n = root.add(langslugs=dict(en=\"en\", nl=\"nl\", fr=\"fr\"))\n t_nl = Type1(node=n, language=\"nl\", title=\"NL\").save()\n t_en = Type1(node=n, language=\"en\", title=\"EN\").save()\n\n\n request = create_request(\"GET\", \"/\")\n\n translation.activate(\"en\")\n toolbar = Toolbar(n, request, \"update\")\n translations = toolbar.translations()\n\n assert translations['active']['id'] == 'en'\n\n ## Do some matching magic using endswith to work around language / base prefixing.\n ## We're mosly interested in create/view/edit actions anyway\n assert translations['translated'][0]['id'] == \"nl\"\n assert translations['translated'][0]['action_url'].endswith('switch_admin_language?path='+n.tree_path + '&language=nl&rest=edit')\n assert translations['untranslated'][0]['id'] == 'fr'\n assert translations['untranslated'][0]['action_url'].endswith('switch_admin_language?path='+n.tree_path + '&language=fr&rest=edit')\n\n def test_translations_list(self, client):\n root = Node.root()\n\n n = root.add(langslugs=dict(en=\"en\", nl=\"nl\", fr=\"fr\"))\n t_nl = Type1(node=n, language=\"nl\", title=\"NL\").save()\n t_en = Type1(node=n, language=\"en\", title=\"EN\").save()\n\n\n request = create_request(\"GET\", \"/\")\n\n translation.activate(\"en\")\n toolbar = Toolbar(n, request, \"list\")\n translations = toolbar.translations()\n\n assert translations['active']['id'] == 'en'\n\n ## Do some matching magic using endswith to work around language / base prefixing.\n ## We're mosly interested in create/view/edit actions anyway\n assert translations['translated'][0]['id'] == \"nl\"\n assert translations['translated'][0]['action_url'].endswith('switch_admin_language?path='+n.tree_path + '&language=nl&rest=list')\n assert translations['untranslated'][0]['id'] == 'fr'\n assert translations['untranslated'][0]['action_url'].endswith('switch_admin_language?path='+n.tree_path + '&language=fr&rest=list')\n\n def test_translations_create(self, client):\n root = Node.root()\n\n n = root.add(langslugs=dict(en=\"en\", nl=\"nl\", fr=\"fr\"))\n t_nl = Type1(node=n, language=\"nl\", title=\"NL\").save()\n t_en = Type1(node=n, language=\"en\", title=\"EN\").save()\n\n\n request = create_request(\"GET\", \"/create\", data=dict(type=\"sometype\"))\n\n translation.activate(\"en\")\n toolbar = Toolbar(n, request, \"create\")\n translations = toolbar.translations()\n\n assert translations['active']['id'] == 'en'\n\n import urllib2\n\n ## Do some matching magic using endswith to work around language / base prefixing.\n ## We're mosly interested in create/view/edit actions anyway\n assert len(translations['translated']) == 0\n assert len(translations['untranslated']) == 3 ## all languages incl 'any', active lang excluded\n\n for ut in translations['untranslated']:\n l = ut['id']\n assert l in ('nl', 'fr', 'en', 'any')\n assert ut['action_url'].endswith('switch_admin_language?path='+n.tree_path + '&language=' + l + '&rest=' + urllib2.quote('create?type=sometype'))\n", "sub_path": "wheelcms_axle/tests/test_toolbar.py", "file_name": "test_toolbar.py", "file_ext": "py", "file_size_in_byte": 13851, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "wheelcms_axle.tests.models.TestTypeRegistry", "line_number": 18, "usage_type": "call"}, {"api_name": "wheelcms_axle.content.type_registry.set", "line_number": 19, "usage_type": "call"}, {"api_name": "wheelcms_axle.content.type_registry", "line_number": 19, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1Type", "line_number": 20, "usage_type": "argument"}, {"api_name": "wheelcms_axle.tests.models.Type2Type", "line_number": 21, "usage_type": "argument"}, {"api_name": "wheelcms_axle.content.type_registry.values", "line_number": 31, "usage_type": "call"}, {"api_name": "wheelcms_axle.content.type_registry", "line_number": 31, "usage_type": "name"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 38, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 38, "usage_type": "name"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 39, "usage_type": "call"}, {"api_name": "test_handler.superuser_request", "line_number": 39, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 47, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 47, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 48, "usage_type": "call"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 50, "usage_type": "call"}, {"api_name": "test_handler.superuser_request", "line_number": 50, "usage_type": "call"}, {"api_name": "wheelcms_axle.spoke.Spoke", "line_number": 69, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1Type", "line_number": 71, "usage_type": "name"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 81, "usage_type": "call"}, {"api_name": "test_handler.superuser_request", "line_number": 81, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type1Type.name", "line_number": 84, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type1Type", "line_number": 84, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1Type.title", "line_number": 85, "usage_type": "attribute"}, {"api_name": "wheelcms_axle.tests.models.Type1Type", "line_number": 85, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1Type.full_type_icon_path", "line_number": 86, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type1Type", "line_number": 86, "usage_type": "name"}, {"api_name": "wheelcms_axle.spoke.Spoke", "line_number": 103, "usage_type": "name"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 115, "usage_type": "call"}, {"api_name": "test_handler.superuser_request", "line_number": 115, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 121, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 121, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 122, "usage_type": "call"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 124, "usage_type": "call"}, {"api_name": "test_handler.superuser_request", "line_number": 124, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 130, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 130, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 131, "usage_type": "call"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 133, "usage_type": "call"}, {"api_name": "test_handler.superuser_request", "line_number": 133, "usage_type": "call"}, {"api_name": "wheelcms_axle.spoke.Spoke", "line_number": 149, "usage_type": "name"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 161, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 161, "usage_type": "name"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 162, "usage_type": "call"}, {"api_name": "test_handler.superuser_request", "line_number": 162, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 167, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 167, "usage_type": "name"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 168, "usage_type": "call"}, {"api_name": "twotest.util.create_request", "line_number": 168, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.get_or_create", "line_number": 172, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 172, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 172, "usage_type": "name"}, {"api_name": "twotest.util.create_request", "line_number": 173, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 176, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 176, "usage_type": "name"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 177, "usage_type": "call"}, {"api_name": "wheelcms_axle.spoke.Spoke", "line_number": 194, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1Type", "line_number": 196, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type2Type", "line_number": 196, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1Type", "line_number": 197, "usage_type": "name"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 207, "usage_type": "call"}, {"api_name": "test_handler.superuser_request", "line_number": 207, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type2Type.name", "line_number": 210, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type2Type", "line_number": 210, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type2Type.title", "line_number": 211, "usage_type": "attribute"}, {"api_name": "wheelcms_axle.tests.models.Type2Type", "line_number": 211, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type2Type.full_type_icon_path", "line_number": 212, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type2Type", "line_number": 212, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1Type.name", "line_number": 216, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type1Type", "line_number": 216, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1Type.title", "line_number": 217, "usage_type": "attribute"}, {"api_name": "wheelcms_axle.tests.models.Type1Type", "line_number": 217, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1Type.full_type_icon_path", "line_number": 218, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type1Type", "line_number": 218, "usage_type": "name"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 222, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 222, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 222, "usage_type": "name"}, {"api_name": "test_handler.superuser_request", "line_number": 222, "usage_type": "call"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 226, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 226, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 226, "usage_type": "name"}, {"api_name": "test_handler.superuser_request", "line_number": 226, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 234, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 234, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 236, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 237, "usage_type": "call"}, {"api_name": "twotest.util.create_request", "line_number": 239, "usage_type": "call"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 242, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 242, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 242, "usage_type": "name"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 250, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 250, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 252, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 253, "usage_type": "call"}, {"api_name": "twotest.util.create_request", "line_number": 255, "usage_type": "call"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 258, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 258, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 258, "usage_type": "name"}, {"api_name": "django.conf.settings.CONTENT_LANGUAGES", "line_number": 272, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 272, "usage_type": "name"}, {"api_name": "django.conf.settings.FALLBACK", "line_number": 273, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 273, "usage_type": "name"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 276, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 276, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 279, "usage_type": "call"}, {"api_name": "django.utils.translation.activate", "line_number": 280, "usage_type": "call"}, {"api_name": "django.utils.translation", "line_number": 280, "usage_type": "name"}, {"api_name": "twotest.util.create_request", "line_number": 281, "usage_type": "call"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 282, "usage_type": "call"}, {"api_name": "django.utils.translation.activate", "line_number": 286, "usage_type": "call"}, {"api_name": "django.utils.translation", "line_number": 286, "usage_type": "name"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 292, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 292, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 295, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 296, "usage_type": "call"}, {"api_name": "twotest.util.create_request", "line_number": 299, "usage_type": "call"}, {"api_name": "django.utils.translation.activate", "line_number": 301, "usage_type": "call"}, {"api_name": "django.utils.translation", "line_number": 301, "usage_type": "name"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 302, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 315, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 315, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 318, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 319, "usage_type": "call"}, {"api_name": "twotest.util.create_request", "line_number": 322, "usage_type": "call"}, {"api_name": "django.utils.translation.activate", "line_number": 324, "usage_type": "call"}, {"api_name": "django.utils.translation", "line_number": 324, "usage_type": "name"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 325, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 338, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 338, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 341, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 342, "usage_type": "call"}, {"api_name": "twotest.util.create_request", "line_number": 345, "usage_type": "call"}, {"api_name": "django.utils.translation.activate", "line_number": 347, "usage_type": "call"}, {"api_name": "django.utils.translation", "line_number": 347, "usage_type": "name"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 348, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node.root", "line_number": 361, "usage_type": "call"}, {"api_name": "wheelcms_axle.node.Node", "line_number": 361, "usage_type": "name"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 364, "usage_type": "call"}, {"api_name": "wheelcms_axle.tests.models.Type1", "line_number": 365, "usage_type": "call"}, {"api_name": "twotest.util.create_request", "line_number": 368, "usage_type": "call"}, {"api_name": "django.utils.translation.activate", "line_number": 370, "usage_type": "call"}, {"api_name": "django.utils.translation", "line_number": 370, "usage_type": "name"}, {"api_name": "wheelcms_axle.toolbar.Toolbar", "line_number": 371, "usage_type": "call"}, {"api_name": "urllib2.quote", "line_number": 386, "usage_type": "call"}]}
+{"seq_id": "464426054", "text": "from __future__ import unicode_literals\nimport logging\nfrom django.template.defaultfilters import truncatewords, striptags\nfrom oscar_pesapal import exceptions\n\nlogger = logging.getLogger('oscar_pesapal')\n\ndef _format_description(description):\n if description:\n return truncatewords(striptags(description), 12)\n return ''\n\n\ndef get_txn_status(txn):\n \"\"\"\n Fetch transaction status from Pesapal\n \"\"\"\n params = {\n 'pesapal_merchant_reference': txn.pesapal_merchant_reference,\n 'pesapal_transaction_tracking_id': txn.pesapal_transaction_id,\n }\n\n # Print easy-to-read version of params for debugging\n\n logger.debug(\"Fetching payment status with params: \\n\"\n \"Transaction ID : %s\\n\"\n \"Merchant Reference : %s\",\n txn.pesapal_merchant_reference,\n txn.pesapal_transaction_id)\n\n response = get_payment_status(**params)\n if response['_comm_success']:\n txn.status = response['_payment_status']\n txn.raw_response = response['_raw_response']\n txn.response_time = response['_response_time']\n txn.save()\n\n logger.debug(\"Successful response :\\n%s\", response['_payment_status'])\n\n else:\n\n msg = \"Error retrieving status %s - %s\" % (txn.pesapal_merchant_reference,\n txn.pesapal_transaction_id)\n logger.error(msg)\n raise exceptions.PesaPalError(msg)\n", "sub_path": "oscar_pesapal/gateway.py", "file_name": "gateway.py", "file_ext": "py", "file_size_in_byte": 1423, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "logging.getLogger", "line_number": 6, "usage_type": "call"}, {"api_name": "django.template.defaultfilters.truncatewords", "line_number": 10, "usage_type": "call"}, {"api_name": "django.template.defaultfilters.striptags", "line_number": 10, "usage_type": "call"}, {"api_name": "oscar_pesapal.exceptions.PesaPalError", "line_number": 45, "usage_type": "call"}, {"api_name": "oscar_pesapal.exceptions", "line_number": 45, "usage_type": "name"}]}
+{"seq_id": "79322067", "text": "import logging\nimport operator\nfrom collections import defaultdict\nfrom copy import deepcopy\nfrom typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Tuple, Union, cast, overload\n\nimport requests\nfrom eth_utils.address import to_checksum_address\nfrom web3.exceptions import BadFunctionCallOutput\n\nfrom rotkehlchen.assets.asset import Asset, EthereumToken\nfrom rotkehlchen.constants.assets import A_BTC, A_ETH\nfrom rotkehlchen.constants.misc import ZERO\nfrom rotkehlchen.db.utils import BlockchainAccounts\nfrom rotkehlchen.errors import (\n EthSyncError,\n InputError,\n InvalidBTCAddress,\n RemoteError,\n UnableToDecryptRemoteData,\n)\nfrom rotkehlchen.fval import FVal\nfrom rotkehlchen.inquirer import Inquirer\nfrom rotkehlchen.logging import RotkehlchenLogsAdapter\nfrom rotkehlchen.typing import (\n BlockchainAddress,\n BTCAddress,\n ChecksumEthAddress,\n EthAddress,\n ListOfBlockchainAddresses,\n Price,\n SupportedBlockchain,\n)\nfrom rotkehlchen.user_messages import MessagesAggregator\nfrom rotkehlchen.utils.interfaces import (\n CacheableObject,\n LockableQueryObject,\n cache_response_timewise,\n protect_with_lock,\n)\nfrom rotkehlchen.utils.misc import request_get_direct, satoshis_to_btc\n\nif TYPE_CHECKING:\n from rotkehlchen.ethchain import Ethchain\n\nlogger = logging.getLogger(__name__)\nlog = RotkehlchenLogsAdapter(logger)\n\n# Type Aliases used in this module\nBalances = Dict[\n Asset,\n Dict[BlockchainAddress, Dict[Union[str, Asset], FVal]],\n]\nTotals = Dict[Asset, Dict[str, FVal]]\nBlockchainBalancesUpdate = Dict[str, Union[Balances, Totals]]\nEthBalances = Dict[ChecksumEthAddress, Dict[str, Union[Dict[Asset, Dict[str, FVal]], FVal]]]\n\n\nclass Blockchain(CacheableObject, LockableQueryObject):\n\n def __init__(\n self,\n blockchain_accounts: BlockchainAccounts,\n owned_eth_tokens: List[EthereumToken],\n ethchain: 'Ethchain',\n msg_aggregator: MessagesAggregator,\n ):\n super().__init__()\n self.ethchain = ethchain\n self.msg_aggregator = msg_aggregator\n self.owned_eth_tokens = owned_eth_tokens\n self.accounts = blockchain_accounts\n\n # Per account balances\n self.balances: Balances = defaultdict(dict)\n # Per asset total balances\n self.totals: Totals = defaultdict(dict)\n\n def __del__(self) -> None:\n del self.ethchain\n\n def set_eth_rpc_endpoint(self, endpoint: str) -> Tuple[bool, str]:\n return self.ethchain.set_rpc_endpoint(endpoint)\n\n @property\n def eth_tokens(self) -> List[EthereumToken]:\n return self.owned_eth_tokens\n\n @protect_with_lock()\n @cache_response_timewise()\n def query_balances(\n self, # pylint: disable=unused-argument\n blockchain: Optional[SupportedBlockchain] = None,\n # Kwargs here is so linters don't complain when the \"magic\" ignore_cache kwarg is given\n **kwargs: Any,\n ) -> Dict[str, Dict]:\n \"\"\"Queries either all, or specific blockchain balances\n\n May raise:\n - RemoteError if an external service such as Etherscan or blockchain.info\n is queried and there is a problem with its query.\n - EthSyncError if querying the token balances through a provided ethereum\n client and the chain is not synced\n \"\"\"\n should_query_eth = not blockchain or blockchain == SupportedBlockchain.ETHEREUM\n should_query_btc = not blockchain or blockchain == SupportedBlockchain.BITCOIN\n\n if should_query_eth:\n self.query_ethereum_balances()\n\n if not blockchain or blockchain == SupportedBlockchain.BITCOIN:\n self.query_btc_balances()\n\n per_account = deepcopy(self.balances)\n totals = deepcopy(self.totals)\n if not should_query_eth:\n per_account.pop(A_ETH, None)\n # only keep BTC, remove ETH and any tokens that may be in the result\n totals = {A_BTC: totals[A_BTC]}\n if not should_query_btc:\n per_account.pop(A_BTC, None)\n totals.pop(A_BTC, None)\n\n return {'per_account': per_account, 'totals': totals}\n\n @staticmethod\n def query_btc_account_balance(account: BTCAddress) -> FVal:\n \"\"\"Queries blockchain.info for the balance of account\n\n May raise:\n - InputError if the given account is not a valid BTC address\n - RemotError if there is a problem querying blockchain.info\n \"\"\"\n try:\n btc_resp = request_get_direct(\n url='https://blockchain.info/q/addressbalance/%s' % account,\n handle_429=True,\n # If we get a 429 then their docs suggest 10 seconds\n # https://blockchain.info/q\n backoff_in_seconds=10,\n )\n except InvalidBTCAddress:\n # TODO: Move this validation into our own code and before the balance query\n raise InputError(f'The given string {account} is not a valid BTC address')\n except (requests.exceptions.ConnectionError, UnableToDecryptRemoteData) as e:\n raise RemoteError(f'blockchain.info API request failed due to {str(e)}')\n\n return satoshis_to_btc(FVal(btc_resp)) # result is in satoshis\n\n def query_btc_balances(self) -> None:\n \"\"\"Queries blockchain.info for the balance of all BTC accounts\n\n May raise:\n - RemotError if there is a problem querying blockchain.info or cryptocompare\n \"\"\"\n if len(self.accounts.btc) == 0:\n return\n\n self.balances[A_BTC] = {}\n btc_usd_price = Inquirer().find_usd_price(A_BTC)\n total = FVal(0)\n for account in self.accounts.btc:\n try:\n balance = self.query_btc_account_balance(account)\n except InputError:\n # This should really never happen.\n self.msg_aggregator.add_error(\n f'While querying BTC balances found invalid BTC account {account} in the DB',\n )\n continue\n total += balance\n self.balances[A_BTC][account] = {\n 'amount': balance,\n 'usd_value': balance * btc_usd_price,\n }\n\n self.totals[A_BTC] = {'amount': total, 'usd_value': total * btc_usd_price}\n\n @overload\n @staticmethod\n def _query_token_balances(\n token_asset: EthereumToken,\n query_callback: Callable[[EthereumToken, ChecksumEthAddress], FVal],\n argument: ChecksumEthAddress,\n ) -> FVal:\n ...\n\n @overload # noqa: F811\n @staticmethod\n def _query_token_balances(\n token_asset: EthereumToken,\n query_callback: Callable[\n [EthereumToken, List[ChecksumEthAddress]],\n Dict[ChecksumEthAddress, FVal],\n ],\n argument: List[ChecksumEthAddress],\n ) -> Dict[ChecksumEthAddress, FVal]:\n ...\n\n @staticmethod # noqa: F811\n def _query_token_balances(\n token_asset: EthereumToken,\n query_callback: Callable,\n argument: Union[List[ChecksumEthAddress], ChecksumEthAddress],\n ) -> Union[FVal, Dict[ChecksumEthAddress, FVal]]:\n \"\"\"Query tokens by checking the eth_tokens mapping and using the respective query callback.\n\n The callback is either self.ethchain.get_multitoken_balance or\n self.ethchain.get_token_balance\"\"\"\n result = query_callback(\n token_asset,\n argument,\n )\n\n return result\n\n def track_new_tokens(self, new_tokens: List[EthereumToken]) -> BlockchainBalancesUpdate:\n \"\"\"\n Adds new_tokens to the state and tracks their balance for each account.\n\n May raise:\n - InputError if some of the tokens already exist\n - RemoteError if an external service such as Etherscan is queried and\n there is a problem with its query.\n - EthSyncError if querying the token balances through a provided ethereum\n client and the chain is not synced\n \"\"\"\n\n intersection = set(new_tokens).intersection(set(self.owned_eth_tokens))\n if intersection != set():\n raise InputError('Some of the new provided tokens to track already exist')\n\n self.owned_eth_tokens.extend(new_tokens)\n eth_balances = cast(EthBalances, self.balances[A_ETH])\n\n if eth_balances == {}:\n # if balances have not been yet queried then we should do the entire\n # balance query first in order to create the eth_balances mappings\n self.query_ethereum_balances()\n else:\n # simply update all accounts with any changes adding the token may have\n self.query_ethereum_tokens(\n tokens=new_tokens,\n eth_balances=eth_balances,\n )\n return {'per_account': self.balances, 'totals': self.totals}\n\n def remove_eth_tokens(self, tokens: List[EthereumToken]) -> BlockchainBalancesUpdate:\n \"\"\"\n Removes tokens from the state and stops their balance from being tracked\n for each account\n\n May raise:\n - RemoteError if an external service such as Etherscan or cryptocompare\n is queried and there is a problem with its query.\n - EthSyncError if querying the token balances through a provided ethereum\n client and the chain is not synced\n \"\"\"\n if self.balances[A_ETH] == {}:\n # if balances have not been yet queried then we should do the entire\n # balance query first in order to create the eth_balances mappings\n self.query_ethereum_balances()\n\n for token in tokens:\n usd_price = Inquirer().find_usd_price(token)\n for account, account_data in self.balances[A_ETH].items():\n if token not in account_data['assets']: # type: ignore\n continue\n\n balance = account_data['assets'][token]['amount'] # type: ignore\n deleting_usd_value = balance * usd_price\n del self.balances[A_ETH][account]['assets'][token] # type: ignore\n self.balances[A_ETH][account]['total_usd_value'] = (\n self.balances[A_ETH][account]['total_usd_value'] -\n deleting_usd_value\n )\n # Remove the token from the totals iff existing. May not exist\n # if the token price is 0 but is still tracked.\n # See https://github.com/rotki/rotki/issues/467\n # for more details\n self.totals.pop(token, None)\n self.owned_eth_tokens.remove(token)\n\n return {'per_account': self.balances, 'totals': self.totals}\n\n def modify_btc_account(\n self,\n account: BTCAddress,\n append_or_remove: str,\n add_or_sub: Callable[[FVal, FVal], FVal],\n ) -> None:\n \"\"\"Either appends or removes a BTC acccount.\n\n Call with 'append', operator.add to add the account\n Call with 'remove', operator.sub to remove the account\n\n May raise:\n - InputError if the given account is not a valid BTC address\n - RemotError if there is a problem querying blockchain.info or cryptocompare\n \"\"\"\n btc_usd_price = Inquirer().find_usd_price(A_BTC)\n remove_with_populated_balance = (\n append_or_remove == 'remove' and len(self.balances[A_BTC]) != 0\n )\n # Query the balance of the account except for the case when it's removed\n # and there is no other account in the balances\n if append_or_remove == 'append' or remove_with_populated_balance:\n balance = self.query_btc_account_balance(account)\n usd_balance = balance * btc_usd_price\n\n if append_or_remove == 'append':\n self.balances[A_BTC][account] = {'amount': balance, 'usd_value': usd_balance}\n elif append_or_remove == 'remove':\n if account in self.balances[A_BTC]:\n del self.balances[A_BTC][account]\n else:\n raise AssertionError('Programmer error: Should be append or remove')\n\n if len(self.balances[A_BTC]) == 0:\n # If the last account was removed balance should be 0\n self.totals[A_BTC]['amount'] = FVal(0)\n self.totals[A_BTC]['usd_value'] = FVal(0)\n else:\n self.totals[A_BTC]['amount'] = add_or_sub(\n self.totals[A_BTC].get('amount', FVal(0)),\n balance,\n )\n self.totals[A_BTC]['usd_value'] = add_or_sub(\n self.totals[A_BTC].get('usd_value', FVal(0)),\n usd_balance,\n )\n # At the very end add/remove it from the accounts\n getattr(self.accounts.btc, append_or_remove)(account)\n\n def modify_eth_account(\n self,\n given_account: EthAddress,\n append_or_remove: str,\n add_or_sub: Callable[[FVal, FVal], FVal],\n ) -> None:\n \"\"\"Either appends or removes an ETH acccount.\n\n Call with 'append', operator.add to add the account\n Call with 'remove', operator.sub to remove the account\n\n May raise:\n - Input error if the given_account is not a valid ETH address\n - BadFunctionCallOutput if a token is queried from a local chain\n and the chain is not synced\n - RemoteError if there is a problem with a query to an external\n service such as Etherscan or cryptocompare\n \"\"\"\n # Make sure account goes into web3.py as a properly checksummed address\n try:\n account = to_checksum_address(given_account)\n except ValueError:\n raise InputError(f'The given string {given_account} is not a valid ETH address')\n eth_usd_price = Inquirer().find_usd_price(A_ETH)\n remove_with_populated_balance = (\n append_or_remove == 'remove' and len(self.balances[A_ETH]) != 0\n )\n # Query the balance of the account except for the case when it's removed\n # and there is no other account in the balances\n if append_or_remove == 'append' or remove_with_populated_balance:\n balance = self.ethchain.get_eth_balance(account)\n usd_balance = balance * eth_usd_price\n\n if append_or_remove == 'append':\n self.accounts.eth.append(account)\n self.balances[A_ETH][account] = {\n 'assets': { # type: ignore\n A_ETH: {'amount': balance, 'usd_value': usd_balance},\n },\n 'total_usd_value': usd_balance,\n }\n elif append_or_remove == 'remove':\n if account not in self.accounts.eth:\n raise InputError('Tried to remove a non existing ETH account')\n self.accounts.eth.remove(account)\n if account in self.balances[A_ETH]:\n del self.balances[A_ETH][account]\n else:\n raise AssertionError('Programmer error: Should be append or remove')\n\n if len(self.balances[A_ETH]) == 0:\n # If the last account was removed balance should be 0\n self.totals[A_ETH]['amount'] = FVal(0)\n self.totals[A_ETH]['usd_value'] = FVal(0)\n else:\n self.totals[A_ETH]['amount'] = add_or_sub(\n self.totals[A_ETH].get('amount', FVal(0)),\n balance,\n )\n self.totals[A_ETH]['usd_value'] = add_or_sub(\n self.totals[A_ETH].get('usd_value', FVal(0)),\n usd_balance,\n )\n\n for token in self.owned_eth_tokens:\n try:\n usd_price = Inquirer().find_usd_price(token)\n except RemoteError:\n usd_price = Price(ZERO)\n if usd_price == ZERO:\n # skip tokens that have no price\n continue\n\n if append_or_remove == 'remove' and token not in self.totals:\n # If we remove an account, and the token has no totals entry skip\n continue\n\n token_balance = Blockchain._query_token_balances(\n token_asset=token,\n query_callback=self.ethchain.get_token_balance,\n argument=account,\n )\n if token_balance == 0:\n continue\n\n usd_value = token_balance * usd_price\n if append_or_remove == 'append':\n account_balance = self.balances[A_ETH][account]\n account_balance['assets'][token] = {'amount': token_balance, 'usd_value': usd_value} # type: ignore # noqa: E501\n account_balance['total_usd_value'] = account_balance['total_usd_value'] + usd_value\n\n self.totals[token] = {\n 'amount': add_or_sub(\n self.totals[token].get('amount', ZERO),\n token_balance,\n ),\n 'usd_value': add_or_sub(\n self.totals[token].get('usd_value', ZERO),\n usd_value,\n ),\n }\n\n def add_blockchain_accounts(\n self,\n blockchain: SupportedBlockchain,\n accounts: ListOfBlockchainAddresses,\n ) -> Tuple[BlockchainBalancesUpdate, ListOfBlockchainAddresses, str]:\n \"\"\"Adds new blockchain accounts and requeries all balances after the addition.\n The accounts are added in the blockchain object and not in the database.\n Returns the new total balances, the actually added accounts (some\n accounts may have been invalid) and also any errors that occured\n during the addition.\n\n May Raise:\n - EthSyncError from modify_blockchain_account\n - InputError if the given accounts list is empty\n - RemoteError if an external service such as Etherscan is queried and\n there is a problem\n \"\"\"\n if len(accounts) == 0:\n raise InputError('Empty list of blockchain accounts to add was given')\n\n # If no blockchain query has happened before then we need to query the relevant\n # chain to populate the self.balances mapping.\n if blockchain.value not in self.balances:\n self.query_balances(blockchain, ignore_cache=True)\n\n added_accounts = []\n full_msg = ''\n\n for account in accounts:\n try:\n result = self.modify_blockchain_account(\n blockchain=blockchain,\n account=account,\n append_or_remove='append',\n add_or_sub=operator.add,\n )\n added_accounts.append(account)\n except InputError as e:\n full_msg += str(e)\n result = {'per_account': self.balances, 'totals': self.totals}\n\n # Ignore type checks here. added_accounts is the same type as accounts\n # but not sure how to show that to mypy\n return result, added_accounts, full_msg # type: ignore\n\n def remove_blockchain_accounts(\n self,\n blockchain: SupportedBlockchain,\n accounts: ListOfBlockchainAddresses,\n ) -> Tuple[BlockchainBalancesUpdate, ListOfBlockchainAddresses, str]:\n \"\"\"Removes blockchain accounts and requeries all balances after the removal.\n\n The accounts are removed from the blockchain object and not from the database.\n Returns the new total balances, the actually removes accounts (some\n accounts may have been invalid) and also any errors that occured\n during the removal.\n\n May Raise:\n - EthSyncError from modify_blockchain_account\n - InputError if the given accounts list is empty\n - RemoteError if an external service such as Etherscan is queried and\n there is a problem\n \"\"\"\n if len(accounts) == 0:\n raise InputError('Empty list of blockchain accounts to add was given')\n\n # If no blockchain query has happened before then we need to query the relevant\n # chain to populate the self.balances mapping. But query has to happen after\n # account removal so as not to query unneeded accounts\n balances_queried_before = True\n if blockchain.value not in self.balances:\n balances_queried_before = False\n\n removed_accounts = []\n full_msg = ''\n for account in accounts:\n try:\n self.modify_blockchain_account(\n blockchain=blockchain,\n account=account,\n append_or_remove='remove',\n add_or_sub=operator.sub,\n )\n removed_accounts.append(account)\n except InputError as e:\n full_msg += '. ' + str(e)\n\n if not balances_queried_before:\n self.query_balances(blockchain, ignore_cache=True)\n\n result: BlockchainBalancesUpdate = {'per_account': self.balances, 'totals': self.totals}\n\n # Ignore type checks here. removed_accounts is the same type as accounts\n # but not sure how to show that to mypy\n return result, removed_accounts, full_msg # type: ignore\n\n def modify_blockchain_account(\n self,\n blockchain: SupportedBlockchain,\n account: BlockchainAddress,\n append_or_remove: str,\n add_or_sub: Callable[[FVal, FVal], FVal],\n ) -> BlockchainBalancesUpdate:\n \"\"\"Add or remove a blockchain account\n\n May raise:\n\n - InputError if accounts to remove do not exist or if the ethereum/BTC\n addresses are not valid.\n - EthSyncError if there is a problem querying the ethereum chain\n - RemoteError if there is a problem querying an external service such\n as etherscan or blockchain.info\n \"\"\"\n if blockchain == SupportedBlockchain.BITCOIN:\n if append_or_remove == 'remove' and account not in self.accounts.btc:\n raise InputError('Tried to remove a non existing BTC account')\n\n # above we check that account is a BTC account\n self.modify_btc_account(\n BTCAddress(account),\n append_or_remove,\n add_or_sub,\n )\n\n elif blockchain == SupportedBlockchain.ETHEREUM:\n try:\n # above we check that account is an ETH account\n self.modify_eth_account(EthAddress(account), append_or_remove, add_or_sub)\n except BadFunctionCallOutput as e:\n log.error(\n 'Assuming unsynced chain. Got web3 BadFunctionCallOutput '\n 'exception: {}'.format(str(e)),\n )\n raise EthSyncError(\n 'Tried to use the ethereum chain of a local client to edit '\n 'an eth account but the chain is not synced.',\n )\n\n else:\n # That should not happen. Should be checked by marshmallow\n raise AssertionError(\n 'Unsupported blockchain {} provided at remove_blockchain_account'.format(\n blockchain),\n )\n\n return {'per_account': self.balances, 'totals': self.totals}\n\n def query_ethereum_tokens(\n self,\n tokens: List[EthereumToken],\n eth_balances: EthBalances,\n ) -> None:\n \"\"\"Queries the ethereum token balances and populates the state\n\n May raise:\n - RemoteError if an external service such as Etherscan or cryptocompare\n is queried and there is a problem with its query.\n - EthSyncError if querying the token balances through a provided ethereum\n client and the chain is not synced\n \"\"\"\n token_balances = {}\n token_usd_price = {}\n for token in tokens:\n try:\n usd_price = Inquirer().find_usd_price(token)\n except RemoteError:\n usd_price = Price(ZERO)\n if usd_price == ZERO:\n # skip tokens that have no price\n continue\n token_usd_price[token] = usd_price\n\n try:\n token_balances[token] = Blockchain._query_token_balances(\n token_asset=token,\n query_callback=self.ethchain.get_multitoken_balance,\n argument=self.accounts.eth,\n )\n except BadFunctionCallOutput as e:\n log.error(\n 'Assuming unsynced chain. Got web3 BadFunctionCallOutput '\n 'exception: {}'.format(str(e)),\n )\n raise EthSyncError(\n 'Tried to use the ethereum chain of the provided client to query '\n 'token balances but the chain is not synced.',\n )\n\n for token, token_accounts in token_balances.items():\n token_total = FVal(0)\n for account, balance in token_accounts.items():\n token_total += balance\n usd_value = balance * token_usd_price[token]\n if balance != ZERO:\n eth_balances[account]['assets'][token] = { # type: ignore\n 'amount': balance,\n 'usd_value': usd_value,\n }\n eth_balances[account]['total_usd_value'] = (\n eth_balances[account]['total_usd_value'] + usd_value # type: ignore\n )\n\n self.totals[token] = {\n 'amount': token_total,\n 'usd_value': token_total * token_usd_price[token],\n }\n\n self.balances[A_ETH] = cast(\n Dict[BlockchainAddress, Dict[Union[str, Asset], FVal]],\n eth_balances,\n )\n\n def query_ethereum_balances(self) -> None:\n \"\"\"Queries the ethereum balances and populates the state\n\n May raise:\n - RemoteError if an external service such as Etherscan or cryptocompare\n is queried and there is a problem with its query.\n - EthSyncError if querying the token balances through a provided ethereum\n client and the chain is not synced\n \"\"\"\n if len(self.accounts.eth) == 0:\n return\n\n eth_accounts = self.accounts.eth\n eth_usd_price = Inquirer().find_usd_price(A_ETH)\n balances = self.ethchain.get_multieth_balance(eth_accounts)\n eth_total = FVal(0)\n eth_balances: EthBalances = {}\n for account, balance in balances.items():\n eth_total += balance\n usd_value = balance * eth_usd_price\n eth_balances[account] = {\n 'assets': {\n A_ETH: {'amount': balance, 'usd_value': usd_value},\n },\n 'total_usd_value': usd_value,\n }\n\n self.totals[A_ETH] = {'amount': eth_total, 'usd_value': eth_total * eth_usd_price}\n # but they are not complete until token query\n self.balances[A_ETH] = cast(\n Dict[BlockchainAddress, Dict[Union[str, Asset], FVal]],\n eth_balances,\n )\n\n # And now for tokens\n self.query_ethereum_tokens(self.owned_eth_tokens, eth_balances)\n", "sub_path": "rotkehlchen/blockchain.py", "file_name": "blockchain.py", "file_ext": "py", "file_size_in_byte": 27257, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "typing.TYPE_CHECKING", "line_number": 43, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 46, "usage_type": "call"}, {"api_name": "rotkehlchen.logging.RotkehlchenLogsAdapter", "line_number": 47, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 50, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.Asset", "line_number": 51, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 52, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.BlockchainAddress", "line_number": 52, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 52, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.Asset", "line_number": 52, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 52, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 54, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.Asset", "line_number": 54, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 54, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 55, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 55, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 56, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ChecksumEthAddress", "line_number": 56, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 56, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.Asset", "line_number": 56, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 56, "usage_type": "name"}, {"api_name": "rotkehlchen.utils.interfaces.CacheableObject", "line_number": 59, "usage_type": "name"}, {"api_name": "rotkehlchen.utils.interfaces.LockableQueryObject", "line_number": 59, "usage_type": "name"}, {"api_name": "rotkehlchen.db.utils.BlockchainAccounts", "line_number": 63, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 64, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.EthereumToken", "line_number": 64, "usage_type": "name"}, {"api_name": "rotkehlchen.user_messages.MessagesAggregator", "line_number": 66, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 75, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 77, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 82, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 86, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.EthereumToken", "line_number": 86, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 93, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain", "line_number": 93, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 95, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain.ETHEREUM", "line_number": 105, "usage_type": "attribute"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain", "line_number": 105, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain.BITCOIN", "line_number": 106, "usage_type": "attribute"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain", "line_number": 106, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain.BITCOIN", "line_number": 111, "usage_type": "attribute"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain", "line_number": 111, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 114, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 115, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 117, "usage_type": "argument"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 119, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 121, "usage_type": "argument"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 122, "usage_type": "argument"}, {"api_name": "rotkehlchen.utils.interfaces.protect_with_lock", "line_number": 89, "usage_type": "call"}, {"api_name": "rotkehlchen.utils.interfaces.cache_response_timewise", "line_number": 90, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 96, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.BTCAddress", "line_number": 127, "usage_type": "name"}, {"api_name": "rotkehlchen.utils.misc.request_get_direct", "line_number": 135, "usage_type": "call"}, {"api_name": "rotkehlchen.errors.InvalidBTCAddress", "line_number": 142, "usage_type": "name"}, {"api_name": "rotkehlchen.errors.InputError", "line_number": 144, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 145, "usage_type": "attribute"}, {"api_name": "rotkehlchen.errors.UnableToDecryptRemoteData", "line_number": 145, "usage_type": "name"}, {"api_name": "rotkehlchen.errors.RemoteError", "line_number": 146, "usage_type": "call"}, {"api_name": "rotkehlchen.utils.misc.satoshis_to_btc", "line_number": 148, "usage_type": "call"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 148, "usage_type": "call"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 127, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 159, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 160, "usage_type": "argument"}, {"api_name": "rotkehlchen.inquirer.Inquirer", "line_number": 160, "usage_type": "call"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 161, "usage_type": "call"}, {"api_name": "rotkehlchen.errors.InputError", "line_number": 165, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 172, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 177, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.EthereumToken", "line_number": 182, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 183, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.EthereumToken", "line_number": 183, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ChecksumEthAddress", "line_number": 183, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 183, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ChecksumEthAddress", "line_number": 184, "usage_type": "name"}, {"api_name": "typing.overload", "line_number": 179, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 185, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.EthereumToken", "line_number": 191, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 192, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.EthereumToken", "line_number": 193, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 193, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ChecksumEthAddress", "line_number": 193, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 194, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ChecksumEthAddress", "line_number": 194, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 194, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 196, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ChecksumEthAddress", "line_number": 196, "usage_type": "name"}, {"api_name": "typing.overload", "line_number": 188, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 197, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ChecksumEthAddress", "line_number": 197, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 197, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.EthereumToken", "line_number": 202, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 203, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 204, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 204, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ChecksumEthAddress", "line_number": 204, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 205, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 205, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 205, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ChecksumEthAddress", "line_number": 205, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 217, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.EthereumToken", "line_number": 217, "usage_type": "name"}, {"api_name": "rotkehlchen.errors.InputError", "line_number": 231, "usage_type": "call"}, {"api_name": "typing.cast", "line_number": 234, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 234, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 248, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.EthereumToken", "line_number": 248, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 259, "usage_type": "name"}, {"api_name": "rotkehlchen.inquirer.Inquirer", "line_number": 265, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 266, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 272, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 273, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 274, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.BTCAddress", "line_number": 288, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 290, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 290, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 301, "usage_type": "argument"}, {"api_name": "rotkehlchen.inquirer.Inquirer", "line_number": 301, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 303, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 312, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 314, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 315, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 319, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 321, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 321, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 322, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 322, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 324, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 325, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 325, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 328, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_BTC", "line_number": 329, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 329, "usage_type": "call"}, {"api_name": "rotkehlchen.typing.EthAddress", "line_number": 337, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 339, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 339, "usage_type": "name"}, {"api_name": "eth_utils.address.to_checksum_address", "line_number": 355, "usage_type": "call"}, {"api_name": "rotkehlchen.errors.InputError", "line_number": 357, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 358, "usage_type": "argument"}, {"api_name": "rotkehlchen.inquirer.Inquirer", "line_number": 358, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 360, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 370, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 372, "usage_type": "name"}, {"api_name": "rotkehlchen.errors.InputError", "line_number": 378, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 380, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 381, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 385, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 387, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 387, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 388, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 388, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 390, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 391, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 391, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 394, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 395, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 395, "usage_type": "call"}, {"api_name": "rotkehlchen.inquirer.Inquirer", "line_number": 401, "usage_type": "call"}, {"api_name": "rotkehlchen.errors.RemoteError", "line_number": 402, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.Price", "line_number": 403, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.misc.ZERO", "line_number": 403, "usage_type": "argument"}, {"api_name": "rotkehlchen.constants.misc.ZERO", "line_number": 404, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 422, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.misc.ZERO", "line_number": 428, "usage_type": "argument"}, {"api_name": "rotkehlchen.constants.misc.ZERO", "line_number": 432, "usage_type": "argument"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain", "line_number": 439, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ListOfBlockchainAddresses", "line_number": 440, "usage_type": "name"}, {"api_name": "rotkehlchen.errors.InputError", "line_number": 455, "usage_type": "call"}, {"api_name": "operator.add", "line_number": 471, "usage_type": "attribute"}, {"api_name": "rotkehlchen.errors.InputError", "line_number": 474, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 441, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ListOfBlockchainAddresses", "line_number": 441, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain", "line_number": 484, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ListOfBlockchainAddresses", "line_number": 485, "usage_type": "name"}, {"api_name": "rotkehlchen.errors.InputError", "line_number": 501, "usage_type": "call"}, {"api_name": "operator.sub", "line_number": 518, "usage_type": "attribute"}, {"api_name": "rotkehlchen.errors.InputError", "line_number": 521, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 486, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.ListOfBlockchainAddresses", "line_number": 486, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain", "line_number": 535, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.BlockchainAddress", "line_number": 536, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 538, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 538, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain.BITCOIN", "line_number": 550, "usage_type": "attribute"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain", "line_number": 550, "usage_type": "name"}, {"api_name": "rotkehlchen.errors.InputError", "line_number": 552, "usage_type": "call"}, {"api_name": "rotkehlchen.typing.BTCAddress", "line_number": 556, "usage_type": "call"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain.ETHEREUM", "line_number": 561, "usage_type": "attribute"}, {"api_name": "rotkehlchen.typing.SupportedBlockchain", "line_number": 561, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.EthAddress", "line_number": 564, "usage_type": "call"}, {"api_name": "web3.exceptions.BadFunctionCallOutput", "line_number": 565, "usage_type": "name"}, {"api_name": "rotkehlchen.errors.EthSyncError", "line_number": 570, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 586, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.EthereumToken", "line_number": 586, "usage_type": "name"}, {"api_name": "rotkehlchen.inquirer.Inquirer", "line_number": 601, "usage_type": "call"}, {"api_name": "rotkehlchen.errors.RemoteError", "line_number": 602, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.Price", "line_number": 603, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.misc.ZERO", "line_number": 603, "usage_type": "argument"}, {"api_name": "rotkehlchen.constants.misc.ZERO", "line_number": 604, "usage_type": "name"}, {"api_name": "web3.exceptions.BadFunctionCallOutput", "line_number": 615, "usage_type": "name"}, {"api_name": "rotkehlchen.errors.EthSyncError", "line_number": 620, "usage_type": "call"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 626, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.misc.ZERO", "line_number": 630, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 644, "usage_type": "name"}, {"api_name": "typing.cast", "line_number": 644, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 645, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.BlockchainAddress", "line_number": 645, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 645, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.Asset", "line_number": 645, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 645, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 662, "usage_type": "argument"}, {"api_name": "rotkehlchen.inquirer.Inquirer", "line_number": 662, "usage_type": "call"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 664, "usage_type": "call"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 671, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 676, "usage_type": "name"}, {"api_name": "rotkehlchen.constants.assets.A_ETH", "line_number": 678, "usage_type": "name"}, {"api_name": "typing.cast", "line_number": 678, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 679, "usage_type": "name"}, {"api_name": "rotkehlchen.typing.BlockchainAddress", "line_number": 679, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 679, "usage_type": "name"}, {"api_name": "rotkehlchen.assets.asset.Asset", "line_number": 679, "usage_type": "name"}, {"api_name": "rotkehlchen.fval.FVal", "line_number": 679, "usage_type": "name"}]}
+{"seq_id": "160687177", "text": "#script: ex-polyfit.py\n#linear regression and polynomial regression\n#(using _polyreg module)\n#author: Luis Paris\n\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nimport _poly\nimport _polyreg\n\nx = [1, 2, 3, 4]\ny = [3, 5, 7, 11] #11, not 9, so the fit isn't perfect\n\n#estimate plot x,y border for plotting\nxgap = (min(x) + max(x)) / 20.\nygap = (min(y) + max(y)) / 20.\nxlo = min(x) - xgap\nxhi = max(x) + xgap\nylo = min(y) - ygap\nyhi = max(y) + ygap\n\nprint(\"Linear regression:\") #find polynomial coeffs for linear regression\npol = _polyreg.curvefit(x, y, order=1, debug=True)\n\nprint(\"pol = {}\\ny = {}\\n\".format(pol, _poly.tostr(pol)))\n\nxregr = np.linspace(min(x), max(x))\nyregr = np.array([_poly.eval(pol, xval) for xval in xregr])\n\nplt.plot(x, y, 'yo', xregr, yregr, '--k')\nplt.xlim(xlo, xhi)\nplt.ylim(ylo, yhi)\nplt.title(\"Linear Regression\")\nplt.show()\n\nprint(\"Quadratic regression:\") #find polynomial coeffs for quadratic regression\npol = _polyreg.curvefit(x, y, order=2, debug=True)\n\nprint(\"pol = {}\\ny = {}\\n\".format(pol, _poly.tostr(pol)))\n\nxregr = np.linspace(min(x), max(x))\nyregr = np.array([_poly.eval(pol, xval) for xval in xregr])\n\nplt.plot(x, y, 'yo', xregr, yregr, '--k')\nplt.xlim(xlo, xhi)\nplt.ylim(ylo, yhi)\nplt.title(\"Quadratic Regression\")\nplt.show()\n\nprint(\"Cubic regression:\") #find polynomial coeffs for cubic regression\npol = _polyreg.curvefit(x, y, order=3, debug=True)\n\nprint(\"pol = {}\\ny = {}\\n\".format(pol, _poly.tostr(pol)))\n\nxregr = np.linspace(min(x), max(x))\nyregr = np.array([_poly.eval(pol, xval) for xval in xregr])\n\nplt.plot(x, y, 'yo', xregr, yregr, '--k')\nplt.xlim(xlo, xhi)\nplt.ylim(ylo, yhi)\nplt.title(\"Cubic Regression\")\nplt.show()\n", "sub_path": "Topic 9/ex-polyfit.py", "file_name": "ex-polyfit.py", "file_ext": "py", "file_size_in_byte": 1675, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "_polyreg.curvefit", "line_number": 24, "usage_type": "call"}, {"api_name": "_poly.tostr", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "_poly.eval", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "_polyreg.curvefit", "line_number": 38, "usage_type": "call"}, {"api_name": "_poly.tostr", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "_poly.eval", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "_polyreg.curvefit", "line_number": 52, "usage_type": "call"}, {"api_name": "_poly.tostr", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 57, "usage_type": "call"}, {"api_name": "_poly.eval", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}]}
+{"seq_id": "169237364", "text": "\"\"\"\n\"\"\"\nimport logging\nimport collections\n\nfrom PySide.QtCore import QObject, QEvent\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass AbstractEventDispatcher(QObject):\n \"\"\"Abstract dispatcher class.\n\n Attributes:\n events (list): Events to be tracked\n \"\"\"\n events = []\n\n def dispatch_events(self, event):\n raise NotImplementedError()\n\n def eventFilter(self, obj, event):\n raise NotImplementedError()\n\n\nclass BaseObserver(AbstractEventDispatcher):\n\n def __init__(self):\n super(BaseObserver, self).__init__()\n self.watches = dict()\n self.handlers = collections.defaultdict(set)\n\n def add_handler_for_watch(self, event_type, event_handler):\n logger.debug('Adding handler: {}, to watcher: {}'.format(event_type,\n event_handler))\n self.handlers[event_type].add(event_handler())\n\n def dispatch_events(self, event):\n for handler in self.handlers[event.type()]:\n result = handler.dispatch(event)\n logger.debug('event filter result: {} from: {}'.format(result, handler))\n return result\n\n def eventFilter(self, obj, event):\n # Adds the observer to new child windows created from the main application.\n if event.type() == QEvent.ChildAdded:\n self._temp_child = event.child()\n elif event.type() == QEvent.ChildPolished:\n if self._temp_child.objectName() not in self.watches:\n self.watches[self._temp_child.objectName()] = event.child()\n\n # Only work on events specified.\n if event.type() not in self.events:\n return False\n else:\n return self.dispatch_events(event)\n", "sub_path": "mamqtkeys/qtkeys/observers.py", "file_name": "observers.py", "file_ext": "py", "file_size_in_byte": 1741, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "PySide.QtCore.QObject", "line_number": 12, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 32, "usage_type": "call"}, {"api_name": "PySide.QtCore.QEvent.ChildAdded", "line_number": 47, "usage_type": "attribute"}, {"api_name": "PySide.QtCore.QEvent", "line_number": 47, "usage_type": "name"}, {"api_name": "PySide.QtCore.QEvent.ChildPolished", "line_number": 49, "usage_type": "attribute"}, {"api_name": "PySide.QtCore.QEvent", "line_number": 49, "usage_type": "name"}]}
+{"seq_id": "398126395", "text": "from datetime import datetime\n\nanoatual = int(datetime.now().year)\nmaior = 0\nmenor = 0\nfor i in range(1, 8):\n nasc = int(input('Digite o ano de nascimento da {}° pessoa:'.format(i)))\n idade = anoatual - nasc\n if idade >= 18:\n maior += 1\n else:\n menor += 1\nprint('''Existem {} pessoas menores de idade.\nExistem {} pessoas maiores de idade.'''.format(menor,maior))\n", "sub_path": "Exercicios_1/ex054.py", "file_name": "ex054.py", "file_ext": "py", "file_size_in_byte": 390, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "datetime.datetime.now", "line_number": 3, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 3, "usage_type": "name"}]}
+{"seq_id": "156861627", "text": "#!/usr/bin/env python3\n \nimport socket # Needed for socket creation\nimport fcntl # Needed libraries\nimport os # Needed libraries\nimport ntplib # Needed for NCP time (syncronize the time between source and destination)\nfrom datetime import datetime # Needed for NCP time\n\nHOST = '10.10.3.2' # local adress (used with r1)\nHOST2 = '10.10.5.2' # local adress (used with r2)\nPORT = 31337 # port that message is received (if sent by r1)\nPORT2 = 31338 # port that message is received (if sent by r2)\n\n\nsocks = [] # list of sockets (sockets will be appended)\n\nsock1 = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\nsock1.bind((HOST, PORT))\nfcntl.fcntl(sock1, fcntl.F_SETFL, os.O_NONBLOCK)\nsocks.append(sock1) # append the first socket to socks (socket with r1)\n\nsock2 = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) \nsock2.bind((HOST2, PORT2))\nfcntl.fcntl(sock2, fcntl.F_SETFL, os.O_NONBLOCK)\nsocks.append(sock2) # append the second socket to socks (socket with r2)\n\nwhile True:\n for s in socks: # for every socket in socks\n try: \n msg = s.recv(1024) # receive message\n c = ntplib.NTPClient() # get the current time from time.google.com\n response = c.request('time.google.com', version=3) # get the current time from time.google.com\n time_d = datetime.fromtimestamp(response.orig_time) # get the current time from time.google.com\n time_s = datetime.strptime(str(msg), \"%Y-%m-%d %H:%M:%S.%f\") # get the source time from the message\n print(time_d-time_s) # print the time it takes for a message to get from source the destination \n # Compare with the time printed by the source\n except: # continue listening\n continue\n", "sub_path": "TP_Part1_02/destination.py", "file_name": "destination.py", "file_ext": "py", "file_size_in_byte": 1774, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "socket.socket", "line_number": 17, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 17, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 17, "usage_type": "attribute"}, {"api_name": "fcntl.fcntl", "line_number": 19, "usage_type": "call"}, {"api_name": "fcntl.F_SETFL", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.O_NONBLOCK", "line_number": 19, "usage_type": "attribute"}, {"api_name": "socket.socket", "line_number": 22, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 22, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 22, "usage_type": "attribute"}, {"api_name": "fcntl.fcntl", "line_number": 24, "usage_type": "call"}, {"api_name": "fcntl.F_SETFL", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.O_NONBLOCK", "line_number": 24, "usage_type": "attribute"}, {"api_name": "ntplib.NTPClient", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 33, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 34, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 34, "usage_type": "name"}]}
+{"seq_id": "473926602", "text": "from __future__ import annotations\n\nfrom math import ceil\nimport logging\n\n\nfrom rich.color import Color\nfrom rich.style import Style\nfrom rich.console import Console, ConsoleOptions, RenderResult, RenderableType\nfrom rich.segment import Segment, Segments\nfrom rich.style import Style\n\nlog = logging.getLogger(\"rich\")\n\nfrom .widget import Widget\n\n\nclass VerticalBar:\n def __init__(\n self,\n lines: list[list[Segment]],\n height: int,\n virtual_height: int,\n position: float,\n overlay: bool = False,\n ) -> None:\n self.lines = lines\n self.height = height\n self.virtual_height = virtual_height\n self.position = position\n self.overlay = overlay\n\n def __rich_console__(\n self, console: Console, options: ConsoleOptions\n ) -> RenderResult:\n bar = render_bar(\n size=self.height,\n window_size=len(self.lines),\n virtual_size=self.virtual_height,\n position=self.position,\n )\n new_line = Segment.line()\n for line, bar_segment in zip(self.lines, bar):\n yield from line\n yield bar_segment\n yield new_line\n\n\nclass ScrollBar(Widget):\n def __init__(self, virtual_size: int = 100, window_size: int = 25) -> None:\n self.position = 0\n self.virtual_size = virtual_size\n self.window_size = window_size\n super().__init__()\n\n def render(self, console: Console, options: ConsoleOptions) -> RenderableType:\n\n height = options.height or console.height\n bar_segments = render_bar(\n height,\n window_size=self.window_size,\n virtual_size=self.virtual_size,\n position=self.position,\n depth=options.max_width,\n )\n return Segments(bar_segments, new_lines=True)\n\n\ndef render_bar(\n size: int = 25,\n virtual_size: float = 50,\n window_size: float = 20,\n position: float = 0,\n back_color: str = \"#555555\",\n bar_color: str = \"bright_magenta\",\n ascii_only: bool = False,\n depth: int = 1,\n vertical: bool = True,\n) -> list[Segment]:\n\n if ascii_only:\n bars = [\"|\", \"|\", \"|\", \"|\", \"|\", \"|\", \"|\", \"|\", \"|\"]\n else:\n bars = [\"▁\", \"▂\", \"▃\", \"▄\", \"▅\", \"▆\", \"▇\", \"█\"]\n\n back = Color.parse(back_color)\n bar = Color.parse(bar_color)\n\n _Segment = Segment\n _Style = Style\n blank = \" \" * depth\n segments = [_Segment(blank, _Style(bgcolor=back))] * int(size)\n\n step_size = virtual_size / size\n\n start = int(position / step_size * 8)\n end = start + max(8, int(window_size / step_size * 8))\n\n start_index, start_bar = divmod(start, 8)\n end_index, end_bar = divmod(end, 8)\n\n segments[start_index:end_index] = [_Segment(blank, _Style(bgcolor=bar))] * (\n end_index - start_index\n )\n\n if start_index < len(segments):\n segments[start_index] = _Segment(\n bars[7 - start_bar] * depth, _Style(bgcolor=back, color=bar)\n )\n if end_index < len(segments):\n segments[end_index] = _Segment(\n bars[7 - end_bar] * depth, _Style(color=back, bgcolor=bar)\n )\n\n return segments\n\n\nif __name__ == \"__main__\":\n from rich.console import Console\n from rich.segment import Segments\n\n console = Console()\n\n bar = render_bar(\n size=10,\n virtual_size=100,\n window_size=20,\n position=20,\n vertical=True,\n ascii_only=False,\n )\n\n console.print(Segments(bar, new_lines=True))\n", "sub_path": "src/textual/scrollbar.py", "file_name": "scrollbar.py", "file_ext": "py", "file_size_in_byte": 3528, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "rich.segment.Segment", "line_number": 21, "usage_type": "name"}, {"api_name": "rich.console.Console", "line_number": 34, "usage_type": "name"}, {"api_name": "rich.console.ConsoleOptions", "line_number": 34, "usage_type": "name"}, {"api_name": "rich.segment.Segment.line", "line_number": 42, "usage_type": "call"}, {"api_name": "rich.segment.Segment", "line_number": 42, "usage_type": "name"}, {"api_name": "rich.console.RenderResult", "line_number": 35, "usage_type": "name"}, {"api_name": "widget.Widget", "line_number": 49, "usage_type": "name"}, {"api_name": "rich.console.Console", "line_number": 56, "usage_type": "name"}, {"api_name": "rich.console.ConsoleOptions", "line_number": 56, "usage_type": "name"}, {"api_name": "rich.segment.Segments", "line_number": 66, "usage_type": "call"}, {"api_name": "rich.console.RenderableType", "line_number": 56, "usage_type": "name"}, {"api_name": "rich.color.Color.parse", "line_number": 86, "usage_type": "call"}, {"api_name": "rich.color.Color", "line_number": 86, "usage_type": "name"}, {"api_name": "rich.color.Color.parse", "line_number": 87, "usage_type": "call"}, {"api_name": "rich.color.Color", "line_number": 87, "usage_type": "name"}, {"api_name": "rich.segment.Segment", "line_number": 89, "usage_type": "name"}, {"api_name": "rich.style.Style", "line_number": 90, "usage_type": "name"}, {"api_name": "rich.segment.Segment", "line_number": 79, "usage_type": "name"}, {"api_name": "rich.console.Console", "line_number": 122, "usage_type": "call"}, {"api_name": "rich.segment.Segments", "line_number": 133, "usage_type": "call"}]}
+{"seq_id": "645528168", "text": "import os \nimport json\nimport numpy as np\nimport pandas as pd\nimport dill as pickle\nfrom sklearn.externals import joblib\nfrom sklearn.model_selection import train_test_split, GridSearchCV\nfrom sklearn.base import BaseEstimator, TransformerMixin\nfrom sklearn.ensemble import RandomForestClassifier\n\nfrom sklearn.pipeline import make_pipeline\n\nimport warnings\nwarnings.filterwarnings(\"ignore\")\n\n\ndef build_and_train():\n \n data = pd.read_csv('Titanic_dataset/train.csv')\n print('training')\n p = PreProcessing()\n Train_set,label = p.transform(data)\n rf = RandomForestClassifier()\n rf.fit(Train_set,label)\n return(rf)\n\nclass PreProcessing():\n \"\"\"Custom Pre-Processing estimator for our use-case\n \"\"\"\n\n def __init__(self):\n pass\n\n def transform(self, df):\n y_label = df['Survived']\n df = df.drop(['Ticket', 'Cabin','Survived'], axis=1)\n df = df.drop(['Ticket', 'Cabin'], axis=1)\n df['Title'] = df.Name.str.extract(' ([A-Za-z]+)\\.', expand=False)\n df['Title'] = df['Title'].replace(['Lady', 'Countess','Capt', 'Col', 'Dr', 'Major', 'Rev', 'Sir', 'Jonkheer', 'Dona'], 'Rare')\n df['Title'] = df['Title'].replace('Mlle', 'Miss')\n df['Title'] = df['Title'].replace('Ms', 'Miss')\n df['Title'] = df['Title'].replace('Mme', 'Mrs')\n \n title_mapping = {\"Mr\": 1, \"Miss\": 2, \"Mrs\": 3, \"Master\": 4, \"Rare\": 5}\n df['Title'] = df['Title'].map(title_mapping)\n df['Title'] = df['Title'].fillna(0)\n df['Title'] = df['Title'].astype(int)\n \n df = df.drop(['Name', 'PassengerId'], axis=1)\n \n gender_mapping = {'female': 1, 'male': 0}\n df['Sex'] = df['Sex'].map(gender_mapping).astype(int)\n \n guess_ages = np.zeros((2,3))\n for i in range(0, 2):\n for j in range(0, 3):\n guess_df = df[(df['Sex'] == i) & \\\n (df['Pclass'] == j+1)]['Age'].dropna()\n age_guess = guess_df.median()\n guess_ages[i,j] = int( age_guess/0.5 + 0.5 ) * 0.5\n \n for i in range(0, 2):\n for j in range(0, 3):\n df.loc[ (df.Age.isnull()) & (df.Sex == i) & (df.Pclass == j+1),'Age'] = guess_ages[i,j]\n \n \n df['Age'] = df['Age'].astype(int)\n df['AgeBand'] = pd.cut(df['Age'], 5) \n df.loc[ df['Age'] <= 16, 'Age'] = 0\n df.loc[(df['Age'] > 16) & (df['Age'] <= 32), 'Age'] = 1\n df.loc[(df['Age'] > 32) & (df['Age'] <= 48), 'Age'] = 2\n df.loc[(df['Age'] > 48) & (df['Age'] <= 64), 'Age'] = 3\n df.loc[ df['Age'] > 64, 'Age'] \n df = df.drop(['AgeBand'], axis=1)\n \n df['FamilySize'] = df['SibSp'] + df['Parch'] + 1\n df['IsAlone'] = 0\n df.loc[df['FamilySize'] == 1, 'IsAlone'] = 1 \n df = df.drop(['Parch', 'SibSp', 'FamilySize'], axis=1)\n df['Age*Class'] = df.Age * df.Pclass\n \n freq_port = df.Embarked.dropna().mode()[0]\n df['Embarked'] = df['Embarked'].fillna(freq_port)\n df['Embarked'] = df['Embarked'].map( {'S': 0, 'C': 1, 'Q': 2} ).astype(int)\n \n df['FareBand'] = pd.qcut(df['Fare'], 4)\n df.loc[ df['Fare'] <= 7.91, 'Fare'] = 0\n df.loc[(df['Fare'] > 7.91) & (df['Fare'] <= 14.454), 'Fare'] = 1\n df.loc[(df['Fare'] > 14.454) & (df['Fare'] <= 31), 'Fare'] = 2\n df.loc[ df['Fare'] > 31, 'Fare'] = 3\n df['Fare'] = df['Fare'].astype(int)\n df = df.drop(['FareBand'], axis=1)\n \n return df.as_matrix(),y_label\n\nif __name__ == '__main__':\n\n model = build_and_train()\n print('Complete')\n filename = 'model_v2.pk'\n with open('model/'+filename, 'wb') as file:\n pickle.dump(model, file)", "sub_path": "young-shore-49534/util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 3784, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "warnings.filterwarnings", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 54, "usage_type": "call"}, {"api_name": "pandas.cut", "line_number": 68, "usage_type": "call"}, {"api_name": "pandas.qcut", "line_number": 86, "usage_type": "call"}, {"api_name": "dill.dump", "line_number": 102, "usage_type": "call"}]}
+{"seq_id": "419624754", "text": "from selenium.webdriver import ActionChains\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.wait import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions as EC\nfrom components.base_component import BaseComponent\n\n\nclass KeySearchFormLocators:\n def __init__(self):\n self.root = '//div[@class=\"main-search-form\"]'\n self.keyword_selector = '//input[@class=\"keywords-search__input\"]'\n self.search_button = '//button[@class=\"main-search-form__btn\"]'\n self.search_check_box = '//div[@class=\"option-type\"]'\n self.search_check_box_input = 'option-type__checkbox'\n self.search_check_box_name = 'option-type__name'\n\n\nclass KeySearchForm(BaseComponent):\n def __init__(self, driver):\n super(KeySearchForm, self).__init__(driver)\n\n self.wait = WebDriverWait(self.driver, 20)\n self.locators = KeySearchFormLocators()\n\n def input_keyword(self, key: str) -> None:\n element = self.wait.until(\n EC.element_to_be_clickable((By.XPATH, self.locators.keyword_selector)))\n element.clear()\n\n element.send_keys(key)\n\n actions = ActionChains(self.driver)\n actions.move_to_element(element).perform()\n\n def click_on_search(self) -> None:\n element = self.wait.until(\n EC.element_to_be_clickable((By.XPATH, self.locators.search_button)))\n element.click()\n\n def click_on_search_checkbox(self) -> str:\n element = self.wait.until(\n EC.visibility_of_element_located((By.XPATH, self.locators.search_check_box)))\n\n checkBoxInput = element.find_element_by_class_name(self.locators.search_check_box_input)\n checkBoxInput.click()\n checkBoxName = element.find_element_by_class_name(self.locators.search_check_box_name)\n return checkBoxName.get_attribute('innerText')\n", "sub_path": "components/key_search_form.py", "file_name": "key_search_form.py", "file_ext": "py", "file_size_in_byte": 1873, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "components.base_component.BaseComponent", "line_number": 18, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.wait.WebDriverWait", "line_number": 22, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.element_to_be_clickable", "line_number": 27, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 27, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 27, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 27, "usage_type": "name"}, {"api_name": "selenium.webdriver.ActionChains", "line_number": 32, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.element_to_be_clickable", "line_number": 37, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 37, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 37, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 37, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.expected_conditions.visibility_of_element_located", "line_number": 42, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 42, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 42, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 42, "usage_type": "name"}]}
+{"seq_id": "148755323", "text": "import numpy as np\nimport argparse\n\nimport tensorflow as tf\nimport os\n\nfrom tensorflow.contrib import rnn\nfrom tensorflow.contrib.layers import l2_regularizer, xavier_initializer\n\nfrom data.tf_recorfd import build_inputs_from_cifar_tf_record_data, build_inputs_from_numpy_text_names_data\nfrom training.image_classification import build_model_specification, build_eval_model_specification, train\nfrom training.scripts.evaluate import evaluate\n\n\ndef build_rnn(samples):\n num_input = 27 # MNIST data input (img shape: 28*28)\n timesteps = 11 # timesteps\n num_hidden = 100 # hidden layer num of features\n num_classes = 2 # MNIST total classes (0-9 digits)\n\n weights = {\n 'out': tf.Variable(tf.random_normal([num_hidden, num_classes]))\n }\n biases = {\n 'out': tf.Variable(tf.random_normal([num_classes]))\n }\n x = tf.transpose(samples, [0, 2, 1])\n x = tf.unstack(samples, timesteps, 1)\n out = samples\n with tf.variable_scope('rnn'):\n lstm_cell = tf.nn.rnn_cell.BasicRNNCell(100)\n\n output, states = rnn.static_rnn(lstm_cell, x, dtype=tf.float32)\n out = tf.matmul(output[-1], weights['out']) + biases['out']\n\n with tf.variable_scope('fc1'):\n out = tf.layers.dense(inputs=out, units=2, activation=tf.nn.softmax)\n\n return out\n\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"--batch_size\", required=True)\nparser.add_argument(\"--learning_rate\", required=True)\n\nargs = vars(parser.parse_args())\n\ntrain_path = '/home/piotr/Workspace/Projects/pokedex/data/train.npz'\ntest_path = '/home/piotr/Workspace/Projects/pokedex/data/test.npz'\n\nbatch_size = int(args[\"batch_size\"])\n\ntrain_inputs = build_inputs_from_numpy_text_names_data(train_path, batch_size)\ntest_inputs = build_inputs_from_numpy_text_names_data(test_path, batch_size)\n\n# print(train_inputs)\n# variable_init_op = tf.group(*[tf.global_variables_initializer(), tf.tables_initializer()])\n# with tf.Session() as sess:\n# sess.run(variable_init_op)\n# sess.run(train_inputs['iterator_init_op'])\n#\n# print(sess.run(train_inputs['labels']))\n\nimages = train_inputs['images']\nlabels = train_inputs['labels']\n\nTRAIN_SIZE = 22467\nEVAL_SIZE = 5617\n\ntrain_steps = TRAIN_SIZE // batch_size\ntest_steps = EVAL_SIZE // batch_size\n\nlearning_rate = float(args[\"learning_rate\"])\n\nreg = 'l2-0.1'\nopt = 'Adam'\n#model_dir = '/DATA/piotr/cifar/learning/{} op: {} lr: {} batch: {:03d} reg: {}'.format([500], opt, learning_rate, batch_size, reg)\nmodel_dir = '/DATA/piotr/cifar/sprawko/88888'\n# model_dir = '/DATA/piotr/cifar/learning/apaktest'\nif not os.path.exists(model_dir):\n os.makedirs(model_dir)\n\nloss_fn = tf.losses.sparse_softmax_cross_entropy\noptimizer = tf.train.AdamOptimizer()\nmodel_fn = build_rnn\ntrain_spec = build_model_specification(train_inputs, model_fn, loss_fn, optimizer)\neval_spec = build_eval_model_specification(test_inputs, model_fn, loss_fn)\ntrain(train_spec, eval_spec, model_dir, 30, train_steps, test_steps)\n\n# model_dir = '/DATA/piotr/cifar/learning/best'\n# evaluate(eval_spec, model_dir)", "sub_path": "training/scripts/train_text_classification.py", "file_name": "train_text_classification.py", "file_ext": "py", "file_size_in_byte": 3046, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "tensorflow.Variable", "line_number": 22, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 22, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 27, "usage_type": "call"}, {"api_name": "tensorflow.unstack", "line_number": 28, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.nn.rnn_cell.BasicRNNCell", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 31, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.rnn.static_rnn", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.contrib.rnn", "line_number": 33, "usage_type": "name"}, {"api_name": "tensorflow.float32", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.layers.dense", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 37, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 37, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 42, "usage_type": "call"}, {"api_name": "data.tf_recorfd.build_inputs_from_numpy_text_names_data", "line_number": 53, "usage_type": "call"}, {"api_name": "data.tf_recorfd.build_inputs_from_numpy_text_names_data", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path", "line_number": 80, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.losses", "line_number": 83, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 84, "usage_type": "attribute"}, {"api_name": "training.image_classification.build_model_specification", "line_number": 86, "usage_type": "call"}, {"api_name": "training.image_classification.build_eval_model_specification", "line_number": 87, "usage_type": "call"}, {"api_name": "training.image_classification.train", "line_number": 88, "usage_type": "call"}]}
+{"seq_id": "489283046", "text": "# Copyright (c) 2015 by the parties listed in the AUTHORS file.\n# All rights reserved. Use of this source code is governed by \n# a BSD-style license that can be found in the LICENSE file.\n\n\nimport os\nimport unittest\nimport ctypes as ct\nfrom ctypes.util import find_library\n\nif 'TOAST_NO_MPI' in os.environ.keys():\n from .. import fakempi as MPI\nelse:\n from mpi4py import MPI\n\nimport healpy as hp\nimport numpy as np\nimport numpy.ctypeslib as npc\n\nfrom . import qarray as qa\nfrom ..dist import Comm, Data\nfrom ..operator import Operator\nfrom ..tod import TOD\nfrom ..tod import Interval\nfrom ..tod import quat2angle\n\n# Define portably the MPI communicator datatype\n\ntry:\n if MPI._sizeof(MPI.Comm) == ct.sizeof(ct.c_int):\n MPI_Comm = ct.c_int\n else:\n MPI_Comm = ct.c_void_p\nexcept Exception as e:\n raise Exception(\n 'Failed to set the portable MPI communicator datatype. MPI4py is '\n 'probably too old. You need to have at least version 2.0. ({})'\n ''.format(e))\n\nlibconviqt = None\nif 'TOAST_NO_MPI' not in os.environ.keys():\n try:\n libconviqt = ct.CDLL('libconviqt.so')\n except:\n path = find_library('conviqt')\n if path is not None:\n libconviqt = ct.CDLL(path)\n\nif libconviqt is not None:\n # Beam functions\n\n libconviqt.conviqt_beam_new.restype = ct.c_void_p\n libconviqt.conviqt_beam_new.argtypes = []\n\n libconviqt.conviqt_beam_del.restype = ct.c_int\n libconviqt.conviqt_beam_del.argtypes = [ct.c_void_p]\n\n libconviqt.conviqt_beam_read.restype = ct.c_int\n libconviqt.conviqt_beam_read.argtypes = [\n ct.c_void_p,\n ct.c_long,\n ct.c_long,\n ct.c_byte,\n ct.c_char_p,\n MPI_Comm\n ]\n\n libconviqt.conviqt_beam_lmax.restype = ct.c_int\n libconviqt.conviqt_beam_lmax.argtypes = [ct.c_void_p]\n\n libconviqt.conviqt_beam_mmax.restype = ct.c_int\n libconviqt.conviqt_beam_mmax.argtypes = [ct.c_void_p]\n\n # Sky functions\n\n libconviqt.conviqt_sky_new.restype = ct.c_void_p\n libconviqt.conviqt_sky_new.argtypes = []\n\n libconviqt.conviqt_sky_del.restype = ct.c_int\n libconviqt.conviqt_sky_del.argtypes = [ct.c_void_p]\n\n libconviqt.conviqt_sky_read.restype = ct.c_int\n libconviqt.conviqt_sky_read.argtypes = [\n ct.c_void_p,\n ct.c_long,\n ct.c_byte,\n ct.c_char_p,\n ct.c_double,\n MPI_Comm\n ]\n\n libconviqt.conviqt_sky_lmax.restype = ct.c_int\n libconviqt.conviqt_sky_lmax.argtypes = [ct.c_void_p]\n\n libconviqt.conviqt_sky_remove_monopole.restype = ct.c_int\n libconviqt.conviqt_sky_remove_monopole.argtypes = [ct.c_void_p]\n\n libconviqt.conviqt_sky_remove_dipole.restype = ct.c_int\n libconviqt.conviqt_sky_remove_dipole.argtypes = [ct.c_void_p]\n\n # Detector functions\n\n libconviqt.conviqt_detector_new.restype = ct.c_void_p\n libconviqt.conviqt_detector_new.argtypes = []\n\n libconviqt.conviqt_detector_new_with_id.restype = ct.c_void_p\n libconviqt.conviqt_detector_new_with_id.argtypes = [ct.c_char_p]\n\n libconviqt.conviqt_detector_del.restype = ct.c_int\n libconviqt.conviqt_detector_del.argtypes = [ct.c_void_p]\n\n libconviqt.conviqt_detector_set_epsilon.restype = ct.c_int\n libconviqt.conviqt_detector_set_epsilon.argtypes = [\n ct.c_void_p,\n ct.c_double\n ]\n\n libconviqt.conviqt_detector_get_epsilon.restype = ct.c_int\n libconviqt.conviqt_detector_get_epsilon.argtypes = [\n ct.c_void_p,\n ct.POINTER(ct.c_double)\n ]\n\n libconviqt.conviqt_detector_get_id.restype = ct.c_int\n libconviqt.conviqt_detector_get_id.argtypes = [\n ct.c_void_p,\n ct.c_char_p\n ]\n\n # Pointing functions\n\n libconviqt.conviqt_pointing_new.restype = ct.c_void_p\n libconviqt.conviqt_pointing_new.argtypes = []\n\n libconviqt.conviqt_pointing_del.restype = ct.c_int\n libconviqt.conviqt_pointing_del.argtypes = [ct.c_void_p]\n\n libconviqt.conviqt_pointing_alloc.restype = ct.c_int\n libconviqt.conviqt_pointing_alloc.argtypes = [\n ct.c_void_p,\n ct.c_long\n ]\n\n libconviqt.conviqt_pointing_data.restype = ct.POINTER(ct.c_double)\n libconviqt.conviqt_pointing_data.argtypes = [ct.c_void_p]\n\n # Convolver functions\n\n libconviqt.conviqt_convolver_new.restype = ct.c_void_p\n libconviqt.conviqt_convolver_new.argtypes = [\n ct.c_void_p,\n ct.c_void_p,\n ct.c_void_p,\n ct.c_byte,\n ct.c_long,\n ct.c_long,\n ct.c_long,\n MPI_Comm\n ]\n\n libconviqt.conviqt_convolver_convolve.restype = ct.c_int\n libconviqt.conviqt_convolver_convolve.argtypes = [\n ct.c_void_p,\n ct.c_void_p,\n ct.c_byte\n ]\n\n libconviqt.conviqt_convolver_del.restype = ct.c_int\n libconviqt.conviqt_convolver_del.argtypes = [ct.c_void_p]\n\n\nclass OpSimConviqt(Operator):\n \"\"\"\n Operator which uses libconviqt to generate beam-convolved timestreams.\n\n This passes through each observation and loops over each detector.\n For each detector, it produces the beam-convolved timestream.\n\n Args:\n lmax (int): Maximum ell (and m). Actual resolution in the Healpix FITS\n file may differ.\n beammmax (int): beam maximum m. Actual resolution in the Healpix FITS file\n may differ.\n detectordata (list): list of (detector_name, detector_sky_file,\n detector_beam_file, epsilon, psipol[radian]) tuples\n pol (bool) : boolean to determine if polarized simulation is needed\n fwhm (float) : width of a symmetric gaussian beam [in arcmin] already\n present in the skyfile (will be deconvolved away).\n order (int) : conviqt order parameter (expert mode)\n calibrate (bool) : Calibrate intensity to 1.0, rather than (1+epsilon)/2\n dxx (bool) : The beam frame is either Dxx or Pxx. Pxx includes the\n rotation to polarization sensitive basis, Dxx does not. When\n Dxx=True, detector orientation from attitude quaternions is\n corrected for the polarization angle.\n out (str): the name of the cache object (_) to\n use for output of the detector timestream.\n \"\"\"\n\n def __init__(\n self, lmax, beammmax, detectordata, pol=True, fwhm=4.0, order=13,\n calibrate=True, dxx=True, out='conviqt', quat_name=None,\n flag_name=None, flag_mask=255, common_flag_name=None,\n common_flag_mask=255, apply_flags=False,\n remove_monopole=False, remove_dipole=False):\n\n # We call the parent class constructor, which currently does nothing\n super().__init__()\n\n self._lmax = lmax\n self._beammmax = beammmax\n self._detectordata = {}\n for entry in detectordata:\n self._detectordata[entry[0]] = entry[1:]\n self._pol = pol\n self._fwhm = fwhm\n self._order = order\n self._calibrate = calibrate\n self._dxx = dxx\n self._quat_name = quat_name\n self._flag_name = flag_name\n self._flag_mask = flag_mask\n self._common_flag_name = common_flag_name\n self._common_flag_mask = common_flag_mask\n self._apply_flags = apply_flags\n self._remove_monopole = remove_monopole\n self._remove_dipole = remove_dipole\n\n self._out = out\n\n @property\n def available(self):\n \"\"\"\n (bool): True if libconviqt is found in the library search path.\n \"\"\"\n return (libconviqt is not None)\n\n def exec(self, data):\n \"\"\"\n Loop over all observations and perform the convolution.\n\n This is done one detector at a time. For each detector, all data\n products are read from disk.\n\n Args:\n data (toast.Data): The distributed data.\n \"\"\"\n if libconviqt is None:\n raise RuntimeError(\"The conviqt library was not found\")\n\n # the two-level pytoast communicator\n #comm = data.comm\n # the global communicator\n #cworld = comm.comm_world\n # the communicator within the group\n #cgroup = comm.comm_group\n # the communicator with all processes with\n # the same rank within their group\n #crank = comm.comm_rank\n\n xaxis, yaxis, zaxis = np.eye(3)\n nullquat = np.array([0, 0, 0, 1], dtype=np.float64)\n\n for obs in data.obs:\n tod = obs['tod']\n intrvl = obs['intervals']\n\n comm_ptr = MPI._addressof(tod.mpicomm)\n comm = MPI_Comm.from_address(comm_ptr)\n\n for det in tod.local_dets:\n try:\n skyfile, beamfile, epsilon, psipol = self._detectordata[det]\n except:\n raise Exception(\n 'ERROR: conviqt object not initialized to convolve '\n 'detector {}. Available detectors are {}'.format(\n det, self._detectordata.keys()))\n \n sky = libconviqt.conviqt_sky_new()\n err = libconviqt.conviqt_sky_read(\n sky, self._lmax, self._pol, skyfile.encode(), self._fwhm,\n comm)\n if err != 0:\n raise RuntimeError('Failed to load ' + skyfile)\n if self._remove_monopole:\n err = libconviqt.conviqt_sky_remove_monopole(sky)\n if err != 0: raise RuntimeError('Failed to remove monopole')\n if self._remove_dipole:\n err = libconviqt.conviqt_sky_remove_dipole(sky)\n if err != 0: raise RuntimeError('Failed to remove dipole')\n\n beam = libconviqt.conviqt_beam_new()\n err = libconviqt.conviqt_beam_read(\n beam, self._lmax, self._beammmax,\n self._pol, beamfile.encode(), comm)\n if err != 0:\n raise Exception('Failed to load ' + beamfile)\n\n detector = libconviqt.conviqt_detector_new_with_id(det.encode())\n libconviqt.conviqt_detector_set_epsilon(detector, epsilon)\n \n # We need the three pointing angles to describe the\n # pointing. read_pntg returns the attitude quaternions.\n if self._quat_name is not None:\n cachename = '{}_{}'.format(self._quat_name, det)\n pdata = tod.cache.reference(cachename).copy()\n else:\n pdata = tod.read_pntg(detector=det).copy()\n\n if self._apply_flags:\n common, flags = None, None\n if self._common_flag_name is not None:\n common = tod.cache.reference(self._common_flag_name)\n if self._flag_name is not None:\n cachename = '{}_{}'.format(self._flag_name, det)\n flags = tod.cache.reference(cachename)\n else:\n flags, common_temp = tod.read_flags(detector=det)\n if common is None: common = common_temp\n if common is None:\n common = tod.read_common_flags()\n common = (common & self._common_flag_mask)\n flags = (flags & self._flag_mask)\n totflags = np.copy(flags)\n totflags |= common\n pdata[totflags != 0] = nullquat\n\n theta, phi, psi = quat2angle(pdata)\n \n # Is the psi angle in Pxx or Dxx? Pxx will include the\n # detector polarization angle, Dxx will not.\n\n if self._dxx:\n psi -= psipol\n\n pnt = libconviqt.conviqt_pointing_new()\n\n err = libconviqt.conviqt_pointing_alloc(pnt,\n tod.local_samples[1]*5)\n if err != 0:\n raise Exception('Failed to allocate pointing array')\n\n ppnt = libconviqt.conviqt_pointing_data(pnt)\n\n for row in range(tod.local_samples[1]):\n ppnt[row*5 + 0] = phi[row]\n ppnt[row*5 + 1] = theta[row]\n ppnt[row*5 + 2] = psi[row]\n # This column will host the convolved data upon exit\n ppnt[row*5 + 3] = 0\n # libconviqt will assign the running indices to this column.\n ppnt[row*5 + 4] = 0\n\n convolver = libconviqt.conviqt_convolver_new(\n sky, beam, detector, self._pol, self._lmax, self._beammmax,\n self._order, comm)\n\n if convolver is None:\n raise Exception(\"Failed to instantiate convolver\")\n\n err = libconviqt.conviqt_convolver_convolve(convolver, pnt,\n self._calibrate)\n if err != 0:\n raise Exception('Convolution FAILED!')\n\n # The pointer to the data will have changed during\n # the convolution call ...\n\n ppnt = libconviqt.conviqt_pointing_data(pnt)\n\n convolved_data = np.zeros(tod.local_samples[1])\n for row in range(tod.local_samples[1]):\n convolved_data[row] = ppnt[row*5 + 3]\n\n libconviqt.conviqt_convolver_del(convolver)\n\n cachename = \"{}_{}\".format(self._out, det)\n if not tod.cache.exists(cachename):\n tod.cache.create(cachename, np.float64,\n (tod.local_samples[1],))\n ref = tod.cache.reference(cachename)\n if ref.size != convolved_data.size:\n raise RuntimeError(\n '{} already exists in tod.cache but has wrong size: {} '\n '!= {}'.format(cachename, ref.size, convolved_data.size))\n ref[:] += convolved_data\n\n libconviqt.conviqt_pointing_del(pnt)\n libconviqt.conviqt_detector_del(detector)\n libconviqt.conviqt_beam_del(beam)\n libconviqt.conviqt_sky_del(sky)\n\n return\n", "sub_path": "toast/tod/conviqt.py", "file_name": "conviqt.py", "file_ext": "py", "file_size_in_byte": 14279, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "os.environ.keys", "line_number": 11, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 11, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI._sizeof", "line_number": 30, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 30, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Comm", "line_number": 30, "usage_type": "attribute"}, {"api_name": "ctypes.sizeof", "line_number": 30, "usage_type": "call"}, {"api_name": "ctypes.c_int", "line_number": 30, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 31, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.environ.keys", "line_number": 41, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 41, "usage_type": "attribute"}, {"api_name": "ctypes.CDLL", "line_number": 43, "usage_type": "call"}, {"api_name": "ctypes.util.find_library", "line_number": 45, "usage_type": "call"}, {"api_name": "ctypes.CDLL", "line_number": 47, "usage_type": "call"}, {"api_name": "ctypes.c_void_p", "line_number": 52, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 55, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 56, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 58, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 60, "usage_type": "attribute"}, {"api_name": "ctypes.c_long", "line_number": 61, "usage_type": "attribute"}, {"api_name": "ctypes.c_long", "line_number": 62, "usage_type": "attribute"}, {"api_name": "ctypes.c_byte", "line_number": 63, "usage_type": "attribute"}, {"api_name": "ctypes.c_char_p", "line_number": 64, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 68, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 69, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 71, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 72, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 76, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 79, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 80, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 82, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 84, "usage_type": "attribute"}, {"api_name": "ctypes.c_long", "line_number": 85, "usage_type": "attribute"}, {"api_name": "ctypes.c_byte", "line_number": 86, "usage_type": "attribute"}, {"api_name": "ctypes.c_char_p", "line_number": 87, "usage_type": "attribute"}, {"api_name": "ctypes.c_double", "line_number": 88, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 92, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 93, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 95, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 96, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 98, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 99, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 103, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 106, "usage_type": "attribute"}, {"api_name": "ctypes.c_char_p", "line_number": 107, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 109, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 110, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 112, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 114, "usage_type": "attribute"}, {"api_name": "ctypes.c_double", "line_number": 115, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 118, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 120, "usage_type": "attribute"}, {"api_name": "ctypes.POINTER", "line_number": 121, "usage_type": "call"}, {"api_name": "ctypes.c_double", "line_number": 121, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 124, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 126, "usage_type": "attribute"}, {"api_name": "ctypes.c_char_p", "line_number": 127, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 132, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 135, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 136, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 138, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 140, "usage_type": "attribute"}, {"api_name": "ctypes.c_long", "line_number": 141, "usage_type": "attribute"}, {"api_name": "ctypes.POINTER", "line_number": 144, "usage_type": "call"}, {"api_name": "ctypes.c_double", "line_number": 144, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 145, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 149, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 151, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 152, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 153, "usage_type": "attribute"}, {"api_name": "ctypes.c_byte", "line_number": 154, "usage_type": "attribute"}, {"api_name": "ctypes.c_long", "line_number": 155, "usage_type": "attribute"}, {"api_name": "ctypes.c_long", "line_number": 156, "usage_type": "attribute"}, {"api_name": "ctypes.c_long", "line_number": 157, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 161, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 163, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 164, "usage_type": "attribute"}, {"api_name": "ctypes.c_byte", "line_number": 165, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 168, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 169, "usage_type": "attribute"}, {"api_name": "operator.Operator", "line_number": 172, "usage_type": "name"}, {"api_name": "numpy.eye", "line_number": 260, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 261, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 261, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI._addressof", "line_number": 267, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 267, "usage_type": "name"}, {"api_name": "tod.mpicomm", "line_number": 267, "usage_type": "attribute"}, {"api_name": "tod.local_dets", "line_number": 270, "usage_type": "attribute"}, {"api_name": "tod.cache.reference", "line_number": 306, "usage_type": "call"}, {"api_name": "tod.cache", "line_number": 306, "usage_type": "attribute"}, {"api_name": "tod.read_pntg", "line_number": 308, "usage_type": "call"}, {"api_name": "tod.cache.reference", "line_number": 313, "usage_type": "call"}, {"api_name": "tod.cache", "line_number": 313, "usage_type": "attribute"}, {"api_name": "tod.cache.reference", "line_number": 316, "usage_type": "call"}, {"api_name": "tod.cache", "line_number": 316, "usage_type": "attribute"}, {"api_name": "tod.read_flags", "line_number": 318, "usage_type": "call"}, {"api_name": "tod.read_common_flags", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 324, "usage_type": "call"}, {"api_name": "tod.quat2angle", "line_number": 328, "usage_type": "call"}, {"api_name": "tod.local_samples", "line_number": 339, "usage_type": "attribute"}, {"api_name": "tod.local_samples", "line_number": 345, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 371, "usage_type": "call"}, {"api_name": "tod.local_samples", "line_number": 371, "usage_type": "attribute"}, {"api_name": "tod.local_samples", "line_number": 372, "usage_type": "attribute"}, {"api_name": "tod.cache.exists", "line_number": 378, "usage_type": "call"}, {"api_name": "tod.cache", "line_number": 378, "usage_type": "attribute"}, {"api_name": "tod.cache.create", "line_number": 379, "usage_type": "call"}, {"api_name": "tod.cache", "line_number": 379, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 379, "usage_type": "attribute"}, {"api_name": "tod.local_samples", "line_number": 380, "usage_type": "attribute"}, {"api_name": "tod.cache.reference", "line_number": 381, "usage_type": "call"}, {"api_name": "tod.cache", "line_number": 381, "usage_type": "attribute"}]}
+{"seq_id": "487382203", "text": "import contextlib\nimport io\nimport unittest\nfrom pathlib import Path\nfrom unittest import mock\n\nfrom sr.comp.cli.for_each_match import (\n command,\n PlaceholderExpander,\n replace_placeholders,\n)\nfrom sr.comp.types import TLA\n\nfrom .factories import build_match\n\n\nclass ForEachMatchTests(unittest.TestCase):\n longMessage = True\n maxDiff = None\n\n def test_smoke(self) -> None:\n compstate_path = str(Path(__file__).parent / 'dummy')\n\n mock_settings = mock.Mock(\n compstate=compstate_path,\n arena=None,\n matches=set([0, 2, 3]),\n command=['spam', '{TYPE}:{ARENA}', '{NUMBER}|{TLAS}'],\n )\n\n with mock.patch('subprocess.check_call') as mock_check_call:\n command(mock_settings)\n\n mock_check_call.assert_has_calls([\n mock.call(['spam', 'league:A', '0|- CLY TTN -']),\n mock.call(['spam', 'league:B', '0|GRS QMC - -']),\n mock.call(['spam', 'league:A', '2|ICE MFG SWI BRN']),\n mock.call(['spam', 'league:B', '2|TBG EMM SGS GYG']),\n mock.call(['spam', 'league:A', '3|MAI2 HSO KDE CCR']),\n mock.call(['spam', 'league:B', '3|SCC LSS HZW MAI']),\n ])\n\n def test_validate_placeholders(self) -> None:\n with contextlib.redirect_stderr(io.StringIO()) as stderr:\n PlaceholderExpander.validate('fine')\n self.assertEqual(\"\", stderr.getvalue())\n\n with contextlib.redirect_stderr(io.StringIO()) as stderr:\n PlaceholderExpander.validate('@unknown')\n self.assertEqual(\n \"Warning: unrecognised value '@unknown'.\\n\",\n stderr.getvalue(),\n )\n\n with contextlib.redirect_stderr(io.StringIO()) as stderr:\n PlaceholderExpander.validate('@TLAS')\n self.assertEqual(\"\", stderr.getvalue())\n\n def test_replace_placeholders(self) -> None:\n match = build_match(num=42, teams=[TLA('ABC'), None])\n\n command = replace_placeholders(match, [\n 'spam',\n '{NUMBER}:{ARENA}',\n '{TLAS}',\n '@TLAS',\n '@TLAS|',\n ])\n\n self.assertEqual(\n ['spam', '42:main', 'ABC -', 'ABC', '-', '@TLAS|'],\n command,\n )\n", "sub_path": "tests/test_for_each_match.py", "file_name": "test_for_each_match.py", "file_ext": "py", "file_size_in_byte": 2272, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "unittest.TestCase", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 22, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 24, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 24, "usage_type": "name"}, {"api_name": "unittest.mock.patch", "line_number": 31, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 31, "usage_type": "name"}, {"api_name": "sr.comp.cli.for_each_match.command", "line_number": 32, "usage_type": "call"}, {"api_name": "unittest.mock.call", "line_number": 35, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 35, "usage_type": "name"}, {"api_name": "unittest.mock.call", "line_number": 36, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 36, "usage_type": "name"}, {"api_name": "unittest.mock.call", "line_number": 37, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 37, "usage_type": "name"}, {"api_name": "unittest.mock.call", "line_number": 38, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 38, "usage_type": "name"}, {"api_name": "unittest.mock.call", "line_number": 39, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 39, "usage_type": "name"}, {"api_name": "unittest.mock.call", "line_number": 40, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 40, "usage_type": "name"}, {"api_name": "contextlib.redirect_stderr", "line_number": 44, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 44, "usage_type": "call"}, {"api_name": "sr.comp.cli.for_each_match.PlaceholderExpander.validate", "line_number": 45, "usage_type": "call"}, {"api_name": "sr.comp.cli.for_each_match.PlaceholderExpander", "line_number": 45, "usage_type": "name"}, {"api_name": "contextlib.redirect_stderr", "line_number": 48, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 48, "usage_type": "call"}, {"api_name": "sr.comp.cli.for_each_match.PlaceholderExpander.validate", "line_number": 49, "usage_type": "call"}, {"api_name": "sr.comp.cli.for_each_match.PlaceholderExpander", "line_number": 49, "usage_type": "name"}, {"api_name": "contextlib.redirect_stderr", "line_number": 55, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 55, "usage_type": "call"}, {"api_name": "sr.comp.cli.for_each_match.PlaceholderExpander.validate", "line_number": 56, "usage_type": "call"}, {"api_name": "sr.comp.cli.for_each_match.PlaceholderExpander", "line_number": 56, "usage_type": "name"}, {"api_name": "factories.build_match", "line_number": 60, "usage_type": "call"}, {"api_name": "sr.comp.types.TLA", "line_number": 60, "usage_type": "call"}, {"api_name": "sr.comp.cli.for_each_match.command", "line_number": 62, "usage_type": "name"}, {"api_name": "sr.comp.cli.for_each_match.replace_placeholders", "line_number": 62, "usage_type": "call"}, {"api_name": "sr.comp.cli.for_each_match.command", "line_number": 72, "usage_type": "argument"}]}
+{"seq_id": "402650195", "text": "#!/usr/bin/python3\n\nimport argparse\nimport sys\n\nimport base64\nimport requests\n\nimport imghdr\nfrom PIL import Image\nfrom io import BytesIO\n\n\ndef encode_image(img_path: str) -> str:\n try:\n with open(img_path, \"rb\") as image_file:\n return base64.b64encode(image_file.read()).decode()\n except Exception as e:\n sys.stderr.write('ERROR: %s\\n\\n' % str(e))\n sys.exit(2)\n\ndef send_request(url: str, data: dict) -> str:\n try:\n response = requests.get(url, json=data)\n response.raise_for_status()\n return response.json()['image']\n except Exception as e:\n sys.stderr.write('ERROR: %s\\n\\n' % str(e))\n sys.exit(2)\n\ndef save_image(encoded_img: str, img_path: str) -> None:\n try:\n decoded_file = base64.b64decode(encoded_img)\n file_format = imghdr.what(None, h=decoded_file)\n img = Image.open(BytesIO(decoded_file))\n img.save(img_path, file_format)\n except Exception as e:\n sys.stderr.write('ERROR: %s\\n\\n' % str(e))\n sys.exit(2)\n\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"-i\", \"--input\", metavar='PATH', help=\"input image\", required=True)\nparser.add_argument(\"-o\", \"--output\", metavar='PATH', help=\"output image\", required=True)\nparser.add_argument(\"-u\", \"--url\", metavar='URL', help=\"URL to REST API\", required=True)\nparser.add_argument(\"--resize\", nargs=2, metavar=('WIDTH', 'HEIGHT'), type=int)\nparser.add_argument(\"--interpolation\", help=\"if resize is selected\")\nparser.add_argument(\"--crop\", nargs=4, metavar=('X_BEGIN', 'X_END', 'Y_BEGIN', 'Y_END'), type=int)\nparser.add_argument(\"--rotate\", metavar='ANGLE_IN_DEGREES', type=int)\nparser.add_argument(\"--scale\", help=\"if rotate is selected\", type=float)\nparser.add_argument(\"--negative\", action=\"store_true\")\nargs = parser.parse_args()\n\nif args.resize:\n url_augmentation = f\"resize?width={args.resize[0]}&height={args.resize[1]}\"\n if args.interpolation:\n url_augmentation += f\"&interpolation={args.interpolation}\"\nelif args.crop:\n url_augmentation = f\"crop?xBegin={args.crop[0]}&xEnd={args.crop[1]}&yBegin={args.crop[2]}&yEnd={args.crop[3]}\"\nelif args.rotate:\n url_augmentation = f\"rotate?angle={args.rotate}\"\n if args.scale:\n url_augmentation += f\"&scale={args.scale}\"\nelif args.negative:\n url_augmentation = \"negative\"\nelse:\n message = 'You need to choose one of the following: --resize, --crop, --rotate, --negative'\n sys.stderr.write('ERROR: %s\\n\\n' % message)\n parser.print_help()\n sys.exit(2)\n\nencoded_img = encode_image(args.input)\n\ndata = {'image': encoded_img}\nurl = f\"{args.url}/augmentation/{url_augmentation}\"\n\nencoded_img = send_request(url, data)\nsave_image(encoded_img, args.output)\n", "sub_path": "testing_script/get_augmentation.py", "file_name": "get_augmentation.py", "file_ext": "py", "file_size_in_byte": 2724, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "base64.b64encode", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 19, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 20, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 24, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 29, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 33, "usage_type": "call"}, {"api_name": "imghdr.what", "line_number": 34, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 35, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 35, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 35, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 38, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 38, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 39, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 42, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 68, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 68, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 70, "usage_type": "call"}]}
+{"seq_id": "416438224", "text": "from car.models import CarModel\nfrom rest_framework import viewsets, serializers, status\nfrom rest_framework.response import Response\nimport datetime\n\nfrom .fine import FineSerializer\nfrom ..models import Contract, Fine\n\n\nclass ContractSerializer(serializers.ModelSerializer):\n status = serializers.SerializerMethodField(method_name='get_status')\n status_color = serializers.SerializerMethodField(method_name='get_status_color')\n car = serializers.SlugRelatedField(slug_field='name', many=False, read_only=True)\n\n @staticmethod\n def get_status(obj):\n return obj.get_status_display()\n\n @staticmethod\n def get_status_color(obj):\n status_color_by_tag = {\n 'En proceso': 'blue',\n 'Activo': 'green',\n 'Finalizado': 'black',\n 'Cancelado': 'red',\n }\n return status_color_by_tag[obj.get_status_display()]\n\n class Meta:\n model = Contract\n fields = ['guid', 'monthly_cost', 'annual_mileage', 'duration', 'start_date', 'reject_date', 'bank_account',\n 'status', 'status_color', 'user', 'car', 'car_color', 'creation_datetime']\n\n\nclass ContractViewSet(viewsets.ModelViewSet):\n lookup_field = 'guid'\n queryset = Contract.objects.all()\n serializer_class = ContractSerializer\n http_method_names = ['get', 'post', 'put', 'delete']\n\n def create(self, request, *args, **kwargs):\n duration_increment = {12: 1.3, 24: 1.2, 36: 1.1, 48: 1, 60: 1}\n km_increment = {15000: 1, 20000: 1, 25000: 1.1, 30000: 1.2, 35000: 1.3, 40000: 1.4, 45000: 1.5}\n user = request.user\n car = CarModel.objects.get(guid=request.data['carGuid'])\n price = car.base_price * duration_increment[request.data['duration']] * km_increment[request.data['km']]\n Contract.objects.create(monthly_cost=price, duration=request.data['duration'],\n annual_mileage=request.data['km'], car_color=request.data['carColor'],\n car_id=car.id, user_id=user.id, status='E', bank_account=request.data['account'])\n return Response(status=status.HTTP_201_CREATED)\n\n def list(self, request, *args, **kwargs):\n user = request.user\n contracts = Contract.objects.filter(user__user_info__guid=user.user_info.guid)\n serializer = self.get_serializer(contracts, many=True)\n return Response(status=status.HTTP_200_OK, data=serializer.data)\n\n def destroy(self, request, *args, **kwargs):\n contract_to_delete = self.get_object()\n if contract_to_delete.status == 'E':\n contract_to_delete.delete()\n return Response(status=status.HTTP_200_OK)\n else:\n contract_to_delete.status = 'C'\n contract_to_delete.reject_date = datetime.date.today()\n today = datetime.date.today()\n duration_to_end = (today.year - contract_to_delete.start_date.year) * 12 \\\n + today.month - contract_to_delete.start_date.month\n fine_price = duration_to_end * contract_to_delete.monthly_cost * 0.75\n fine = Fine.objects.create(description=request.query_params['description'],\n contract_id=contract_to_delete.id,\n cost=fine_price)\n contract_to_delete.save()\n serializer = {\n 'description': fine.description,\n 'cost': fine.cost,\n 'creation_datetime': fine.creation_datetime,\n 'contract': fine.contract.guid,\n 'pay_date': ''\n }\n return Response(status=status.HTTP_200_OK, data=serializer)\n", "sub_path": "backend/contract/rest/contract.py", "file_name": "contract.py", "file_ext": "py", "file_size_in_byte": 3689, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 10, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 10, "usage_type": "name"}, {"api_name": "rest_framework.status", "line_number": 11, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 11, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 11, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 12, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 12, "usage_type": "name"}, {"api_name": "car.models", "line_number": 13, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SlugRelatedField", "line_number": 13, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 13, "usage_type": "name"}, {"api_name": "models.Contract", "line_number": 30, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 35, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 35, "usage_type": "name"}, {"api_name": "models.Contract.objects.all", "line_number": 37, "usage_type": "call"}, {"api_name": "models.Contract.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.Contract", "line_number": 37, "usage_type": "name"}, {"api_name": "car.models", "line_number": 45, "usage_type": "name"}, {"api_name": "car.models.CarModel.objects.get", "line_number": 45, "usage_type": "call"}, {"api_name": "car.models.CarModel.objects", "line_number": 45, "usage_type": "attribute"}, {"api_name": "car.models.CarModel", "line_number": 45, "usage_type": "name"}, {"api_name": "car.models.base_price", "line_number": 46, "usage_type": "attribute"}, {"api_name": "car.models", "line_number": 46, "usage_type": "name"}, {"api_name": "models.Contract.objects.create", "line_number": 47, "usage_type": "call"}, {"api_name": "models.Contract.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "models.Contract", "line_number": 47, "usage_type": "name"}, {"api_name": "car.models.id", "line_number": 49, "usage_type": "attribute"}, {"api_name": "car.models", "line_number": 49, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 50, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 50, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 50, "usage_type": "name"}, {"api_name": "models.Contract.objects.filter", "line_number": 54, "usage_type": "call"}, {"api_name": "models.Contract.objects", "line_number": 54, "usage_type": "attribute"}, {"api_name": "models.Contract", "line_number": 54, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 56, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 56, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 56, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 62, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 62, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 62, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 65, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 65, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 66, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 66, "usage_type": "attribute"}, {"api_name": "models.Fine.objects.create", "line_number": 70, "usage_type": "call"}, {"api_name": "models.Fine.objects", "line_number": 70, "usage_type": "attribute"}, {"api_name": "models.Fine", "line_number": 70, "usage_type": "name"}, {"api_name": "fine.description", "line_number": 75, "usage_type": "attribute"}, {"api_name": "fine.cost", "line_number": 76, "usage_type": "attribute"}, {"api_name": "fine.creation_datetime", "line_number": 77, "usage_type": "attribute"}, {"api_name": "fine.contract", "line_number": 78, "usage_type": "attribute"}, {"api_name": "rest_framework.response.Response", "line_number": 81, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 81, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 81, "usage_type": "name"}]}
+{"seq_id": "381799441", "text": "import math\nimport random\nimport numpy as np\nfrom Utils.utils import Util as util\nfrom matplotlib import pyplot as plt\nfrom sklearn.metrics import accuracy_score as acc\n#TPR: (Sensitivity, hit rate, recall)\nfrom sklearn.metrics import recall_score as tpr\n#TNR=SPC: (Specificity)\n#PPV: Pos Pred Value (Precision)\nfrom sklearn.metrics import precision_score as ppv\n\nclass ELM:\n def __init__(self, x_data, y_data, activation='logistic', g_search=False, hidden_layer=10):\n self.x_data = x_data\n self.y_data = y_data\n self.n_classes = np.unique(self.y_data, axis=0)\n self.g_search = g_search\n self.attributes = x_data.shape[1]\n self.hidden_layer = hidden_layer\n self.output_layer = y_data.shape[1]\n self.epochs = 500\n self.realizations = 1\n self.precision = 10**(-5)\n self.train_size = 0.8\n self.activation = activation\n self.hit_rate = []\n self.tpr = []\n self.spc = []\n self.ppv = []\n \n def initWeigths(self, hidden_layer):\n params = {}\n a = -1.5\n b = 1.5\n params['w'] = (b - a) * np.random.random_sample((self.attributes+1, hidden_layer)) + a\n params['m'] = (b - a) * np.random.random_sample((hidden_layer+1, self.output_layer)) + a\n return params\n\n def updateEta(self, epoch):\n eta_i = 0.1\n eta_f = 0.05\n eta = eta_i * ((eta_f / eta_i) ** (epoch / self.epochs))\n self.eta = eta\n\n def function(self, u):\n if self.activation == 'logistic':\n y = 1.0/(1.0 + np.exp(-u))\n elif self.activation == 'tanh':\n y = (np.exp(u) - np.exp(-u))/(np.exp(u) + np.exp(-u))\n else:\n raise ValueError('Error in function!')\n y = 0\n return y \n\n def derivate(self, u):\n if self.activation == 'logistic':\n y_ = u * (1.0 - u)\n elif self.activation == 'tanh':\n y_ = 0.5 * (1.0 - (u * u))\n else:\n raise ValueError('Error in derivate!')\n y_ = 0\n return y_\n\n def activationFunction(self, u):\n value = np.amax(u)\n y = np.where(u == value, 1, 0)\n return y\n\n def predict(self, xi, params):\n w = params['w']\n m = params['m']\n \n H = np.dot(xi, w)\n H = self.function(H)\n H = np.concatenate(([-1], H), axis=None)\n H = H.reshape(1,-1)\n\n Y = np.dot(H, m)\n Y = self.function(Y)\n y = self.activationFunction(Y)\n\n return y[0]\n\n def train(self, x_train, y_train, hidden_layer):\n error_old = 0\n cont_epochs = 0\n mse_vector = []\n params = self.initWeigths(hidden_layer)\n w = params['w']\n m = params['m']\n\n H = np.dot(x_train, w)\n H = self.function(H)\n\n # Bias\n (m, _) = H.shape\n bias = -1 * np.ones((m, 1))\n H = np.concatenate((bias, H), axis=1)\n\n H_pinv = np.linalg.pinv(H)\n m = np.dot(H_pinv, y_train)\n\n params['w'] = w\n params['m'] = m\n return params\n\n def test(self, x_test, y_test, params):\n y_true = []\n y_pred = []\n (p, _) = x_test.shape\n for k in range(p):\n x_k = x_test[k]\n y = self.predict(x_k, params)\n d = y_test[k]\n \n # Confusion Matrix\n y_true.append(list(d))\n y_pred.append(list(y))\n\n a = util.inverse_transform(y_true, self.n_classes)\n b = util.inverse_transform(y_pred, self.n_classes)\n return acc(a,b), tpr(a,b, average='macro'), 0, ppv(a,b, average='weighted')\n #return acc(a,b), 0, 0, 0\n\n def grid_search(self, x_train, y_train):\n (n, _) = x_train.shape\n hidden_layer = [4,6,8,10,12,14,16,18,20,22,24,26,28,30]\n k_fold = 5\n slice_ = int(n/k_fold)\n\n grid_accuracy = []\n for q in hidden_layer:\n scores = []\n # cross validation\n for j in range(k_fold):\n # set range\n a = j*slice_\n b = (j+1)*slice_\n\n X_tra_aux = np.concatenate((x_train[0:a], x_train[b:n]), axis=0)\n X_test_aux = x_train[a:b]\n Y_tra_aux = np.concatenate((y_train[0:a], y_train[b:n]), axis=0)\n Y_test_aux = y_train[a:b]\n\n params = self.train(X_tra_aux, Y_tra_aux, q)\n acc, _, _, _ = self.test(X_test_aux, Y_test_aux, params)\n scores.append(acc)\n grid_accuracy.append(np.mean(scores))\n print('Grid search:', grid_accuracy)\n index_max = np.argmax(grid_accuracy)\n return hidden_layer[index_max]\n\n def execute(self):\n x_data = util.normalizeData(self.x_data)\n x_data = util.insertBias(x_data)\n y_data = self.y_data\n\n for i in range(self.realizations):\n x_data_aux, y_data_aux = util.shuffleData(x_data, y_data)\n x_train, x_test, y_train, y_test = util.splitData(x_data_aux, y_data_aux, self.train_size)\n \n if self.g_search:\n best_hidden_layer = self.grid_search(x_train, y_train)\n print('Hidden Layer:', best_hidden_layer)\n else:\n best_hidden_layer = self.hidden_layer\n\n params = self.train(x_train, y_train, best_hidden_layer)\n acc, tpr, spc, ppv = self.test(x_test, y_test, params)\n \n self.hit_rate.append(acc)\n self.tpr.append(tpr)\n self.spc.append(spc)\n self.ppv.append(ppv)\n\n self.acc = np.mean(self.hit_rate)\n self.std = np.std(self.hit_rate)\n self.tpr = np.mean(self.tpr)\n self.spc = np.mean(self.spc)\n self.ppv = np.mean(self.ppv)\n\n print('Hit rate: {}'.format(self.hit_rate))\n print('Accuracy: {:.2f}'.format(self.acc*100))\n print('Minimum: {:.2f}'.format(np.amin(self.hit_rate)*100))\n print('Maximum: {:.2f}'.format(np.amax(self.hit_rate)*100))\n print('Standard Deviation: {:.2f}'.format(self.std))\n print('Sensitivity: {:.2f}'.format(self.tpr*100))\n print('Specificity: {:.2f}'.format(self.spc*100))\n print('Precision: {:.2f}'.format(self.ppv*100))\n\n #self.plotColorMap_3C(x_train, x_test, y_train, self.predict, params)\n #self.plotColorMap_2C(x_train, x_test, y_train, self.predict, params)\n\n def plotColorMap_3C(self, x_train, x_test, y_train, predict, params):\n color1_x = []\n color1_y = []\n color2_x = []\n color2_y = []\n color3_x = []\n color3_y = []\n for i in np.arange(0,1.0,0.005):\n for j in np.arange(0,1.0,0.005):\n xi = np.array([-1, i, j])\n y = predict(xi, params)\n if np.array_equal(y, [0,0,1]):\n color1_x.append(i)\n color1_y.append(j)\n elif np.array_equal(y, [0,1,0]):\n color2_x.append(i)\n color2_y.append(j)\n elif np.array_equal(y, [1,0,0]):\n color3_x.append(i)\n color3_y.append(j)\n else:\n raise ValueError('Error color!\\n')\n \n # Split a train class\n i = []\n j = []\n k = []\n for index,y in enumerate(y_train):\n if np.array_equal(y, [0,0,1]):\n i.append(index)\n elif np.array_equal(y, [0,1,0]):\n j.append(index)\n elif np.array_equal(y, [1,0,0]):\n k.append(index)\n else:\n raise ValueError('Error!\\n')\n train1 = x_train[i]\n train2 = x_train[j]\n train3 = x_train[k]\n\n fig, ax = plt.subplots()\n plt.title('ELM Color Map')\n plt.xlabel('Eixo X')\n plt.ylabel('Eixo y')\n\n ax.scatter(color1_x, color1_y, color=[0.80, 0.88, 0.97])\n ax.scatter(color2_x, color2_y, color=[0.80, 0.80, 0.80])\n ax.scatter(color3_x, color3_y, color=[0.95, 0.87, 0.73])\n ax.scatter(train1[:,1], train1[:,2], label='Classe 1', color=[0.00, 0.45, 0.74])\n ax.scatter(train2[:,1], train2[:,2], label='Classe 2', color=[0.31, 0.31, 0.31])\n ax.scatter(train3[:,1], train3[:,2], label='Classe 3', color=[0.60, 0.20, 0.00])\n ax.scatter(x_test[:,1], x_test[:,2], label='Test Data', color='green')\n\n ax.legend()\n ax.grid(True)\n plt.show()\n\n #fig.savefig('.\\MultilayerPerceptron\\Results\\color_map.png')\n \n def plotColorMap_2C(self, x_train, x_test, y_train, predict, params):\n color1_x = []\n color1_y = []\n color2_x = []\n color2_y = []\n for i in np.arange(0,1,0.005):\n for j in np.arange(0,1,0.005):\n xi = np.array([-1, i, j])\n y = predict(xi, params)\n if np.array_equal(y, [0,1]):\n color1_x.append(i)\n color1_y.append(j)\n elif np.array_equal(y, [1,0]):\n color2_x.append(i)\n color2_y.append(j)\n else:\n raise ValueError('Error color!\\n')\n \n # Split a train class\n i = []\n j = []\n for index,y in enumerate(y_train):\n if np.array_equal(y, [0,1]):\n i.append(index)\n elif np.array_equal(y, [1,0]):\n j.append(index)\n else:\n raise ValueError('Error!\\n')\n train1 = x_train[i]\n train2 = x_train[j]\n\n fig, ax = plt.subplots()\n plt.title('ELM Color Map')\n plt.xlabel('Eixo X')\n plt.ylabel('Eixo y')\n\n ax.scatter(color1_x, color1_y, color=[0.80, 0.88, 0.97])\n ax.scatter(color2_x, color2_y, color=[0.80, 0.80, 0.80])\n ax.scatter(train1[:,1], train1[:,2], label='Classe 1', color=[0.00, 0.45, 0.74])\n ax.scatter(train2[:,1], train2[:,2], label='Classe 2', color=[0.31, 0.31, 0.31])\n ax.scatter(x_test[:,1], x_test[:,2], label='Test Data', color='green')\n\n ax.legend()\n ax.grid(True)\n plt.show()\n\n #fig.savefig('.\\MultilayerPerceptron\\Results\\color_map.png')\n #fig.savefig('.\\MultilayerPerceptron\\Results\\color_map.png', dpi=fig.dpi, bbox_inches='tight')\n", "sub_path": "ExtremeLearningMachine/ELM.py", "file_name": "ELM.py", "file_ext": "py", "file_size_in_byte": 10334, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "numpy.unique", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.random.random_sample", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.random.random_sample", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.linalg.pinv", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 102, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 103, "usage_type": "call"}, {"api_name": "Utils.utils.Util.inverse_transform", "line_number": 122, "usage_type": "call"}, {"api_name": "Utils.utils.Util", "line_number": 122, "usage_type": "name"}, {"api_name": "Utils.utils.Util.inverse_transform", "line_number": 123, "usage_type": "call"}, {"api_name": "Utils.utils.Util", "line_number": 123, "usage_type": "name"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 124, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 124, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 144, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 148, "usage_type": "name"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 149, "usage_type": "argument"}, {"api_name": "numpy.mean", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 152, "usage_type": "call"}, {"api_name": "Utils.utils.Util.normalizeData", "line_number": 156, "usage_type": "call"}, {"api_name": "Utils.utils.Util", "line_number": 156, "usage_type": "name"}, {"api_name": "Utils.utils.Util.insertBias", "line_number": 157, "usage_type": "call"}, {"api_name": "Utils.utils.Util", "line_number": 157, "usage_type": "name"}, {"api_name": "Utils.utils.Util.shuffleData", "line_number": 161, "usage_type": "call"}, {"api_name": "Utils.utils.Util", "line_number": 161, "usage_type": "name"}, {"api_name": "Utils.utils.Util.splitData", "line_number": 162, "usage_type": "call"}, {"api_name": "Utils.utils.Util", "line_number": 162, "usage_type": "name"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 171, "usage_type": "name"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 171, "usage_type": "name"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 171, "usage_type": "name"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 173, "usage_type": "argument"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 174, "usage_type": "argument"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 176, "usage_type": "argument"}, {"api_name": "numpy.mean", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 226, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 236, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 237, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 238, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 238, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 239, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 251, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 251, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 260, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 261, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 267, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 279, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 286, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 286, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 287, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 287, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 288, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 288, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 289, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 289, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 299, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 299, "usage_type": "name"}]}
+{"seq_id": "576717805", "text": "from django import forms\nfrom django.forms.fields import CharField\nfrom django.forms.models import ModelForm\nfrom adm.endereco.models import *\n__author__ = 'mattyws'\n\nclass EstadoForm(ModelForm):\n\n class Meta:\n model = Estado\n\nclass CidadeForm(ModelForm) :\n\n class Meta:\n model = Cidade\n\n def __init__(self, *args, **kwargs):\n super(CidadeForm, self).__init__(*args, **kwargs)\n self.fields['estado'].choices = [[x.id, x.sigla] for x in Estado.objects.all()]\n\nclass EnderecoForm(ModelForm) :\n\n class Meta:\n model = Endereco\n\n def __init__(self, *args, **kwargs):\n super(EnderecoForm, self).__init__(*args, **kwargs)\n self.fields['cidade'].choices = [[x.id, x.nome] for x in Cidade.objects.all()]", "sub_path": "adm/endereco/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 758, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "django.forms.models.ModelForm", "line_number": 7, "usage_type": "name"}, {"api_name": "django.forms.models.ModelForm", "line_number": 12, "usage_type": "name"}, {"api_name": "django.forms.models.ModelForm", "line_number": 21, "usage_type": "name"}]}
+{"seq_id": "131783668", "text": "import signal\nimport sys\nimport json\nimport jsonrpclib\nsys.path.insert(0, '../')\nimport makesets\nimport pickle\nsys.path.insert(0, '../treebuilder')\nfrom Solver import Solver\nfrom sympy.solvers.solvers import solve\nfrom random import randint\n\nclass StanfordNLP:\n def __init__(self, port_number=8080):\n self.server = jsonrpclib.Server(\"http://localhost:%d\" % port_number)\n\n def parse(self, text):\n return json.loads(self.server.parse(text))\n\nnlp = StanfordNLP()\n\ndef cleannum(n):\n return ''.join([x for x in n if x.isdigit() or x=='.' or x=='x' or x=='x*'])\n\ndef kill(signum, frame):\n raise Exception(\"end of time\")\n\ndef training(trips,problem,target):\n #this function take the trips and creates positive and negative training instances from them\n \n texamples = {x:([],[]) for x in [\"+\",\"*\",'/','-','=']}\n for op,a,b in trips:\n vec = makesets.vector(a,b,problem,target)\n texamples[op][0].append(vec)\n\n return texamples\n\ndef parse(q):\n eqs = []\n wps = []\n with open(q) as f:\n g = f.read()\n g = g.split(\"___\")\n for line in g:\n line = line.split('\\n')\n eq = [x for x in line if '=' in x]\n prblm = [x for x in line if ' how ' in x or ' what ' in x]\n print(prblm,eq)\n eqs.append(eq)\n wps.append(prblm)\n return wps,eqs\n\n\ndef make_eq(q):\n bigtexamples = {x:([],[]) for x in [\"+\",\"*\",'/','-','=']}\n wps,eqs= parse(q)\n \n\n for k in range(len(wps)):\n if len(wps[k])==0:continue\n problem = wps[k][0].lower()\n story = nlp.parse(problem)\n sets = makesets.makesets(story['sentences'])\n i = 0\n print(sets)\n while i < len(sets):\n dups = [y for y in sets if y[1].idx != None]\n dups = [y for y in dups if y[1].idx == sets[i][1].idx]\n if len(dups)>1:\n good = [y for y in dups if len([x for x in y[1].num if x.isdigit()])>0]\n if good:\n others = [x for x in dups if x!=good[0]]\n for x in others:\n sets.remove(x)\n else:\n # just pick 1\n for x in dups[1:]:\n sets.remove(x)\n i+=1\n\n\n xidx = [x for x in sets if x[1].num=='x']\n if not xidx:\n problematic.write('no x :'+problem); continue\n\n #TODO look for 2 xes\n '''\n xidx = xidx[0][0]\n postx = [x for x in numbs if x[0]>=xidx]\n if len(postx)>1:\n # 2 vals to right\n twoToRight = True\n else:\n twoToRight = False\n '''\n\n \n\n\n\n numlist = [(cleannum(v.num),v) for k,v in sets]\n numlist = [x for x in numlist if x[0]!='']\n\n allnumbs = {str(k):v for k,v in numlist}\n \n\n objs = {k:(0,v) for k,v in numlist}\n answers = eqs[k]\n\n for j,eq in enumerate(answers):\n trips = []\n print(j,eq)\n l,r = [x.strip().split(' ') for x in eq.split('=')]\n \n compound = r if len(l)==1 else l\n simplex = l if len(l)==1 else r\n target = simplex[0]\n target = (target,objs[target])\n\n #find innermost parens?\n while len(compound)>1:\n if \"(\" in compound:\n rpidx = (len(compound) - 1) - compound[::-1].index('(')\n lpidx = rpidx+compound[rpidx:].index(\")\")\n subeq = compound[rpidx+1:lpidx]\n substr = \"(\"+''.join(subeq)+\")\"\n compound = compound[:rpidx]+[substr]+compound[lpidx+1:]\n else:\n subeq = compound[0:3]\n substr = \"(\"+''.join(subeq)+\")\"\n compound = [substr]+compound[3:]\n if True:\n p,op,e = subeq\n #print(p,op,e)\n p = objs[p]\n e = objs[e]\n op = op.strip()\n trips.append((op,p,e))\n pute = (0,makesets.combine(p[1],e[1],op))\n #print(\"OPERATION SELECTED: \",op)\n #p.details()\n #e.details()\n #print(substr,pute[1].num)\n objs[substr]=pute\n if pute == -1:\n exit()\n if simplex == l:\n trips.append((\"=\",objs[simplex[0]],objs[compound[0]]))\n else:\n trips.append((\"=\",objs[compound[0]],objs[simplex[0]]))\n t = training(trips,problem,target)\n for op in t:\n bigtexamples[op][0].extend(t[op][0])\n bigtexamples[op][1].extend(t[op][1])\n print(op,len(bigtexamples[op][0]))\n pickle.dump(bigtexamples,open('data/gold_training.pickle','wb'))\n\n\n\n\nif __name__==\"__main__\":\n q = sys.argv[1]\n make_eq(q)\n\n\n", "sub_path": "no_lexo/make_gold_equations.py", "file_name": "make_gold_equations.py", "file_ext": "py", "file_size_in_byte": 4957, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "sys.path.insert", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "jsonrpclib.Server", "line_number": 15, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 18, "usage_type": "call"}, {"api_name": "makesets.vector", "line_number": 33, "usage_type": "call"}, {"api_name": "makesets.makesets", "line_number": 63, "usage_type": "call"}, {"api_name": "makesets.combine", "line_number": 139, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 156, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 162, "usage_type": "attribute"}]}
+{"seq_id": "653899604", "text": "from network.huawei import *\nfrom datetime import datetime\nimport argparse\nimport threading\nimport time\n\n# create parser\n# parser = argparse.ArgumentParser('Monitor Huawei E5576c')\n\n# # add arguments\n# parser.add_argument(\n# 'SessionID',\n# metavar='sid',\n# type=str,\n# help='SessionID used to access huaweu e5776c dashboard'\n# )\n\n# args = parser.parse_args()\n\n\n\ndef get_usage(api):\n while True:\n resp = api.get_traffic()\n if resp is not None:\n total_upload = int(resp.TotalUpload)\n total_download = int(resp.TotalDownload)\n\n ttotal_upload = total_upload\n ttotal_download = total_download\n\n udtype = 'B'\n if len(str(total_upload)) >= 7:\n ttotal_upload = total_upload/1024/1024\n udtype = 'MB'\n elif len(str(total_upload)) >= 4:\n ttotal_upload = total_upload/1024\n udtype = 'KB'\n\n ddtype = 'B'\n if len(str(total_download)) >= 7:\n ttotal_download = total_download/1024/1024\n ddtype = 'MB'\n elif len(str(total_download)) >= 4:\n ttotal_download = total_download/1024\n ddtype = 'KB'\n\n print(f'U{ttotal_upload:.2f}{udtype} : D{ttotal_download:.2f}{ddtype}\\n' + '*'*24)\n\n with open(f'traffic_{datetime.now().strftime(\"%d-%m-%Y\")}.log', 'at+') as f:\n f.write(f'{datetime.now()}:{total_upload}:{total_download}\\n')\n else: \n break\n time.sleep(30)\n\napi = HuaweiApi()\nthreading.Thread(target=get_usage, args=[api, ], daemon=True).start()\nwhile True:\n traffic_resp = api.get_traffic()\n\n if traffic_resp is not None:\n up_rate = int(traffic_resp.CurrentUploadRate)\n down_rate = int(traffic_resp.CurrentDownloadRate)\n\n udtype = 'B'\n if len(str(up_rate)) >= 7:\n up_rate = up_rate/1024/1024\n udtype = 'MB'\n elif len(str(up_rate)) >= 4:\n up_rate = up_rate/1024\n udtype = 'KB'\n\n ddtype = 'B'\n if len(str(down_rate)) >= 7:\n down_rate = down_rate/1024/1024\n ddtype = 'MB'\n elif len(str(down_rate)) >= 4:\n down_rate = down_rate/1024\n ddtype = 'KB'\n\n out = (\n f'{up_rate:.2f}{udtype}/s - {down_rate:.2f}{ddtype}/s{\" \"*12}'\n )\n print(out, end='\\r')\n time.sleep(1)\n# else:\n# api.session_id = input(f\"{HuaweiApi.BASE_URL}:SESSION_ID\")\n\n\n\n\n\n\n", "sub_path": "monitor.py", "file_name": "monitor.py", "file_ext": "py", "file_size_in_byte": 2530, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "datetime.datetime.now", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 51, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 54, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 57, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 85, "usage_type": "call"}]}
+{"seq_id": "230369662", "text": "# This toolbox contains functions for prepping, tuning and updating when using the enkf-c\n# for assimilation.\n\n# There are still som uncertainties regarding the tuning that needs to be solve when the\n# format of the observation files are decided, but only one sst and one ice conc file is \n# used as the present plan is, the tuning should be working well.\n\nimport glob\nimport netCDF4 as nc\nimport xarray as xr # Egentlig bruke denne istedenfor nc ettersom jeg tror den er raskere\nimport datetime\nimport numpy as np\nimport os\nimport sys\nimport copy\nimport subprocess\nfrom scipy.ndimage import uniform_filter\nfrom pyresample.geometry import GridDefinition\nfrom pyresample import geometry, image, SwathDefinition\n\ndef cmd(command):\n \"\"\"Function runs provided command in the system shell.\n\n Args:\n command (string) : Command to be executed in shell\n Returns:\n result (integer) : Shell returned status of command\n \"\"\"\n print(\"> \" + command)\n result = subprocess.call(command, shell = True)\n\n if result != 0:\n print(\"Command failed: %d\" % result)\n else:\n return result\n\n# Rewrite the model output to files recognized by the enkf-c\ndef Prep_ensemble(ens_date, grid_dir, ens_inn_dir, enkf_c_dir, res_type, EnKF_var,Nens):\n \n # This should return the number of ensembles found!!!!\n\n # grid_dir: file containing longitude and latitude of the model\n # ens_inn_dir: folder containning the model output data\n # enkf-c_dir: folder containting the netcdf files used by enkf-c\n # ens_inn_file_dummy: file name shared by all the ensemble members\n ens_count = 0\n \n file_cord_handle = nc.Dataset(grid_dir, mode='r')\n \n ice_halo_cells = False\n # Disse leses fra grid fila! \n lat_rho = file_cord_handle.variables['lat_rho']\n #lon_rho = file_cord_handle.variables['lon_rho']\n\n # Write the files that are used to disk\n #glob.glob(folder+file_dummy)\n #for ff in glob.glob(folder+file_dummy):\n # file1.writelines(ff+'\\n')\n prescripts = ['iced.','ocean.']\n syear = str(ens_date.year)\n smnd = str(ens_date.month)\n if ens_date.month < 10:\n smnd = '0' + smnd\n sday = str(ens_date.day)\n if ens_date.day < 10:\n sday = '0' + sday\n\n # Write the files that are used to disk\n #glob.glob(folder+file_dummy)\n file_ens = open(enkf_c_dir+'files_in_ensemble', \"w\")\n\n for pre in prescripts:\n iens2 = 0\n for iens in range(1,Nens+1):\n s1ens = str(iens)\n if iens < 100:\n s1ens = '0'+s1ens\n if iens < 10:\n s1ens = '0'+s1ens\n file = ens_inn_dir+pre+syear+smnd+sday+'_'+s1ens+'.nc'\n print(file)\n #This accounts for the possibilty that some ensemble members have not finished in time\n if os.path.exists(file):\n \n\n iens2 += 1\n sens = str(iens2)\n if iens2 < 100:\n sens = '0'+sens\n if iens2 < 10:\n sens = '0'+sens\n \n file_handle = xr.open_dataset(file)\n if pre == 'ocean.':\n ens_count +=1\n # Write the file name to the member files folder\n # No need to wrote it two times as both files are assumed to exist\n file_ens.writelines(s1ens+'\\n')\n # The ocean restart files can contain several restart times.\n #file_time = file_handle.variables['ocean_time']\n #target_num = nc.date2num(ens_date, units=file_time.units,\n # calendar=file_time.calendar)\n #date_index = np.where(np.array(file_time[:]) == target_num)[0]\n # Må bare konverte Xarray output til num også tror jeg\n date_index = 0 # Denne må fikses!!!!!!\n \n elif pre == 'iced.':\n date_index = 0\n test = file_handle.variables['uvel']\n # Check the if the sea ice restart files is using haloc cells for boundary conditions\n if test.shape[0]>lat_rho.shape[0]:\n ice_halo_cells = True\n\n\n \n\n\n\n # read the important variables from the file\n for var in file_handle.variables.keys():\n #print(var)\n\n if var in EnKF_var:\n\n \n\n fn = enkf_c_dir+'ensemble_6565/mem'+sens+'_'+var+'.nc'\n ds = nc.Dataset(fn, 'w', format='NETCDF4')\n\n var_inn = file_handle[var]\n\n # Litt usikker på om jeg skal ha time, \n # teste om det har noe å si på assimilasjonen\n time = ds.createDimension('time', None)\n\n if len(var_inn.shape) == 4:\n \n dx = ds.createDimension('dx', var_inn.shape[2])\n dy = ds.createDimension('dy', var_inn.shape[3])\n dz = ds.createDimension('dz', var_inn.shape[1])\n\n times = ds.createVariable('time', 'f4', ('time',))\n dxs = ds.createVariable('dx', 'f4', ('dx',))\n dys = ds.createVariable('dy', 'f4', ('dy',))\n dzs = ds.createVariable('dz', 'f4', ('dz',))\n\n temps = ds.createVariable(var, 'f4', ('time', 'dz','dx', 'dy',))\n\n dxs[:] = np.arange(0, var_inn.shape[2], 1.0)\n dys[:] = np.arange(0, var_inn.shape[3], 1.0)\n dzs[:] = np.arange(0, var_inn.shape[1], 1.0)\n \n if pre == 'ocean.':\n # a bit uncertain if it is nan or very high so include both\n var_inn = var_inn.fillna(0)\n var_inn.values[:][var_inn.values[:] > 100] = 0\n temps[0,:,:,:] = var_inn[date_index,:,:,:]\n else:\n sys.exit('Not implemented for ice!')\n\n elif len(var_inn.shape) == 3:\n \n\n times = ds.createVariable('time', 'f4', ('time',))\n if ice_halo_cells:\n dx = ds.createDimension('dx', var_inn.shape[1]-2)\n dy = ds.createDimension('dy', var_inn.shape[2]-2)\n\n dxs = ds.createVariable('dx', 'f4', ('dx',))\n dys = ds.createVariable('dy', 'f4', ('dy',))\n \n dxs[:] = np.arange(0, var_inn.shape[1]-2, 1.0)\n dys[:] = np.arange(0, var_inn.shape[2]-2, 1.0)\n else:\n dx = ds.createDimension('dx', var_inn.shape[1])\n dy = ds.createDimension('dy', var_inn.shape[2])\n\n dxs = ds.createVariable('dx', 'f4', ('dx',))\n dys = ds.createVariable('dy', 'f4', ('dy',))\n \n dxs[:] = np.arange(0, var_inn.shape[1], 1.0)\n dys[:] = np.arange(0, var_inn.shape[2], 1.0)\n\n if pre == 'ocean.':\n temps = ds.createVariable(var, 'f4', ('time', 'dx', 'dy',))\n temps[:,:,:] = var_inn[date_index,:,:]\n elif pre == 'iced.':\n dz = ds.createDimension('dz', var_inn.shape[0])\n dzs = ds.createVariable('dz', 'f4', ('dz',))\n temps = ds.createVariable(var, 'f4', ('time','dz', 'dx', 'dy',))\n dzs[:] = np.arange(0, var_inn.shape[0], 1.0)\n \n \n if ice_halo_cells:\n temps[0,:,:,:] = var_inn[:,1:-1,1:-1]\n else: \n temps[0,:,:,:] = var_inn[:,:,:]\n\n elif len(var_inn.shape) == 2:\n time = ds.createDimension('time', None)\n \n\n times = ds.createVariable('time', 'f4', ('time',))\n \n if ice_halo_cells:\n dx = ds.createDimension('dx', var_inn.shape[0]-2)\n dy = ds.createDimension('dy', var_inn.shape[1]-2)\n\n dxs = ds.createVariable('dx', 'f4', ('dx',))\n dys = ds.createVariable('dy', 'f4', ('dy',))\n\n dxs[:] = np.arange(0, var_inn.shape[0]-2, 1.0)\n dys[:] = np.arange(0, var_inn.shape[1]-2, 1.0)\n else: \n dx = ds.createDimension('dx', var_inn.shape[0])\n dy = ds.createDimension('dy', var_inn.shape[1])\n\n dxs = ds.createVariable('dx', 'f4', ('dx',))\n dys = ds.createVariable('dy', 'f4', ('dy',))\n\n dxs[:] = np.arange(0, var_inn.shape[0], 1.0)\n dys[:] = np.arange(0, var_inn.shape[1], 1.0)\n\n temps = ds.createVariable(var, 'f4', ('time', 'dx', 'dy',))\n\n \n\n\n\n if pre == 'ocean.':\n sys.exit('Not implemented for ocean!')\n elif pre == 'iced.':\n if ice_halo_cells:\n temps[0,:,:] = var_inn[1:-1,1:-1]\n else: \n \n temps[0,:,:] = var_inn[:,:]\n \n\n #lat = ds.createVariable('lat', 'f4', ('dx', 'dy',))\n #lon = ds.createVariable('lon', 'f4', ('dx', 'dy',))\n\n times[:] = nc.date2num(ens_date, units='days since 1990-01-01',\n calendar='gregorian')\n times.units = 'days since 1990-01-01'\n times.calendar='gregorian'\n\n if var == 'temp':\n temps.units = 'degrees C'\n\n #lat[:,:] = lat_rho[:,:]\n #lon[:,:] = lon_rho[:,:]\n\n #value[0,:,:] = 3\n\n ds.close()\n\n if var == 'temp' or var == 'salt':\n # Write also sst variable, must figure out which depth it is and then calculate the variable to\n # the surface if possible, Johannes should know more about this\n\n #print('Also make an sst parameter')\n \n if var == 'temp':\n var2 = 'sst'\n fn = enkf_c_dir+'ensemble_6565/mem'+sens+'_sst.nc'\n if var == 'salt':\n var2 = 'sss'\n fn = enkf_c_dir+'ensemble_6565/mem'+sens+'_sss.nc'\n ds = nc.Dataset(fn, 'w', format='NETCDF4')\n\n #print(var)\n time = ds.createDimension('time', None)\n dx = ds.createDimension('dx', var_inn.shape[2])\n dy = ds.createDimension('dy', var_inn.shape[3])\n\n\n times = ds.createVariable('time', 'f4', ('time',))\n dxs = ds.createVariable('dx', 'f4', ('dx',))\n dys = ds.createVariable('dy', 'f4', ('dy',))\n sst = ds.createVariable(var2, 'f4', ('time', 'dx', 'dy',))\n # lat = ds.createVariable('lat', 'f4', ('dx', 'dy',))\n # lon = ds.createVariable('lon', 'f4', ('dx', 'dy',))\n if var == 'temp':\n sst.units = 'degrees C'\n\n times[:] = nc.date2num(ens_date, units='days since 1990-01-01',\n calendar='gregorian')\n times.units = 'days since 1990-01-01'\n times.calendar='gregorian'\n\n #times[:] = file_time[date_index]\n #times.units = file_time.units\n\n dxs[:] = np.arange(0, var_inn.shape[2], 1.0)\n dys[:] = np.arange(0, var_inn.shape[3], 1.0)\n sst[0,:,:] = var_inn[0,41,:,:]\n #lat[:,:] = lat_rho[:,:]\n #lon[:,:] = lon_rho[:,:]\n\n #value[0,:,:] = 3\n\n ds.close()\n\n if var == 'aicen' or var == 'vicen':\n # Write also integrated variables for assimilation\n \n if var == 'aicen':\n var2 = 'aice'\n fn = enkf_c_dir+'ensemble_6565/mem'+sens+'_aice.nc'\n elif var == 'vicen':\n var2 = 'vice'\n fn = enkf_c_dir+'ensemble_6565/mem'+sens+'_vice.nc'\n ds = nc.Dataset(fn, 'w', format='NETCDF4')\n\n time = ds.createDimension('time', None)\n times = ds.createVariable('time', 'f4', ('time',))\n\n\n\n if ice_halo_cells:\n dx = ds.createDimension('dx', var_inn.shape[1]-2)\n dy = ds.createDimension('dy', var_inn.shape[2]-2)\n\n dxs = ds.createVariable('dx', 'f4', ('dx',))\n dys = ds.createVariable('dy', 'f4', ('dy',))\n\n dxs[:] = np.arange(0, var_inn.shape[1]-2, 1.0)\n dys[:] = np.arange(0, var_inn.shape[2]-2, 1.0)\n else: \n dx = ds.createDimension('dx', var_inn.shape[1])\n dy = ds.createDimension('dy', var_inn.shape[2])\n\n dxs = ds.createVariable('dx', 'f4', ('dx',))\n dys = ds.createVariable('dy', 'f4', ('dy',))\n\n dxs[:] = np.arange(0, var_inn.shape[1], 1.0)\n dys[:] = np.arange(0, var_inn.shape[2], 1.0)\n\n sst = ds.createVariable(var2, 'f4', ('time', 'dx', 'dy',))\n \n if ice_halo_cells:\n sst[0,:,:] = np.sum(var_inn[:,1:-1,1:-1],axis=0)\n else:\n sst[0,:,:] = np.sum(var_inn[:,:,:],axis=0)\n #lat[:,:] = lat_rho[:,:]\n #lon[:,:] = lon_rho[:,:]\n\n #value[0,:,:] = 3\n\n times[:] = nc.date2num(ens_date, units='days since 1990-01-01',\n calendar='gregorian')\n times.units = 'days since 1990-01-01'\n times.calendar='gregorian'\n\n ds.close()\n\n if var == 'aicen':\n # Write smoothed variable for low resolution assimilation\n \n fn = enkf_c_dir+'ensemble_6565/mem'+sens+'_aiceosi.nc'\n ds = nc.Dataset(fn, 'w', format='NETCDF4')\n\n time = ds.createDimension('time', None)\n times = ds.createVariable('time', 'f4', ('time',))\n\n\n\n if ice_halo_cells:\n dx = ds.createDimension('dx', var_inn.shape[1]-2)\n dy = ds.createDimension('dy', var_inn.shape[2]-2)\n\n dxs = ds.createVariable('dx', 'f4', ('dx',))\n dys = ds.createVariable('dy', 'f4', ('dy',))\n\n dxs[:] = np.arange(0, var_inn.shape[1]-2, 1.0)\n dys[:] = np.arange(0, var_inn.shape[2]-2, 1.0)\n else: \n dx = ds.createDimension('dx', var_inn.shape[1])\n dy = ds.createDimension('dy', var_inn.shape[2])\n\n dxs = ds.createVariable('dx', 'f4', ('dx',))\n dys = ds.createVariable('dy', 'f4', ('dy',))\n\n dxs[:] = np.arange(0, var_inn.shape[1], 1.0)\n dys[:] = np.arange(0, var_inn.shape[2], 1.0)\n\n sst = ds.createVariable('aiceosi', 'f4', ('time', 'dx', 'dy',))\n \n if ice_halo_cells:\n sst[0,:,:] = uniform_filter(np.sum(var_inn[:,1:-1,1:-1],axis=0), size=16, mode='constant')\n else:\n sst[0,:,:] = uniform_filter(np.sum(var_inn[:,:,:],axis=0), size=16, mode='constant')\n #lat[:,:] = lat_rho[:,:]\n #lon[:,:] = lon_rho[:,:]\n\n #value[0,:,:] = 3\n\n times[:] = nc.date2num(ens_date, units='days since 1990-01-01',\n calendar='gregorian')\n times.units = 'days since 1990-01-01'\n times.calendar='gregorian'\n\n ds.close()\n\n \n # if var equals aicen or vicen\n \n file_handle.close()\n file_ens.close()\n\n file_cord_handle.close()\n\n return ens_count\n \ndef set_number_ensembles(prm_file, ens_count):\n file1 = open(prm_file, 'r') \n Lines = file1.readlines()\n ii = -1\n for l in Lines:\n ii += 1\n if l[0:7] == 'ENSSIZE':\n #print('Fant den!')\n Lines[ii] = 'ENSSIZE = '+str(ens_count)+'\\n'\n #l = 'ENSSIZE = '+str(ens_count)+'\\n'\n\n file1 = open('/Users/sindrefritzner/enkf-c/Assimilation/enkf2.prm', \"w\")\n file1.writelines(Lines)\n file1.close()\n\n\ndef check_dfs_srf(ens_count, diag_file):\n # Check the DFS and SRF values to see if any exceed the recomended values\n # for example if more than 5% of the data has values larger than the limit\n \n file_handle = nc.Dataset(diag_file, mode='r')\n psrf = file_handle.variables['psrf']\n pdfs = file_handle.variables['pdfs']\n\n\n srf_lim = 2 #'må sjekke'\n dfs_lim = ens_count/3 #'må sjekke'\n \n update_srf=np.zeros(psrf.shape[0])\n update_dfs=np.zeros(psrf.shape[0])\n for i in range(0, psrf.shape[0]):\n # more than 5% of the data larger than the limit then reduce\n if len(psrf[i,:,:][psrf[i,:,:]>srf_lim])/(psrf.shape[1]*psrf.shape[2]) > 0.05:\n print('Must increase Rfactor for obstype ' + str(i))\n update_srf[i] = 1\n elif len(psrf[i,:,:][psrf[i,:,:] 0.99:\n print('Must decrease R_factor for obstype ' + str(i))\n update_srf[i] = -1\n\n if len(pdfs[i,:,:][pdfs[i,:,:]>dfs_lim])/(pdfs.shape[1]*pdfs.shape[2]) > 0.05:\n print('Must reduce loc_rad for obstype ' + str(i))\n update_dfs[i] = 1\n elif len(pdfs[i,:,:][pdfs[i,:,:] 0.99:\n print('Can increase loc_rad for obstype ' + str(i))\n update_dfs[i] = -1\n\n # If any must be updated run update prm with information regarding which obstypes that should be updated\n # and how they should be updated. Consider to change back to default values when the assimilation is finished \n # with satisfying results\n file_handle.close()\n \n return update_srf, update_dfs\n \n \ndef update_tuning(tuning_file, update_srf, update_dfs):\n# To check each locrad and R_factor obstypes.prm should be searched from top to bottom and each time it passes a name\n# this name should be rembered such that the next locrad and rfactor encoutered belongs to this name\n# It is also important that each obstypes has a locrad specified and an rfractor specified.\n# Must investigate a bit more how this should be done in practice, have a list with 0 and 1 per obs type is probably\n# the easiest.\n\n file1 = open(tuning_file, 'r') \n Lines = file1.readlines()\n ii = -1\n obs_num = -1\n\n for l in Lines:\n ii += 1\n \n if l[0:4] == 'NAME':\n print(l[7:])\n current_obs = l[7:]\n obs_num += 1\n print(obs_num)\n if obs_num < len(update_srf):\n if update_dfs[obs_num] == 1 and l[0:7] == 'RFACTOR':\n rf_old = float(l[10:-1])\n Lines[ii] = 'RFACTOR = '+str(round(rf_old*1.5))+'\\n'\n elif update_dfs[obs_num] == -1 and l[0:7] == 'RFACTOR':\n rf_old = float(l[10:-1])\n Lines[ii] = 'RFACTOR = '+str(round(rf_old*0.75))+'\\n'\n elif update_dfs[obs_num] == -1 and l[0:6] == 'LOCRAD':\n lr_old = float(l[9:-1])\n Lines[ii] = 'RFACTOR = '+str(round(lr_old*0.75))+'\\n'\n elif update_dfs[obs_num] == -1 and l[0:6] == 'LOCRAD':\n lr_old = float(l[9:-1])\n Lines[ii] = 'RFACTOR = '+str(round(lr_old*0.75))+'\\n'\n\ndef update_the_ensemble(enkf_c_dir, EnKF_var,ens_out_dir,ens_date):\n # Update the ensemble, in practise the whole ensemble does not need to be updated, only those that require \n # new initial states, but for now everything can be updated. values should also be checked for consistence, \n # potential large errors should possibly lead to an error, or at least they should be flagged for investigation.\n\n # Get the list of files that was used as input in the correct order, this file should ideally be written to disk.\n file_ens = open(enkf_c_dir+'files_in_ensemble', 'r') \n Lines = file_ens.readlines()\n file_count = 0\n prescripts = ['iced.','ocean.']\n syear = str(ens_date.year)\n smnd = str(ens_date.month)\n if ens_date.month < 10:\n smnd = '0' + smnd\n sday = str(ens_date.day)\n if ens_date.day < 10:\n sday = '0' + sday\n\n zero_checks_0 = ['alvl','qice001','qice002','qice003','qice004','qice005','qice006','qice007'\n 'qsno001','sice001','sice002','sice003','sice004','sice005','sice006','sice007',\n 'vlvl','vsnon']\n qices = ['qice001','qice002','qice003','qice004','qice005','qice006','qice007']\n sices = ['sice001','sice002','sice003','sice004','sice005','sice006','sice007']\n\n #not_update = ['aicen','vicen','vsnon','temp','salt','qice001','qice002','qice003',\n # 'qice004','qice005','qice006','qice007','sice001','sice002','sice003', \n # 'sice004','sice005','sice006','sice007','Tsfcn']\n not_update = []\n\n # Make sure that aicen is first in the list of variables such that this can be used for the other variables \n if EnKF_var.index(\"aicen\") != 0:\n # Switch aicen with the variable that is first in the list\n tlist = copy.deepcopy(EnKF_var)\n tlist[0] = EnKF_var[EnKF_var.index(\"aicen\")]\n tlist[EnKF_var.index(\"aicen\")] = EnKF_var[0]\n EnKF_var = tlist\n\n\n\n\n for ll in Lines:\n file_count += 1\n for pre in prescripts:\n file = ens_out_dir+pre+syear+smnd+sday+'_'+ll[0:-1]+'.nc' \n print(file)\n org_ds = nc.Dataset(file, 'r+', format='NETCDF4') \n num = str(file_count)\n halo_cells = False\n if file_count < 10:\n num = '0'+num\n \n for var in org_ds.variables.keys():\n if var in EnKF_var:\n if var in not_update:\n fn = enkf_c_dir+'ensemble_6565/mem0'+num+'_'+var+'.nc'\n else:\n fn = enkf_c_dir+'ensemble_6565/mem0'+num+'_'+var+'.nc.analysis'\n mem_ds = nc.Dataset(fn, 'r', format='NETCDF4')\n #print(fn)\n #print(file)\n new_var = mem_ds.variables[var]\n old_var = org_ds.variables[var]\n\n # Check bounds for this file, but what should the bounds be? \n # With the dfs and srf checks it is not expected that the updates are too large, but could probalby\n # check just to make sure.\n # At least SST should never be less than -2 and ice conc should be between 0 and 1.\n\n \n #print(new_var.shape)\n #print(old_var.shape)\n\n if old_var.shape[2]>new_var.shape[3] and pre == 'iced.':\n halo_cells = True\n temp = new_var[:]\n if halo_cells:\n if len(temp.shape) == 3:\n temp2 = np.zeros((temp.shape[0],temp.shape[1]+2, \n temp.shape[2]+2))\n temp2[0,1:-1,1:-1] = temp\n temp = temp2\n elif len(temp.shape) == 4:\n temp2 = np.zeros((temp.shape[0],temp.shape[1], \n temp.shape[2]+2, temp.shape[3]+2))\n temp2[0,:,1:-1,1:-1] = temp\n temp = temp2\n\n if var == 'temp':\n # Temperature cannot be below minus 2 \n temp[temp < -2] = -2\n if var == 'salt':\n # Salinity cannot be less than 0 \n temp[temp < 0] = 0\n\n if var == 'aicen': \n # Check ice boundaries\n temp[temp < 0.01] = 0\n temp[temp > 1] = 1\n # Check that aggregated concentraion is less than 1\n temsum = np.sum(temp,axis=1)+0.01\n temsum[temsum < 1] = 1\n \n for i in range(temp.shape[1]):\n #print(temsum.shape)\n #print(temp[:,i,:,:].shape)\n temp[:,i,:,:] = temp[:,i,:,:]/temsum\n #print(temp[0,:,:,:].shape)\n #print(old_var[:].shape)\n\n #old_var[:] = temp[0,:,:,:]\n\n # These varaibles are to be used for checking the other ice variables,\n # especially zero checks and new ice checks\n aicen = temp[:,:,:,:]\n aice = np.sum(temp,axis=1)\n aice1 = np.ceil(aice)\n \n # Set all variables in zero_checks to zero if there is no ice.\n if var in zero_checks_0:\n temp[aicen == 0] = 0\n \n if var == 'Tsfcn':\n temp[aicen == 0] = 0\n temp[temp > 0] = 0\n temp[temp < -20] = -20\n\n if var in qices:\n # set value of new data to -1e8\n temp[temp > 0] = 0\n temp[temp < -3.6e8] = -3.6e8\n for i in range(temp.shape[1]):\n temp[:,i,:,:] = np.minimum(temp[:,i,:,:],aice1*-1.2e8)\n\n if var in sices:\n temp[temp < 0] = 0\n temp[temp > 31] = 31\n # set value of new data to 4\n for i in range(temp.shape[1]):\n temp[:,i,:,:] = np.maximum(temp[:,i,:,:],aice1*4)\n \n if var == 'qsno001':\n temp[temp > 0] = 0\n temp[temp < -1.4e8] = -1.4e8\n for i in range(temp.shape[1]):\n temp[:,i,:,:] = np.minimum(temp[:,i,:,:],aice1*-1.2e8)\n \n\n if var == 'vicen': \n # Check thickness boundaries\n temp[temp < 0] = 0\n # Set thickness to zero for areas without ice\n temp[aicen == 0] = 0\n \n # Make sure that new ice is also updated if missed by the assimilation\n # Assume that the new thickness is very thin, vicen=aicen just for simplicity,\n # this is not really expected to happen, but can cause numerical errors\n temp = np.maximum(temp,aicen) \n\n if var == 'vsnon': \n temp[temp < 0] = 0\n temp[aicen == 0] = 0 \n \n if pre == 'iced.':\n old_var[:] = temp[0]\n elif pre == 'ocean.':\n old_var[:] = temp\n\n \n mem_ds.close()\n org_ds.close()\n\n file_ens.close()\n #old_var = new_var\n\n\ndef get_osisaf_obs(date,obs_dir,Assim_dir):\n #osisaf_pre = 'ice_conc_nh_ease-125_multi_'\n osisaf_pre = 'ice_conc_nh_polstere-100_multi_'\n osisaf_post = '1200.nc'\n smnd = str(date.month) if date.month > 9 else '0'+str(date.month)\n sday = str(date.day) if date.day > 9 else '0'+str(date.day)\n\n obs_file = obs_dir+str(date.year)+'/'+smnd+'/'+osisaf_pre+str(date.year)+smnd+sday+osisaf_post\n\n file_out = Assim_dir +'/obs/OSISAF/this_day.nc'\n\n # Export concentraion and uncertainty\n cmd('ncks -O -v ice_conc,total_uncertainty '+obs_file+' temp_osisaf1.nc')\n # Rename uncertainty to that read by enkf-c\n cmd('ncrename -v total_uncertainty,error_std temp_osisaf1.nc')\n # Change dimension from percent to decimal concentration\n cmd('ncap2 -O -s \"ice_conc=ice_conc/100\" temp_osisaf1.nc temp_osisaf2.nc')\n cmd('ncap2 -O -s \"error_std=error_std/100\" temp_osisaf2.nc '+file_out)\n\n # Fix nan values \n cmd('ncatted -O -h -a _FillValue,error_std,m,f,-1 '+file_out)\n cmd('ncatted -O -h -a _FillValue,ice_conc,m,f,-1 '+file_out)\n\n # delete temporary files\n cmd('rm temp_osisaf1.nc temp_osisaf2.nc')\n\ndef write_results(date,enkf_c_dir,ens_out_dir,Nens):\n \n smnd = str(date.month) if date.month > 9 else '0'+str(date.month)\n sday = str(date.day) if date.day > 9 else '0'+str(date.day)\n file = enkf_c_dir +'Assim_summary_'+str(date.year)+smnd+sday+'.nc'\n\n # Generate the netcdf, shoudl contain aice, vice, before and after in addition,\n # mem1 aicen before and after and sst and vice\n # Can add more later on\n\n # Read in temp file and use this as template for the dimensions\n print(enkf_c_dir+'ensemble_6565/mem001_temp.nc') \n tt = xr.open_dataset(enkf_c_dir+'ensemble_6565/mem001_temp.nc')\n temp = tt['temp']\n \n print(file)\n ds = nc.Dataset(file, 'w', format='NETCDF4')\n\n time = ds.createDimension('time', None)\n times = ds.createVariable('time', 'f4', ('time',))\n times[:] = nc.date2num(date, units='days since 1990-01-01',\n calendar='gregorian')\n times.units = 'days since 1990-01-01'\n times.calendar='gregorian'\n\n de = ds.createDimension('de', 10) # Ens\n di = ds.createDimension('di', 5) # Ice categories\n dz = ds.createDimension('dz', temp.shape[1]) # Depth levels\n dx = ds.createDimension('dx', temp.shape[2])\n dy = ds.createDimension('dy', temp.shape[3])\n \n\n des = ds.createVariable('de', 'f4', ('de',))\n dis = ds.createVariable('di', 'f4', ('di',))\n dzs = ds.createVariable('dz', 'f4', ('dz',))\n dxs = ds.createVariable('dx', 'f4', ('dx',))\n dys = ds.createVariable('dy', 'f4', ('dy',))\n\n\n des[:] = np.arange(0, Nens, 1.0)\n dis[:] = np.arange(0, 5, 1.0)\n dzs[:] = np.arange(0, temp.shape[1], 1.0)\n dxs[:] = np.arange(0, temp.shape[2], 1.0)\n dys[:] = np.arange(0, temp.shape[3], 1.0)\n\n\n\n tt.close()\n\n\n aicen_mem1_before = ds.createVariable('aicen1_inn', 'f4', ('time','di', 'dx', 'dy',))\n aicen_mem1_after = ds.createVariable('aicen1_out', 'f4', ('time','di', 'dx', 'dy',))\n\n vicen_mem1_before = ds.createVariable('vicen1_inn', 'f4', ('time','di', 'dx', 'dy',))\n vicen_mem1_after = ds.createVariable('vicen1_out', 'f4', ('time','di', 'dx', 'dy',))\n\n aice_before = ds.createVariable('aice_inn', 'f4', ('time','de', 'dx', 'dy',))\n aice_after = ds.createVariable('aice_out', 'f4', ('time','de', 'dx', 'dy',))\n\n vice_before = ds.createVariable('vice_inn', 'f4', ('time','de', 'dx', 'dy',))\n vice_after = ds.createVariable('vice_out', 'f4', ('time','de', 'dx', 'dy',))\n\n temp_mem1_before = ds.createVariable('temp1_inn', 'f4', ('time','dz', 'dx', 'dy',))\n temp_mem1_after = ds.createVariable('temp1_out', 'f4', ('time','dz', 'dx', 'dy',))\n\n sst_before = ds.createVariable('sst_inn', 'f4', ('time','de', 'dx', 'dy',))\n sst_after = ds.createVariable('sst_out', 'f4', ('time','de', 'dx', 'dy',))\n \n\n file_ens = open(enkf_c_dir+'files_in_ensemble', 'r') \n Lines = file_ens.readlines()\n file_count = 0\n\n for ll in Lines:\n file_count += 1\n sens = str(file_count) if file_count > 9 else '0'+str(file_count)\n file_out_ice = ens_out_dir+'iced.'+str(date.year)+smnd+sday+'_'+ll[0:-1]+'.nc' \n file_out_ocn = ens_out_dir+'ocean.'+str(date.year)+smnd+sday+'_'+ll[0:-1]+'.nc'\n\n ############ Write the inn first ###################\n # Write aice_inn to res \n file_inn = enkf_c_dir+'ensemble_6565/mem0'+sens+'_aice.nc'\n handle = xr.open_dataset(file_inn)\n aice_before[0,int(ll[0:-1])-1,:,:] = handle['aice'][0,:,:]\n handle.close()\n\n # Write vice_inn to res \n file_inn = enkf_c_dir+'ensemble_6565/mem0'+sens+'_vice.nc'\n handle = xr.open_dataset(file_inn)\n vice_before[0,int(ll[0:-1])-1,:,:] = handle['vice'][0,:,:]\n handle.close()\n\n # Write sst_inn to res \n file_inn = enkf_c_dir+'ensemble_6565/mem0'+sens+'_sst.nc'\n handle = xr.open_dataset(file_inn)\n sst_before[0,int(ll[0:-1])-1,:,:] = handle['sst'][0,:,:]\n handle.close()\n \n # Write the full member 1 states\n if file_count == 1:\n file_inn = enkf_c_dir+'ensemble_6565/mem0'+sens+'_aicen.nc'\n handle = xr.open_dataset(file_inn)\n aicen_mem1_before[0,:,:,:] = handle['aicen'][0,:,:,:]\n nx_size = handle['aicen'].shape[2]\n handle.close()\n\n file_inn = enkf_c_dir+'ensemble_6565/mem0'+sens+'_vicen.nc'\n handle = xr.open_dataset(file_inn)\n vicen_mem1_before[0,:,:,:] = handle['vicen'][0,:,:,:]\n handle.close()\n \n file_inn = enkf_c_dir+'ensemble_6565/mem0'+sens+'_temp.nc'\n handle = xr.open_dataset(file_inn)\n temp_mem1_before[0,:,:,:] = handle['temp'][0,:,:,:]\n handle.close()\n ###################################################################\n\n ##################### Write the out results #####################\n handle = xr.open_dataset(file_out_ocn)\n sst_after[0,int(ll[0:-1])-1,:,:] = handle['temp'][0,41,:,:]\n if file_count == 1:\n temp_mem1_after[0,:,:,:] = handle['temp'][0,:,:,:]\n handle.close()\n\n \n handle = xr.open_dataset(file_out_ice)\n nx_size2 = handle['aicen'].shape[1]\n ice_halo_cells = True if nx_size2 > nx_size else False\n\n if ice_halo_cells:\n aice_after[0,int(ll[0:-1])-1,:,:] = np.sum(handle['aicen'][:,1:-1,1:-1],axis=0)\n vice_after[0,int(ll[0:-1])-1,:,:] = np.sum(handle['vicen'][:,1:-1,1:-1],axis=0)\n if file_count == 1:\n aicen_mem1_after[0,:,:,:] = handle['aicen'][:,1:-1,1:-1]\n vicen_mem1_after[0,:,:,:] = handle['vicen'][:,1:-1,1:-1]\n else:\n aice_after[0,int(ll[0:-1])-1,:,:] = np.sum(handle['aicen'][:,:,:],axis=0)\n vice_after[0,int(ll[0:-1])-1,:,:] = np.sum(handle['vicen'][:,:,:],axis=0)\n if file_count == 1:\n aicen_mem1_after[0,:,:,:] = handle['aicen'][:,:,:]\n vicen_mem1_after[0,:,:,:] = handle['vicen'][:,:,:]\n\n\n\n ####################################################################\n\n ##################### Also write the observations for easy reference? #####\n # Might convert it first so it might be easier to compare? ################\n # With several observations potentially a list of string could be used here,\n #OSISAF\n file_osisaf = enkf_c_dir+'obs/OSISAF/this_day.nc'\n grid_file = enkf_c_dir+'conf/new_grid_ice.nc'\n\n handle = xr.open_dataset(file_osisaf)\n ice_conc = handle['ice_conc']\n lon_obs = handle['lon']\n lat_obs = handle['lat']\n obs_grid_def = geometry.GridDefinition(lons=lon_obs, lats=lat_obs)\n\n\n handle2 = xr.open_dataset(grid_file)\n lon_mod = handle2['lon']\n lat_mod = handle2['lat']\n mod_grid_def = geometry.GridDefinition(lons=lon_mod, lats=lat_mod)\n\n # Fix future warning!\n obs_container = image.ImageContainerNearest(ice_conc[0,:,:].values, \n obs_grid_def, radius_of_influence=20000)\n obs_modelgrid = obs_container.resample(mod_grid_def)\n res = obs_modelgrid.image_data\n\n\n Obs1 = ds.createVariable('Obs1', 'f4', ('time','dx', 'dy',))\n Obs1[0,:,:] = res[:]\n\n\n ds.close()\n", "sub_path": "python_tools/enkf_c_toolbox.py", "file_name": "enkf_c_toolbox.py", "file_ext": "py", "file_size_in_byte": 38513, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "subprocess.call", "line_number": 30, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "xarray.open_dataset", "line_number": 93, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 151, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 222, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 231, "usage_type": "call"}, {"api_name": "netCDF4.date2num", "line_number": 243, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 270, "usage_type": "call"}, {"api_name": "netCDF4.date2num", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 295, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 296, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 314, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 328, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 329, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 337, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 338, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 343, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 345, "usage_type": "call"}, {"api_name": "netCDF4.date2num", "line_number": 351, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 362, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 377, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 385, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 386, "usage_type": "call"}, {"api_name": "scipy.ndimage.uniform_filter", "line_number": 391, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 391, "usage_type": "call"}, {"api_name": "scipy.ndimage.uniform_filter", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 393, "usage_type": "call"}, {"api_name": "netCDF4.date2num", "line_number": 399, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 436, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 444, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 445, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 536, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 549, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 561, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 581, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 586, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 603, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 618, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 619, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 635, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 642, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 648, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 660, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 717, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 721, "usage_type": "call"}, {"api_name": "netCDF4.date2num", "line_number": 725, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 744, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 745, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 746, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 747, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 748, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 787, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 793, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 799, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 806, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 812, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 817, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 823, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 830, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 835, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 836, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 841, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 842, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 858, "usage_type": "call"}, {"api_name": "pyresample.geometry.GridDefinition", "line_number": 862, "usage_type": "call"}, {"api_name": "pyresample.geometry", "line_number": 862, "usage_type": "name"}, {"api_name": "xarray.open_dataset", "line_number": 865, "usage_type": "call"}, {"api_name": "pyresample.geometry.GridDefinition", "line_number": 868, "usage_type": "call"}, {"api_name": "pyresample.geometry", "line_number": 868, "usage_type": "name"}, {"api_name": "pyresample.image.ImageContainerNearest", "line_number": 871, "usage_type": "call"}, {"api_name": "pyresample.image", "line_number": 871, "usage_type": "name"}]}
+{"seq_id": "109136351", "text": "from sunpy.extern.sunkit_instruments.lyra import (\n get_lytaf_event_types,\n get_lytaf_events,\n remove_lytaf_events_from_timeseries,\n split_series_using_lytaf,\n)\n\n__all__ = ['remove_lytaf_events_from_timeseries',\n 'get_lytaf_events',\n 'get_lytaf_event_types',\n 'split_series_using_lytaf']\n\n# Trick the docs into thinking these functions are defined in here.\nfor _a in (get_lytaf_event_types,\n get_lytaf_events,\n remove_lytaf_events_from_timeseries,\n split_series_using_lytaf):\n _a.__module__ = __name__\n", "sub_path": "sunpy/instr/lyra.py", "file_name": "lyra.py", "file_ext": "py", "file_size_in_byte": 577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "sunpy.extern.sunkit_instruments.lyra.get_lytaf_event_types", "line_number": 14, "usage_type": "name"}, {"api_name": "sunpy.extern.sunkit_instruments.lyra.get_lytaf_events", "line_number": 15, "usage_type": "name"}, {"api_name": "sunpy.extern.sunkit_instruments.lyra.remove_lytaf_events_from_timeseries", "line_number": 16, "usage_type": "name"}, {"api_name": "sunpy.extern.sunkit_instruments.lyra.split_series_using_lytaf", "line_number": 17, "usage_type": "name"}]}
+{"seq_id": "85549034", "text": "# import ptvsd\n# ptvsd.enable_attach(address = ('0.0.0.0', 5678))\n# ptvsd.wait_for_attach()\nimport os\nimport compress_model as cm\nimport argparse\nimport numpy as np\n\ndef get_opt():\n parser = argparse.ArgumentParser()\n parser.add_argument(\"--compress_layer\", type=int, nargs='+', default = [0,1,2,3,4])\n parser.add_argument(\"--compress_block\", type=int, nargs='+', default = [1,2], help='range from 1 to 2')\n parser.add_argument(\"--compress_rate\", type=float, nargs='+', default = [0.5, 0.5])\n parser.add_argument(\"--gpu\", type=int, default = 4)\n parser.add_argument('--method', type=str, default='single')\n\n opt = parser.parse_args()\n return opt\n\nif __name__ == '__main__':\n gpu = 0\n while True:\n os.system('nvidia-smi -q -d Memory |grep -A4 GPU|grep Free >tmp')\n memory_gpu=[int(x.split()[2]) for x in open('tmp','r').readlines()]\n memory_max = max(memory_gpu)\n if memory_max>5000:\n gpu = np.argmax(memory_gpu)\n os.environ[\"CUDA_VISIBLE_DEVICES\"] = str(np.argmax(memory_gpu))\n os.system('rm tmp')\n print('Find vacant GPU: %d' % gpu)\n break\n opt = get_opt()\n opt.gpu = int(gpu)\n cm.pipeline(opt)\n\n model_dir = os.path.join('/ssd/yqian/prune/model/ResNet50', '-'.join([str(i) for i in opt.compress_layer])+'_'+'-'.join([str(i) for i in opt.compress_block])+'_'+'-'.join([str(i) for i in opt.compress_rate]))\n os.system('/root/caffe/build/tools/caffe_parallel train --solver %s/solver.prototxt --weights=%s/prune.caffemodel' % (model_dir, model_dir))\n os.system('/root/caffe/build/tools/caffe_parallel test --model %s/trainval.prototxt --weights=%s/snapshot/_iter_10000.caffemodel.h5' % (model_dir, model_dir))\n os.system('python ThiNet_Code/ToolKit/FLOPs_and_size.py %s/trainval.prototxt' % model_dir)\n ", "sub_path": "ThiNet_TPAMI/ResNet50/prune_finetune.py", "file_name": "prune_finetune.py", "file_ext": "py", "file_size_in_byte": 1839, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "os.system", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 27, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 28, "usage_type": "call"}, {"api_name": "os.system", "line_number": 29, "usage_type": "call"}, {"api_name": "compress_model.pipeline", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 37, "usage_type": "call"}, {"api_name": "os.system", "line_number": 38, "usage_type": "call"}, {"api_name": "os.system", "line_number": 39, "usage_type": "call"}]}
+{"seq_id": "517530024", "text": "import spotify_client\r\nimport threading\r\nimport requests\r\nimport lxml\r\nfrom bs4 import BeautifulSoup\r\nimport re\r\nimport tkinter as tk\r\nfrom tkinter import ttk\r\n\r\n### Connect to Spotify API\r\n\r\nsp = spotify_client.SpotifyClient()\r\n\r\nif sp.refresh_token == \" \":\r\n sp.get_authentication_code()\r\n sp.get_access_token()\r\n\r\n\r\ndef auto_refresh_access_token():\r\n threading.Timer(3500, auto_refresh_access_token).start()\r\n sp.refresh_access_token()\r\n print (\"Access token was refreshed\")\r\n\r\nauto_refresh_access_token()\r\n\r\n\r\n### Get the artist and track name\r\n\r\ndef get_currently_playing_data():\r\n data = sp.get_currently_playing()\r\n if data == None:\r\n print (\"No music is currently playing\")\r\n return data\r\n \r\ncurrent_playing_data = get_currently_playing_data()\r\nprint (current_playing_data)\r\n\r\n\r\n### Scrape Genius' lyrics page\r\n\r\ndef get_lyrics(data):\r\n genius_artist = data[\"artist\"]\r\n genius_track = data[\"track\"]\r\n\r\n # String processing for genius \r\n if \"(\" in genius_track:\r\n indx = genius_track.find(\"(\")\r\n genius_track = genius_track[:indx-1]\r\n if \"-\" in genius_track:\r\n indx = genius_track.find(\"-\")\r\n genius_track = genius_track[:indx-1]\r\n print (genius_artist,genius_track) \r\n genius_artist = genius_artist.replace(\" \", \"-\").replace(\"/\", \"-\").replace(\"'\", \"\").replace(\"&\", \"and\")\r\n genius_track = genius_track.replace(\" \",\"-\").replace(\"/\", \"-\").replace(\"'\", \"\").replace(\",\",\"\").replace(\"&\", \"and\")\r\n\r\n res = requests.get(f\"https://genius.com/{genius_artist}-{genius_track}-lyrics\")\r\n if res.status_code not in range(200,299):\r\n raise Exception(\"Could not get lyrics\")\r\n pass\r\n soup = BeautifulSoup(res.text,\"lxml\")\r\n page_text = soup.find_all(\"div\", class_=\"Lyrics__Container-sc-1ynbvzw-2 jgQsqn\")\r\n text_with_spaces = \"\"\r\n for part in page_text:\r\n text_with_spaces = text_with_spaces + part.prettify()\r\n\r\n def remove_html_tags(text):\r\n\r\n \"\"\"Remove html tags from a string\"\"\"\r\n clean = re.compile('<.*?>')\r\n return re.sub(clean, '', text)\r\n \r\n lyrics = remove_html_tags(text_with_spaces)\r\n return lyrics\r\n\r\n\r\nlyrics = get_lyrics(current_playing_data)\r\n\r\n\r\n### Display lyrics with tkinter\r\n\r\ndef display_lyrics(lyrics):\r\n\r\n root = tk.Tk()\r\n root.title(\"Lyrics\")\r\n root.geometry(\"500x300\")\r\n\r\n main_frame = tk.Frame(root)\r\n main_frame.pack(fill=\"both\", expand=1)\r\n\r\n canvas = tk.Canvas(main_frame)\r\n canvas.pack(side=\"left\", fill=\"both\", expand= 1)\r\n\r\n scrollbar = ttk.Scrollbar(main_frame, orient= \"vertical\", command=canvas.yview)\r\n scrollbar.pack(side=\"right\", fill=\"y\")\r\n\r\n canvas.configure(yscrollcommand=scrollbar.set)\r\n canvas.bind(\"\", lambda e: canvas.configure(scrollregion=canvas.bbox(\"all\")))\r\n\r\n second_frame = tk.Frame(canvas)\r\n\r\n canvas.create_window((0,0), window=second_frame, anchor=\"nw\")\r\n\r\n def refresh_lyrics():\r\n refreshed_data = get_currently_playing_data()\r\n new_lyrics = get_lyrics(refreshed_data)\r\n str_var.set(new_lyrics)\r\n \r\n button = tk.Button(second_frame, text=\"Refresh lyrics\", command=refresh_lyrics).pack()\r\n \r\n str_var = tk.StringVar()\r\n str_var.set(lyrics)\r\n\r\n label = tk.Label(second_frame, textvariable=str_var).pack(padx=60)\r\n \r\n def _on_mousewheel(event):\r\n canvas.yview_scroll(int(-1*(event.delta/70)), \"units\")\r\n\r\n canvas.bind_all(\"\", _on_mousewheel)\r\n\r\n root.mainloop()\r\n \r\ndisplay_lyrics(lyrics)\r\n", "sub_path": "spotify_lyrics/run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 3539, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "spotify_client.SpotifyClient", "line_number": 12, "usage_type": "call"}, {"api_name": "threading.Timer", "line_number": 20, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 56, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 60, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 69, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 70, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 83, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 87, "usage_type": "call"}, {"api_name": "tkinter.Canvas", "line_number": 90, "usage_type": "call"}, {"api_name": "tkinter.ttk.Scrollbar", "line_number": 93, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 93, "usage_type": "name"}, {"api_name": "tkinter.Frame", "line_number": 99, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 108, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 110, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 113, "usage_type": "call"}]}
+{"seq_id": "96084023", "text": "from fastapi import FastAPI\r\nfrom books_route import router as books_router\r\n\r\napp = FastAPI()\r\n\r\napp.include_router(books_router)\r\n\r\n@app.get(\"/\")\r\nasync def read_main():\r\n return {\"message\": \"Hello Bigger Applications!\"}", "sub_path": "mongodb app/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 225, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "fastapi.FastAPI", "line_number": 4, "usage_type": "call"}, {"api_name": "books_route.router", "line_number": 6, "usage_type": "argument"}]}
+{"seq_id": "366275708", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('movies', '0005_movie_image1'),\n ]\n\n operations = [\n migrations.AlterField(\n model_name='movie',\n name='image1',\n field=models.ImageField(upload_to=b'movies/images/icons/', blank=True),\n preserve_default=True,\n ),\n ]\n", "sub_path": "movies/migrations/0006_auto_20161018_1712.py", "file_name": "0006_auto_20161018_1712.py", "file_ext": "py", "file_size_in_byte": 462, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}]}
+{"seq_id": "565146079", "text": "#!/usr/bin/env python2.7\n#coding:utf-8\n\nimport sys\nimport time\nimport socket\nimport logging\nimport subprocess\nfrom StringIO import StringIO\nfrom daemonize import Daemonize\n\nreload(sys)\nsys.setdefaultencoding('utf-8')\n\n\nclass CheckAdmin(object):\n def __init__(self):\n self.address = '127.0.0.1' \n self.port = 9090\n self.request = 'GET /login HTTP/1.1\\r\\nHost:127.0.0.1\\r\\n\\r\\n'\n \n def check(self, logger):\n try:\n s = socket.socket()\n s.connect((self.address, self.port))\n s.send(self.request)\n line = s.recv(100)\n status = line.split()[1]\n except Exception as e:\n logger.info(str(e))\n return\n else:\n if status != '200':\n logger.info('admin down {0:=<{1}} '.format(time.strftime(\"%Y-%m-%d %H:%M:%S\", time.localtime(int(time.time()))), 50))\n cmd = '''\n cd /data/www/center-new/http/ ;\n sh shutdown.sh ;\n sh startup.sh\n '''\n logger.info(subprocess.Popen(cmd,shell=True,stdout=subprocess.PIPE,stderr=subprocess.PIPE).communicate())\n finally:\n s.close()\n return\n\n __call__ = check \n\ndef main():\n logger.info('check admin up/down')\n while True:\n time.sleep(120)\n check_admin(logger)\n\npid = \"/data/tmp/check_admin.pid\"\nlogger = logging.getLogger(__name__)\nlogger.setLevel(logging.DEBUG)\nlogger.propagate = False\nfh = logging.FileHandler(\"/data/tmp/check_admin.log\", \"w\")\nfh.setLevel(logging.DEBUG)\nformatter = logging.Formatter('%(asctime)-8s: %(message)s')\nfh.setFormatter(formatter)\nlogger.addHandler(fh)\nkeep_fds = [fh.stream.fileno()]\n\ncheck_admin = CheckAdmin()\ndaemon = Daemonize(app='check_admin', pid=pid, action=main, keep_fds=keep_fds)\ndaemon.start()\n", "sub_path": "check_admin.py", "file_name": "check_admin.py", "file_ext": "py", "file_size_in_byte": 1869, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "sys.setdefaultencoding", "line_number": 13, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 24, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 34, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 34, "usage_type": "call"}, {"api_name": "time.time", "line_number": 34, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 40, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 40, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 50, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 54, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 55, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 57, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 58, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 59, "usage_type": "call"}, {"api_name": "daemonize.Daemonize", "line_number": 65, "usage_type": "call"}]}
+{"seq_id": "119765943", "text": "######################################################################################\n# Copyright (c) muhanzhang, D-VAE, NeurIPS 2019 [GitHub D-VAE]\n# Modified by Hayeon Lee, Eunyoung Hyung, MetaD2A, ICLR2021, 2021. 03 [GitHub MetaD2A]\n######################################################################################\nimport math\nimport random\nimport torch\nfrom torch import nn\nfrom torch.nn import functional as F\nimport numpy as np\nimport igraph\nimport pdb\nfrom set_encoder.setenc_models import SetPool\n\n\nclass GeneratorModel(nn.Module):\n def __init__(self, args, graph_config):\n super(GeneratorModel, self).__init__()\n self.max_n = graph_config['max_n'] # maximum number of vertices\n self.nvt = args.nvt # number of vertex types\n self.START_TYPE = graph_config['START_TYPE']\n self.END_TYPE = graph_config['END_TYPE']\n self.hs = args.hs # hidden state size of each vertex\n self.nz = args.nz # size of latent representation z\n self.gs = args.hs # size of graph state\n self.bidir = True # whether to use bidirectional encoding\n self.vid = True\n self.device = None\n self.num_sample = args.num_sample\n\n if self.vid:\n self.vs = self.hs + self.max_n # vertex state size = hidden state + vid\n else:\n self.vs = self.hs\n \n # 0. encoding-related\n self.grue_forward = nn.GRUCell(self.nvt, self.hs) # encoder GRU\n self.grue_backward = nn.GRUCell(self.nvt, self.hs) # backward encoder GRU\n self.enc_g_mu = nn.Linear(self.gs, self.nz) # latent mean\n self.enc_g_var = nn.Linear(self.gs, self.nz) # latent var\n self.fc1 = nn.Linear(self.gs, self.nz) # latent mean\n self.fc2 = nn.Linear(self.gs, self.nz) # latent logvar\n\n # 1. decoding-related\n self.grud = nn.GRUCell(self.nvt, self.hs) # decoder GRU\n self.fc3 = nn.Linear(self.nz, self.hs) # from latent z to initial hidden state h0\n self.add_vertex = nn.Sequential(\n nn.Linear(self.hs, self.hs * 2),\n nn.ReLU(),\n nn.Linear(self.hs * 2, self.nvt)\n ) # which type of new vertex to add f(h0, hg)\n self.add_edge = nn.Sequential(\n nn.Linear(self.hs * 2, self.hs * 4),\n nn.ReLU(),\n nn.Linear(self.hs * 4, 1)\n ) # whether to add edge between v_i and v_new, f(hvi, hnew)\n self.decoding_gate = nn.Sequential(\n nn.Linear(self.vs, self.hs),\n nn.Sigmoid()\n )\n self.decoding_mapper = nn.Sequential(\n nn.Linear(self.vs, self.hs, bias=False),\n ) # disable bias to ensure padded zeros also mapped to zeros\n\n # 2. gate-related\n self.gate_forward = nn.Sequential(\n nn.Linear(self.vs, self.hs),\n nn.Sigmoid()\n )\n self.gate_backward = nn.Sequential(\n nn.Linear(self.vs, self.hs),\n nn.Sigmoid()\n )\n self.mapper_forward = nn.Sequential(\n nn.Linear(self.vs, self.hs, bias=False),\n ) # disable bias to ensure padded zeros also mapped to zeros\n self.mapper_backward = nn.Sequential(\n nn.Linear(self.vs, self.hs, bias=False),\n )\n \n # 3. bidir-related, to unify sizes\n if self.bidir:\n self.hv_unify = nn.Sequential(\n nn.Linear(self.hs * 2, self.hs),\n )\n self.hg_unify = nn.Sequential(\n nn.Linear(self.gs * 2, self.gs),\n )\n\n # 4. other\n self.relu = nn.ReLU()\n self.sigmoid = nn.Sigmoid()\n self.tanh = nn.Tanh()\n self.logsoftmax1 = nn.LogSoftmax(1)\n \n # 6. predictor\n np = self.gs\n self.intra_setpool = SetPool(dim_input=512, \n num_outputs=1, \n dim_output=self.nz, \n dim_hidden=self.nz, \n mode='sabPF')\n self.inter_setpool = SetPool(dim_input=self.nz, \n num_outputs=1, \n dim_output=self.nz, \n dim_hidden=self.nz, \n mode='sabPF')\n self.set_fc = nn.Sequential(\n nn.Linear(512, self.nz),\n nn.ReLU())\n\n def get_device(self):\n if self.device is None:\n self.device = next(self.parameters()).device\n return self.device\n \n def _get_zeros(self, n, length):\n return torch.zeros(n, length).to(self.get_device()) # get a zero hidden state\n \n def _get_zero_hidden(self, n=1):\n return self._get_zeros(n, self.hs) # get a zero hidden state\n \n def _one_hot(self, idx, length):\n if type(idx) in [list, range]:\n if idx == []:\n return None\n idx = torch.LongTensor(idx).unsqueeze(0).t()\n x = torch.zeros((len(idx), length)\n ).scatter_(1, idx, 1).to(self.get_device())\n else:\n idx = torch.LongTensor([idx]).unsqueeze(0)\n x = torch.zeros((1, length)\n ).scatter_(1, idx, 1).to(self.get_device())\n return x\n \n def _gated(self, h, gate, mapper):\n return gate(h) * mapper(h)\n \n def _collate_fn(self, G):\n return [g.copy() for g in G]\n \n def _propagate_to(self, G, v, propagator, \n H=None, reverse=False, gate=None, mapper=None):\n # propagate messages to vertex index v for all graphs in G\n # return the new messages (states) at v\n G = [g for g in G if g.vcount() > v]\n if len(G) == 0:\n return\n if H is not None:\n idx = [i for i, g in enumerate(G) if g.vcount() > v]\n H = H[idx]\n v_types = [g.vs[v]['type'] for g in G]\n X = self._one_hot(v_types, self.nvt)\n H_name = 'H_forward' # name of the hidden states attribute\n H_pred = [[g.vs[x][H_name] for x in g.predecessors(v)] for g in G]\n if self.vid:\n vids = [self._one_hot(g.predecessors(v), self.max_n) for g in G]\n if reverse:\n H_name = 'H_backward' # name of the hidden states attribute\n H_pred = [[g.vs[x][H_name] for x in g.successors(v)] for g in G]\n if self.vid:\n vids = [self._one_hot(g.successors(v), self.max_n) for g in G]\n gate, mapper = self.gate_backward, self.mapper_backward\n else:\n H_name = 'H_forward' # name of the hidden states attribute\n H_pred = [\n [g.vs[x][H_name] for x in g.predecessors(v)] for g in G]\n if self.vid:\n vids = [\n self._one_hot(g.predecessors(v), self.max_n) for g in G]\n if gate is None:\n gate, mapper = self.gate_forward, self.mapper_forward\n if self.vid:\n H_pred = [[torch.cat(\n [x[i], y[i:i + 1]], 1) for i in range(len(x))\n ] for x, y in zip(H_pred, vids)]\n # if h is not provided, use gated sum of v's predecessors' states as the input hidden state\n if H is None:\n max_n_pred = max([len(x) for x in H_pred]) # maximum number of predecessors\n if max_n_pred == 0:\n H = self._get_zero_hidden(len(G))\n else:\n H_pred = [torch.cat(h_pred + \n [self._get_zeros(max_n_pred - len(h_pred), \n self.vs)], 0).unsqueeze(0)\n for h_pred in H_pred] # pad all to same length\n H_pred = torch.cat(H_pred, 0) # batch * max_n_pred * vs\n H = self._gated(H_pred, gate, mapper).sum(1) # batch * hs\n Hv = propagator(X, H)\n for i, g in enumerate(G):\n g.vs[v][H_name] = Hv[i:i + 1]\n return Hv\n \n def _propagate_from(self, G, v, propagator, H0=None, reverse=False):\n # perform a series of propagation_to steps starting from v following a topo order\n # assume the original vertex indices are in a topological order\n if reverse:\n prop_order = range(v, -1, -1)\n else:\n prop_order = range(v, self.max_n)\n Hv = self._propagate_to(G, v, propagator, H0, reverse=reverse) # the initial vertex\n for v_ in prop_order[1:]:\n self._propagate_to(G, v_, propagator, reverse=reverse)\n return Hv\n \n def _update_v(self, G, v, H0=None):\n # perform a forward propagation step at v when decoding to update v's state\n # self._propagate_to(G, v, self.grud, H0, reverse=False)\n self._propagate_to(G, v, self.grud, H0, \n reverse=False, gate=self.decoding_gate, \n mapper=self.decoding_mapper)\n return\n \n def _get_vertex_state(self, G, v):\n # get the vertex states at v\n Hv = []\n for g in G:\n if v >= g.vcount():\n hv = self._get_zero_hidden()\n else:\n hv = g.vs[v]['H_forward']\n Hv.append(hv)\n Hv = torch.cat(Hv, 0)\n return Hv\n\n def _get_graph_state(self, G, decode=False):\n # get the graph states\n # when decoding, use the last generated vertex's state as the graph state\n # when encoding, use the ending vertex state or unify the starting and ending vertex states\n Hg = []\n for g in G:\n hg = g.vs[g.vcount() - 1]['H_forward']\n if self.bidir and not decode: # decoding never uses backward propagation\n hg_b = g.vs[0]['H_backward']\n hg = torch.cat([hg, hg_b], 1)\n Hg.append(hg)\n Hg = torch.cat(Hg, 0)\n if self.bidir and not decode:\n Hg = self.hg_unify(Hg)\n return Hg\n\n def graph_encode(self, G):\n # encode graphs G into latent vectors\n if type(G) != list:\n G = [G]\n self._propagate_from(G, 0, self.grue_forward, \n H0=self._get_zero_hidden(len(G)), reverse=False)\n if self.bidir:\n self._propagate_from(G, self.max_n - 1, self.grue_backward,\n H0=self._get_zero_hidden(len(G)), reverse=True)\n Hg = self._get_graph_state(G)\n mu, logvar = self.enc_g_mu(Hg), self.enc_g_var(Hg)\n return mu, logvar\n\n\n def set_encode(self, X):\n proto_batch = []\n for x in X: # X.shape: [32, 400, 512]\n cls_protos = self.intra_setpool(\n x.view(-1, self.num_sample, 512)).squeeze(1)\n proto_batch.append(\n self.inter_setpool(cls_protos.unsqueeze(0)))\n v = torch.stack(proto_batch).squeeze()\n mu, logvar = self.fc1(v), self.fc2(v)\n return mu, logvar\n\n\n def reparameterize(self, mu, logvar, eps_scale=0.01):\n # return z ~ N(mu, std)\n if self.training:\n std = logvar.mul(0.5).exp_()\n eps = torch.randn_like(std) * eps_scale\n return eps.mul(std).add_(mu)\n else:\n return mu\n\n def _get_edge_score(self, Hvi, H, H0):\n # compute scores for edges from vi based on Hvi, H (current vertex) and H0\n # in most cases, H0 need not be explicitly included since Hvi and H contain its information\n return self.sigmoid(self.add_edge(torch.cat([Hvi, H], -1)))\n\n def graph_decode(self, z, stochastic=True):\n # decode latent vectors z back to graphs\n # if stochastic=True, stochastically sample each action from the predicted distribution;\n # otherwise, select argmax action deterministically.\n H0 = self.tanh(self.fc3(z)) # or relu activation, similar performance\n G = [igraph.Graph(directed=True) for _ in range(len(z))]\n for g in G:\n g.add_vertex(type=self.START_TYPE)\n self._update_v(G, 0, H0)\n finished = [False] * len(G)\n for idx in range(1, self.max_n):\n # decide the type of the next added vertex\n if idx == self.max_n - 1: # force the last node to be end_type\n new_types = [self.END_TYPE] * len(G)\n else:\n Hg = self._get_graph_state(G, decode=True)\n type_scores = self.add_vertex(Hg)\n if stochastic:\n type_probs = F.softmax(type_scores, 1\n ).cpu().detach().numpy()\n new_types = [np.random.choice(range(self.nvt), \n p=type_probs[i]) for i in range(len(G))]\n else:\n new_types = torch.argmax(type_scores, 1)\n new_types = new_types.flatten().tolist()\n for i, g in enumerate(G):\n if not finished[i]:\n g.add_vertex(type=new_types[i])\n self._update_v(G, idx)\n \n # decide connections\n edge_scores = []\n for vi in range(idx - 1, -1, -1):\n Hvi = self._get_vertex_state(G, vi)\n H = self._get_vertex_state(G, idx)\n ei_score = self._get_edge_score(Hvi, H, H0)\n if stochastic:\n random_score = torch.rand_like(ei_score)\n decisions = random_score < ei_score\n else:\n decisions = ei_score > 0.5\n for i, g in enumerate(G):\n if finished[i]:\n continue\n if new_types[i] == self.END_TYPE:\n # if new node is end_type, connect it to all loose-end vertices (out_degree==0)\n end_vertices = set([\n v.index for v in g.vs.select(_outdegree_eq=0)\n if v.index != g.vcount() - 1])\n for v in end_vertices:\n g.add_edge(v, g.vcount() - 1)\n finished[i] = True\n continue\n if decisions[i, 0]:\n g.add_edge(vi, g.vcount() - 1)\n self._update_v(G, idx)\n \n for g in G:\n del g.vs['H_forward'] # delete hidden states to save GPU memory\n return G\n \n\n def loss(self, mu, logvar, G_true, beta=0.005):\n # compute the loss of decoding mu and logvar to true graphs using teacher forcing\n # ensure when computing the loss of step i, steps 0 to i-1 are correct\n z = self.reparameterize(mu, logvar)\n H0 = self.tanh(self.fc3(z)) # or relu activation, similar performance\n G = [igraph.Graph(directed=True) for _ in range(len(z))]\n for g in G:\n g.add_vertex(type=self.START_TYPE)\n self._update_v(G, 0, H0)\n res = 0 # log likelihood\n for v_true in range(1, self.max_n):\n # calculate the likelihood of adding true types of nodes\n # use start type to denote padding vertices since start type only appears for vertex 0\n # and will never be a true type for later vertices, thus it's free to use\n true_types = [g_true.vs[v_true]['type'] \n if v_true < g_true.vcount()\n else self.START_TYPE for g_true in G_true]\n Hg = self._get_graph_state(G, decode=True)\n type_scores = self.add_vertex(Hg)\n # vertex log likelihood\n vll = self.logsoftmax1(type_scores)[\n np.arange(len(G)), true_types].sum()\n res = res + vll\n for i, g in enumerate(G):\n if true_types[i] != self.START_TYPE:\n g.add_vertex(type=true_types[i])\n self._update_v(G, v_true)\n \n # calculate the likelihood of adding true edges\n true_edges = []\n for i, g_true in enumerate(G_true):\n true_edges.append(g_true.get_adjlist(igraph.IN)[v_true] \n if v_true < g_true.vcount() else [])\n edge_scores = []\n for vi in range(v_true - 1, -1, -1):\n Hvi = self._get_vertex_state(G, vi)\n H = self._get_vertex_state(G, v_true)\n ei_score = self._get_edge_score(Hvi, H, H0)\n edge_scores.append(ei_score)\n for i, g in enumerate(G):\n if vi in true_edges[i]:\n g.add_edge(vi, v_true)\n self._update_v(G, v_true)\n edge_scores = torch.cat(edge_scores[::-1], 1)\n \n ground_truth = torch.zeros_like(edge_scores)\n idx1 = [i for i, x in enumerate(true_edges) \n for _ in range(len(x))]\n idx2 = [xx for x in true_edges for xx in x]\n ground_truth[idx1, idx2] = 1.0\n \n # edges log-likelihood\n ell = - F.binary_cross_entropy(\n edge_scores, ground_truth, reduction='sum')\n res = res + ell\n \n res = -res # convert likelihood to loss\n kld = -0.5 * torch.sum(1 + logvar - mu.pow(2) - logvar.exp())\n return res + beta * kld, res, kld", "sub_path": "MetaD2A_nas_bench_201/generator/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 15219, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "torch.nn.Module", "line_number": 16, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.GRUCell", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.GRUCell", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "name"}, {"api_name": "torch.nn.GRUCell", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 47, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 48, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 50, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 54, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 55, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 57, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 58, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 61, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 62, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 66, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 67, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 68, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 70, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 71, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 72, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 74, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 75, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 77, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 83, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 84, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 86, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 87, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 91, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 92, "usage_type": "name"}, {"api_name": "torch.nn.Tanh", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 93, "usage_type": "name"}, {"api_name": "torch.nn.LogSoftmax", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 94, "usage_type": "name"}, {"api_name": "set_encoder.setenc_models.SetPool", "line_number": 98, "usage_type": "call"}, {"api_name": "set_encoder.setenc_models.SetPool", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.nn.Sequential", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 108, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 109, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 110, "usage_type": "name"}, {"api_name": "torch.zeros", "line_number": 118, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 128, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 174, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 183, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 187, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 223, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 235, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 237, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 263, "usage_type": "call"}, {"api_name": "torch.randn_like", "line_number": 272, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 280, "usage_type": "call"}, {"api_name": "igraph.Graph", "line_number": 287, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 300, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 300, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 302, "usage_type": "attribute"}, {"api_name": "torch.argmax", "line_number": 305, "usage_type": "call"}, {"api_name": "torch.rand_like", "line_number": 319, "usage_type": "call"}, {"api_name": "igraph.Graph", "line_number": 349, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 365, "usage_type": "call"}, {"api_name": "igraph.IN", "line_number": 375, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 387, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 389, "usage_type": "call"}, {"api_name": "torch.nn.functional.binary_cross_entropy", "line_number": 396, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 396, "usage_type": "name"}, {"api_name": "torch.sum", "line_number": 401, "usage_type": "call"}]}
+{"seq_id": "512334895", "text": "import math\nimport pygame\n\natraction = 0.02\n\nclass Link(object):\n def __init__(self, node1, node2, strength=1):\n self.node1 = node1\n self.node2 = node2\n self.strength = strength\n node1.add_link(self)\n node2.add_link(self)\n\n def update(self):\n if self.node1.pos != None and self.node2.pos != None:\n rel = (self.node2.pos[0]-self.node1.pos[0], self.node2.pos[1]-self.node1.pos[1])\n self.node1.velocity = (self.node1.velocity[0]+(rel[0]*atraction*self.strength), self.node1.velocity[1]+(rel[1]*atraction*self.strength))\n self.node2.velocity = (self.node2.velocity[0]-(rel[0]*atraction*self.strength), self.node2.velocity[1]-(rel[1]*atraction*self.strength))\n\n def draw(self, surface):\n rel = (self.node2.pos[0]-self.node1.pos[0], self.node2.pos[1]-self.node1.pos[1])\n distance = math.sqrt(rel[0]**2 + rel[1]**2)\n rel_distance1 = (self.node1.radius if not (self.node1.selected or self.node1.hovering) else self.node1.selected_radius) / (distance+1)\n rel_distance2 = (self.node2.radius if not (self.node2.selected or self.node2.hovering) else self.node2.selected_radius) / (distance+1)\n start_point = (self.node1.pos[0]+(rel[0]*rel_distance1), self.node1.pos[1]+(rel[1]*rel_distance1))\n end_point = (self.node2.pos[0]-(rel[0]*rel_distance2), self.node2.pos[1]-(rel[1]*rel_distance2))\n pygame.draw.line(surface, (0, 0, 0), start_point, end_point, int(self.strength))\n", "sub_path": "link.py", "file_name": "link.py", "file_ext": "py", "file_size_in_byte": 1495, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "math.sqrt", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.draw.line", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 27, "usage_type": "attribute"}]}
+{"seq_id": "71213311", "text": "from functools import reduce\n\nimport datajoint as dj\nimport numpy as np\nimport scipy.stats as stats\nimport warnings\nimport spikeinterface as si\nfrom spikeinterface.core.segmentutils import AppendSegmentRecording\n\nfrom .common_interval import IntervalList\nfrom .common_spikesorting import SpikeSortingRecording\nfrom .common_session import Session\nfrom .nwb_helper_fn import get_valid_intervals\n\nschema = dj.schema('common_artifact')\n\n@schema\nclass ArtifactDetectionParameters(dj.Manual):\n definition = \"\"\"\n # Parameters for detecting artifact times within a sort group.\n artifact_params_name: varchar(200)\n ---\n artifact_params: blob # dictionary of parameters\n \"\"\"\n def insert_default(self):\n \"\"\"Insert the default artifact parameters with an appropriate parameter dict.\n \"\"\"\n artifact_params = {}\n artifact_params['zscore_thresh'] = None # must be None or >= 0\n artifact_params['amplitude_thresh'] = 3000 # must be None or >= 0\n artifact_params['proportion_above_thresh'] = 1.0 # all electrodes of sort group\n artifact_params['removal_window_ms'] = 1.0 # in milliseconds\n self.insert1(['default', artifact_params], skip_duplicates=True) \n\n artifact_params_none = {}\n artifact_params_none['zscore_thresh'] = None \n artifact_params_none['amplitude_thresh'] = None \n self.insert1(['none', artifact_params_none], skip_duplicates=True) \n\n@schema\nclass ArtifactDetectionSelection(dj.Manual):\n definition = \"\"\"\n # Specifies artifact detection parameters to apply to a sort group's recording.\n -> SpikeSortingRecording\n -> ArtifactDetectionParameters\n ---\n \"\"\"\n\n@schema\nclass ArtifactDetection(dj.Computed):\n definition = \"\"\"\n # Stores artifact times and valid no-artifact times as intervals.\n -> ArtifactDetectionSelection\n ---\n artifact_times: longblob # np array of artifact intervals\n artifact_removed_valid_times: longblob # np array of valid no-artifact intervals\n artifact_removed_interval_list_name: varchar(200) # name of the array of no-artifact valid time intervals\n \"\"\"\n\n def make(self, key):\n # get the dict of artifact params associated with this artifact_params_name\n artifact_params = (ArtifactDetectionParameters & key).fetch1(\"artifact_params\")\n \n recording_path = (SpikeSortingRecording & key).fetch1('recording_path')\n recording = si.load_extractor(recording_path) \n \n artifact_removed_valid_times, artifact_times = _get_artifact_times(recording, **artifact_params)\n \n # NOTE: decided not to do this but to just create a single long segment; keep for now\n # get artifact times by segment\n # if AppendSegmentRecording, get artifact times for each segment\n # if isinstance(recording, AppendSegmentRecording):\n # artifact_removed_valid_times = []\n # artifact_times = []\n # for rec in recording.recording_list:\n # rec_valid_times, rec_artifact_times = _get_artifact_times(rec, **artifact_params)\n # for valid_times in rec_valid_times:\n # artifact_removed_valid_times.append(valid_times)\n # for artifact_times in rec_artifact_times:\n # artifact_times.append(artifact_times)\n # artifact_removed_valid_times = np.asarray(artifact_removed_valid_times)\n # artifact_times = np.asarray(artifact_times)\n # else:\n # artifact_removed_valid_times, artifact_times = _get_artifact_times(recording, **artifact_params)\n\n key['artifact_times'] = artifact_times\n key['artifact_removed_valid_times'] = artifact_removed_valid_times\n \n # set up a name for no-artifact times using recording id\n key['artifact_removed_interval_list_name'] = key['recording_id'] + '_' + key['artifact_params_name'] + '_artifact_removed_valid_times'\n \n # insert artifact times and valid times into ArtifactRemovedIntervalList with an appropriate name\n tmp_key = {}\n tmp_key['nwb_file_name'] = key['nwb_file_name']\n tmp_key['artifact_removed_interval_list_name'] = key['artifact_removed_interval_list_name']\n tmp_key['artifact_removed_valid_times'] = key['artifact_removed_valid_times']\n tmp_key['artifact_times'] = key['artifact_times']\n ArtifactRemovedIntervalList.insert1(tmp_key, skip_duplicates = True)\n \n # also insert into IntervalList\n tmp_key = {}\n tmp_key['nwb_file_name'] = key['nwb_file_name']\n tmp_key['interval_list_name'] = key['artifact_removed_interval_list_name']\n tmp_key['valid_times'] = key['artifact_removed_valid_times']\n IntervalList.insert1(tmp_key, skip_duplicates=True)\n \n # insert into computed table\n self.insert1(key)\n\n@schema\nclass ArtifactRemovedIntervalList(dj.Manual):\n definition = \"\"\"\n # Stores intervals without detected artifacts.\n # Note that entries can come from either ArtifactDetection() or alternative artifact removal analyses.\n -> Session\n artifact_removed_interval_list_name: varchar(200)\n ---\n artifact_removed_valid_times: longblob\n artifact_times: longblob # np array of artifact intervals\n \"\"\"\n \ndef _get_artifact_times(recording, zscore_thresh=None, amplitude_thresh=None,\n proportion_above_thresh=1.0, removal_window_ms=1.0):\n \"\"\"Detects times during which artifacts do and do not occur.\n Artifacts are defined as periods where the absolute value of the recording signal exceeds one\n OR both specified amplitude or zscore thresholds on the proportion of channels specified,\n with the period extended by the removal_window_ms/2 on each side. Z-score and amplitude\n threshold values of None are ignored.\n\n Parameters\n ----------\n recording : si.Recording\n zscore_thresh : float, optional\n Stdev threshold for exclusion, should be >=0, defaults to None\n amplitude_thresh : float, optional\n Amplitude threshold for exclusion, should be >=0, defaults to None\n proportion_above_thresh : float, optional, should be>0 and <=1\n Proportion of electrodes that need to have threshold crossings, defaults to 1 \n removal_window_ms : float, optional\n Width of the window in milliseconds to mask out per artifact (window/2 removed on each side of threshold crossing), defaults to 1 ms\n \n Returns\n ------_\n artifact_intervals : np.ndarray\n Intervals in which artifacts are detected (including removal windows), unit: seconds\n artifact_removed_valid_times : np.ndarray\n Intervals of valid times where artifacts were not detected, unit: seconds\n \"\"\"\n \n valid_timestamps = SpikeSortingRecording._get_recording_timestamps(recording)\n if recording.get_num_segments()>1:\n recording = si.concatenate_recordings(recording.recording_list)\n \n # if both thresholds are None, we essentially skip artifract detection and\n # return an array with the times of the first and last samples of the recording\n if (amplitude_thresh is None) and (zscore_thresh is None):\n recording_interval = np.asarray([valid_timestamps[0], valid_timestamps[-1]])\n artifact_times_empty = np.asarray([])\n print(\"Amplitude and zscore thresholds are both None, skipping artifact detection\")\n return recording_interval, artifact_times_empty\n \n # verify threshold parameters\n amplitude_thresh, zscore_thresh, proportion_above_thresh = _check_artifact_thresholds(amplitude_thresh, zscore_thresh, proportion_above_thresh)\n\n # turn ms to remove total into s to remove from either side of each detected artifact\n half_removal_window_s = removal_window_ms * (1/1000) * (1/2)\n \n # TODO: load by chunk to avoid memory problems\n data = recording.get_traces()\n\n # compute the number of electrodes that have to be above threshold\n nelect_above = np.ceil(proportion_above_thresh * len(recording.get_channel_ids()))\n\n # find the artifact occurrences using one or both thresholds, across channels\n if ((amplitude_thresh is not None) and (zscore_thresh is None)):\n above_a = np.abs(data) > amplitude_thresh\n above_thresh = np.ravel(np.argwhere(np.sum(above_a, axis=0) >= nelect_above))\n elif ((amplitude_thresh is None) and (zscore_thresh is not None)):\n dataz = np.abs(stats.zscore(data, axis=1))\n above_z = dataz > zscore_thresh\n above_thresh = np.ravel(np.argwhere(np.sum(above_z, axis=0) >= nelect_above))\n else:\n above_a = np.abs(data) > amplitude_thresh\n dataz = np.abs(stats.zscore(data, axis=1))\n above_z = dataz > zscore_thresh\n above_thresh = np.ravel(np.argwhere(\n np.sum(np.logical_or(above_z, above_a), axis=0) >= nelect_above))\n \n if len(above_thresh) == 0:\n recording_interval = np.asarray([[valid_timestamps[0], valid_timestamps[-1]]])\n artifact_times_empty = np.asarray([])\n print(\"No artifacts detected.\")\n return recording_interval, artifact_times_empty\n\n above_thresh_times = valid_timestamps[above_thresh] # find timestamps of initial artifact threshold crossings\n \n # keep track of all the artifact timestamps within each artifact removal window and the indices of those timestamps\n artifact_times = []\n artifact_indices = []\n for a in above_thresh_times:\n a_times = np.copy(valid_timestamps[(valid_timestamps > (a - half_removal_window_s)) & (valid_timestamps <= (a + half_removal_window_s))])\n a_indices = np.argwhere((valid_timestamps > (a - half_removal_window_s)) & (valid_timestamps <= (a + half_removal_window_s)))\n artifact_times.append(a_times)\n artifact_indices.append(a_indices)\n all_artifact_times = reduce(np.union1d, artifact_times)\n all_artifact_indices = reduce(np.union1d, artifact_indices)\n # turn artifact detected times into intervals\n if not np.all(all_artifact_times[:-1] <= all_artifact_times[1:]): #should be faster than diffing and comparing to zero\n warnings.warn(\"Warning: sorting artifact timestamps; all_artifact_times was not strictly increasing\")\n all_artifact_times = np.sort(all_artifact_times)\n artifact_intervals = get_valid_intervals(all_artifact_times, recording.get_sampling_frequency(), 1.5, .000001)\n\n artifact_percent_of_times = 100 * len(all_artifact_times) / len(valid_timestamps)\n print(f\"{len(artifact_intervals)} artifact intervals detected;\\\n {artifact_percent_of_times} % of the recording's valid_timestamps removed as artifact\")\n \n # turn all artifact detected times into -1 to easily find non-artifact intervals\n valid_timestamps[all_artifact_indices] = -1\n artifact_removed_valid_times = get_valid_intervals(valid_timestamps[valid_timestamps != -1], \n recording.get_sampling_frequency(), 1.5, 0.000001) \n \n return artifact_removed_valid_times, artifact_intervals\n\ndef _check_artifact_thresholds(amplitude_thresh, zscore_thresh, proportion_above_thresh):\n \"\"\"Alerts user to likely unintended parameters. Not an exhaustive verification.\n\n Parameters\n ----------\n zscore_thresh: float\n amplitude_thresh: float\n proportion_above_thresh: float\n\n Return\n ------\n zscore_thresh: float\n amplitude_thresh: float\n proportion_above_thresh: float\n\n Raise\n ------\n ValueError: if signal thresholds are negative \n \"\"\"\n # amplitude or zscore thresholds should be negative, as they are applied to an absolute signal\n signal_thresholds = [t for t in [amplitude_thresh, zscore_thresh] if t is not None]\n for t in signal_thresholds:\n if t < 0:\n raise ValueError(\"Amplitude and Z-Score thresholds must be >= 0, or None\")\n \n # proportion_above_threshold should be in [0:1] inclusive\n if proportion_above_thresh < 0:\n warnings.warn(\"Warning: proportion_above_thresh must be a proportion >0 and <=1. Using proportion_above_thresh = 0.01 instead of \"+str(proportion_above_thresh))\n proportion_above_thresh = 0.01\n elif proportion_above_thresh > 1:\n warnings.warn(\"Warning: proportion_above_thresh must be a proportion >0 and <=1. Using proportion_above_thresh = 1 instead of \"+str(proportion_above_thresh))\n proportion_above_thresh = 1\n return amplitude_thresh, zscore_thresh, proportion_above_thresh\n", "sub_path": "src/nwb_datajoint/common/common_artifact.py", "file_name": "common_artifact.py", "file_ext": "py", "file_size_in_byte": 12488, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "datajoint.schema", "line_number": 15, "usage_type": "call"}, {"api_name": "datajoint.Manual", "line_number": 18, "usage_type": "attribute"}, {"api_name": "datajoint.Manual", "line_number": 41, "usage_type": "attribute"}, {"api_name": "datajoint.Computed", "line_number": 50, "usage_type": "attribute"}, {"api_name": "common_spikesorting.SpikeSortingRecording", "line_number": 64, "usage_type": "name"}, {"api_name": "spikeinterface.load_extractor", "line_number": 65, "usage_type": "call"}, {"api_name": "common_interval.IntervalList.insert1", "line_number": 105, "usage_type": "call"}, {"api_name": "common_interval.IntervalList", "line_number": 105, "usage_type": "name"}, {"api_name": "datajoint.Manual", "line_number": 111, "usage_type": "attribute"}, {"api_name": "common_spikesorting.SpikeSortingRecording._get_recording_timestamps", "line_number": 150, "usage_type": "call"}, {"api_name": "common_spikesorting.SpikeSortingRecording", "line_number": 150, "usage_type": "name"}, {"api_name": "spikeinterface.concatenate_recordings", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 179, "usage_type": "call"}, {"api_name": "scipy.stats.zscore", "line_number": 179, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 179, "usage_type": "name"}, {"api_name": "numpy.ravel", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 184, "usage_type": "call"}, {"api_name": "scipy.stats.zscore", "line_number": 184, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 184, "usage_type": "name"}, {"api_name": "numpy.ravel", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.logical_or", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 202, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.union1d", "line_number": 205, "usage_type": "attribute"}, {"api_name": "functools.reduce", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.union1d", "line_number": 206, "usage_type": "attribute"}, {"api_name": "numpy.all", "line_number": 208, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 210, "usage_type": "call"}, {"api_name": "nwb_helper_fn.get_valid_intervals", "line_number": 211, "usage_type": "call"}, {"api_name": "nwb_helper_fn.get_valid_intervals", "line_number": 219, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 251, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 254, "usage_type": "call"}]}
+{"seq_id": "362785562", "text": "#helpful code: http://brianfarris.me/static/digit_recognizer.html\nfrom __future__ import print_function\nimport numpy as np\nimport time\nimport psutil\nimport sklearn\nfrom sklearn import cross_validation\nfrom sklearn.svm import LinearSVC\nfrom sklearn.metrics import accuracy_score\nimport keras\nfrom keras.datasets import fashion_mnist\nfrom sklearn.model_selection import StratifiedShuffleSplit\n\nstart_time=time.time()\n\n# load the MNIST digits dataset\n(x_train, y_train), (x_test, y_test) = fashion_mnist.load_data()\n\n#format data\nx_train = x_train.reshape(60000, 784)\nx_test = x_test.reshape(10000, 784)\nx_train = x_train.astype('float32')\nx_test = x_test.astype('float32')\nx_train /= 255\nx_test /= 255\nprint(\"Training matrix shape\", x_train.shape)\nprint(\"Testing matrix shape\", x_test.shape)\n\nresults=[]\nlist1=[] \nlist2=[] \nlist3=[]\n\nfor q in range(0,7):\n\trng = [2,3,4,5,10,20,40,60,80,100,200,400,600,800,1000,2000,3000,4000,5000] #limit is 5000: 50000 train, 10000 valid\t\n\tfor i in rng:\n\t\t################ CREATING TRAIN AND VAL SETS #################\n\t\tprint(\"########################\")\t\t\n\n\t\t#TRAINING DATA: Creating i instances of each class, VAL DATA: taking 20% of Train\n\t\tnum_train=i*10\n\t\tif(i>=5):\t\t\t\n\t\t\tnum_test=int(0.2*num_train)\n\t\telse:\n\t\t\tnum_test=10\n\n\t\ttrain_sss = StratifiedShuffleSplit(n_splits=5, \n test_size=num_test, train_size=num_train, random_state=0) \n\t\tfor train_index, test_index in train_sss.split(x_train, y_train):\n\t\t tempx_train, tempx_val = x_train[train_index], x_train[test_index]\n\t\t tempy_train, tempy_val = y_train[train_index], y_train[test_index]\n\t\tprint(\"size of svm balanced training set:\", len(tempx_train))\n\t\tprint(\"size of svm balanced val set:\", len(tempx_val))\n\t\n\t\t#verifying correct numbers\n\t\tprint(tempx_train.shape[0], 'train samples')\n\t\tprint(tempx_val.shape[0], 'valid samples')\n\t\tprint(x_test.shape[0], 'test samples')\n\t\t#################################################################\n\n\t\tprint(\"EVALUATION ON TESTING DATA FOR\", num_train, \"TRAINING DATA POINTS\")\n\t\tclf_svm = LinearSVC()\n\t\tclf_svm.fit(tempx_train, tempy_train)\n\t\ty_pred_svm = clf_svm.predict(x_test)\n\t\tacc_svm = accuracy_score(y_test, y_pred_svm)\n\t\tprint(\"Linear SVM accuracy: \",acc_svm)\n\t\tresults=[i,acc_svm]\n\t\tif(q==0):\n\t\t\tlist1.append(results)\n\t\telif(q==1):\n\t\t\tlist2.append(results)\n\t\telse:\n\t\t\tlist3.append(results)\n\ndatapoints=len(list1)\n######finding the variance of three tests#######\nvariance=np.zeros((datapoints,2))\neach_pt=np.zeros((3,2))\nfor num in range(0,datapoints):\n\tvariance[num][0]=list1[num][0]\n\teach_pt[0][0]=list1[num][0]\n\teach_pt[0][1]=list1[num][1]\n\teach_pt[1][0]=list2[num][0]\n\teach_pt[1][1]=list2[num][1]\n\teach_pt[2][0]=list3[num][0]\n\teach_pt[2][1]=list3[num][1]\n\teach_var=np.var(each_pt,axis=0)\n\tvariance[num][1]=each_var[1]\nprint(variance)\n\n######finding the average of three tests########\noverall=np.zeros((datapoints,2))\t\t\nfor num in range(0,datapoints):\n\toverall[num][0]=list1[num][0]\t\n\toverall[num][1]=list1[num][1]+list2[num][1]+list3[num][1]\nfor num in range(0,datapoints):\n\toverall[num][1]=(overall[num][1])/3.0\n\nprint(overall)\n\n######finding the change in accuracy############\nacc_change=np.zeros((datapoints-1,2))\nfor num in range(1,datapoints):\n\tacc_change[num-1][0]=num-1\n\tacc_change[num-1][1]=overall[num][1]-overall[num-1][1]\nprint(acc_change)\n\nnp.savetxt(\"svm_VariableTrainSet.csv\",overall,delimiter=\",\")\nnp.savetxt(\"svm_Variance.csv\",variance,delimiter=\",\")\nnp.savetxt(\"svm_AccuracyChange.csv\",acc_change,delimiter=\",\")\n\nprint(\"Total Time Elapsed: \", round(time.time()-start_time,1), \"seconds\")\nprint(psutil.virtual_memory())\n", "sub_path": "mnist_fashion_svm.py", "file_name": "mnist_fashion_svm.py", "file_ext": "py", "file_size_in_byte": 3616, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "time.time", "line_number": 14, "usage_type": "call"}, {"api_name": "keras.datasets.fashion_mnist.load_data", "line_number": 17, "usage_type": "call"}, {"api_name": "keras.datasets.fashion_mnist", "line_number": 17, "usage_type": "name"}, {"api_name": "sklearn.model_selection.StratifiedShuffleSplit", "line_number": 47, "usage_type": "call"}, {"api_name": "sklearn.svm.LinearSVC", "line_number": 62, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 110, "usage_type": "call"}, {"api_name": "time.time", "line_number": 112, "usage_type": "call"}, {"api_name": "psutil.virtual_memory", "line_number": 113, "usage_type": "call"}]}
+{"seq_id": "594519762", "text": "\nimport open3d as o3d\nimport numpy as np\nimport pickle as pkl\nimport os\nimport shutil\n\ndef clean_info(filename):\n \"\"\"\n some fbx downloaded from the internet has strange pattern, clean that\n \"\"\"\n with open(filename, \"r\") as f:\n content = f.read().strip()\n start = content.find('mix')\n end = content.find(':')\n if start == -1 or end == -1:\n return\n pattern = content[start:end+1]\n # print(pattern)\n content = content.replace(pattern, \"\")\n new_filename = filename.replace(\".txt\",\"_clean.txt\")\n with open(new_filename, \"w\") as f:\n f.write(content)\n print('clean finished')\n\nsmpl_joint_names = [\n \"hips\",\n \"leftUpLeg\",\n \"rightUpLeg\",\n \"spine\",\n \"leftLeg\",\n \"rightLeg\",\n \"spine1\",\n \"leftFoot\",\n \"rightFoot\",\n \"spine2\",\n \"leftToeBase\",\n \"rightToeBase\",\n \"neck\",\n \"leftShoulder\",\n \"rightShoulder\",\n \"head\",\n \"leftArm\",\n \"rightArm\",\n \"leftForeArm\",\n \"rightForeArm\",\n \"leftHand\",\n \"rightHand\",\n \"leftHandIndex1\",\n \"rightHandIndex1\",\n]\n\ndef print_joint2(infoname,save = True):\n if not os.path.exists(infoname):\n clean_info(infoname.replace(\"_clean\",\"\"))\n\n\n lines = open(infoname).readlines()\n\n\n\n infoname = infoname.replace(\"_clean\",\"\")\n meshname = infoname.replace(\".txt\", \".obj\")\n inmesh = o3d.io.read_triangle_mesh(meshname)\n # v_posed = np.array(inmesh.vertices)\n\n hier = {}\n joint2index = {}\n index = 0\n # parse rig info file and obtain kinematic chain(hierarchical structure)\n for line in lines:\n line = line.strip('\\n').strip()\n if line[:4] != 'hier':\n continue\n splits = line.split(' ')\n parent_name = splits[1]\n child_name = splits[2]\n if parent_name not in joint2index:\n joint2index[parent_name] = index\n index += 1\n if child_name not in joint2index:\n joint2index[child_name] = index\n index += 1\n if parent_name not in hier:\n hier[parent_name] = [child_name]\n else:\n hier[parent_name].append(child_name)\n\n index2joint = {v: k for k, v in joint2index.items()}\n hier_index = {}\n for k, v in hier.items():\n hier_index[joint2index[k]] = [joint2index[vv] for vv in v]\n parents = list(hier_index.keys())\n children = []\n for v in hier_index.values():\n children.extend(v)\n root = [item for item in parents if item not in children]\n assert len(root) == 1\n root = root[0]\n\n # reorganize the index mapping to ensure that along each chain,\n # from root node to leaf node, the index number increases\n new_hier = {}\n new_joint2index = {index2joint[root]: 0}\n top_level = [root]\n index = 1\n for item in top_level:\n if item not in hier_index:\n # print('continue')\n continue\n for child in hier_index[item]:\n child_name = index2joint[child]\n if child_name not in new_joint2index:\n new_joint2index[child_name] = index\n index += 1\n if new_joint2index[index2joint[item]] not in new_hier:\n new_hier[new_joint2index[index2joint[item]]] = []\n new_hier[new_joint2index[index2joint[item]]].append(new_joint2index[child_name])\n top_level.append(child)\n if save:\n savejoint_dir = infoname.replace(\".txt\", \"_joint.txt\")\n savejoint_dir = os.path.join(\"model\", \"match_list\", os.path.split(savejoint_dir)[-1])\n if os.path.exists(savejoint_dir):\n print(\"It has been jointed!\")\n return\n names = [s.lower() for s in smpl_joint_names]\n count = 0\n with open(savejoint_dir, 'w') as f:\n for joint, id in new_joint2index.items():\n if joint.lower() in names:\n f.write(\"%s : %d:%d\\n\" % (joint, id, names.index(joint.lower())))\n count += 1\n else:\n f.write(\"%s : %d\\n\" % (joint, id))\n f.write(\"\\n[%d/%d] TO MATCH\" % (count, len(smpl_joint_names)))\n print(\"finish joint!\")\n return None\n else:\n return new_joint2index\n\ndef read_match(path:str):\n #clean\n path = path.replace(\"_clean\",'').replace('_intermediate','')\n # add joint\n path = path.replace(\".txt\",\"_joint.txt\")\n match_path = os.path.join(\"model\",\"match_list\",os.path.split(path)[-1])\n assert os.path.exists(match_path),\"match_file doesn't exist!\"\n match = {}\n with open(match_path,'r') as f:\n s = f.read().split('\\n')\n for line in s:\n if line == '':\n break\n line = line[line.find(':') + 1:]\n if ':' in line:\n i = line.find(':')\n match[int(line[:i])] = int(line[i + 1:])\n\n return match\n\ndef perfect_matching():\n path = os.path.join(\"model\",\"group_0\",\"fbx\")\n for file in os.listdir(path):\n if 'txt' in file:\n s = print_joint2(os.path.join(path,file),False)\n count = 0\n for joint, id in s.items():\n if joint.lower() == 'spine':\n count +=1\n if joint.lower() == 'spine1':\n count +=1\n if joint.lower() == 'spine2':\n count +=1\n if count >= 3:\n print(file)\n\ndef load_pkl(path):\n with open(path,'rb') as f:\n m = pkl.load(f)\n print(m['model2smpl'])\n\ndef save_result(good_match_list):\n if not os.path.exists(\"submit_results\"):\n os.mkdir(\"submit_results\")\n for good_match in good_match_list:\n if not os.path.exists(os.path.join(\"submit_results\",good_match)):\n os.mkdir(os.path.join(\"submit_results\",good_match))\n else:\n pass\n path_obj = os.path.join(\"model\",\"group_0\",\"fbx\",good_match+\".obj\")\n path_fbx = os.path.join(\"model\", \"group_0\", \"fbx\", good_match + \".fbx\")\n path_txt = os.path.join(\"model\", \"group_0\", \"fbx\", good_match + \".txt\")\n path_vis = os.path.join(\"results\",good_match,\"vis.mp4\")\n for file in os.listdir(os.path.join(\"results\",good_match)):\n if 'pkl' in file:\n pkl_name = file\n break\n path_pkl = os.path.join(\"results\", good_match, pkl_name)\n shutil.copy(path_obj,os.path.join(\"submit_results\",good_match,good_match+\".obj\"))\n shutil.copy(path_fbx, os.path.join(\"submit_results\", good_match, good_match + \".fbx\"))\n shutil.copy(path_txt, os.path.join(\"submit_results\", good_match, good_match + \".txt\"))\n shutil.copy(path_vis, os.path.join(\"submit_results\", good_match, \"vis.mp4\"))\n shutil.copy(path_pkl, os.path.join(\"submit_results\", good_match,pkl_name))\n\ndef add_mtl():\n for file in os.listdir(\"model/obj_seq_5_3dmodel\"):\n if \"obj\" in file:\n with open(os.path.join(\"model\",\"obj_seq_5_3dmodel\",file),'r') as f:\n record = f.readlines()\n with open(os.path.join(\"model\",\"obj_seq_5_3dmodel\",file),'w') as f:\n f.write(record[0])\n f.write(\"g skinCluster1Set tweakSet1\\n\")\n f.writelines(record[1:])", "sub_path": "util_cat.py", "file_name": "util_cat.py", "file_ext": "py", "file_size_in_byte": 7158, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "os.path.exists", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "open3d.io.read_triangle_mesh", "line_number": 64, "usage_type": "call"}, {"api_name": "open3d.io", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path", "line_number": 122, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path", "line_number": 146, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path", "line_number": 147, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path", "line_number": 162, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 165, "usage_type": "call"}, {"api_name": "os.path", "line_number": 165, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path", "line_number": 183, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 184, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 186, "usage_type": "call"}, {"api_name": "os.path", "line_number": 186, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 186, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 187, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 187, "usage_type": "call"}, {"api_name": "os.path", "line_number": 187, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 190, "usage_type": "call"}, {"api_name": "os.path", "line_number": 190, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 191, "usage_type": "call"}, {"api_name": "os.path", "line_number": 191, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path", "line_number": 192, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 193, "usage_type": "call"}, {"api_name": "os.path", "line_number": 193, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 194, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 194, "usage_type": "call"}, {"api_name": "os.path", "line_number": 194, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 198, "usage_type": "call"}, {"api_name": "os.path", "line_number": 198, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 199, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 199, "usage_type": "call"}, {"api_name": "os.path", "line_number": 199, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path", "line_number": 200, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path", "line_number": 201, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path", "line_number": 202, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path", "line_number": 203, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 208, "usage_type": "call"}, {"api_name": "os.path", "line_number": 208, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path", "line_number": 210, "usage_type": "attribute"}]}
+{"seq_id": "362184854", "text": "# encoding: utf-8\r\n\r\nimport re\r\nimport os\r\nimport sys\r\nimport tkinter\r\nfrom datetime import date\r\nfrom readjson import JsonReader\r\nfrom newdishdialog import NewDishDialog\r\nfrom newentrydialog import NewEntryDialog\r\nfrom newdishhtml import DishesHTMLReport\r\nfrom tkinter.ttk import Frame, Label, Button, Entry, Scrollbar, Treeview, Style, Checkbutton\r\nfrom tkinter import N, S, W, E, X, NO, RIGHT, CENTER, VERTICAL, END, messagebox, filedialog, PhotoImage, DISABLED, IntVar\r\nfrom autoentry import AutocompleteEntry\r\n\r\nDEFAULT_MINSIZE_WIDTH = 900\r\nDEFAULT_MINSIZE_HEIGHT = 750\r\n\r\njson_path = os.getcwd() + '\\\\files\\\\dishes.json'\r\nlogo_path = os.getcwd() + '\\\\images\\\\logo.png'\r\n\r\nclass Program(Frame):\r\n \"\"\"Class to represent a main window\"\"\"\r\n def __init__(self, parent):\r\n # Some top-level settings.\r\n Frame.__init__(self, parent)\r\n self.content = Frame(self, padding=(5, 5, 5, 20))\r\n self.parent = parent\r\n self.parent.title('Restaurant manager')\r\n self.parent.minsize(DEFAULT_MINSIZE_WIDTH, DEFAULT_MINSIZE_HEIGHT)\r\n self.parent.protocol('WM_DELETE_WINDOW', self.onQuit)\r\n\r\n # Some variables.\r\n self.track_weight = IntVar()\r\n self.custom_entries = []\r\n self.jsonReader = JsonReader(json_path)\r\n self.frames = []\r\n self.entries = {}\r\n self.dishes = self.jsonReader.getDishesDict()\r\n self.dishes_names = self.jsonReader.getDishesNames()\r\n self.parent.iconphoto(\r\n True,\r\n PhotoImage(\r\n file=os.path.join(\r\n sys.path[0], \r\n logo_path\r\n )\r\n )\r\n )\r\n\r\n # Initialize all widgets.\r\n self.initWidgets()\r\n\r\n def initWidgets(self):\r\n \"\"\"\r\n Initialize all widgets in window here.\r\n Entries are saved in [self.entries] list.\r\n ----------------------------------------------------------\r\n 5 frames are initialized and saved into [self.frames] list here:\r\n 1) [About frame] - Frame to contain all non-food entries.\r\n 2) [Separator frame] - Frame to contain a simple separator.\r\n 3) [New order frame] - Frame to contain entries for adding orders to the main table.\r\n 4) [Orders table frame] - Frame to contain table with list of added orders.\r\n 5) [Save report frame] - Frame to contain buttons, which save a report.\r\n ----------------------------------------------------------\r\n \"\"\"\r\n # Create 5 frames here.\r\n for i in range(5):\r\n self.frames.append(\r\n Frame(self.content)\r\n )\r\n # Set the weights of cols and rows in the grid.\r\n self.configureGrid()\r\n # Center a window.\r\n self.centerWindow()\r\n\r\n # Place 4 frames to the window.\r\n self.frames[0].grid(column=0, row=0, sticky=N+E+W, padx=5, pady=3)\r\n self.frames[1].grid(column=0, row=1, sticky=N+S+E+W, padx=5, pady=3)\r\n self.frames[2].grid(column=0, row=2, sticky=N+S+E+W, padx=5, pady=3)\r\n self.frames[3].grid(column=0, row=3, sticky=N+S+E+W, padx=5, pady=10)\r\n self.frames[4].grid(column=0, row=4, sticky=S+E+W, padx=5, pady=3)\r\n\r\n # About frame widgets.\r\n Label(self.frames[0], text='Заказчик').grid(row=0, column=0, sticky=E, pady=5)\r\n Label(self.frames[0], text='Менеджер').grid(row=0, column=2, sticky=E, pady=5)\r\n Label(self.frames[0], text='Вид мероприятия', justify=RIGHT, wraplength=90\r\n ).grid(row=0, column=4, sticky=E, pady=5)\r\n Label(self.frames[0], text='Дата').grid(row=1, column=0, sticky=E, pady=5)\r\n Label(self.frames[0], text='Время').grid(row=1, column=2, sticky=E, pady=5)\r\n Label(self.frames[0], text='Место проведения', justify=RIGHT, wraplength=90\r\n ).grid(row=1, column=4, sticky=E, pady=5)\r\n Label(self.frames[0], text='Количество персон', justify=RIGHT, wraplength=90\r\n ).grid(row=1, column=6, sticky=E, pady=5)\r\n\r\n self.entries['client'] = Entry(self.frames[0])\r\n self.entries['manager'] = Entry(self.frames[0])\r\n self.entries['type'] = Entry(self.frames[0], width=10)\r\n self.entries['date'] = Entry(self.frames[0])\r\n self.entries['time'] = Entry(self.frames[0])\r\n self.entries['location'] = Entry(self.frames[0], width=10)\r\n self.entries['persons'] = Entry(self.frames[0], width=10)\r\n\r\n self.entries['client'].focus_set()\r\n today = date.today().isoformat().split('-')[::-1]\r\n self.entries['date'].insert(0, '.'.join(today))\r\n\r\n self.entries['client'].grid(row=0, column=1, sticky=E+W, padx=(3, 13), pady=5)\r\n self.entries['manager'].grid(row=0, column=3, sticky=E+W, padx=(3, 13), pady=5)\r\n self.entries['type'].grid(row=0, column=5, columnspan=3, sticky=E+W, padx=(3, 13), pady=5)\r\n self.entries['date'].grid(row=1, column=1, sticky=E+W, padx=(3, 13), pady=5)\r\n self.entries['time'].grid(row=1, column=3, sticky=E+W, padx=(3, 13), pady=5)\r\n self.entries['location'].grid(row=1, column=5, sticky=E + W, padx=(3, 13), pady=5)\r\n self.entries['persons'].grid(row=1, column=7, sticky=E+W, padx=(3, 13), pady=5)\r\n\r\n # Add a separator between [about] and [new order] frames\r\n sep1 = Frame(self.frames[1], height=2, borderwidth=1, relief='sunken')\r\n sep1.pack(fill=X, padx=1, pady=10)\r\n\r\n # New Order frame widgets.\r\n Label(self.frames[2], text='Название', anchor=E).grid(row=0, column=0, sticky=E)\r\n Label(self.frames[2], text='Комментарий', anchor=E).grid(row=1, column=0, sticky=E)\r\n Label(self.frames[2], text='Количество', anchor=E).grid(row=2, column=0, sticky=E)\r\n self.sum_lbl = Label(self.frames[2], text='Текущая сумма заказа:\\n 0 грн', anchor=E, justify=RIGHT)\r\n\r\n self.entries['name'] = AutocompleteEntry(self.frames[2])\r\n self.entries['comment'] = Entry(self.frames[2])\r\n self.entries['amount'] = Entry(self.frames[2])\r\n addOrder_btn = Button(self.frames[2], text='Добавить блюдо в отчет')\r\n\r\n self.entries['name'].set_completion_list(self.dishes_names)\r\n self.entries['amount'].insert(0, '1')\r\n addOrder_btn['command'] = lambda: self.addDish()\r\n\r\n self.entries['name'].grid(row=0, column=1, columnspan=5, sticky=W+E, pady=3, padx=(3, 15))\r\n self.entries['comment'].grid(row=1, column=1, columnspan=5, sticky=W+E, pady=3, padx=(3, 15))\r\n self.entries['amount'].grid(row=2, column=1, columnspan=2, sticky=W+E, pady=3, padx=(3, 15))\r\n addOrder_btn.grid(row=3, column=1, sticky=W, pady=3, padx=3)\r\n self.sum_lbl.grid(row=4, column=5, pady=0, padx=(190, 0))\r\n\r\n customEntry_btn = Button(self.frames[2], text='Добавить собственную строку в отчет')\r\n customEntry_btn['command'] = lambda: self.addCustomEntry()\r\n customEntry_btn.grid(row=4, column=1, sticky=W, pady=3, padx=3)\r\n\r\n # Orders Table frame widgets.\r\n self.orders_view = Treeview(self.frames[3])\r\n self.orders_view['columns'] = ('Weight', 'Amount', 'Comment', 'Price', 'Sum')\r\n self.orders_view.bind('', lambda e: self.deleteEntry(e, self.orders_view))\r\n\r\n self.orders_view.heading('#0', text='Название')\r\n self.orders_view.column('#0', anchor='w', minwidth=307, width=307)\r\n self.orders_view.heading('Weight', text='Выход')\r\n self.orders_view.column('Weight', anchor=CENTER, minwidth=100, width=100, stretch=NO)\r\n self.orders_view.heading('Amount', text='Количество')\r\n self.orders_view.column('Amount', anchor=CENTER, minwidth=100, width=100, stretch=NO)\r\n self.orders_view.heading('Comment', text='Комментарий')\r\n self.orders_view.column('Comment', anchor='w', minwidth=130, width=130)\r\n self.orders_view.heading('Price', text='Цена, грн')\r\n self.orders_view.column('Price', anchor=CENTER, minwidth=110, width=110, stretch=NO)\r\n self.orders_view.heading('Sum', text='Сумма, грн')\r\n self.orders_view.column('Sum', anchor=CENTER, minwidth=108, width=108, stretch=NO)\r\n\r\n self.orders_view.grid(row=0, column=0, sticky=N+S+E+W, padx=3, pady=3)\r\n\r\n # NOTE: next [for] block is for testing purposes only.\r\n \"\"\"\r\n for dish in self.dishes:\r\n self.orders_view.insert('', 'end',\r\n text=dish,\r\n values=[\r\n self.dishes[dish]['weight'],\r\n 5, # [amount] column\r\n 'wats up?', # [comment] column\r\n self.dishes[dish]['price'],\r\n 5 * float(self.dishes[dish]['price'])\r\n ]\r\n )\r\n \"\"\"\r\n\r\n orders_scrlbar = Scrollbar(self.frames[3], orient=VERTICAL, command=self.orders_view.yview)\r\n self.orders_view['yscrollcommand'] = orders_scrlbar.set\r\n orders_scrlbar.grid(row=0, column=1, sticky=N+S)\r\n\r\n # Save Report frame widgets.\r\n saveWeb_btn = Button(self.frames[4], text='Сохранить отчет', width=20)\r\n trackweight_chkbox = Checkbutton(self.frames[4], text='Учитывать средний вес', variable=self.track_weight)\r\n\r\n saveWeb_btn['command'] = lambda: self.saveWeb()\r\n\r\n saveWeb_btn.pack(side='left', anchor=CENTER, padx=5, pady=3)\r\n trackweight_chkbox.pack(side='left', anchor=CENTER, pady=3, padx=3)\r\n\r\n def configureGrid(self):\r\n \"\"\"Configure weights of grids columns and rows\"\"\"\r\n # Top-level configuration.\r\n self.grid(sticky=N+S+E+W)\r\n top = self.winfo_toplevel()\r\n top.rowconfigure(0, weight=1)\r\n top.columnconfigure(0, weight=1)\r\n self.rowconfigure(0, weight=1)\r\n self.columnconfigure(0, weight=1)\r\n\r\n # Configuration of main content frame.\r\n self.configureFrame(self.content, [1], [0, 0, 0, 3, 0])\r\n # Configuration of about frame.\r\n self.configureFrame(self.frames[0], [0, 1, 0, 1, 0, 0, 0, 0], [0, 0, 0])\r\n # Configuration of new order frame.\r\n self.configureFrame(self.frames[2], [0, 1, 1, 1, 1, 1], [0, 0, 0, 0])\r\n # Configuration of orders table frame.\r\n self.configureFrame(self.frames[3], [1, 0], [1])\r\n\r\n def centerWindow(self):\r\n \"\"\"Place the main window in the center of screen\"\"\"\r\n window_w = DEFAULT_MINSIZE_WIDTH\r\n window_h = DEFAULT_MINSIZE_HEIGHT\r\n screen_w = self.winfo_screenwidth()\r\n screen_h = self.winfo_screenheight()\r\n x = (screen_w - window_w) / 2\r\n y = (screen_h - window_h) / 2\r\n self.parent.geometry('%dx%d+%d+%d' % (window_w, window_h, x, y))\r\n\r\n def onQuit(self):\r\n \"\"\"Before the program exits, ask user, if he really wants it\"\"\"\r\n if messagebox.askyesno('Выход из программы', 'Вы действительно хотите выйти из программы?'):\r\n self.quit()\r\n\r\n @staticmethod\r\n def deleteEntry(e, tree):\r\n \"\"\"Event to handle deletion in TreeView woth [Delete] button\"\"\"\r\n selected_item = tree.selection()[0] # get selected item\r\n tree.delete(selected_item)\r\n\r\n @staticmethod\r\n def configureFrame(frame, columns_wght, rows_wght):\r\n \"\"\"Function to organize frame configuration routine\"\"\"\r\n frame.grid(column=0, row=0, sticky=N + S + E + W)\r\n for i, weight in enumerate(columns_wght):\r\n frame.columnconfigure(i, weight=weight)\r\n for i, weight in enumerate(rows_wght):\r\n frame.rowconfigure(i, weight=weight)\r\n\r\n def saveWeb(self):\r\n \"\"\"Save data from the TreeView to the html report\"\"\"\r\n # Validation: if there is no items in tree_view\r\n if not self.orders_view.get_children():\r\n messagebox.showerror(\r\n 'Ошибка',\r\n 'Ошибка создания отчета: нет блюд.'\r\n )\r\n return\r\n\r\n # Open file dialog to choose saving path.\r\n file = filedialog.asksaveasfile(\r\n mode='w',\r\n defaultextension='.html',\r\n filetypes=[('Веб страница', '.html'), ('Все файлы', '.*')]\r\n )\r\n # If user pressed [Cancel] or closed a file dialog.\r\n if file is None:\r\n return\r\n\r\n # Pack files to send to HTML saver module.\r\n # Dictionary looks like: { name:{weight,amount,comment,price,total}, name:... }\r\n packed_dishes = {\r\n 'global': {\r\n 'trackw': self.track_weight.get(), # flag - is weight tracked\r\n 'totalsum': self.sum_lbl['text'][21:-4], # str - total sum of dishes\r\n 'centries': self.custom_entries # [5*str] - custom entries\r\n },\r\n 'about': {\r\n 'client': self.entries['client'].get(),\r\n 'manager': self.entries['manager'].get(),\r\n 'type': self.entries['type'].get(),\r\n 'date': self.entries['date'].get(),\r\n 'time': self.entries['time'].get(),\r\n 'location': self.entries['location'].get(),\r\n 'persons': self.entries['persons'].get()\r\n },\r\n 'dishes': {}\r\n }\r\n for child in self.orders_view.get_children():\r\n child_content = self.orders_view.item(child)\r\n child_values = child_content['values']\r\n child_name = child_content['text']\r\n child_type = self.dishes[child_name]['type']\r\n # Pack data about a certain dish.\r\n packed_child = {\r\n child_name: {\r\n 'weight': child_values[0],\r\n 'amount': child_values[1],\r\n 'comment': child_values[2],\r\n 'price': child_values[3],\r\n 'total': child_values[4],\r\n 'type': child_type\r\n }\r\n }\r\n packed_dishes['dishes'].update(packed_child)\r\n # Save the HTML report\r\n html_writer = DishesHTMLReport(file, data=packed_dishes)\r\n html_writer.create_html()\r\n messagebox.showinfo(\r\n 'Успешное сохранение',\r\n 'Данные были успешно сохранены.'\r\n )\r\n\r\n def addCustomEntry(self):\r\n \"\"\"This functions opens a window, which gives access for creating custom entries to report\"\"\"\r\n newentry_dialog = NewEntryDialog(self, title='Добавить собственную строку в отчет')\r\n if newentry_dialog.result:\r\n self.custom_entries.append(newentry_dialog.result)\r\n\r\n def addDish(self):\r\n \"\"\"This function adds an entry to the order_view TreeView widget.\"\"\"\r\n # Get packed_info as a dictionary.\r\n packed_info = self.packInfo()\r\n # If info packed successfully, add a new entry to orders TreeView.\r\n if not packed_info:\r\n return\r\n # Add a dish to the TreeView.\r\n self.orders_view.insert(\r\n '', 'end',\r\n text=packed_info['dish'],\r\n values=packed_info['values']\r\n )\r\n # Clear all entries.\r\n entries_toclear = ['name', 'comment', 'amount']\r\n for entry_key in entries_toclear:\r\n self.entries[entry_key].delete(0, END)\r\n self.entries[entry_key].insert(0, '')\r\n self.entries['amount'].insert(0, '1')\r\n # Change [sum] label.\r\n total_sum = float(self.sum_lbl['text'][21:-4]) + float(packed_info['values'][4])\r\n self.sum_lbl['text'] = 'Текущая сумма заказа:\\n %.2f грн' % total_sum\r\n # Set focus to ['name'] entry.\r\n self.entries['name'].focus_set()\r\n\r\n def packInfo(self):\r\n \"\"\"Pack values, that were inserted into Entries and Text, into dictionary\"\"\"\r\n msg = self.validateForm()\r\n if msg == 'OK':\r\n name = self.entries['name'].get()\r\n amount = self.entries['amount'].get()\r\n total_price = float(amount) * float(self.dishes[name]['price'])\r\n pack = {\r\n 'dish': name,\r\n 'values': [\r\n self.dishes[name]['weight'],\r\n str(amount),\r\n self.entries['comment'].get(),\r\n self.dishes[name]['price'],\r\n '%.2f' % total_price\r\n ]\r\n }\r\n return pack\r\n elif msg == 'NEWDISH_CANCELED':\r\n return False\r\n else:\r\n messagebox.showerror(\r\n 'Ошибка ввода',\r\n 'При вводе случились ошибки: \\n%s' % msg\r\n )\r\n return False\r\n\r\n def validateForm(self):\r\n \"\"\"Validate all Entry and Text Widgets\"\"\"\r\n msg = ''\r\n index = 1\r\n translations = {\r\n 'client': 'Клиент',\r\n 'manager': 'Менеджер',\r\n 'type': 'Вид мероприятия',\r\n 'date': 'Дата',\r\n 'time': 'Время',\r\n 'location': 'Место проведения',\r\n 'persons': 'Количество персон',\r\n 'name': 'Название',\r\n 'comment': 'Комментарий',\r\n 'amount': 'Количество',\r\n }\r\n\r\n for k in self.entries:\r\n # Check if some entry is empty.\r\n if self.entries[k].get() == '':\r\n if k == 'comment':\r\n continue\r\n msg += '%i) Поле [%s] пустое.\\n' % (index, translations[k])\r\n index += 1\r\n\r\n # Persons entry should contain only digits.\r\n if not all(letter.isdigit() for letter in self.entries['persons'].get()):\r\n msg += '%i) Поле [Количество персон] должно содержать только цифры.\\n' % index\r\n index += 1\r\n # Check date entry.\r\n match = re.search(r'(\\d{2})[.](\\d{2})[.](\\d{4})$', self.entries['date'].get())\r\n if not match:\r\n msg += '%i) Неправильный формат даты.\\nПравильный формат: дд.мм.гггг\\n' % index\r\n index += 1\r\n # Check amount entry.\r\n check = all(letter.isdigit() or letter == '.' for letter in self.entries['amount'].get())\r\n if not check:\r\n msg += '%i) Поле [Количество] должно содержать только цифры, или точки.\\n' % index\r\n index += 1\r\n\r\n # Ask about [name] entry only if there are no more errors left.\r\n if msg != '':\r\n return msg\r\n\r\n # Check name entry.\r\n if (self.entries['name'].get() not in self.dishes_names and self.entries['name'].get() != ''):\r\n if messagebox.askyesno('Добавление нового блюда', 'Блюда %s нет в списке. Добавить?' % self.entries['name'].get()):\r\n # Create a new window, which will contain a [result] dictionary {name,type,weight,price}\r\n newdish_dialog = NewDishDialog(self, self.entries['name'].get(), 'Добавить новое блюдо')\r\n if newdish_dialog.result:\r\n dishToJSON = {\r\n newdish_dialog.result['name']: {\r\n 'weight': newdish_dialog.result['weight'],\r\n 'price': newdish_dialog.result['price'],\r\n 'type': newdish_dialog.result['type']\r\n }\r\n }\r\n self.jsonReader.writeDish(dishToJSON)\r\n self.jsonReader.prepInformation()\r\n self.dishes = self.jsonReader.getDishesDict()\r\n self.dishes_names = self.jsonReader.getDishesNames()\r\n self.entries['name'].set_completion_list(self.dishes_names)\r\n else:\r\n # if [exit] was pressed\r\n return 'NEWDISH_CANCELED'\r\n else:\r\n return 'NEWDISH_CANCELED'\r\n # Check if the same dish is in the orderslist.\r\n for child in self.orders_view.get_children():\r\n child_content = self.orders_view.item(child)\r\n if self.entries['name'].get() == child_content['text']:\r\n msg += '%i) Ошибка в поле [Название] - такое блюдо уже есть в списке.' % index\r\n index += 1\r\n\r\n # If all tests passed correctly, msg is 'OK'\r\n if msg == '':\r\n msg = 'OK'\r\n return msg\r\n\r\n\r\ndef main():\r\n root = tkinter.Tk()\r\n app = Program(root)\r\n root.mainloop()\r\n\r\n\r\nif __name__ == '__main__':\r\n main()\r\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 21060, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "os.getcwd", "line_number": 19, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 20, "usage_type": "call"}, {"api_name": "tkinter.ttk.Frame", "line_number": 22, "usage_type": "name"}, {"api_name": "tkinter.ttk.Frame.__init__", "line_number": 26, "usage_type": "call"}, {"api_name": "tkinter.ttk.Frame", "line_number": 26, "usage_type": "name"}, {"api_name": "tkinter.ttk.Frame", "line_number": 27, "usage_type": "call"}, {"api_name": "tkinter.IntVar", "line_number": 34, "usage_type": "call"}, {"api_name": "readjson.JsonReader", "line_number": 36, "usage_type": "call"}, {"api_name": "tkinter.PhotoImage", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "sys.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "tkinter.ttk.Frame", "line_number": 70, "usage_type": "call"}, {"api_name": "tkinter.N", "line_number": 78, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 78, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 78, "usage_type": "name"}, {"api_name": "tkinter.N", "line_number": 79, "usage_type": "name"}, {"api_name": "tkinter.S", "line_number": 79, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 79, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 79, "usage_type": "name"}, {"api_name": "tkinter.N", "line_number": 80, "usage_type": "name"}, {"api_name": "tkinter.S", "line_number": 80, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 80, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 80, "usage_type": "name"}, {"api_name": "tkinter.N", "line_number": 81, "usage_type": "name"}, {"api_name": "tkinter.S", "line_number": 81, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 81, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 81, "usage_type": "name"}, {"api_name": "tkinter.S", "line_number": 82, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 82, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 82, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 85, "usage_type": "call"}, {"api_name": "tkinter.E", "line_number": 85, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 86, "usage_type": "call"}, {"api_name": "tkinter.E", "line_number": 86, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 87, "usage_type": "call"}, {"api_name": "tkinter.RIGHT", "line_number": 87, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 88, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 89, "usage_type": "call"}, {"api_name": "tkinter.E", "line_number": 89, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 90, "usage_type": "call"}, {"api_name": "tkinter.E", "line_number": 90, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 91, "usage_type": "call"}, {"api_name": "tkinter.RIGHT", "line_number": 91, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 92, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 93, "usage_type": "call"}, {"api_name": "tkinter.RIGHT", "line_number": 93, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 94, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 96, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 97, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 98, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 99, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 100, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 101, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 102, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 105, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 105, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 108, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 108, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 109, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 109, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 110, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 110, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 111, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 111, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 112, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 112, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 113, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 113, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 114, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 114, "usage_type": "name"}, {"api_name": "tkinter.ttk.Frame", "line_number": 117, "usage_type": "call"}, {"api_name": "tkinter.X", "line_number": 118, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 121, "usage_type": "call"}, {"api_name": "tkinter.E", "line_number": 121, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 122, "usage_type": "call"}, {"api_name": "tkinter.E", "line_number": 122, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 123, "usage_type": "call"}, {"api_name": "tkinter.E", "line_number": 123, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 124, "usage_type": "call"}, {"api_name": "tkinter.E", "line_number": 124, "usage_type": "name"}, {"api_name": "tkinter.RIGHT", "line_number": 124, "usage_type": "name"}, {"api_name": "autoentry.AutocompleteEntry", "line_number": 126, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 127, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 128, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 129, "usage_type": "call"}, {"api_name": "tkinter.W", "line_number": 135, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 135, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 136, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 136, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 137, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 137, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 138, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 141, "usage_type": "call"}, {"api_name": "tkinter.W", "line_number": 143, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 146, "usage_type": "call"}, {"api_name": "tkinter.CENTER", "line_number": 153, "usage_type": "name"}, {"api_name": "tkinter.NO", "line_number": 153, "usage_type": "name"}, {"api_name": "tkinter.CENTER", "line_number": 155, "usage_type": "name"}, {"api_name": "tkinter.NO", "line_number": 155, "usage_type": "name"}, {"api_name": "tkinter.CENTER", "line_number": 159, "usage_type": "name"}, {"api_name": "tkinter.NO", "line_number": 159, "usage_type": "name"}, {"api_name": "tkinter.CENTER", "line_number": 161, "usage_type": "name"}, {"api_name": "tkinter.NO", "line_number": 161, "usage_type": "name"}, {"api_name": "tkinter.N", "line_number": 163, "usage_type": "name"}, {"api_name": "tkinter.S", "line_number": 163, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 163, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 163, "usage_type": "name"}, {"api_name": "tkinter.ttk.Scrollbar", "line_number": 180, "usage_type": "call"}, {"api_name": "tkinter.VERTICAL", "line_number": 180, "usage_type": "name"}, {"api_name": "tkinter.N", "line_number": 182, "usage_type": "name"}, {"api_name": "tkinter.S", "line_number": 182, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 185, "usage_type": "call"}, {"api_name": "tkinter.ttk.Checkbutton", "line_number": 186, "usage_type": "call"}, {"api_name": "tkinter.CENTER", "line_number": 190, "usage_type": "name"}, {"api_name": "tkinter.CENTER", "line_number": 191, "usage_type": "name"}, {"api_name": "tkinter.N", "line_number": 196, "usage_type": "name"}, {"api_name": "tkinter.S", "line_number": 196, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 196, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 196, "usage_type": "name"}, {"api_name": "tkinter.messagebox.askyesno", "line_number": 224, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 224, "usage_type": "name"}, {"api_name": "tkinter.N", "line_number": 236, "usage_type": "name"}, {"api_name": "tkinter.S", "line_number": 236, "usage_type": "name"}, {"api_name": "tkinter.E", "line_number": 236, "usage_type": "name"}, {"api_name": "tkinter.W", "line_number": 236, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 246, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 246, "usage_type": "name"}, {"api_name": "tkinter.filedialog.asksaveasfile", "line_number": 253, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 253, "usage_type": "name"}, {"api_name": "newdishhtml.DishesHTMLReport", "line_number": 299, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 301, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 301, "usage_type": "name"}, {"api_name": "newentrydialog.NewEntryDialog", "line_number": 308, "usage_type": "call"}, {"api_name": "tkinter.END", "line_number": 328, "usage_type": "argument"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 358, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 358, "usage_type": "name"}, {"api_name": "re.search", "line_number": 394, "usage_type": "call"}, {"api_name": "tkinter.messagebox.askyesno", "line_number": 410, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 410, "usage_type": "name"}, {"api_name": "newdishdialog.NewDishDialog", "line_number": 412, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 445, "usage_type": "call"}]}
+{"seq_id": "281175453", "text": "#!/usr/bin/python3\n\nimport pytest\n\nfrom brownie import accounts\n\n\n@pytest.fixture(scope=\"module\", autouse=True)\ndef setup(approve_many, issuer, token):\n token.mint(issuer, 100000, {'from': accounts[0]})\n token.transfer(accounts[1], 1000, {'from': accounts[0]})\n\n\ndef test_transfer_from(token):\n '''investor transferFrom - approved'''\n token.approve(accounts[3], 500, {'from': accounts[1]})\n assert token.allowance(accounts[1], accounts[3]) == 500\n token.transferFrom(accounts[1], accounts[2], 400, {'from': accounts[3]})\n assert token.allowance(accounts[1], accounts[3]) == 100\n token.transferFrom(accounts[1], accounts[2], 100, {'from': accounts[3]})\n assert token.allowance(accounts[1], accounts[3]) == 0\n\n\ndef test_transfer_from_investor_no_approval(token):\n '''transferFrom - no approval'''\n with pytest.reverts(\"Insufficient allowance\"):\n token.transferFrom(accounts[1], accounts[2], 1000, {'from': accounts[3]})\n\n\ndef test_transfer_from_investor_insufficient_approval(token):\n '''transferFrom - insufficient approval'''\n token.approve(accounts[3], 500, {'from': accounts[1]})\n with pytest.reverts(\"Insufficient allowance\"):\n token.transferFrom(accounts[1], accounts[2], 1000, {'from': accounts[3]})\n\n\ndef test_transfer_from_same_id(kyc, token):\n '''transferFrom - same investor ID'''\n kyc.registerAddresses(kyc.getID(accounts[1]), [accounts[-1]], {'from': accounts[0]})\n token.transferFrom(accounts[1], accounts[2], 500, {'from': accounts[-1]})\n\n\ndef test_transfer_from_issuer(token):\n '''issuer transferFrom'''\n token.transferFrom(accounts[1], accounts[2], 1000, {'from': accounts[0]})\n\n\ndef test_authority_permission(issuer, token):\n '''authority transferFrom permission'''\n issuer.addAuthority([accounts[-1]], [\"0x23b872dd\"], 2000000000, 1, {'from': accounts[0]})\n token.transferFrom(accounts[1], accounts[2], 500, {'from': accounts[-1]})\n issuer.setAuthoritySignatures(\n issuer.getID(accounts[-1]),\n [\"0x23b872dd\"],\n False,\n {'from': accounts[0]}\n )\n with pytest.reverts(\"Authority not permitted\"):\n token.transferFrom(accounts[1], accounts[2], 500, {'from': accounts[-1]})\n", "sub_path": "tests/SecurityToken/transfer/test_token_transfer_from.py", "file_name": "test_token_transfer_from.py", "file_ext": "py", "file_size_in_byte": 2208, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "brownie.accounts", "line_number": 10, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 11, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 8, "usage_type": "call"}, {"api_name": "brownie.accounts", "line_number": 16, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 17, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 18, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 19, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 20, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 21, "usage_type": "name"}, {"api_name": "pytest.reverts", "line_number": 26, "usage_type": "call"}, {"api_name": "brownie.accounts", "line_number": 27, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 32, "usage_type": "name"}, {"api_name": "pytest.reverts", "line_number": 33, "usage_type": "call"}, {"api_name": "brownie.accounts", "line_number": 34, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 39, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 40, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 45, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 50, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 51, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 53, "usage_type": "name"}, {"api_name": "brownie.accounts", "line_number": 56, "usage_type": "name"}, {"api_name": "pytest.reverts", "line_number": 58, "usage_type": "call"}, {"api_name": "brownie.accounts", "line_number": 59, "usage_type": "name"}]}
+{"seq_id": "79742813", "text": "import os\nfrom pyspark.sql import SparkSession\nfrom pyspark.sql import DataFrame\nfrom time import sleep\n\ndef main():\n print(\"Started aggregation example\")\n spark = get_spark_session()\n\n dfHeights: DataFrame = spark.read.option(\"header\", \"true\")\\\n .csv(\"./sample_datasets/heights.csv\").alias(\"heights\")\n dfNames: DataFrame = spark.read.option(\"header\", \"true\")\\\n .csv(\"./sample_datasets/names.csv\").alias(\"names\")\n joined_df: DataFrame = dfHeights.join(dfNames, \"id\")\n joined_df.show()\n\n print(\"Finished aggregation example\")\n\ndef get_spark_session():\n return SparkSession\\\n .builder.master(\"local\")\\\n .appName('SparkMapReduceExample')\\\n .getOrCreate()\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "HeightsExampleJoin.py", "file_name": "HeightsExampleJoin.py", "file_ext": "py", "file_size_in_byte": 751, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "pyspark.sql.DataFrame", "line_number": 10, "usage_type": "name"}, {"api_name": "pyspark.sql.DataFrame", "line_number": 12, "usage_type": "name"}, {"api_name": "pyspark.sql.DataFrame", "line_number": 14, "usage_type": "name"}, {"api_name": "pyspark.sql.SparkSession.builder.master", "line_number": 20, "usage_type": "call"}, {"api_name": "pyspark.sql.SparkSession.builder", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pyspark.sql.SparkSession", "line_number": 20, "usage_type": "name"}]}
+{"seq_id": "443792115", "text": "from typing import Any, Callable, List, Optional, cast\r\nimport torch\r\nimport argparse\r\nimport numpy as np\r\nimport shutil\r\nimport glob\r\nimport time\r\nimport random\r\nimport os\r\nfrom torch import nn\r\nimport wandb\r\nimport pdb\r\nfrom perceptual_advex import evaluationdelta\r\nfrom perceptual_advex.utilities import add_dataset_model_arguments, \\\r\n get_dataset_model, calculate_accuracy\r\nfrom perceptual_advex.attacks import *\r\nfrom perceptual_advex.ci_attacks2 import *\r\nfrom perceptual_advex.models import FeatureModel\r\nfrom perceptual_advex.hidden_attacks import *\r\nfrom perceptual_advex.distances import L2Distance, LinfDistance\r\nfrom perceptual_advex.vae import *\r\nVAL_ITERS = 100\r\nfrom torch.optim.optimizer import Optimizer\r\nimport torchvision\r\nclass AdamW(Optimizer):\r\n \"\"\"Implements Adam algorithm.\r\n It has been proposed in `Adam: A Method for Stochastic Optimization`_.\r\n Arguments:\r\n params (iterable): iterable of parameters to optimize or dicts defining\r\n parameter groups\r\n lr (float, optional): learning rate (default: 1e-3)\r\n betas (Tuple[float, float], optional): coefficients used for computing\r\n running averages of gradient and its square (default: (0.9, 0.999))\r\n eps (float, optional): term added to the denominator to improve\r\n numerical stability (default: 1e-8)\r\n weight_decay (float, optional): weight decay (L2 penalty) (default: 0)\r\n amsgrad (boolean, optional): whether to use the AMSGrad variant of this\r\n algorithm from the paper `On the Convergence of Adam and Beyond`_\r\n .. _Adam\\: A Method for Stochastic Optimization:\r\n https://arxiv.org/abs/1412.6980\r\n .. _On the Convergence of Adam and Beyond:\r\n https://openreview.net/forum?id=ryQu7f-RZ\r\n \"\"\"\r\n\r\n def __init__(self, params, lr=1e-3, betas=(0.9, 0.999), eps=1e-8,\r\n weight_decay=0, amsgrad=False):\r\n if not 0.0 <= lr:\r\n raise ValueError(\"Invalid learning rate: {}\".format(lr))\r\n if not 0.0 <= eps:\r\n raise ValueError(\"Invalid epsilon value: {}\".format(eps))\r\n if not 0.0 <= betas[0] < 1.0:\r\n raise ValueError(\"Invalid beta parameter at index 0: {}\".format(betas[0]))\r\n if not 0.0 <= betas[1] < 1.0:\r\n raise ValueError(\"Invalid beta parameter at index 1: {}\".format(betas[1]))\r\n defaults = dict(lr=lr, betas=betas, eps=eps,\r\n weight_decay=weight_decay, amsgrad=amsgrad)\r\n super(AdamW, self).__init__(params, defaults)\r\n\r\n def __setstate__(self, state):\r\n super(AdamW, self).__setstate__(state)\r\n for group in self.param_groups:\r\n group.setdefault('amsgrad', False)\r\n\r\n def step(self, closure=None):\r\n \"\"\"Performs a single optimization step.\r\n Arguments:\r\n closure (callable, optional): A closure that reevaluates the model\r\n and returns the loss.\r\n \"\"\"\r\n loss = None\r\n if closure is not None:\r\n loss = closure()\r\n\r\n for group in self.param_groups:\r\n for p in group['params']:\r\n if p.grad is None:\r\n continue\r\n grad = p.grad.data\r\n if grad.is_sparse:\r\n raise RuntimeError('Adam does not support sparse gradients, please consider SparseAdam instead')\r\n amsgrad = group['amsgrad']\r\n\r\n state = self.state[p]\r\n\r\n # State initialization\r\n if len(state) == 0:\r\n state['step'] = 0\r\n # Exponential moving average of gradient values\r\n state['exp_avg'] = torch.zeros_like(p.data)\r\n # Exponential moving average of squared gradient values\r\n state['exp_avg_sq'] = torch.zeros_like(p.data)\r\n if amsgrad:\r\n # Maintains max of all exp. moving avg. of sq. grad. values\r\n state['max_exp_avg_sq'] = torch.zeros_like(p.data)\r\n\r\n exp_avg, exp_avg_sq = state['exp_avg'], state['exp_avg_sq']\r\n if amsgrad:\r\n max_exp_avg_sq = state['max_exp_avg_sq']\r\n beta1, beta2 = group['betas']\r\n\r\n state['step'] += 1\r\n\r\n # if group['weight_decay'] != 0:\r\n # grad = grad.add(group['weight_decay'], p.data)\r\n\r\n # Decay the first and second moment running average coefficient\r\n exp_avg.mul_(beta1).add_(1 - beta1, grad)\r\n exp_avg_sq.mul_(beta2).addcmul_(1 - beta2, grad, grad)\r\n if amsgrad:\r\n # Maintains the maximum of all 2nd moment running avg. till now\r\n torch.max(max_exp_avg_sq, exp_avg_sq, out=max_exp_avg_sq)\r\n # Use the max. for normalizing running avg. of gradient\r\n denom = max_exp_avg_sq.sqrt().add_(group['eps'])\r\n else:\r\n denom = exp_avg_sq.sqrt().add_(group['eps'])\r\n\r\n bias_correction1 = 1 - beta1 ** state['step']\r\n bias_correction2 = 1 - beta2 ** state['step']\r\n step_size = group['lr'] * math.sqrt(bias_correction2) / bias_correction1\r\n\r\n # p.data.addcdiv_(-step_size, exp_avg, denom)\r\n p.data.add_(-step_size, torch.mul(p.data, group['weight_decay']).addcdiv_(1, exp_avg, denom))\r\n\r\n return loss\r\n\r\nCIFAR_MEAN = [0.4914, 0.4822, 0.4465]\r\nCIFAR_STD = [0.2470, 0.2435, 0.2616]\r\n\r\ndef get_vae_template(vae_path, featured_dim, CNN_embed_dim):\r\n\r\n vae = CVAE_Normalize(d=featured_dim, z=CNN_embed_dim)\r\n vae = nn.DataParallel(vae)\r\n save_model = torch.load(vae_path)\r\n model_dict = vae.state_dict()\r\n state_dict = {k: v for k, v in save_model.items() if k in model_dict.keys()}\r\n print(state_dict.keys())\r\n model_dict.update(state_dict)\r\n vae.load_state_dict(model_dict)\r\n vae.eval()\r\n return vae\r\n\r\ndef get_eps_params(base_eps, resol):\r\n eps_list = []\r\n max_list = []\r\n min_list = []\r\n for i in range(3):\r\n eps_list.append(torch.full((resol, resol), base_eps, device='cuda'))\r\n min_list.append(torch.full((resol, resol), 0., device='cuda'))\r\n max_list.append(torch.full((resol, resol), 255., device='cuda'))\r\n\r\n eps_t = torch.unsqueeze(torch.stack(eps_list), 0)\r\n max_t = torch.unsqueeze(torch.stack(max_list), 0)\r\n min_t = torch.unsqueeze(torch.stack(min_list), 0)\r\n return eps_t, max_t, min_t\r\n\r\ndef get_cifar_params(resol):\r\n mean_list = []\r\n std_list = []\r\n for i in range(3):\r\n mean_list.append(torch.full((resol, resol), CIFAR_MEAN[i], device='cuda'))\r\n std_list.append(torch.full((resol, resol), CIFAR_STD[i], device='cuda'))\r\n return torch.unsqueeze(torch.stack(mean_list), 0), torch.unsqueeze(torch.stack(std_list), 0)\r\n\r\nclass CIFARNORMALIZE(nn.Module):\r\n def __init__(self, resol):\r\n super().__init__()\r\n self.mean, self.std = get_cifar_params(resol)\r\n\r\n def forward(self, x):\r\n '''\r\n Parameters:\r\n x: input image with pixels normalized to ([0, 1] - IMAGENET_MEAN) / IMAGENET_STD\r\n '''\r\n x = x.sub(self.mean)\r\n x = x.div(self.std)\r\n return x\r\n\r\nclass CIFARINNORMALIZE(nn.Module):\r\n def __init__(self, resol):\r\n super().__init__()\r\n self.mean, self.std = get_cifar_params(resol)\r\n\r\n def forward(self, x):\r\n '''\r\n Parameters:\r\n x: input image with pixels normalized to ([0, 1] - IMAGENET_MEAN) / IMAGENET_STD\r\n '''\r\n x = x.mul(self.std)\r\n x = x.add(*self.mean)\r\n return x\r\n\r\nif __name__ == '__main__':\r\n parser = argparse.ArgumentParser()\r\n\r\n add_dataset_model_arguments(parser)\r\n\r\n parser.add_argument('--num_epochs', type=int, required=False,\r\n help='number of epochs trained')\r\n parser.add_argument('--batch_size', type=int, default=100,\r\n help='number of examples/minibatch')\r\n parser.add_argument('--val_batches', type=int, default=10,\r\n help='number of batches to validate on')\r\n parser.add_argument('--log_dir', type=str, default='data/logs')\r\n parser.add_argument('--parallel', type=int, default=1,\r\n help='number of GPUs to train on')\r\n\r\n\r\n parser.add_argument('--only_attack_correct', action='store_true',\r\n default=False, help='only attack examples that '\r\n 'are classified correctly')\r\n parser.add_argument('--randomize_attack', action='store_true',\r\n default=False,\r\n help='randomly choose an attack at each step')\r\n parser.add_argument('--maximize_attack', action='store_true',\r\n default=False,\r\n help='choose the attack with maximum loss')\r\n\r\n parser.add_argument('--seed', type=int, default=0, help='RNG seed')\r\n parser.add_argument('--continue', default=False, action='store_true',\r\n help='continue previous training')\r\n parser.add_argument('--keep_every', type=int, default=1,\r\n help='only keep a checkpoint every X epochs')\r\n\r\n parser.add_argument('--optim', type=str, default='sgd')\r\n parser.add_argument('--lr', type=float, metavar='LR', required=False,\r\n help='learning rate')\r\n parser.add_argument('--lr_schedule', type=str, required=False,\r\n help='comma-separated list of epochs when learning '\r\n 'rate should drop')\r\n parser.add_argument('--clip_grad', type=float, default=1.0,\r\n help='clip gradients to this value')\r\n\r\n parser.add_argument('--attack', type=str, action='append',\r\n help='attack(s) to harden against')\r\n parser.add_argument('--alpha', type=float, default=1.0,\r\n help='alpha')\r\n parser.add_argument('--re', type=float, default=1.0,\r\n help='reconstruction weight')\r\n parser.add_argument('--vae_dir', type=str, default='data/emb2048/model_epoch172.pth')\r\n args = parser.parse_args()\r\n wandb.init(config=args, name=args.log_dir.replace(\"data/\", ''))\r\n if args.optim == 'adam':\r\n if args.lr is None:\r\n args.lr = 1e-3\r\n if args.lr_schedule is None:\r\n args.lr_schedule = '50'\r\n if args.num_epochs is None:\r\n args.num_epochs = 100\r\n elif args.optim == 'sgd':\r\n if args.dataset.startswith('cifar'):\r\n if args.lr is None:\r\n args.lr = 1e-1\r\n if args.lr_schedule is None:\r\n args.lr_schedule = '75,90,100'\r\n if args.num_epochs is None:\r\n args.num_epochs = 100\r\n elif (\r\n args.dataset.startswith('imagenet')\r\n or args.dataset == 'bird_or_bicycle'\r\n ):\r\n if args.lr is None:\r\n args.lr = 1e-1\r\n if args.lr_schedule is None:\r\n args.lr_schedule = '30,60,80'\r\n if args.num_epochs is None:\r\n args.num_epochs = 90\r\n\r\n torch.manual_seed(args.seed)\r\n np.random.seed(args.seed)\r\n random.seed(args.seed)\r\n\r\n normalize = CIFARNORMALIZE(32)\r\n innormalize = CIFARINNORMALIZE(32)\r\n\r\n dataset, model = get_dataset_model(args)\r\n vae = get_vae_template(args.vae_dir, 32, 2048)\r\n\r\n if isinstance(model, FeatureModel):\r\n model.allow_train()\r\n if torch.cuda.is_available():\r\n model.cuda()\r\n vae.cuda()\r\n\r\n\r\n l2_distance = L2Distance()\r\n linf_distance = LinfDistance()\r\n l2_distance.cuda()\r\n linf_distance.cuda()\r\n\r\n train_loader, val_loader = dataset.make_loaders(\r\n workers=4, batch_size=args.batch_size)\r\n\r\n attacks = [eval(attack_str) for attack_str in args.attack]\r\n validation_attacks = [\r\n NoAttack(),\r\n DeltaAttack(model, vae, num_iterations=100, norm='linf',eps_max=8 / 255),\r\n DeltaAttack(model, vae, num_iterations=100, norm='l2',eps_max=1.0),\r\n XAttack(model, vae, num_iterations=100, norm='linf', eps_max=8 / 255),\r\n XAttack(model, vae, num_iterations=100, norm='l2', eps_max=1.0)\r\n ]\r\n\r\n iteration = 0\r\n log_dir = os.path.join(args.log_dir)\r\n if os.path.exists(log_dir):\r\n shutil.rmtree(log_dir)\r\n time.sleep(5)\r\n os.makedirs(log_dir)\r\n\r\n # optimizer\r\n optimizer: optim.Optimizer\r\n if args.optim == 'sgd':\r\n weight_decay = 1e-4 if (\r\n args.dataset.startswith('imagenet')\r\n or args.dataset == 'bird_or_bicycle'\r\n ) else 2e-4\r\n optimizer = optim.SGD([{'params': model.parameters()},\r\n {'params': vae.parameters(), 'lr': 0.1*args.lr}],\r\n lr=args.lr,\r\n momentum=0.9,\r\n weight_decay=weight_decay)\r\n elif args.optim == 'adam':\r\n optimizer = AdamW([{'params': model.parameters()},\r\n {'params': vae.parameters()}\r\n ], lr=args.lr, betas=(0.9, 0.999), weight_decay=1.e-6)\r\n scheduler = torch.optim.lr_scheduler.StepLR(optimizer, step_size=30, gamma=0.1)\r\n else:\r\n raise ValueError(f'invalid optimizer {args.optim}')\r\n\r\n # check for checkpoints\r\n def get_checkpoint_fnames():\r\n for checkpoint_fname in glob.glob(os.path.join(glob.escape(log_dir),\r\n '*.ckpt.pth')):\r\n epoch = int(os.path.basename(checkpoint_fname).split('.')[0])\r\n if epoch < args.num_epochs:\r\n yield epoch, checkpoint_fname\r\n\r\n start_epoch = 0\r\n\r\n # parallelize\r\n if torch.cuda.is_available():\r\n device_ids = list(range(args.parallel))\r\n model = nn.DataParallel(model, device_ids)\r\n vae = nn.DataParallel(vae, device_ids)\r\n attacks = [nn.DataParallel(attack, device_ids) for attack in attacks]\r\n validation_attacks = [nn.DataParallel(attack, device_ids)\r\n for attack in validation_attacks]\r\n\r\n\r\n def reconst_images(model, vae, val_loader, attacks, batch_num=2, num_samples=10):\r\n\r\n if isinstance(attacks[0], nn.DataParallel):\r\n attack_name = attacks[0].module.__class__.__name__\r\n else:\r\n attack_name = attacks[0].__class__.__name__\r\n for batch_idx, (inputs, y) in enumerate(val_loader):\r\n inputs = inputs.cuda()\r\n y = y.cuda()\r\n\r\n if batch_idx >= batch_num:\r\n break\r\n else:\r\n adv_inputs = attacks[0](inputs, y)\r\n _, _, adv_inputs_i = vae(adv_inputs)\r\n _, _, inputs_i = vae(inputs)\r\n successful_attacks = []\r\n successful_l2_distance = []\r\n successful_linf_distance = []\r\n successful_l2_distance.extend(l2_distance(\r\n inputs,\r\n adv_inputs,\r\n ).detach())\r\n successful_linf_distance.extend(linf_distance(\r\n inputs,\r\n adv_inputs,\r\n ).detach())\r\n\r\n for lpips_name, successful_lpips in [\r\n ('l2', successful_l2_distance),\r\n ('linf', successful_linf_distance)\r\n ]:\r\n wandb.log({f'train-{attack_name}-distance/{lpips_name}':\r\n wandb.Histogram(torch.stack(successful_lpips)\r\n .cpu().detach().numpy())}, commit=False)\r\n for idx in range(num_samples):\r\n successful_attacks.append(torch.cat([\r\n inputs[idx],\r\n adv_inputs[idx],\r\n torch.clamp((adv_inputs[idx] -\r\n inputs[idx]) * 3 + 0.5,\r\n 0, 1),\r\n adv_inputs_i[idx],\r\n inputs_i[idx],\r\n torch.clamp((adv_inputs_i[idx] -\r\n inputs_i[idx]) * 3 + 0.5,\r\n 0, 1),\r\n torch.clamp((inputs[idx] -\r\n inputs_i[idx]) * 3 + 0.5,\r\n 0, 1),\r\n torch.clamp((adv_inputs[idx] -\r\n adv_inputs_i[idx]) * 3 + 0.5,\r\n 0, 1),\r\n torch.clamp(((adv_inputs[idx] -\r\n adv_inputs_i[idx]) - (inputs[idx] -\r\n inputs_i[idx])) * 3 + 0.5,\r\n 0, 1),\r\n ], dim=1).detach())\r\n wandb.log({f'train-{attack_name}-images': [\r\n wandb.Image(torch.cat(successful_attacks, dim=2))]}, step=iteration)\r\n\r\n # necessary to put training loop in a function because otherwise we get\r\n # huge memory leaks\r\n def run_iter(\r\n inputs: torch.Tensor,\r\n labels: torch.Tensor,\r\n iteration: int,\r\n train: bool = True\r\n ):\r\n model.eval() # set model to eval to generate adversarial examples\r\n vae.eval()\r\n\r\n if torch.cuda.is_available():\r\n inputs = inputs.cuda()\r\n labels = labels.cuda()\r\n\r\n if args.only_attack_correct:\r\n with torch.no_grad():\r\n _, _, inputs_i = vae(inputs)\r\n orig_logits = model(inputs - inputs_i)\r\n to_attack = orig_logits.argmax(1) == labels\r\n else:\r\n to_attack = torch.ones_like(labels).bool()\r\n\r\n adv_inputs = inputs.clone()\r\n if to_attack.sum() > 0:\r\n adv_inputs[to_attack] = attacks[0](inputs[to_attack],\r\n labels[to_attack])\r\n # FORWARD PASS\r\n if train:\r\n optimizer.zero_grad()\r\n model.train() # now we set the model to train mode\r\n vae.train()\r\n\r\n mu, logvar, inputs_i = vae(inputs)\r\n adv_mu, adv_logvar, adv_inputs_i = vae(adv_inputs)\r\n output = model(torch.cat((inputs-inputs_i, inputs_i), dim=0))\r\n logits = output[0:inputs.size(0)]\r\n logits2 = output[inputs.size(0):]\r\n\r\n adv_output = model(torch.cat((adv_inputs - adv_inputs_i, adv_inputs_i), dim=0))\r\n adv_logits = adv_output[0:inputs.size(0)]\r\n adv_logits2 = adv_output[inputs.size(0):]\r\n\r\n # CONSTRUCT LOSS\r\n loss1 = F.mse_loss(normalize(adv_inputs_i), normalize(adv_inputs)) + \\\r\n F.mse_loss(normalize(inputs_i), normalize(inputs)) + \\\r\n F.mse_loss(normalize(adv_inputs_i), normalize(inputs_i)+normalize(adv_inputs)-normalize(inputs))\r\n cross_entropy = F.cross_entropy(logits, labels, reduction='none').mean() + F.cross_entropy(adv_logits, labels, reduction='none').mean()\r\n entropy = (F.softmax(logits2, dim=1) * F.log_softmax(logits2, dim=1)).sum(dim=1).mean() + (F.softmax(adv_logits2, dim=1) * F.log_softmax(adv_logits2, dim=1)).sum(dim=1).mean()\r\n loss2 = cross_entropy + entropy\r\n loss3 = -0.5 * torch.sum(1 + logvar - mu.pow(2) - logvar.exp()) -0.5 * torch.sum(1 + adv_logvar - adv_mu.pow(2) - adv_logvar.exp())\r\n loss3 /= args.batch_size * 3 * 2048\r\n loss = args.re * loss1 + loss2 + loss3\r\n with torch.no_grad():\r\n norm = torch.norm(torch.abs( (normalize(adv_inputs)-normalize(inputs)+normalize(inputs_i)).view(inputs.size(0), -1)), p=2, dim=1)\r\n acc_adv = 1 - F.mse_loss(torch.div(normalize(adv_inputs)- normalize(inputs) + normalize(inputs_i), norm.unsqueeze(1).unsqueeze(2).unsqueeze(3)), \\\r\n torch.div(normalize(adv_inputs_i), norm.unsqueeze(1).unsqueeze(2).unsqueeze(3)), \\\r\n reduction = 'sum')/100\r\n # LOGGING\r\n orig_accuracy = calculate_accuracy(logits, labels)\r\n accuracy = calculate_accuracy(adv_logits, labels)\r\n wandb.log({'loss': loss.item()}, step=iteration)\r\n wandb.log({'loss1': loss1.item()}, step=iteration)\r\n wandb.log({'loss2': loss2.item()}, step=iteration)\r\n wandb.log({'loss3': loss3.item()}, step=iteration)\r\n wandb.log({'cross_entropy': cross_entropy.item()}, step=iteration)\r\n #wandb.log({'entropy': entropy.item()}, step=iteration)\r\n wandb.log({'accuracy': accuracy.item()}, step=iteration)\r\n wandb.log({'acc_adv': acc_adv.data.item()}, step=iteration)\r\n wandb.log({'orig_accuracy': orig_accuracy.item()}, step=iteration)\r\n\r\n\r\n if train:\r\n print(f'ITER {iteration:06d}',\r\n f'accuracy: {accuracy.item() * 100:5.1f}%',\r\n f'orig_accuracy: {orig_accuracy.item() * 100:5.1f}%',\r\n f'loss: {loss.item():.2f}',\r\n f'loss1: {loss1.item():.2f}',\r\n f'loss2: {loss2.item():.2f}',\r\n f'loss3: {loss3.item():.2f}',\r\n f'cross_entropy: {cross_entropy.item():.2f}',\r\n #f'entropy: {entropy.item():.2f}',\r\n f'acc_adv: {acc_adv.data.item():.2f}',\r\n sep='\\t')\r\n\r\n # OPTIMIZATION\r\n if train:\r\n loss.backward()\r\n\r\n # clip gradients and optimize\r\n nn.utils.clip_grad_value_(model.parameters(), args.clip_grad)\r\n nn.utils.clip_grad_value_(vae.parameters(), args.clip_grad)\r\n optimizer.step()\r\n\r\n for epoch in range(start_epoch, args.num_epochs):\r\n\r\n lr = optimizer.param_groups[0]['lr']\r\n print(f'START EPOCH {epoch:04d}(lr={lr:0e})')\r\n for batch_index, (inputs, labels) in enumerate(train_loader):\r\n # ramp-up learning rate for SGD\r\n run_iter(inputs, labels, iteration)\r\n iteration += 1\r\n print(f'END EPOCH {epoch:04d}')\r\n scheduler.step()\r\n if torch.cuda.is_available():\r\n torch.cuda.empty_cache()\r\n \r\n if epoch % 10 == 0:\r\n # VALIDATION\r\n print('BEGIN VALIDATION')\r\n model.eval()\r\n vae.eval()\r\n\r\n reconst_images(model, vae, val_loader, attacks)\r\n\r\n evaluationdelta.evaluate_against_attacks(\r\n model, vae, validation_attacks, val_loader, parallel=args.parallel,\r\n wandb=wandb, iteration=iteration, num_batches=args.val_batches,\r\n )\r\n\r\n checkpoint_fname = os.path.join(log_dir, f'{epoch:04d}.ckpt.pth')\r\n print(f'CHECKPOINT {checkpoint_fname}')\r\n checkpoint_model = model\r\n if isinstance(checkpoint_model, nn.DataParallel):\r\n checkpoint_model = checkpoint_model.module\r\n if isinstance(checkpoint_model, FeatureModel):\r\n checkpoint_model = checkpoint_model.model\r\n checkpoint_vae = vae\r\n if isinstance(checkpoint_vae, nn.DataParallel):\r\n checkpoint_vae = checkpoint_vae.module\r\n state = {\r\n 'model': checkpoint_model.state_dict(),\r\n 'vae': checkpoint_vae.state_dict(),\r\n 'optimizer': optimizer.state_dict(),\r\n 'iteration': iteration,\r\n 'arch': args.arch,\r\n }\r\n torch.save(state, checkpoint_fname)\r\n\r\n # delete extraneous checkpoints\r\n last_keep_checkpoint = (epoch // args.keep_every) * args.keep_every\r\n for epoch, checkpoint_fname in get_checkpoint_fnames():\r\n if epoch < last_keep_checkpoint and epoch % args.keep_every != 0:\r\n print(f' remove {checkpoint_fname}')\r\n os.remove(checkpoint_fname)\r\n\r\n print('BEGIN EVALUATION')\r\n model.eval()\r\n vae.eval()\r\n\r\n evaluationdelta.evaluate_against_attacks(\r\n model, vae, validation_attacks, val_loader, parallel=args.parallel,\r\n )\r\n print('END EVALUATION')\r\n", "sub_path": "adv_disen.py", "file_name": "adv_disen.py", "file_ext": "py", "file_size_in_byte": 24192, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "torch.optim.optimizer.Optimizer", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.zeros_like", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 122, "usage_type": "call"}, {"api_name": "torch.nn.DataParallel", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 132, "usage_type": "name"}, {"api_name": "torch.load", "line_number": 133, "usage_type": "call"}, {"api_name": "torch.full", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.full", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.full", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 151, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 151, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 152, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 152, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 153, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 153, "usage_type": "call"}, {"api_name": "torch.full", "line_number": 160, "usage_type": "call"}, {"api_name": "torch.full", "line_number": 161, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 164, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 164, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 178, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 178, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 193, "usage_type": "call"}, {"api_name": "perceptual_advex.utilities.add_dataset_model_arguments", "line_number": 195, "usage_type": "call"}, {"api_name": "wandb.init", "line_number": 241, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 268, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 269, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 269, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 270, "usage_type": "call"}, {"api_name": "perceptual_advex.utilities.get_dataset_model", "line_number": 275, "usage_type": "call"}, {"api_name": "perceptual_advex.models.FeatureModel", "line_number": 278, "usage_type": "argument"}, {"api_name": "torch.cuda.is_available", "line_number": 280, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 280, "usage_type": "attribute"}, {"api_name": "perceptual_advex.distances.L2Distance", "line_number": 285, "usage_type": "call"}, {"api_name": "perceptual_advex.distances.LinfDistance", "line_number": 286, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 303, "usage_type": "call"}, {"api_name": "os.path", "line_number": 303, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 304, "usage_type": "call"}, {"api_name": "os.path", "line_number": 304, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 305, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 306, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 307, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 325, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 325, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 331, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 331, "usage_type": "call"}, {"api_name": "os.path", "line_number": 331, "usage_type": "attribute"}, {"api_name": "glob.escape", "line_number": 331, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 333, "usage_type": "call"}, {"api_name": "os.path", "line_number": 333, "usage_type": "attribute"}, {"api_name": "torch.cuda.is_available", "line_number": 340, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 340, "usage_type": "attribute"}, {"api_name": "torch.nn.DataParallel", "line_number": 342, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 342, "usage_type": "name"}, {"api_name": "torch.nn.DataParallel", "line_number": 343, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 343, "usage_type": "name"}, {"api_name": "torch.nn.DataParallel", "line_number": 344, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 344, "usage_type": "name"}, {"api_name": "torch.nn.DataParallel", "line_number": 345, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 345, "usage_type": "name"}, {"api_name": "torch.nn.DataParallel", "line_number": 351, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 351, "usage_type": "name"}, {"api_name": "wandb.log", "line_number": 381, "usage_type": "call"}, {"api_name": "wandb.Histogram", "line_number": 382, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 382, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 385, "usage_type": "call"}, {"api_name": "torch.clamp", "line_number": 388, "usage_type": "call"}, {"api_name": "torch.clamp", "line_number": 393, "usage_type": "call"}, {"api_name": "torch.clamp", "line_number": 396, "usage_type": "call"}, {"api_name": "torch.clamp", "line_number": 399, "usage_type": "call"}, {"api_name": "torch.clamp", "line_number": 402, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 407, "usage_type": "call"}, {"api_name": "wandb.Image", "line_number": 408, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 408, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 413, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 414, "usage_type": "attribute"}, {"api_name": "torch.cuda.is_available", "line_number": 421, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 421, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 426, "usage_type": "call"}, {"api_name": "torch.ones_like", "line_number": 431, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 445, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 449, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 460, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 463, "usage_type": "call"}, {"api_name": "torch.norm", "line_number": 464, "usage_type": "call"}, {"api_name": "torch.abs", "line_number": 464, "usage_type": "call"}, {"api_name": "torch.div", "line_number": 465, "usage_type": "call"}, {"api_name": "torch.div", "line_number": 466, "usage_type": "call"}, {"api_name": "perceptual_advex.utilities.calculate_accuracy", "line_number": 469, "usage_type": "call"}, {"api_name": "perceptual_advex.utilities.calculate_accuracy", "line_number": 470, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 471, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 472, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 473, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 474, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 475, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 477, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 478, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 479, "usage_type": "call"}, {"api_name": "torch.nn.utils.clip_grad_value_", "line_number": 500, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 500, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 500, "usage_type": "name"}, {"api_name": "torch.nn.utils.clip_grad_value_", "line_number": 501, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 501, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 501, "usage_type": "name"}, {"api_name": "torch.cuda.is_available", "line_number": 514, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 514, "usage_type": "attribute"}, {"api_name": "torch.cuda.empty_cache", "line_number": 515, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 515, "usage_type": "attribute"}, {"api_name": "perceptual_advex.evaluationdelta.evaluate_against_attacks", "line_number": 525, "usage_type": "call"}, {"api_name": "perceptual_advex.evaluationdelta", "line_number": 525, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 530, "usage_type": "call"}, {"api_name": "os.path", "line_number": 530, "usage_type": "attribute"}, {"api_name": "torch.nn.DataParallel", "line_number": 533, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 533, "usage_type": "name"}, {"api_name": "perceptual_advex.models.FeatureModel", "line_number": 535, "usage_type": "argument"}, {"api_name": "torch.nn.DataParallel", "line_number": 538, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 538, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 547, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 554, "usage_type": "call"}, {"api_name": "perceptual_advex.evaluationdelta.evaluate_against_attacks", "line_number": 560, "usage_type": "call"}, {"api_name": "perceptual_advex.evaluationdelta", "line_number": 560, "usage_type": "name"}]}
+{"seq_id": "413759432", "text": "\n# -*- coding:utf-8 -*- \n# 设置中文utf-8解释环境标识\n# No例1:气象站长期年平均风速的条形图绘制方法示例及代码详解\n\nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\nimport pandas as pd #导入pandas包\nimport pylab as pl\nimport numpy as np\n\nmpl.rcParams[\"font.sans-serif\"]=[\"SimHei\"] # 设置坐标轴标识字符串中的中文标识的字体为黑体\nmpl.rcParams['axes.unicode_minus'] = False # 不使用默认的unicode minus模式来处理坐标轴轴线的刻度标签为负数的情况。\n \ndata = pd.read_csv(\"no1_anex_bar.csv\") #pandas 中输入EXCEL表格数据CSV的方法\nprint(data)#测试数据输入是否成功\nprint(\"===================================\")\n\nprint(data[\"year\"])#测试数据框的以头标识为year的一维向量的数值调用方法\nprint(\"===================================\")\nXyear=data[\"year\"] #将年数值向量附值给Xyear\n\nYspeed=data[\"speed\"] #将风速向量附值给Yspeed\n\nplt.figure(figsize=(9, 6)) #创建figure窗口画布大小\npl.xticks(rotation=45) #设置条形图X轴的刻度标签的旋转方向\n# 绘制BAR图\nplt.bar(Xyear,Yspeed,align=\"center\",color=\"lightsteelblue\",tick_label=Xyear,hatch=\"\",alpha=0.9,width=0.4,label=\"气象站年平均风速A\")\n# 条形图的各参数请参加bar 参数设置手册\n\n#绘制气象站风速平均值水平参考线\navg=np.mean(Yspeed)\nplt.axhline(y=avg,C=\"r\",ls=\"--\",lw=2)\n#绘制图纸区域的指向性文本描述,如:多年平均风速值\nx1=2003.5\ny1=avg+0.05\nplt.text(x1,y1,\"多年平均风速值\",weight=\"bold\",color=\"black\",fontsize=12)\n#设置XY坐标轴的中文标签\nplt.xlabel(\"年份\",fontsize=12)\nplt.ylabel(\"风速\",fontsize=12)\n\n\n#设置Y轴在网格线\nplt.grid(True,axis=\"y\",ls=\":\",color=\"green\",alpha=0.3) #注意这里\"y\"指Y坐标轴,不能采用变量Yspeed写入。\n#设置Y轴的坐标取值范围\nplt.ylim(1,3) #Y轴起点1,终点3.\n#设置图例标签\nplt.title(\"气象站多年年平均风速分布图\")\n\n# 关于人工站与自动站数据的叠加分析,图形叠加分析\ndata1 = pd.read_csv(\"no1_anex_bar2.csv\") #pandas 中输入EXCEL表格后10年自动站数据CSV的方法\nXyear1=data1[\"year\"] #将年数值向量附值给Xyear1\nYspeed1=data1[\"speed\"] #将风速向量附值给Yspeed1\nwidth=0.4 #设定后10年柱状图与原来柱状图在X的位置偏移量\nplt.bar(Xyear1+width,Yspeed1,width,align=\"center\",color=\"#FFA500\",hatch=\"\",alpha=0.9,label=\"气象站年平均风速B\")\n\n# 设置X轴的不同情况的刻度位置及刻度标签\n\n#plt.xticks([index+(width/2) for index in Xyear],Xyear)\n\t\nplt.legend() #设定不同线条的标签标识\n#显示图形\nplt.show()\n", "sub_path": "test/出图/no1_anex_bar.py", "file_name": "no1_anex_bar.py", "file_ext": "py", "file_size_in_byte": 2673, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "matplotlib.rcParams", "line_number": 12, "usage_type": "attribute"}, {"api_name": "matplotlib.rcParams", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "pylab.xticks", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}]}
+{"seq_id": "98947850", "text": "import importlib\nimport os\nimport sys\nimport getopt\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport random\nimport ast\n\nargs,_ = getopt.getopt(sys.argv[1:], \"i:t:m:a:p:o:\")\nargs = dict(args)\nif not (\"-i\" in args and \"-m\" in args and \"-t\" in args):\n\tprint(\"USAGE : train.py -i inputFile -t target -m model\")\n\tquit()\n\n\n\ntarget = args[\"-t\"]\npreprocessing = args[\"-p\"]\nmodelParameters = ast.literal_eval(args[\"-a\"])\nmodel = getattr(importlib.import_module(\"Models.\"+args[\"-m\"]), args[\"-m\"])(**modelParameters)\npre = getattr(importlib.import_module(\"Preprocessing.\"+args[\"-p\"]), \"preprocess\")\ndata = pre(pd.read_csv(args[\"-i\"]))\noutputFile = args[\"-o\"]\n\nos.system(\"clear\")\n\n\nprint(\"\\nInput file : \"+args[\"-i\"])\nprint(\"Target : \"+args[\"-t\"])\nprint(\"Model : \"+args[\"-m\"])\nprint(\"Models parameters : \"+args[\"-a\"])\nprint(\"Preprocessing method : \"+args[\"-p\"])\nprint(\"Output file : \"+args[\"-o\"])\n\nY = data[target]\nX = data.drop(target, axis=1)\n\nmsk = np.random.rand(len(X)) < 0.5\n\ntrainX = X[msk]\ntestX = X[~msk]\ntrainY = Y[msk]\ntestY = Y[~msk]\n\nmodel.fit(trainX, trainY)\n\nprint(\"Train score = \" + str(model.score(trainX, trainY)))\nprint(\"Test score = \" + str(model.score(testX, testY)))\n\npd.DataFrame(model.predict(X), columns=[target]).to_csv(outputFile)\n", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 1275, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "getopt.getopt", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 11, "usage_type": "attribute"}, {"api_name": "ast.literal_eval", "line_number": 21, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 22, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 24, "usage_type": "call"}, {"api_name": "os.system", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 52, "usage_type": "call"}]}
+{"seq_id": "602308161", "text": "from .base import FunctionalTest\nfrom selenium.webdriver.support.ui import WebDriverWait\nimport time\n\n\nTEST_EMAIL = 'astinchoi@mockmyid.com'\n\nclass LoginTest(FunctionalTest):\n\n def switch_to_new_window(self, text_in_title):\n retries = 60\n while retries > 0:\n for handle in self.browser.window_handles:\n self.browser.switch_to_window(handle)\n if text_in_title in self.browser.title:\n return\n retries -= 1\n time.sleep(0.5)\n self.fail('could not find window')\n\n def wait_for_element_with_id(self, element_id):\n WebDriverWait(self.browser, timeout=30).until(\n lambda b: b.find_element_by_id(element_id),\n 'Could not find element with id {}. Page text was {}'.format(\n element_id, self.browser.find_element_by_tag_name('body').text\n )\n )\n\n # def wait_to_be_logged_in(self):\n # self.wait_for_element_with_id('id_logout')\n # navbar = self.browser.find_element_by_css_selector('.navbar')\n # self.assertIn('astinchoi@mockmyid.com', navbar.text)\n\n # def wait_to_be_logged_out(self):\n # self.wait_for_element_with_id('id_login')\n # navbar = self.browser.find_element_by_css_selector('.navbar')\n # self.assertNotIn('astinchoi@mockmyid.com', navbar.text) \n\n def test_login_with_persona(self):\n self.browser.get(self.server_url)\n self.browser.find_element_by_id('id_login').click()\n\n self.switch_to_new_window('Mozilla Persona')\n\n self.browser.find_element_by_id(\n 'authentication_email'\n ).send_keys(TEST_EMAIL)\n self.browser.find_element_by_tag_name('button').click()\n\n self.switch_to_new_window('To-Do')\n\n # self.wait_for_element_with_id('id_logout')\n # navbar = self.browser.find_element_by_css_selector('.navbar')\n # self.assertIn('astinchoi@mockmyid.com', navbar.text)\n self.wait_to_be_logged_in(email=TEST_EMAIL)\n\n self.browser.refresh()\n self.wait_to_be_logged_in(email=TEST_EMAIL)\n\n self.browser.find_element_by_id('id_logout').click()\n self.wait_to_be_logged_out(email=TEST_EMAIL)\n\n self.browser.refresh()\n self.wait_to_be_logged_out(email=TEST_EMAIL)", "sub_path": "functional_tests/test_login.py", "file_name": "test_login.py", "file_ext": "py", "file_size_in_byte": 2306, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "base.FunctionalTest", "line_number": 8, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 18, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.ui.WebDriverWait", "line_number": 22, "usage_type": "call"}]}
+{"seq_id": "55197206", "text": "from keras import layers\nfrom keras import models\nfrom keras import optimizers\nfrom keras import backend as K\n\n\nimport tensorflow as tf\n\nimport numpy as np\nimport os\n\n\nclass ActorCriticAgent(object):\n \"\"\"\n Actor Critic model with continuous action space\n Input: \n N-dimensional continuous state with dimensions normalized in the range[-1,1]\n 1-dimensional reward normalized in the range [0, 1]\n Output: \n M-dimensional continuous action with dimensions normalized in the range[-1,1]\n \"\"\"\n _use_K_actor_update_function = False\n\n def __init__(self,\n state_dim,\n action_dim,\n discount_rate,\n hidden_dims,\n learning_rate,\n replay_memory_sz,\n batch_sz):\n \"\"\"\n\n :param state_dim:\n :param action_dim:\n :param discount_rate:\n :param hidden_dims:\n :param learning_rate:\n \"\"\"\n self.state_dim = state_dim\n self.action_dim = action_dim\n self.discount_rate = discount_rate\n self.replay_memory_sz = replay_memory_sz\n self.batch_size = batch_sz\n self.state_memory = np.zeros(shape=(self.replay_memory_sz, self.state_dim), dtype=float)\n self.action_memory = np.zeros(shape=(self.replay_memory_sz, self.action_dim), dtype=float)\n self.reward_memory = np.zeros(shape=self.replay_memory_sz, dtype=float)\n self.memory_index = 0\n self.step_index = 0\n # build models\n if ActorCriticAgent._use_K_actor_update_function:\n self.policy, self.critic, self.policy_training_function, self.state_ph = self._build_models(hidden_dims, learning_rate)\n else:\n self.policy, self.critic, self.actor, self.advantage_ph = self._build_models(hidden_dims, learning_rate)\n\n self.critic_training_memory = list()\n\n def _build_models(self, hidden_dims, learning_rate):\n #\n # build policy\n state_ph = layers.Input(shape=(self.state_dim,))\n hidden = layers.Dense(units=hidden_dims[0], activation='relu')(state_ph)\n for jh in np.arange(1, len(hidden_dims)):\n hidden = layers.Dense(units=hidden_dims[jh], activation='relu')(hidden)\n action = layers.Dense(units=self.action_dim, activation='tanh', use_bias=False)(hidden)\n policy_model = models.Model(input=[state_ph], output=[action], name=\"policy_model\")\n policy_model.summary()\n #\n # build critic\n action_ph = layers.Input(shape=(self.action_dim,))\n state_action = layers.concatenate([state_ph, action_ph])\n hidden = layers.Dense(units=hidden_dims[0], activation='relu')(state_action)\n for jh in np.arange(1, len(hidden_dims)):\n hidden = layers.Dense(units=hidden_dims[jh], activation='relu')(hidden)\n value = layers.Dense(units=1, activation='sigmoid', name=\"value_l\", use_bias=False)(hidden)\n critic_model = models.Model(input=[state_ph, action_ph], output=[value], name='critic_model')\n critic_model.compile(optimizer=optimizers.Adam(learning_rate=learning_rate), loss='mse')\n critic_model.summary()\n #\n # build actor\n if ActorCriticAgent._use_K_actor_update_function:\n value_ = critic_model([state_ph, action])\n policy_learn_loss = -K.log(K.clip(value_, 1e-8, 1.0))\n actor_gradients = K.gradients(loss=policy_learn_loss, variables=policy_model.trainable_weights)\n opt = tf.keras.optimizers.Adam(learning_rate=learning_rate)\n policy_train_function = opt.apply_gradients(zip(actor_gradients, policy_model.trainable_weights))\n return policy_model, critic_model, policy_train_function, state_ph\n else:\n advantage_ph = layers.Input(shape=(1,), name=\"advantage\")\n\n def custom_loss(y_true, y_pred):\n learn_loss = K.mean(K.square(y_true - y_pred)) * advantage_ph\n return learn_loss\n\n actor_model = models.Model(input=[state_ph, advantage_ph], output=[action], name='actor_model')\n actor_model.compile(optimizer=optimizers.Adam(learning_rate=learning_rate), loss=custom_loss)\n actor_model.summary()\n return policy_model, critic_model, actor_model, advantage_ph\n\n def choose_action(self, state: np.array):\n j = (self.memory_index - 1) % self.replay_memory_sz\n current_value = self.reward_memory[j]\n value_th = current_value + (1.0 - current_value) / 3.0 # I want to make at least 1/3 of my way to 1\n # try policy\n action = self.policy.predict(x=state[np.newaxis, :])[0]\n expected_value = self.critic.predict(x=[state[np.newaxis, :], action[np.newaxis, :]])[0]\n if expected_value >= value_th:\n return action\n # try random action\n action = np.array(np.random.uniform(-1.0, +1.0, size=self.action_dim))\n #\n return action\n\n def train(self, state: np.array, action: np.array, reward: np.float):\n \"\"\"\n\n :type reward: object\n \"\"\"\n # update memory and apply discount\n j = self.memory_index\n assert self.step_index % self.replay_memory_sz == j\n self.state_memory[j, :] = np.copy(state)\n self.action_memory[j, :] = np.copy(action)\n self.reward_memory[j] = 0\n discounted_reward = reward\n while discounted_reward > 1e-2:\n self.reward_memory[j] = discounted_reward\n discounted_reward *= self.discount_rate\n j = (j - 1) % self.replay_memory_sz\n if j == self.memory_index or discounted_reward < self.reward_memory[j]:\n break\n # training\n if self.step_index >= self.batch_size:\n # select training batch\n js = np.random.uniform(\n low=0,\n high=self.step_index,\n size=self.batch_size).astype(int) % self.replay_memory_sz\n state_batch = self.state_memory[js, :]\n action_batch = self.action_memory[js, :]\n reward_batch = self.reward_memory[js]\n # train critic\n cost = self.critic.train_on_batch(x=[state_batch, action_batch], y=reward_batch)\n self.critic_training_memory.append(cost)\n # train actor based on actual gain\n if ActorCriticAgent._use_K_actor_update_function:\n K.get_session().run(fetches=[self.policy_training_function], feed_dict={self.state_ph: state_batch})\n else:\n cost = self.actor.train_on_batch(x=[state_batch, reward_batch], y=action_batch)\n #predicted_value = self.critic.predict(x=[state[np.newaxis, :], action[np.newaxis, :]])[0]\n #if predicted_value >= reward:\n # advantage = np.divide(predicted_value, (1.0 - reward + 1e-8))\n # cost = self.actor.train_on_batch(x=[state[np.newaxis, :], advantage[np.newaxis, :]], y=action[np.newaxis, :])\n #\n self.memory_index = (self.memory_index + 1) % self.replay_memory_sz\n self.step_index += 1\n\n def save(self, folder, game_name):\n os.makedirs(folder, exist_ok=True)\n filename = os.path.join(folder, game_name + '_critic.h5')\n self.critic.save(filename)\n print('saved model({})'.format(filename))\n filename = os.path.join(folder, game_name + '_actor.h5')\n self.actor.save(filename)\n print('saved model({})'.format(filename))\n\n\n def load(self, folder, game_name):\n filename = os.path.join(folder, game_name + '_critic.h5')\n if os.path.exists(filename):\n self.critic.load_weights(filename)\n print('loaded model({})'.format(filename))\n filename = os.path.join(folder, game_name + '_actor.h5')\n if os.path.exists(filename):\n self.actor.load_weights(filename)\n print('loaded model({})'.format(filename))\n\n def reset(self):\n self.state_memory = np.zeros(shape=(self.replay_memory_sz, self.state_dim), dtype=float)\n self.action_memory = np.zeros(shape=(self.replay_memory_sz, self.action_dim), dtype=float)\n self.reward_memory = np.zeros(shape=self.replay_memory_sz, dtype=float)\n self.memory_index = 0\n self.step_index = 0\n", "sub_path": "openai_gym/actor_critic_agent.py", "file_name": "actor_critic_agent.py", "file_ext": "py", "file_size_in_byte": 8391, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "numpy.zeros", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 47, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 61, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 62, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 63, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 64, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 64, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 65, "usage_type": "name"}, {"api_name": "keras.models.Model", "line_number": 66, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 66, "usage_type": "name"}, {"api_name": "keras.layers.Input", "line_number": 70, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 70, "usage_type": "name"}, {"api_name": "keras.layers.concatenate", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 71, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 72, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 72, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 73, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 74, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 75, "usage_type": "name"}, {"api_name": "keras.models.Model", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 76, "usage_type": "name"}, {"api_name": "keras.optimizers.Adam", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.optimizers", "line_number": 77, "usage_type": "name"}, {"api_name": "keras.backend.log", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 83, "usage_type": "name"}, {"api_name": "keras.backend.clip", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.backend.gradients", "line_number": 84, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 84, "usage_type": "name"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 85, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 85, "usage_type": "attribute"}, {"api_name": "keras.layers.Input", "line_number": 89, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 89, "usage_type": "name"}, {"api_name": "keras.backend.mean", "line_number": 92, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 92, "usage_type": "name"}, {"api_name": "keras.backend.square", "line_number": 92, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 95, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 95, "usage_type": "name"}, {"api_name": "keras.optimizers.Adam", "line_number": 96, "usage_type": "call"}, {"api_name": "keras.optimizers", "line_number": 96, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 100, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 105, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 106, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 110, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 114, "usage_type": "attribute"}, {"api_name": "numpy.float", "line_number": 114, "usage_type": "attribute"}, {"api_name": "numpy.copy", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 135, "usage_type": "attribute"}, {"api_name": "keras.backend.get_session", "line_number": 147, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 147, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path", "line_number": 160, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 163, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 169, "usage_type": "call"}, {"api_name": "os.path", "line_number": 169, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 173, "usage_type": "call"}, {"api_name": "os.path", "line_number": 173, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 174, "usage_type": "call"}, {"api_name": "os.path", "line_number": 174, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 181, "usage_type": "call"}]}
+{"seq_id": "194507866", "text": "import os\nimport sys\nimport glob\nimport nltk\nimport sqlite3\nimport unittest\n\n\nconnection = None\n\n'''\n# In this unit test file, we have three test functions:\n\n# 1. check if all words in database been printed\n# 2. check if the word \"angle\" is printed correctly\n# 3. check if the word \"year\" is printed correctly\n'''\nclass TestStringMethods(unittest.TestCase):\n\n\tdef test_totalNum(self):\n\t\tfp = open(\"./output.txt\")\n\t\tcount = 0\n\t\tfor line in fp:\n\t\t\tcount +=1\n\t\tfp.close()\n\t\tquery = \"select count(distinct word) from posting \"\t\t\n\t\tself.assertEqual(executeQuery(query)[0][0],count)\n\n\tdef test_angle_info(self):\t\n\t\tfp = open(\"./output.txt\")\n\t\tangle_list = None\n\t\tfor line in fp:\n\t\t\ta_list = line.split(\"\t\")\n\t\t\tif a_list[0] == \"angle\":\n\t\t\t\tangle_list = a_list[1].split(\";\")[:-1]\t\n\t\tfp.close()\n\t\tself.assertTrue((angle_list)!= None)\n\t\tself.assertEqual(len(angle_list), 1)\n\t\tself.assertEqual(int(angle_list[0].split(\":\")[0]),1971)\n\t\tself.assertEqual(int(angle_list[0].split(\":\")[1]),250)\n\n\tdef test_years_info(self):\t\n\t\tfp = open(\"./output.txt\")\n\t\tyear_list = None\n\t\tfor line in fp:\n\t\t\ta_list = line.split(\"\t\")\n\t\t\tif a_list[0] == \"year\":\n\t\t\t\tyear_list = a_list[1].split(\";\")[:-1]\t\n\t\tfp.close()\n\t\tself.assertTrue((year_list)!= None)\n\t\tself.assertEqual(len(year_list), 2)\n\t\tself.assertTrue( \"1972:5\" in year_list)\n\t\tself.assertTrue(\"1971:1\" in year_list)\n\n\n'''\n# this function is to execute query\n'''\ndef executeQuery(query):\n\tcur = connection.cursor() #get cursor\n\treturn [i for i in cur.execute(query)]\n\n\n'''\n# this function is to connect to database\n'''\ndef connectionDataBase(data_file_name):\n\tglobal connection\n\t#create a connection to database with file name \"data_file_name\", if error print it out\n\ttry:\n\t\tconnection = sqlite3.connect(data_file_name)\n\t\treturn connection\n\texcept Exception as e:\n\t\tprint(e)\n\t\texit(\"Error,unit_test file can not connect database\")\n\n\n'''\n# this function is to check if create_index.py exist\n# check if Documents contains correct docuemnts\n'''\ndef checkFiles():\n\tif len(sys.argv) != 1:\n\t\texit(\"Error, command line error..\")\n\telse:\n\t\tif not os.path.isfile(\"./../create_index.py\") or not os.path.isfile(\"../print_index.py\"):\n\t\t\texit(\"Error, create_index.py or print_index.py does not exit..\") \n\ttry:\n\t\tfor filepath in glob.glob(os.path.join(\"./Documents\", '*.txt')):\n\t\t\tif int(filepath.split(\"_\")[1]) != 1971 and int(filepath.split(\"_\")[1]) != 1972:\n\t\t\t\texit(\"Error, Document not correct\")\n\texcept Exception:\n\t\traise\n\t\texit(\"Error, Documents' files not correct\")\n\n\n\tprint(\"The documents and python script are correct...\\nchecking...\")\n\n\nif __name__ == '__main__':\n\tcheckFiles()\n\tos.system(\"python3 ./../print_index.py Documents > output.txt\")\n\tconnection = connectionDataBase(\"./Documents/data.db\")\n\tunittest.main(verbosity=2)\n", "sub_path": "assignment1/Part1/Part1_test/print_index_unit_test.py", "file_name": "print_index_unit_test.py", "file_ext": "py", "file_size_in_byte": 2750, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "unittest.TestCase", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 71, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 102, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 104, "usage_type": "call"}]}
+{"seq_id": "558765098", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nfrom sklearn import datasets, linear_model\nfrom sklearn.metrics import mean_squared_error, r2_score\n\n\n# Load the diabetes dataset\ndiabetes = datasets.load_diabetes()\n# Use only one feature\ndiabetes_X = diabetes.data[:, np.newaxis, 2]\n\n# 데이터 설정\nX_train, X_test = diabetes_X[:-20],diabetes_X[-20:]\ny_train, y_test = diabetes.target[:-20], diabetes.target[-20:]\n\nprint(X_train, '----------',X_test)\nprint()\nprint(y_train, '-------------',y_test)\ndef Linear_model_def(X_train,y_train,X_test):\n Linear_model = linear_model.LinearRegression() # Linear 모델 선언\n Linear_model.fit(X_train, y_train) # 학습 데이터 fit\n y_pred = Linear_model.predict(X_test) # 예측 데이터\n # # The coefficients\n # print('Coefficients: \\n', Linear_model.coef_)\n # # The mean squared error\n # print(\"Mean squared error: %.2f\" %mean_squared_error(y_test, y_pred))\n # # Explained variance score: 1 is perfect prediction\n # print('Variance score: %.2f' %r2_score(y_test, y_pred))\n # Plot outputs\n plt.scatter(X_test, y_test, color='red')\n plt.plot(X_test, y_pred, color='blue', linewidth=6)\n plt.scatter(X_train, y_train, color='black')\n plt.plot(X_train, Linear_model.predict(X_train), color='blue', linewidth=3)\n plt.xticks(())\n plt.yticks(())\n\nLinear_model_def(X_train,y_train,X_test)\nplt.show()", "sub_path": "Bigcon_package/temp/lin.py", "file_name": "lin.py", "file_ext": "py", "file_size_in_byte": 1391, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "sklearn.datasets.load_diabetes", "line_number": 8, "usage_type": "call"}, {"api_name": "sklearn.datasets", "line_number": 8, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 20, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}]}
+{"seq_id": "132568366", "text": "\"\"\"\nDefinition of views.\n\"\"\"\n\nfrom app.models import Choice, Poll\nfrom datetime import datetime\nfrom django.contrib.auth.decorators import login_required\nfrom django.core.urlresolvers import reverse\nfrom django.http import HttpRequest, HttpResponseRedirect\nfrom django.shortcuts import get_object_or_404, render\nfrom django.template import RequestContext\nfrom django.utils import timezone\nfrom django.views.generic import ListView, DetailView\nfrom os import path\n\nimport json\n\nfrom .forms import ActorForm\n\nif form.is_valid():\n actor = form.cleaned_data['actor']\n\n\n\ndef get_actor(request):\n # if this is a POST request we need to process the form data\n if request.method == 'POST':\n # create a form instance and populate it with data from the request:\n form = ActorForm(request.POST)\n # check whether it's valid:\n if form.is_valid():\n # process the data in form.cleaned_data as required\n # ...\n # redirect to a new URL:\n return HttpResponseRedirect('/answer')\n\n # if a GET (or any other method) we'll create a blank form\n else:\n form = ActorForm()\n\n return render(request, 'answer.html', {'form': form})\n\nclass PollListView(ListView):\n \"\"\"Renders the home page, with a list of all polls.\"\"\"\n model = Poll\n\n def get_context_data(self, **kwargs):\n context = super(PollListView, self).get_context_data(**kwargs)\n context['title'] = 'Polls'\n context['year'] = datetime.now().year\n return context\n\nclass PollDetailView(DetailView):\n \"\"\"Renders the poll details page.\"\"\"\n model = Poll\n\n def get_context_data(self, **kwargs):\n context = super(PollDetailView, self).get_context_data(**kwargs)\n context['title'] = 'Poll'\n context['year'] = datetime.now().year\n return context\n\nclass PollResultsView(DetailView):\n \"\"\"Renders the results page.\"\"\"\n model = Poll\n\n def get_context_data(self, **kwargs):\n context = super(PollResultsView, self).get_context_data(**kwargs)\n context['title'] = 'Results'\n context['year'] = datetime.now().year\n return context\n\ndef contact(request):\n \"\"\"Renders the contact page.\"\"\"\n assert isinstance(request, HttpRequest)\n return render(\n request,\n 'app/contact.html',\n {\n 'title':'Contact',\n 'message':'Your contact page.',\n 'year':datetime.now().year,\n \n }\n )\n\ndef about(request):\n \"\"\"Renders the about page.\"\"\"\n assert isinstance(request, HttpRequest)\n return render(\n request,\n 'app/about.html',\n {\n 'title':'About',\n 'message':'Your application description page.',\n 'year':datetime.now().year,\n }\n )\n\ndef vote(request, poll_id):\n \"\"\"Handles voting. Validates input and updates the repository.\"\"\"\n poll = get_object_or_404(Poll, pk=poll_id)\n try:\n selected_choice = poll.choice_set.get(pk=request.POST['choice'])\n except (KeyError, Choice.DoesNotExist):\n return render(request, 'app/details.html', {\n 'title': 'Poll',\n 'year': datetime.now().year,\n 'poll': poll,\n 'error_message': \"Please make a selection.\",\n })\n else:\n selected_choice.votes += 1\n selected_choice.save()\n return HttpResponseRedirect(reverse('app:results', args=(poll.id,)))\n\n@login_required\ndef seed(request):\n \"\"\"Seeds the database with sample polls.\"\"\"\n samples_path = path.join(path.dirname(__file__), 'samples.json')\n with open(samples_path, 'r') as samples_file:\n samples_polls = json.load(samples_file)\n\n for sample_poll in samples_polls:\n poll = Poll()\n poll.text = sample_poll['text']\n poll.pub_date = timezone.now()\n poll.save()\n\n for sample_choice in sample_poll['choices']:\n choice = Choice()\n choice.poll = poll\n choice.text = sample_choice\n choice.votes = 0\n choice.save()\n\n return HttpResponseRedirect(reverse('app:home'))\n\ndef answer(request):\n \"\"\"Renders the answer page.\"\"\"\n assert isinstance(request, HttpRequest)\n return render(\n request,\n 'app/answer.html',\n {\n 'title':'Answer',\n 'message':'ANSWER:',\n 'year':datetime.now().year,\n }\n )", "sub_path": "KevinBaconFrontEnd/app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4392, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "forms.ActorForm", "line_number": 29, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 35, "usage_type": "call"}, {"api_name": "forms.ActorForm", "line_number": 39, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 41, "usage_type": "call"}, {"api_name": "django.views.generic.ListView", "line_number": 43, "usage_type": "name"}, {"api_name": "app.models.Poll", "line_number": 45, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "name"}, {"api_name": "django.views.generic.DetailView", "line_number": 53, "usage_type": "name"}, {"api_name": "app.models.Poll", "line_number": 55, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "name"}, {"api_name": "django.views.generic.DetailView", "line_number": 63, "usage_type": "name"}, {"api_name": "app.models.Poll", "line_number": 65, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 70, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 70, "usage_type": "name"}, {"api_name": "django.http.HttpRequest", "line_number": 75, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 76, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 82, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 82, "usage_type": "name"}, {"api_name": "django.http.HttpRequest", "line_number": 89, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 90, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 96, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 96, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 102, "usage_type": "call"}, {"api_name": "app.models.Poll", "line_number": 102, "usage_type": "argument"}, {"api_name": "app.models.Choice.DoesNotExist", "line_number": 105, "usage_type": "attribute"}, {"api_name": "app.models.Choice", "line_number": 105, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 106, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 108, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 108, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 115, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path", "line_number": 120, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 120, "usage_type": "call"}, {"api_name": "json.load", "line_number": 122, "usage_type": "call"}, {"api_name": "app.models.Poll", "line_number": 125, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 127, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 127, "usage_type": "name"}, {"api_name": "app.models.Choice", "line_number": 131, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 137, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 137, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 117, "usage_type": "name"}, {"api_name": "django.http.HttpRequest", "line_number": 141, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 142, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 148, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 148, "usage_type": "name"}]}
+{"seq_id": "560619993", "text": "# -*- coding: utf-8 -*-\n\n\"\"\"\nTest power plants and chp.\n\nSPDX-FileCopyrightText: 2016-2019 Uwe Krien \n\nSPDX-License-Identifier: MIT\n\"\"\"\n__copyright__ = \"Uwe Krien \"\n__license__ = \"MIT\"\n\nfrom nose.tools import eq_, assert_raises_regexp\nfrom deflex import basic_scenario, geometries, powerplants, config as cfg\n\n\nclass TestScenarioPowerplantsAndCHP:\n @classmethod\n def setUpClass(cls):\n cls.regions = geometries.deflex_regions(rmap=\"de21\")\n cls.pp = basic_scenario.scenario_powerplants(\n dict(), cls.regions, 2014, \"de21\", 1\n )\n\n def test_01_deflex_power_plants_by_year(self):\n pp = powerplants.get_deflex_pp_by_year(\n self.regions, 2014, \"de21\", overwrite_capacity=True\n )\n eq_(int(pp[\"capacity\"].sum()), 181489)\n\n def scenario_pp_test(self):\n eq_(float(self.pp[\"volatile_source\"][\"DE03\", \"wind\"]), 3052.8)\n eq_(\n float(self.pp[\"transformer\"].loc[\"capacity\", (\"DE03\", \"lignite\")]),\n 1135.6,\n )\n\n def test_scenario_transmission(self):\n lines = basic_scenario.scenario_transmission(\n self.pp, self.regions, \"de22\"\n )\n eq_(int(lines.loc[\"DE07-DE05\", (\"electrical\", \"capacity\")]), 1978)\n eq_(int(lines.loc[\"DE07-DE05\", (\"electrical\", \"distance\")]), 199)\n eq_(float(lines.loc[\"DE07-DE05\", (\"electrical\", \"efficiency\")]), 0.9)\n lines = basic_scenario.scenario_transmission(\n self.pp, self.regions, \"de22\", copperplate=True\n )\n eq_(\n float(lines.loc[\"DE07-DE05\", (\"electrical\", \"capacity\")]),\n float(\"inf\"),\n )\n eq_(str(lines.loc[\"DE07-DE05\", (\"electrical\", \"distance\")]), \"nan\")\n eq_(float(lines.loc[\"DE07-DE05\", (\"electrical\", \"efficiency\")]), 1.0)\n\n def test_scenario_transmisson_error(self):\n old_value = cfg.get(\"transmission\", \"general_efficiency\")\n cfg.tmp_set(\"transmission\", \"general_efficiency\", \"None\")\n msg = \"The calculation of the efficiency by distance is not yet\"\n with assert_raises_regexp(NotImplementedError, msg):\n basic_scenario.scenario_transmission(self.pp, self.regions, \"de22\")\n cfg.tmp_set(\"transmission\", \"general_efficiency\", old_value)\n\n def test_scenario_commodity_sources(self):\n src = basic_scenario.scenario_commodity_sources(self.pp, 2013)[\n \"commodity_source\"\n ]\n eq_(round(src.loc[\"costs\", (\"DE\", \"hard coal\")], 2), 9.71)\n eq_(round(src.loc[\"emission\", (\"DE\", \"natural gas\")], 2), 201.24)\n\n def test_chp(self):\n eq_(\n int(self.pp[\"transformer\"].loc[\"capacity\", (\"DE01\", \"hard coal\")]),\n 1291,\n )\n transf = basic_scenario.scenario_chp(\n self.pp, self.regions, 2014, \"de21\"\n )[\"transformer\"]\n eq_(int(transf.loc[\"capacity\", (\"DE01\", \"hard coal\")]), 623)\n eq_(int(transf.loc[\"capacity_elec_chp\", (\"DE01\", \"hard coal\")]), 667)\n", "sub_path": "tests/test_scenario_powerplant_and_chp.py", "file_name": "test_scenario_powerplant_and_chp.py", "file_ext": "py", "file_size_in_byte": 3005, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "deflex.geometries.deflex_regions", "line_number": 20, "usage_type": "call"}, {"api_name": "deflex.geometries", "line_number": 20, "usage_type": "name"}, {"api_name": "deflex.basic_scenario.scenario_powerplants", "line_number": 21, "usage_type": "call"}, {"api_name": "deflex.basic_scenario", "line_number": 21, "usage_type": "name"}, {"api_name": "deflex.powerplants.get_deflex_pp_by_year", "line_number": 26, "usage_type": "call"}, {"api_name": "deflex.powerplants", "line_number": 26, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 29, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 32, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 33, "usage_type": "call"}, {"api_name": "deflex.basic_scenario.scenario_transmission", "line_number": 39, "usage_type": "call"}, {"api_name": "deflex.basic_scenario", "line_number": 39, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 42, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 43, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 44, "usage_type": "call"}, {"api_name": "deflex.basic_scenario.scenario_transmission", "line_number": 45, "usage_type": "call"}, {"api_name": "deflex.basic_scenario", "line_number": 45, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 48, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 52, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 53, "usage_type": "call"}, {"api_name": "deflex.config.get", "line_number": 56, "usage_type": "call"}, {"api_name": "deflex.config", "line_number": 56, "usage_type": "name"}, {"api_name": "deflex.config.tmp_set", "line_number": 57, "usage_type": "call"}, {"api_name": "deflex.config", "line_number": 57, "usage_type": "name"}, {"api_name": "nose.tools.assert_raises_regexp", "line_number": 59, "usage_type": "call"}, {"api_name": "deflex.basic_scenario.scenario_transmission", "line_number": 60, "usage_type": "call"}, {"api_name": "deflex.basic_scenario", "line_number": 60, "usage_type": "name"}, {"api_name": "deflex.config.tmp_set", "line_number": 61, "usage_type": "call"}, {"api_name": "deflex.config", "line_number": 61, "usage_type": "name"}, {"api_name": "deflex.basic_scenario.scenario_commodity_sources", "line_number": 64, "usage_type": "call"}, {"api_name": "deflex.basic_scenario", "line_number": 64, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 67, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 68, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 71, "usage_type": "call"}, {"api_name": "deflex.basic_scenario.scenario_chp", "line_number": 75, "usage_type": "call"}, {"api_name": "deflex.basic_scenario", "line_number": 75, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 78, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 79, "usage_type": "call"}]}
+{"seq_id": "483153949", "text": "\"\"\"\nRRT* for simulations\nAuthor: Ellie Cho\nEditor: Yashwanth Nakka\n\"\"\"\n\n\nimport math\nimport os\nimport random\nimport sys\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom RRT import RRT\nfrom Node import Node\n\nshow_animation = True\nshow_graph = False\n\n\nclass RRTStar(RRT):\n \"\"\"\n Class for RRT Star planning\n \"\"\"\n def __init__(self, start, goal, obstacle_list, rand_area, rand_angle, vrange,\n expand_dis=3.0, path_resolution=0.5, goal_sample_rate=5, max_iter=500,\n connect_circle_dist=50.0\n ):\n \"\"\"\n Setting Parameter\n start:Start Position [x,y,0,0,0,0]\n goal:Goal Position [x,y,0,0,0,0]\n obstacleList:obstacle Positions [[x,y,size],...]\n randArea:Random Sampling Area [min,max]\n \"\"\"\n self.start = Node([start[0], start[1], 0, 0, 0, 0])\n self.end = Node([goal[0], goal[1], 0, 0, 0, 0])\n self.min_rand = rand_area[0]\n self.max_rand = rand_area[1]\n self.expand_dis = expand_dis\n self.path_resolution = path_resolution\n self.goal_sample_rate = goal_sample_rate\n self.max_iter = max_iter\n self.obstacle_list = obstacle_list\n self.node_list = []\n \n \n self.connect_circle_dist = connect_circle_dist\n self.goal_node = Node([goal[0], goal[1], 0, 0, 0, 0])\n\n def planning(self, animation=True, search_until_max_iter=True):\n \"\"\"\n rrt star path planning\n animation: flag for animation on or off\n search_until_max_iter: search until max iteration for path improving or not\n \"\"\"\n\n self.node_list = [self.start]\n for i in range(self.max_iter):\n print(\"Iter:\", i, \", number of nodes:\", len(self.node_list))\n rnd_node = self.get_random_node()\n nearest_ind = self.get_nearest_node_index(self.node_list, rnd_node)\n nearest_node = self.node_list[nearest_ind]\n new_node = self.steer(nearest_node, rnd_node, self.expand_dis)\n #print(\"x:\", new_node.x[0], \"y:\", new_node.x[1])\n if self.check_collision(new_node, self.obstacle_list):\n near_inds = self.find_near_nodes(new_node)\n new_node = self.choose_parent(new_node, near_inds)\n if new_node:\n self.node_list.append(new_node)\n self.rewire(new_node, near_inds)\n # print(\"x:\", new_node.x[0], \"y:\", new_node.x[1])\n\n if animation and i % 5 == 0:\n self.draw_graph1(rnd_node)\n\n if (not search_until_max_iter) and new_node: # check reaching the goal\n last_index = self.search_best_goal_node()\n if last_index:\n print (\"last\")\n return self.generate_final_course(last_index)\n\n\n print(\"reached max iteration\")\n\n\n last_index = self.search_best_goal_node()\n if last_index:\n print (\"last\")\n return self.generate_final_course(last_index)\n\n return None\n\n def choose_parent(self, new_node, near_inds):\n \"\"\"\n chooses the parent with minimum cost and returns the node\n \n \"\"\"\n\n\n if not near_inds:\n return None\n\n # search nearest cost in near_inds\n costs = []\n for i in near_inds:\n near_node = self.node_list[i]\n t_node = self.steer(near_node, new_node)\n if t_node and self.check_collision(t_node, self.obstacle_list):\n costs.append(self.calc_new_cost(near_node, new_node))\n else:\n costs.append(float(\"inf\")) # the cost of collision node\n min_cost = min(costs)\n # print(\"min cost:\", min_cost)\n\n if min_cost == float(\"inf\"):\n print(\"There is no good path.(min_cost is inf)\")\n return None\n\n min_ind = near_inds[costs.index(min_cost)]\n new_node = self.steer(self.node_list[min_ind], new_node)\n new_node.parent = self.node_list[min_ind]\n new_node.cost = min_cost\n\n return new_node\n def search_best_goal_node(self):\n \"\"\"\n searches for the goal node with the least cost and returns the index\n\n \"\"\"\n dist_to_goal_list = [self.calc_dist_to_goal(n.x[0], n.x[1]) for n in self.node_list]\n goal_inds = [dist_to_goal_list.index(i) for i in dist_to_goal_list if i <= self.expand_dis]\n\n safe_goal_inds = []\n for goal_ind in goal_inds:\n t_node = self.steer(self.node_list[goal_ind], self.goal_node)\n if self.check_collision(t_node, self.obstacle_list):\n safe_goal_inds.append(goal_ind)\n\n if not safe_goal_inds:\n return None\n\n min_cost = min([self.node_list[i].cost for i in safe_goal_inds])\n for i in safe_goal_inds:\n if self.node_list[i].cost == min_cost:\n return i\n\n\n return None\n\n def find_near_nodes(self, new_node):\n \"\"\"\n finds nodes near the new node\n \n \"\"\"\n \n nnode = len(self.node_list) + 1\n r = self.connect_circle_dist * math.sqrt((math.log(nnode) / nnode))\n # if expand_dist exists, search vertices in a range no more than expand_dist\n if hasattr(self, 'expand_dis'): \n r = min(r, self.expand_dis)\n dist_list = [(node.x[0] - new_node.x[0]) ** 2 +\n (node.x[1] - new_node.x[1]) ** 2 +\n (node.x[2] - new_node.x[2]) ** 2 for node in self.node_list]\n near_inds = [dist_list.index(i) for i in dist_list if i <= r ** 2]\n return near_inds\n\n\n def rewire(self, new_node, near_inds):\n \n \n for i in near_inds:\n near_node = self.node_list[i]\n edge_node = self.steer(new_node, near_node)\n if not edge_node:\n continue\n edge_node.cost = self.calc_new_cost(new_node, near_node)\n\n no_collision = self.check_collision(edge_node, self.obstacle_list)\n improved_cost = near_node.cost > edge_node.cost\n\n if no_collision and improved_cost:\n self.node_list[i] = edge_node\n self.propagate_cost_to_leaves(new_node)\n\n def calc_new_cost(self, from_node, to_node):\n \"\"\"\n calculates cost from node to node\n \n \"\"\"\n\n d, _ = self.calc_distance_and_angle(from_node, to_node)\n # print(\"d:\", d)\n return (from_node.cost + d)\n\n def propagate_cost_to_leaves(self, parent_node):\n \"\"\"\n gives cost to nodes in tree\n \n \"\"\"\n \n for node in self.node_list:\n if node.parent == parent_node:\n node.cost = self.calc_new_cost(parent_node, node)\n self.propagate_cost_to_leaves(node)\n\n\ndef main(gx=8.0, gy=4.0):\n print(\"Start \" + __file__)\n\n # ====Search Path with RRT====\n obstacleList = [\n (2, 0, 2),\n (7, 0, 2),\n (4, 5, 2)\n ] # [x, y, radius]\n\n # Set Initial parameters\n rrt_star = RRTStar(start=[0, 2],\n goal=[gx, gy],\n rand_area=[-2, 15],\n rand_angle=[-3, 3],\n vrange=[0, 5, -.5, .5],\n obstacle_list=obstacleList)\n path = rrt_star.planning(animation=show_animation)\n\n if path is None:\n print(\"Cannot find path\")\n else:\n print(\"found path!!\")\n\n # Draw final path\n if show_animation:\n rrt_star.draw_graph1()\n plt.plot([x for (x, y) in path], [y for (x, y) in path], '-r')\n plt.grid(True)\n plt.pause(0.01) # Need for Mac\n plt.show()\n if show_graph:\n rrt_star.draw_graph2()\n plt.show()\n\n\nif __name__ == '__main__':\n main()\n \n\n\n\n\n\n", "sub_path": "src/mstar_guidance/src/rrt_mstar/RRT_Star.py", "file_name": "RRT_Star.py", "file_ext": "py", "file_size_in_byte": 7808, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "RRT.RRT", "line_number": 21, "usage_type": "name"}, {"api_name": "Node.Node", "line_number": 36, "usage_type": "call"}, {"api_name": "Node.Node", "line_number": 37, "usage_type": "call"}, {"api_name": "Node.Node", "line_number": 49, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 158, "usage_type": "call"}, {"api_name": "math.log", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 235, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 235, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 236, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 237, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 238, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 238, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 241, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 241, "usage_type": "name"}]}
+{"seq_id": "156202646", "text": "# -*- coding: utf-8 -*-\nfrom io import BytesIO\nfrom os import SEEK_SET\nfrom os.path import splitext\nfrom typing import Any, Dict, IO, List\nfrom unicodedata import normalize\n\nimport pandas as pd\nimport platiagro\nfrom chardet.universaldetector import UniversalDetector\nfrom pandas.io.common import infer_compression\nfrom platiagro import save_dataset, stat_dataset\nfrom platiagro.featuretypes import infer_featuretypes, validate_featuretypes\nfrom werkzeug.exceptions import BadRequest, NotFound\n\n\ndef list_datasets() -> List[Dict[str, Any]]:\n \"\"\"Lists all datasets from our object storage.\n\n Returns:\n A list of all datasets.\n \"\"\"\n datasets = platiagro.list_datasets()\n return [get_dataset(name) for name in datasets]\n\n\ndef create_dataset(files: Dict[str, IO]) -> Dict[str, Any]:\n \"\"\"Creates a new dataset in our object storage.\n\n Args:\n files (dict): file objects.\n\n Returns:\n The dataset details: name, columns, and filename.\n\n Raises:\n BadRequest: when incoming files are missing or valid.\n \"\"\"\n # checks if the post request has the file part\n if \"file\" not in files:\n raise BadRequest(\"No file part\")\n file = files[\"file\"]\n\n # if user does not select file, the browser also\n # submits an empty part without filename\n if file.filename == \"\":\n raise BadRequest(\"No selected file\")\n\n # generate a dataset name from filename\n name = generate_name(file.filename)\n\n try:\n # reads file into a DataFrame\n df = read_into_dataframe(file, file.filename)\n except Exception as e:\n # if read fails, then uploads raw file\n save_dataset(name, file)\n return {\"name\": name, \"filename\": file.filename}\n\n columns = df.columns.values.tolist()\n\n # checks if the post request has the 'featuretypes' part\n if \"featuretypes\" in files:\n try:\n ftype_file = files[\"featuretypes\"]\n featuretypes = list(map(lambda s: s.strip().decode(\"utf8\"), ftype_file.readlines()))\n validate_featuretypes(featuretypes)\n except ValueError as e:\n raise BadRequest(str(e))\n\n if len(columns) != len(featuretypes):\n raise BadRequest(\"featuretypes must be the same length as the DataFrame columns\")\n else:\n featuretypes = infer_featuretypes(df)\n\n metadata = {\n \"featuretypes\": featuretypes,\n \"original-filename\": file.filename,\n }\n\n # uses PlatIAgro SDK to save the dataset\n save_dataset(name, df, metadata=metadata)\n\n columns = [{\"name\": col, \"featuretype\": ftype} for col, ftype in zip(columns, featuretypes)]\n return {\"name\": name, \"columns\": columns, \"filename\": file.filename}\n\n\ndef get_dataset(name: str) -> Dict[str, Any]:\n \"\"\"Details a dataset from our object storage.\n\n Args:\n name (str): the dataset name to look for in our object storage.\n\n Returns:\n The dataset details: name, columns, and filename.\n\n Raises:\n NotFound: when the dataset does not exist.\n \"\"\"\n try:\n metadata = stat_dataset(name)\n\n filename = metadata.get(\"original-filename\")\n\n if \"columns\" in metadata and \"featuretypes\" in metadata:\n columns = metadata[\"columns\"]\n featuretypes = metadata[\"featuretypes\"]\n columns = [{\"name\": col, \"featuretype\": ftype} for col, ftype in zip(columns, featuretypes)]\n return {\"name\": name, \"columns\": columns, \"filename\": filename}\n\n return {\"name\": name, \"filename\": filename}\n except FileNotFoundError:\n raise NotFound(\"The specified dataset does not exist\")\n\n\ndef read_into_dataframe(file: IO, filename: str = \"\", nrows: int = 100,max_characters: int = 50) -> pd.DataFrame:\n \"\"\"Reads a file into a DataFrame.\n Infers the file encoding and whether a header column exists\n Args:\n file (IO): file buffer.\n filename (str): filename. Used to infer compression.\n nrows (int, optional): number of rows to peek. Default: 100.\n max_characters (int, optional): max characters a column name can have to be distinguished from a real text value\n Returns:\n A pandas.DataFrame.\n \"\"\"\n detector = UniversalDetector()\n for line, text in enumerate(file):\n detector.feed(text)\n if detector.done or line > nrows:\n break\n detector.close()\n encoding = detector.result.get(\"encoding\")\n\n compression = infer_compression(filename, \"infer\")\n\n file.seek(0, SEEK_SET)\n contents = file.read()\n\n with BytesIO(contents) as file:\n df0 = pd.read_csv(\n file,\n encoding=encoding,\n compression=compression,\n sep=None,\n engine=\"python\",\n header=\"infer\",\n nrows=nrows,\n )\n \n df0_cols = list(df0.columns)\n \n #Check if all columns are strins and short strings(text values tend to be long)\n column_names_checker = all([type(item) == str for item in df0_cols])\n if column_names_checker:\n column_names_checker = all([len(item) < max_characters for item in df0_cols]) \n \n \n #Check if any column can be turned to float\n conversion_checker= True\n for item in df0_cols:\n try:\n item = float(item)\n conversion_checker = False\n break\n except:\n pass\n \n\n #Prefix and header \n final_checker = True if (column_names_checker and conversion_checker) else False\n header = \"infer\" if final_checker else None\n prefix = None if header else \"col\"\n\n with BytesIO(contents) as file:\n df = pd.read_csv(\n file,\n encoding=encoding,\n compression=compression,\n sep=None,\n engine=\"python\",\n header=header,\n prefix=prefix,\n )\n return df\n\n\n\n\n\ndef generate_name(filename: str, attempt: int = 1) -> str:\n \"\"\"Generates a dataset name from a given filename.\n\n Args:\n filename (str): source filename.\n attempt (int): the current attempt of generating a new name.\n\n Return:\n str: new generated dataset name.\n \"\"\"\n # normalize filename to ASCII characters\n # replace spaces by dashes\n name = normalize('NFKD', filename) \\\n .encode('ASCII', 'ignore') \\\n .replace(b' ', b'-') \\\n .decode()\n\n if attempt > 1:\n # adds a suffix '-NUMBER' to filename\n name, extension = splitext(name)\n name = f\"{name}-{attempt}{extension}\"\n\n try:\n # check if final_name is already in use\n stat_dataset(name)\n except FileNotFoundError:\n return name\n\n # if it is already in use,\n return generate_name(filename, attempt + 1)\n", "sub_path": "datasets/datasets.py", "file_name": "datasets.py", "file_ext": "py", "file_size_in_byte": 6701, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "platiagro.list_datasets", "line_number": 23, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 27, "usage_type": "name"}, {"api_name": "typing.IO", "line_number": 27, "usage_type": "name"}, {"api_name": "werkzeug.exceptions.BadRequest", "line_number": 41, "usage_type": "call"}, {"api_name": "werkzeug.exceptions.BadRequest", "line_number": 47, "usage_type": "call"}, {"api_name": "platiagro.save_dataset", "line_number": 57, "usage_type": "call"}, {"api_name": "platiagro.featuretypes.validate_featuretypes", "line_number": 67, "usage_type": "call"}, {"api_name": "werkzeug.exceptions.BadRequest", "line_number": 69, "usage_type": "call"}, {"api_name": "werkzeug.exceptions.BadRequest", "line_number": 72, "usage_type": "call"}, {"api_name": "platiagro.featuretypes.infer_featuretypes", "line_number": 74, "usage_type": "call"}, {"api_name": "platiagro.save_dataset", "line_number": 82, "usage_type": "call"}, {"api_name": "typing.Any", "line_number": 27, "usage_type": "name"}, {"api_name": "platiagro.stat_dataset", "line_number": 101, "usage_type": "call"}, {"api_name": "werkzeug.exceptions.NotFound", "line_number": 113, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 88, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 88, "usage_type": "name"}, {"api_name": "typing.IO", "line_number": 116, "usage_type": "name"}, {"api_name": "chardet.universaldetector.UniversalDetector", "line_number": 127, "usage_type": "call"}, {"api_name": "pandas.io.common.infer_compression", "line_number": 135, "usage_type": "call"}, {"api_name": "os.SEEK_SET", "line_number": 137, "usage_type": "argument"}, {"api_name": "io.BytesIO", "line_number": 140, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 141, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 175, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 176, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 116, "usage_type": "attribute"}, {"api_name": "unicodedata.normalize", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 210, "usage_type": "call"}, {"api_name": "platiagro.stat_dataset", "line_number": 215, "usage_type": "call"}]}
+{"seq_id": "585140772", "text": "import argparse\nimport RaceRandom as random\nimport os\nfrom pathlib import Path\n\nimport urllib.request\nimport urllib.parse\nimport yaml\n\n\ndef get_weights(path):\n if os.path.exists(Path(path)):\n with open(path, \"r\", encoding=\"utf-8\") as f:\n return yaml.load(f, Loader=yaml.SafeLoader)\n elif urllib.parse.urlparse(path).scheme in ['http', 'https']:\n return yaml.load(urllib.request.urlopen(path), Loader=yaml.FullLoader)\n\ndef roll_settings(weights):\n def get_choice(option, root=None):\n root = weights if root is None else root\n if option not in root:\n return None\n if type(root[option]) is not dict:\n return root[option]\n if not root[option]:\n return None\n return random.choices(list(root[option].keys()), weights=list(map(int, root[option].values())))[0]\n\n def get_choice_default(option, root=weights, default=None):\n choice = get_choice(option, root)\n if choice is None and default is not None:\n return default\n return choice\n\n while True:\n subweights = weights.get('subweights', {})\n if len(subweights) == 0:\n break\n chances = ({k: int(v['chance']) for (k, v) in subweights.items()})\n subweight_name = random.choices(list(chances.keys()), weights=list(chances.values()))[0]\n subweights = weights.get('subweights', {}).get(subweight_name, {}).get('weights', {})\n subweights['subweights'] = subweights.get('subweights', {})\n weights = {**weights, **subweights}\n\n ret = argparse.Namespace()\n\n ret.algorithm = get_choice('algorithm')\n\n glitch_map = {'none': 'noglitches', 'no_logic': 'nologic', 'owglitches': 'owglitches',\n 'owg': 'owglitches', 'minorglitches': 'minorglitches'}\n glitches_required = get_choice('glitches_required')\n if glitches_required is not None:\n if glitches_required not in glitch_map.keys():\n print(f'Logic did not match one of: {\", \".join(glitch_map.keys())}')\n glitches_required = 'none'\n ret.logic = glitch_map[glitches_required]\n\n # item_placement = get_choice('item_placement')\n # not supported in ER\n\n dungeon_items = get_choice('dungeon_items')\n dungeon_items = '' if dungeon_items == 'standard' or dungeon_items is None else dungeon_items\n dungeon_items = 'mcsb' if dungeon_items == 'full' else dungeon_items\n ret.mapshuffle = get_choice('map_shuffle') == 'on' if 'map_shuffle' in weights else 'm' in dungeon_items\n ret.compassshuffle = get_choice('compass_shuffle') == 'on' if 'compass_shuffle' in weights else 'c' in dungeon_items\n if 'smallkey_shuffle' in weights:\n ret.keyshuffle = get_choice('smallkey_shuffle')\n else:\n if 's' in dungeon_items:\n ret.keyshuffle = 'wild'\n if 'u' in dungeon_items:\n ret.keyshuffle = 'universal'\n ret.bigkeyshuffle = get_choice('bigkey_shuffle') == 'on' if 'bigkey_shuffle' in weights else 'b' in dungeon_items\n\n ret.accessibility = get_choice('accessibility')\n ret.restrict_boss_items = get_choice('restrict_boss_items')\n\n overworld_shuffle = get_choice('overworld_shuffle')\n ret.ow_shuffle = overworld_shuffle if overworld_shuffle != 'none' else 'vanilla'\n ret.ow_terrain = get_choice('overworld_terrain') == 'on'\n valid_options = {'none', 'polar', 'grouped', 'limited', 'chaos'}\n ret.ow_crossed = get_choice('overworld_crossed')\n ret.ow_crossed = ret.ow_crossed if ret.ow_crossed in valid_options else 'none'\n ret.ow_keepsimilar = get_choice('overworld_keepsimilar') == 'on'\n ret.ow_mixed = get_choice('overworld_swap') == 'on'\n ret.ow_whirlpool = get_choice('whirlpool_shuffle') == 'on'\n overworld_flute = get_choice('flute_shuffle')\n ret.ow_fluteshuffle = overworld_flute if overworld_flute != 'none' else 'vanilla'\n ret.bonk_drops = get_choice('bonk_drops') == 'on'\n entrance_shuffle = get_choice('entrance_shuffle')\n ret.shuffle = entrance_shuffle if entrance_shuffle != 'none' else 'vanilla'\n overworld_map = get_choice('overworld_map')\n ret.overworld_map = overworld_map if overworld_map != 'default' else 'default'\n door_shuffle = get_choice('door_shuffle')\n ret.door_shuffle = door_shuffle if door_shuffle != 'none' else 'vanilla'\n ret.intensity = get_choice('intensity')\n ret.door_type_mode = get_choice('door_type_mode')\n ret.trap_door_mode = get_choice('trap_door_mode')\n ret.key_logic_algorithm = get_choice('key_logic_algorithm')\n ret.decoupledoors = get_choice('decoupledoors') == 'on'\n ret.door_self_loops = get_choice('door_self_loops') == 'on'\n ret.experimental = get_choice('experimental') == 'on'\n ret.collection_rate = get_choice('collection_rate') == 'on'\n\n ret.dungeon_counters = get_choice('dungeon_counters') if 'dungeon_counters' in weights else 'default'\n if ret.dungeon_counters == 'default':\n ret.dungeon_counters = 'pickup' if ret.door_shuffle != 'vanilla' or ret.compassshuffle == 'on' else 'off'\n\n ret.pseudoboots = get_choice('pseudoboots') == 'on'\n ret.shopsanity = get_choice('shopsanity') == 'on'\n keydropshuffle = get_choice('keydropshuffle') == 'on'\n ret.dropshuffle = get_choice('dropshuffle') == 'on' or keydropshuffle\n ret.pottery = get_choice('pottery') if 'pottery' in weights else 'none'\n ret.pottery = 'keys' if ret.pottery == 'none' and keydropshuffle else ret.pottery\n ret.colorizepots = get_choice_default('colorizepots', default='on') == 'on'\n ret.shufflepots = get_choice('pot_shuffle') == 'on'\n ret.mixed_travel = get_choice('mixed_travel') if 'mixed_travel' in weights else 'prevent'\n ret.standardize_palettes = (get_choice('standardize_palettes') if 'standardize_palettes' in weights\n else 'standardize')\n\n goal = get_choice('goals')\n if goal is not None:\n ret.goal = {'ganon': 'ganon',\n 'fast_ganon': 'crystals',\n 'dungeons': 'dungeons',\n 'pedestal': 'pedestal',\n 'triforce-hunt': 'triforcehunt',\n 'trinity': 'trinity',\n 'ganonhunt': 'ganonhunt',\n 'completionist': 'completionist'\n }[goal]\n\n ret.openpyramid = get_choice('open_pyramid') if 'open_pyramid' in weights else 'auto'\n\n ret.shuffleganon = get_choice('shuffleganon') == 'on'\n ret.shufflelinks = get_choice('shufflelinks') == 'on'\n ret.shuffletavern = get_choice('shuffletavern') == 'on'\n \n ret.crystals_gt = get_choice('tower_open')\n ret.crystals_ganon = get_choice('ganon_open')\n\n ret.triforce_pool = get_choice_default('triforce_pool', default=0)\n ret.triforce_goal = get_choice_default('triforce_goal', default=0)\n ret.triforce_pool_min = get_choice_default('triforce_pool_min', default=0)\n ret.triforce_pool_max = get_choice_default('triforce_pool_max', default=0)\n ret.triforce_goal_min = get_choice_default('triforce_goal_min', default=0)\n ret.triforce_goal_max = get_choice_default('triforce_goal_max', default=0)\n ret.triforce_min_difference = get_choice_default('triforce_min_difference', default=0)\n ret.triforce_max_difference = get_choice_default('triforce_max_difference', default=10000)\n\n ret.mode = get_choice('world_state')\n if ret.mode == 'retro':\n ret.mode = 'open'\n ret.retro = True\n ret.retro = get_choice('retro') == 'on' # this overrides world_state if used\n ret.take_any = get_choice_default('take_any', default='none')\n\n ret.bombbag = get_choice('bombbag') == 'on'\n\n ret.hints = get_choice('hints') == 'on'\n\n swords = get_choice('weapons')\n if swords is not None:\n ret.swords = {'randomized': 'random',\n 'assured': 'assured',\n 'vanilla': 'vanilla',\n 'swordless': 'swordless'\n }[swords]\n\n ret.difficulty = get_choice('item_pool')\n ret.flute_mode = get_choice_default('flute_mode', default='normal')\n ret.bow_mode = get_choice_default('bow_mode', default='progressive')\n\n ret.item_functionality = get_choice('item_functionality')\n\n old_style_bosses = {'basic': 'simple',\n 'normal': 'full',\n 'chaos': 'random'}\n boss_choice = get_choice('boss_shuffle')\n if boss_choice in old_style_bosses.keys():\n boss_choice = old_style_bosses[boss_choice]\n ret.shufflebosses = boss_choice\n\n enemy_choice = get_choice('enemy_shuffle')\n if enemy_choice == 'chaos':\n enemy_choice = 'random'\n ret.shuffleenemies = enemy_choice\n\n old_style_damage = {'none': 'default',\n 'chaos': 'random'}\n damage_choice = get_choice('enemy_damage')\n if damage_choice in old_style_damage:\n damage_choice = old_style_damage[damage_choice]\n ret.enemy_damage = damage_choice\n\n ret.enemy_health = get_choice('enemy_health')\n\n ret.beemizer = get_choice('beemizer') if 'beemizer' in weights else '0'\n\n inventoryweights = weights.get('startinventory', {})\n startitems = []\n for item in inventoryweights.keys():\n if get_choice(item, inventoryweights) == 'on':\n startitems.append(item)\n ret.startinventory = ','.join(startitems)\n if len(startitems) > 0:\n ret.usestartinventory = True\n\n if 'rom' in weights:\n romweights = weights['rom']\n ret.sprite = get_choice('sprite', romweights)\n ret.disablemusic = get_choice('disablemusic', romweights) == 'on'\n ret.quickswap = get_choice('quickswap', romweights) == 'on'\n ret.reduce_flashing = get_choice('reduce_flashing', romweights) == 'on'\n ret.msu_resume = get_choice('msu_resume', romweights) == 'on'\n ret.fastmenu = get_choice('menuspeed', romweights)\n ret.heartcolor = get_choice('heartcolor', romweights)\n ret.heartbeep = get_choice('heartbeep', romweights)\n ret.ow_palettes = get_choice('ow_palettes', romweights)\n ret.uw_palettes = get_choice('uw_palettes', romweights)\n ret.shuffle_sfx = get_choice('shuffle_sfx', romweights) == 'on'\n ret.msu_resume = get_choice('msu_resume', romweights) == 'on'\n\n return ret\n", "sub_path": "source/tools/MysteryUtils.py", "file_name": "MysteryUtils.py", "file_ext": "py", "file_size_in_byte": 10232, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "os.path.exists", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 12, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 14, "usage_type": "call"}, {"api_name": "yaml.SafeLoader", "line_number": 14, "usage_type": "attribute"}, {"api_name": "urllib.request.parse.urlparse", "line_number": 15, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 15, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 15, "usage_type": "name"}, {"api_name": "yaml.load", "line_number": 16, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 16, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 16, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 16, "usage_type": "name"}, {"api_name": "yaml.FullLoader", "line_number": 16, "usage_type": "attribute"}, {"api_name": "RaceRandom.choices", "line_number": 27, "usage_type": "call"}, {"api_name": "RaceRandom.choices", "line_number": 40, "usage_type": "call"}, {"api_name": "argparse.Namespace", "line_number": 45, "usage_type": "call"}]}
+{"seq_id": "238803972", "text": "import numpy as np\nimport time\nimport datetime\nimport pandas as pd\nimport pdb\nfrom tqdm import tqdm\n\nclass OuterRadiation:\n\n def __init__(self,\n # ex) around Tokyo\n latitude = np.array([35.695, 35.705, 35.715]), \n longitude = np.array([139.725, 139.735, 139.745]), \n date = np.array([2016,1,1,12,0])\n # [year, month, day, hoor, second]\n ):\n\n self.rownames = list(map(str, latitude[::-1]))\n self.colnames = list(map(str, longitude))\n self.dp = pd.Panel(major_axis = self.rownames, minor_axis = self.colnames)\n \n self.latitude = np.radians(latitude)\n self.longitude = np.radians(longitude)\n\n \n\n self.y, self.m, self.d, self.h, self.s = date\n\n days = datetime.datetime(self.y, self.m, self.d) - \\\n datetime.datetime(self.y, 1, 1)\n self.dn = days.days + 1\n\n self.theta = 2 * np.pi * (self.dn - 1)/365\n\n # sun declination\n self.delta = 0.006918 - 0.399912 * np.cos(self.theta) \\\n + 0.070257 * np.sin(self.theta) \\\n - 0.006758 * np.cos(2 * self.theta) \\\n + 0.000907 * np.sin(2 * self.theta) \\\n - 0.002697 * np.cos(3 * self.theta) \\\n + 0.001480 * np.sin(3 * self.theta)\n\n # Geocentric solar distance\n self.r = 1 / np.sqrt(1.000110 + 0.034221 * np.cos(self.theta) \\\n + 0.001280 * np.sin(self.theta) \\\n + 0.000719 * np.cos(2 * self.theta) \\\n + 0.000077 * np.sin(2 * self.theta))\n\n # Uniform time difference\n \n self.Eq = - 0.0002786049 + 0.1227715 * np.cos(self.theta + 1.498311) \\\n - 0.1654575 * np.cos(2 * self.theta - 1.261546) \\\n - 0.0053538 * np.cos(3 * self.theta - 1.1571)\n\n self.StandardLatitude = np.radians(135.0)\n\n # tenmoral solar angle\n self.angle = None\n\n # sun orientation\n self.psi = None\n # sun altitude\n self.alpha = None\n\n # radiation out of air\n self.Q = None\n\n def compute(self, interval = 2.5, number = 10, save = False):\n\n yname = str(self.y)\n if(self.m < 10):\n mname = '0' + str(self.m)\n elif():\n mname = str(self.m)\n if(self.d < 10):\n dname = '0' + str(self.d)\n elif():\n dname = str(self.d)\n \n\n # for each time\n for i in tqdm(range(number)):\n\n df = pd.DataFrame(index = self.latitude[::-1])\n\n interval_sum = interval * i\n \n h = self.h + (self.s + interval_sum)//60\n if(h < 10):\n hname = '0' + str(h)\n elif():\n hname = str(h)\n\n s = (10 * interval_sum//10)%60\n if(s < 10):\n sname = '0' + str(s)\n elif():\n sname = str(s)\n\n a = 1 + i%4\n aname = '0' + str(a)\n \n # for each longtitude\n for lon in self.longitude:\n \n self.angle = (self.h + (self.s + interval * i)/60 - 12) \\\n * np.pi / 12 \\\n + (lon - self.StandardLatitude) \\\n + self.Eq * np.pi / 12\n\n Qseries = []\n # for each latitude\n for lat in self.latitude:\n \n self.alpha = np.arcsin(np.sin(lat) * np.sin(self.delta) \\\n + np.cos(lat) \\\n * np.cos(self.delta) \\\n * np.cos(self.angle))\n self.psi = np.arctan(np.cos(lat) * np.cos(self.delta) \\\n * np.sin(self.angle) \\\n / (np.sin(lat) * np.sin(self.alpha))\\\n - np.sin(self.delta))\n self.Q = 1367 * self.r**2 * np.sin(self.alpha)\n\n Qseries.append(self.Q)\n \n Qseries = Qseries[::-1]\n\n df[str(np.rad2deg(lon))] = Qseries\n\n index = [str(np.rad2deg(l)) for l in self.latitude[::-1]]\n df.index = index\n # pdb.set_trace()\n\n self.dp[i] = df\n\n return self.dp\n", "sub_path": "datautil/OuterRadiation.py", "file_name": "OuterRadiation.py", "file_ext": "py", "file_size_in_byte": 4503, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "83", "api": [{"api_name": "numpy.array", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.Panel", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 33, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 55, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 82, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 107, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 109, "usage_type": "attribute"}, {"api_name": "numpy.arcsin", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.arctan", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.rad2deg", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.rad2deg", "line_number": 131, "usage_type": "call"}]}