diff --git "a/6244.jsonl" "b/6244.jsonl"
new file mode 100644--- /dev/null
+++ "b/6244.jsonl"
@@ -0,0 +1,661 @@
+{"seq_id":"219957056","text":"from bs4 import BeautifulSoup\nimport ctypes\nimport urllib.request\nimport urllib\n\nimport sys\n\nfrom collections import Counter\nimport random\nimport webbrowser\nfrom konlpy.tag import Hannanum\nfrom lxml import html\nimport pytagcloud\nimport sys\n\n\nif sys.version_info[0] >= 3:\n urlopen = urllib.request.urlopen\nelse:\n urlopen = urllib.urlopen\n\nr = lambda: random.randint(0,255)\ncolor = lambda: (r(),r(),r())\n\n\ndef get_tag(text, ntags=50, multiplier=10):\n h = Hannanum()\n nouns = h.nouns(text)\n count = Counter(nouns)\n return [{ 'color' : color(), 'tag' : n, 'size': c*multiplier} for n,c in count.most_common(ntags)]\n\n\ndef draw_cloud(tags, filename, fontname='Noto Sans CJK', size=(800,600)):\n pytagcloud.create_tag_image(tags, filename, fontname=fontname, size=size)\n webbrowser.open(filename)\n\nindex = 0\nmy_list = []\ns = \"\"\ninit_number = 63844\nwhile True:\n\tif index == 20:\n\t\tbreak\n\tno = init_number+index\n\tpage_url = 'http://www.thisisgame.com/webzine/game/nboard/16/?n='+str(no)\n\turl_open = urllib.request.urlopen(page_url)\n\tsoup = BeautifulSoup(url_open, 'html.parser', from_encoding = 'utf-8')\n\tp_list = soup.select('.content-line span')\n\n\tfor i in range(0, len(p_list)):\n \t\tmy_list.append(p_list[i].text)\n\tindex = index+1\nfor i in range(0, len(my_list)):\n s += my_list[i]\ntags = get_tag(s)\nprint(tags)\ndraw_cloud(tags, 'thisisgame.png') \n","sub_path":"crawling/thisisgmae.py","file_name":"thisisgmae.py","file_ext":"py","file_size_in_byte":1370,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"140541967","text":"import random\n\nimport pygame\n\nfrom menu.Item import Item\nfrom menu.Menu import Menu\nfrom playobjects.Block import Block\nfrom playobjects.Player import Player\n\nwindow = pygame.display.set_mode((800, 400))\npygame.display.set_caption(\"RUN!\")\nicon = pygame.image.load(\"img/block.png\")\npygame.display.set_icon(icon)\n\nscreen = pygame.Surface((800, 400))\n\nx = 0\ny = 0\nsquare_go_right = True\nsquare_go_down = True\n\nw, h = pygame.display.get_surface().get_size()\n\nmenu = Menu(w, h, [])\nmenu.menu(window)\n\nplayer = Player(0, 360)\nblock = Block(800, 360)\npygame.font.init()\nfont = pygame.font.Font(\"fonts/Mono.ttf\", 32)\n\ndone = True\npygame.key.set_repeat(1, 1)\ntime_delay = pygame.time.Clock()\ndt = 0\n\nwhile done:\n for e in pygame.event.get():\n if e.type == pygame.QUIT:\n done = False\n if e.type == pygame.KEYDOWN:\n if e.key == pygame.K_SPACE:\n player.jump()\n screen.fill((255, 255, 255))\n\n # render\n player.render(screen)\n f = font.render(\"points %d\" % player.points, 1, (0, 0, 0))\n screen.blit(f, (0, 0))\n if block is not None:\n block.render(screen)\n window.blit(screen, (0, 0))\n\n # Relation logic\n if block.visible:\n if player.check_collision(block):\n menu = Menu(w, h, [Item(\"Points = %d\" % player.points, w, h / 2)])\n menu.menu(window)\n block = Block(800, 360)\n player = Player(0, 360)\n if block.x < player.x and player.y + player.h > block.y:\n player.y = min(360, block.y)\n\n player.move(dt)\n\n # Block logic\n\n if block.visible:\n block.move()\n if block.x < -50:\n block.y = 400\n block.visible = False\n block.real_speed = 0\n player.points += 1\n if player.points % 5 == 0:\n block.increase_speed()\n\n if not block.visible:\n if random.randint(0, 1000) > 990:\n block.reset()\n\n pygame.display.flip()\n dt = time_delay.tick(200)\n","sub_path":"Main.py","file_name":"Main.py","file_ext":"py","file_size_in_byte":1989,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"631054728","text":"import cv2\nimport numpy as np\nimport random\nimport matplotlib.pyplot as plt\nimport os\nfrom tqdm import tqdm\nfrom initializers import *\nfrom auto_encoder import Autoencoder\nimport pandas as pd\n\nrandom.seed(1447)\nnp.random.seed(1447)\n\n\ndef random_input(path, size=256):\n image_paths = [os.path.join(path, img_path) for img_path in os.listdir(path)]\n # img_path = image_paths[0] # beee :)\n img_path = random.choice(image_paths)\n print(\"Loading image: {}\".format(img_path))\n img = cv2.imread(img_path, 0)\n img = cv2.resize(img, (size, size))\n img = 2 * img / 255 - 1\n img = limit_values(img)\n return img.copy()\n\n\ndef loss_sqe(x, y):\n return (x - y) ** 2\n\n\nDATA_TRAIN_DIR = './data/train/'\nDATA_TRAIN_DIR = './data/test/'\nDATA_TEST_DIR = './data/test/'\nRESULTS_DIR = './results/'\n\nMOMENTUM = 0.95\nTRAIN_ITERS = 1000\n\n# 1e-4 for bee lr\n\n\ndef train(min_error=1000., max_iter_count=1000, data_dir=DATA_TRAIN_DIR, learn_rate=0.0001, print_w=False,\n use_norm=True, loss=loss_sqe, mid_layers=64, use_adapt_lr=True):\n global TRAIN_ITERS\n model = Autoencoder(use_norm=use_norm, initializer=glorot_uniform, loss=loss, mid_layers=mid_layers, lr=learn_rate,\n use_adapt_lr=use_adapt_lr)\n train_sample = random_input(data_dir)\n errors = []\n best_errors = []\n moving_average = None\n best_error = np.inf\n best_weights = None\n is_good = False\n try:\n for _ in tqdm(range(max_iter_count)):\n if is_good:\n continue\n error = model(train_sample)\n moving_average = MOMENTUM * moving_average + error * (1. - MOMENTUM) if moving_average else error\n errors.append(moving_average)\n if best_errors:\n best_errors.append([_, min(moving_average, best_errors[-1][1])])\n else:\n best_errors.append([_, moving_average])\n if moving_average < best_error:\n print(moving_average)\n best_error = moving_average\n best_weights = model.get_weights()\n\n if best_error < min_error and not is_good:\n TRAIN_ITERS = _\n is_good = True\n except RuntimeError as e:\n print(e)\n finally:\n print(\"Number of iteration to get e={}: {}\".format(min_error, TRAIN_ITERS))\n x = np.arange(len(errors))\n plt.plot(x, np.array(errors))\n idx = np.argmin(errors)\n print(\"BEST ERROR {}\".format(errors[idx]))\n plt.plot(x[idx], errors[idx], 'rx--', linewidth=2, markersize=12)\n plt.xlabel('Iteration')\n plt.ylabel('Current Error')\n plt.savefig('error_{}.png'.format(mid_layers))\n plt.show()\n model.set_phase('test')\n model.set_weights(best_weights)\n if print_w:\n print(best_weights)\n print(best_errors)\n s_e = sorted(best_errors, key=lambda x: x[1])\n x = [_[1] for _ in s_e]\n y = [_[0] for _ in s_e]\n plt.plot(x, y, 'ro--', linewidth=0.5, markersize=3)\n plt.xlabel('Error')\n plt.ylabel('Min Iterations')\n plt.savefig('error_deps.png')\n plt.show()\n print(s_e)\n\n return model\n\n\nif __name__ == \"__main__\":\n\n model = train(mid_layers=64, learn_rate=0.0001, min_error=1000.0)\n flower_path = os.path.join(DATA_TEST_DIR, '1.jpg')\n img, res = model.predict(flower_path)\n fig, ax = plt.subplots(1, 2)\n ax[0].imshow(img, cmap='gray')\n ax[0].set_xlabel('original')\n ax[1].imshow(res, cmap='gray')\n ax[1].set_xlabel('modified')\n plt.savefig('32.png')\n plt.show()\n","sub_path":"IAI/autoencoder/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":3518,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"259021157","text":"#!/usr/bin/env python3\n\nimport rsa\n'''\n refer to https://github.com/zhanghe06/python/blob/master/tools/rsa_encrypt.py\n'''\nclass MRsa(object):\n def __init__(self,pubfile,prifile):\n self.pubf=pubfile\n self.prif=prifile\n\n def CreateKeys(self):\n (_pub,_pri)=rsa.newkeys(1024)\n\n pub=_pub.save_pkcs1()\n with open(self.pubf,'w+') as pub_file:\n pub_file.write(pub)\n\n pri=_pri.save_pkcs1()\n with open(self.prif,'w+') as pri_file:\n pri_file.write(pri)\n\n def GetPri(self):\n with open(self.prif) as pri_file:\n p=pri_file.read()\n pri_key=rsa.PrivateKey.load_pkcs1(p)\n return pri_key\n\n def GetPub(self):\n with open(self.pubf) as pub_file:\n p=pub_file.read()\n pub_key=rsa.PublicKey.load_pkcs1(p)\n return pub_key\n\n def Encrypt(self,plaintext):\n pub_key=self.GetPub()\n ciphertext=rsa.encrypt(plaintext,pub_key)\n return ciphertext\n\n def Decrypt(self,ciphertext):\n pri_key=self.GetPri()\n plaintext=rsa.decrypt(ciphertext,pri_key)\n return plaintext\n\n def Signature(self,message):\n pri_key=self.GetPri()\n sign=rsa.sign(message,pri_key,'SHA-256')\n return sign\n\n def VerifySign(self,plaintext,ciphertext):\n pub_key=self.GetPub()\n try:\n rsa.verify(plaintext,ciphertext,pub_key)\n return True\n except:\n return False\n\n\n\nif __name__=='__main__':\n m=MRsa('./pub.pem','./pri.pem')\n# m.CreateKeys()\n print(m.Decrypt(m.Encrypt('helloworld')))\n\n","sub_path":"python/rsahandler.py","file_name":"rsahandler.py","file_ext":"py","file_size_in_byte":1605,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"79618480","text":"# -*- coding: utf-8 -*-\n\"\"\"\n This spider is a AsusJobs spider created on top of the ATSSpider\n scrapy crawl asusjobs -a mining_job_id=9999 -a iteration=1 -a extract=1 -a url=\"https://hr-recruit.asus.com/\"\n\n sample job url:\n https://hr-recruit.asus.com/showPositionDesc.aspx?fb=1837\n\"\"\"\n\nfrom re import compile\nfrom scrapy.http import Request\nfrom scrapy.http import FormRequest\nfrom scrapy.selector import Selector\nfrom urlparse import urljoin\n\nfrom brightcorp.base.atsspiders import ATSSpider\nfrom brightcorp.items import BrightcorpItemLoader\nfrom brightcorp.processors import Prefix\n\n\nclass AsusJobs(ATSSpider):\n\n name = \"asusjobs\"\n Job_Url = compile(r\"\\('(\\S+)','.*\\)\")\n\n def parse(self, response):\n sel = Selector(response)\n redirect_url = sel.xpath('//frame/@src').extract()\n if redirect_url:\n yield Request(\n callback=self.parse_redirect_page,\n url=urljoin(response.url, redirect_url[0])\n )\n\n def parse_redirect_page(self, response):\n sel = Selector(response)\n event_target = sel.xpath('//select[@name=\"dropSelZLD00\"]/@id').extract()\n categories = sel.xpath('//select[@name=\"dropSelZLD00\"]/option/@value').extract()\n for category in categories:\n yield FormRequest.from_response(\n callback=self.parse_job_list,\n formname='form1',\n formdata={\n '__EVENTTARGET': ''.join(event_target),\n 'dropSelZLD00': category,\n },\n response=response\n )\n\n def parse_job_list(self, response):\n sel = Selector(response)\n jobs = sel.xpath('//table[@id=\"itemTable\"]//tr/td/label/@onclick').extract()\n for job in jobs:\n url = self.Job_Url.search(job)\n if url:\n yield Request(\n callback=self.parse_job_callback(),\n url=urljoin(response.url, '/%s' % url.group(1))\n )\n\n def parse_job(self, response):\n loader = BrightcorpItemLoader(response=response)\n\n loader.add_xpath(\n 'description',\n '//td[contains(text(), \"%s\")]/../following-sibling::tr[@class=\"cart_oii\"][1]' % unicode('職位說明', 'utf-8')\n )\n loader.add_xpath(\n 'jobcategory',\n '//td[contains(text(), \"%s\")]/../following-sibling::tr[@class=\"cart_oii\"][1]/td/text()' % unicode('需求部門/單位', 'utf-8')\n )\n loader.add_xpath(\n 'location',\n '//td[contains(text(), \"%s\")]/../following-sibling::tr[@class=\"cart_oii\"][1]/td/text()' % unicode('工作地點', 'utf-8')\n )\n loader.add_xpath(\n 'referencenumber',\n '//td[contains(text(), \"%s\")]/../following-sibling::tr[@class=\"cart_oii\"][1]/td/text()' % unicode('職位代號', 'utf-8'),\n Prefix('%s-' % self.name)\n )\n loader.add_xpath(\n 'title',\n '//td[contains(text(), \"%s\")]/../following-sibling::tr[@class=\"cart_oii\"][1]/td/text()' % unicode('職位名稱', 'utf-8')\n )\n loader.add_xpath(\n 'other',\n '//td[contains(text(), \"%s\")]/../following-sibling::tr[@class=\"cart_oii\"][1]/td/text()' % unicode('其他工作條件', 'utf-8')\n )\n\n loader.add_value('apply_url', response.url)\n loader.add_value('url', response.url)\n\n yield loader.load_item()\n","sub_path":"brightcorp/brightcorp/spiders/asusjobs.py","file_name":"asusjobs.py","file_ext":"py","file_size_in_byte":3471,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"106179007","text":"import cv2\nimport numpy as np\n\ndef nothing(x):\n pass\n\nimage_x, image_y = 64,64\n\nfrom keras.models import load_model\nclassifier = load_model('Trained_model.h5')\n\ndef predictor():\n import numpy as np\n from keras.preprocessing import image\n test_image = image.load_img('1.png', target_size=(64, 64))\n test_image = image.img_to_array(test_image)\n test_image = np.expand_dims(test_image, axis = 0)\n result = classifier.predict(test_image)\n \n if result[0][0] == 1:\n return 'Book'\n elif result[0][1] == 1:\n return 'Clock'\n elif result[0][2] == 1:\n return 'Fork'\n elif result[0][3] == 1:\n return 'Friend'\n elif result[0][4] == 1:\n return 'Lamp'\n elif result[0][5] == 1:\n return 'Marriage'\n elif result[0][6] == 1:\n return 'Shoes'\n elif result[0][7] == 1:\n return 'Stop'\n elif result[0][8] == 1:\n return 'Toilet'\n elif result[0][9] == 1:\n return 'with'\n\n\ncam = cv2.VideoCapture(0)\n\ncv2.namedWindow(\"Trackbars\")\n\ncv2.createTrackbar(\"L - H\", \"Trackbars\", 0, 179, nothing)\ncv2.createTrackbar(\"L - S\", \"Trackbars\", 0, 255, nothing)\ncv2.createTrackbar(\"L - V\", \"Trackbars\", 0, 255, nothing)\ncv2.createTrackbar(\"U - H\", \"Trackbars\", 179, 179, nothing)\ncv2.createTrackbar(\"U - S\", \"Trackbars\", 255, 255, nothing)\ncv2.createTrackbar(\"U - V\", \"Trackbars\", 255, 255, nothing)\n\ncv2.namedWindow(\"test\")\n\nimg_counter = 0\n\nimg_text = ''\nwhile True:\n ret, frame = cam.read()\n frame = cv2.flip(frame,1)\n l_h = cv2.getTrackbarPos(\"L - H\", \"Trackbars\")\n l_s = cv2.getTrackbarPos(\"L - S\", \"Trackbars\")\n l_v = cv2.getTrackbarPos(\"L - V\", \"Trackbars\")\n u_h = cv2.getTrackbarPos(\"U - H\", \"Trackbars\")\n u_s = cv2.getTrackbarPos(\"U - S\", \"Trackbars\")\n u_v = cv2.getTrackbarPos(\"U - V\", \"Trackbars\")\n\n\n img = cv2.rectangle(frame, (425,100),(625,300), (0,255,0), thickness=2, lineType=8, shift=0)\n\n lower_blue = np.array([l_h, l_s, l_v])\n upper_blue = np.array([u_h, u_s, u_v])\n imcrop = img[102:298, 427:623]\n hsv = cv2.cvtColor(imcrop, cv2.COLOR_BGR2HSV)\n mask = cv2.inRange(hsv, lower_blue, upper_blue)\n \n cv2.putText(frame, img_text, (30, 400), cv2.FONT_HERSHEY_TRIPLEX, 1.5, (0, 255, 0))\n cv2.imshow(\"test\", frame)\n cv2.imshow(\"mask\", mask)\n \n #if cv2.waitKey(1) == ord('c'):\n \n img_name = \"1.png\"\n save_img = cv2.resize(mask, (image_x, image_y))\n cv2.imwrite(img_name, save_img)\n print(\"{} written!\".format(img_name))\n img_text = predictor()\n \n\n if cv2.waitKey(1) == 27:\n break\n\n\ncam.release()\ncv2.destroyAllWindows()","sub_path":"recognise.py","file_name":"recognise.py","file_ext":"py","file_size_in_byte":2722,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"478780663","text":"import sys\nsys.path.append('../')\n\nimport argparse\nimport mediascrapers\nfrom mediascrapers import TwitterImageScraper\n\nif __name__ == \"__main__\":\n ap = argparse.ArgumentParser()\n ap.add_argument(\"--url-list\", required=True,\n help=\"Path of file\")\n args = vars(ap.parse_args())\n filepath = args[\"url_list\"]\n\n scraper = TwitterImageScraper(driver='chrome', mode='verbose', debug=True)\n file = open(filepath, 'r')\n URLs = file.readlines()\n\n for url in URLs:\n print(url)\n tasks = scraper.scrape(url)\n scraper.download(tasks=tasks, path='.\\\\download\\\\general')\n","sub_path":"mediascraper/twitterimagescraper.py","file_name":"twitterimagescraper.py","file_ext":"py","file_size_in_byte":626,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"24999974","text":"import ldap\n\nfrom functools import wraps\n\nfrom flask import redirect, url_for, abort, request, current_app\n\n\ndef check_ldap(remote_user):\n try:\n l = ldap.initialize(current_app.config[\"LDAPS_URL\"])\n l.set_option(ldap.OPT_X_TLS, 1)\n l.bind_s(\n current_app.config[\"LDAP_DN\"],\n current_app.config[\"LDAP_PW\"])\n admin = l.search_s(\n current_app.config[\"LDAP_ADMIN_FQDN\"],\n ldap.SCOPE_SUBTREE,\n \"(userPrincipalName={})\".format(remote_user),\n [\"memberOf\"])\n if admin and \"memberOf\" in admin[0][1]:\n results = admin[0][1][\"memberOf\"]\n return results\n\n staff = l.search_s(\n current_app.config[\"LDAP_STAFF_FQDN\"],\n ldap.SCOPE_SUBTREE,\n \"(userPrincipalName={})\".format(remote_user),\n [\"memberOf\"])\n if staff and \"memberOf\" in staff[0][1]:\n results = staff[0][1][\"memberOf\"]\n return results\n except:\n return []\n return []\n\n\ndef permissions(needed, return_to=False):\n def check_permissions(function):\n @wraps(function)\n def wrapper(*args, **kwargs):\n my_perms = check_ldap(request.environ[\"REMOTE_USER\"])\n if set(needed).intersection(set(my_perms)):\n return function(*args, **kwargs)\n if current_app.config[\"LDAP_ADMIN_FQDN\"] in my_perms:\n return function(*args, **kwargs)\n if return_to:\n return redirect(url_for(return_to))\n return abort(401)\n return wrapper\n return check_permissions\n","sub_path":"app/security.py","file_name":"security.py","file_ext":"py","file_size_in_byte":1619,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"196380527","text":"from math import log\nfrom math import pow\nfrom math import sqrt\n\nimport sys\n\ndef nsqrt(x, n):\n\treturn x**(1/float(n))\n\n\ndef d(a0, e):\n\tcnt = 0\n\teroot = nsqrt(a0, e)\n\twhile eroot == int(eroot):\n\t\tcnt+=1\n\t\teroot = nsqrt(eroot, e)\n\treturn cnt\n\n\ndef subd(a0, e, N):\n\tk = d(a0, e)\n\tcandidat = int(nsqrt(a0, e))\n\twhile pow(a0, e)+candidat > N and k >= 0:\n\t\tcandidat = int(nsqrt(candidat, e))\n\t\tk -= 1\n\t\n\treturn k\n\ndef main(argv):\n\tN = int(argv[0])\n\tresultat = 0\n\temax = int(log(N)/log(2))\n\tprint (\"emax = \", emax)\n\tfor e in range(2, emax+1):\n\t\ta0 = 2\n\t\twhile pow(a0, e)+a0 <= N:\n\t\t\tk = d(a0, e)+1\n\t\t\tresultat += k*k\n\t\t\ta0+=1\n\t\tif d(a0, e) > 0: # Border case where the condition a0^e+a0 <= N might not hold anymore.\n\t\t\tprint (a0, e)\n\t\t\tl = subd(a0, e, N) # If a0 = x^(e*k), we find the highest l s.t. a0^e+x^(e*l) <= N.\n\t\t\tif l >= 0:\n\t\t\t\tresultat += (l+1)**2\n\n\tprint (resultat)\n\nif __name__ == \"__main__\":\n\tmain(sys.argv[1:])\n","sub_path":"src/pb617.py","file_name":"pb617.py","file_ext":"py","file_size_in_byte":920,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"620301227","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Mon Jun 11 15:02:09 2018\r\n\r\n@author: Balasubramaniam\r\n\"\"\"\r\n\r\nfrom openpyxl import load_workbook\r\nfilePath=\"F:/citi_ml_jun2018/day1/GDP.xlsx\"\r\nwb=load_workbook(filePath,read_only=True,data_only=True)\r\nsheetRef=wb.get_sheet_by_name(\"GDP\")\r\nfor row in range(6,70):\r\n print(sheetRef.cell(column=1,row=row).value,end=\"\\t\")\r\n\r\n","sub_path":"ExcelFileReading.py","file_name":"ExcelFileReading.py","file_ext":"py","file_size_in_byte":365,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"286539549","text":"class Node:\r\n def __init__(self, val):\r\n self.data = val\r\n self.next = None\r\n self.previous = None\r\n\r\n\r\nclass DLinkedList:\r\n \"\"\"\r\n DLinkedList class contains the main methods of a doubly linked list\r\n \"\"\"\r\n def __init__(self):\r\n self.head = None\r\n self.tail = None\r\n self.size = 0\r\n\r\n def get_head(self):\r\n \"\"\"\r\n get the value of the first node\r\n \"\"\"\r\n # Time complexity: O(1)\r\n # Space complexity: O(1)\r\n if self.head is None:\r\n return None\r\n return self.head.data\r\n\r\n def get_tail(self):\r\n \"\"\"\r\n get the value of the last node\r\n \"\"\"\r\n # Time complexity: O(1)\r\n # Space complexity: O(1)\r\n if self.tail is None:\r\n return None\r\n return self.tail.data\r\n\r\n def get_size(self):\r\n return self.size\r\n\r\n def prepend(self, val):\r\n \"\"\"\r\n add a leading node at the beginning of the linked list giving its value\r\n \"\"\"\r\n # Time complexity: O(1)\r\n # Space complexity: O(1)\r\n newNode = Node(val)\r\n newNode.next = self.head\r\n newNode.previous = None\r\n if self.tail is None:\r\n self.tail = newNode\r\n if self.head:\r\n self.head.previous = newNode\r\n self.head = newNode\r\n self.size += 1\r\n\r\n def pop_first(self):\r\n \"\"\"\r\n delete the leading node of the linked list\r\n \"\"\"\r\n # Time complexity: O(1)\r\n # Space complexity: O(1)\r\n if self.head is None:\r\n raise IndexError\r\n else:\r\n del_val = self.head.data\r\n self.head = self.head.next\r\n if self.head:\r\n self.head.previous = None\r\n self.size -= 1\r\n if self.size == 0:\r\n self.tail = None\r\n return del_val\r\n\r\n def append(self, val):\r\n \"\"\"\r\n add a node at the end of the linked list giving its value\r\n \"\"\"\r\n # Time complexity: O(1)\r\n # Space complexity: O(1)\r\n new_node = Node(val)\r\n if self.head is None:\r\n self.tail = new_node\r\n self.head = new_node\r\n else:\r\n old_tail = self.tail\r\n self.tail.next = new_node\r\n self.tail = new_node\r\n self.tail.previous = old_tail\r\n self.size += 1\r\n\r\n def pop_last(self):\r\n \"\"\"\r\n delete the last node of the linked list\r\n \"\"\"\r\n # Time complexity: O(1)\r\n # Space complexity: O(1)\r\n if self.head is None:\r\n raise IndexError\r\n last_node = self.tail.data\r\n self.tail = self.tail.previous\r\n if self.tail:\r\n self.tail.next = None\r\n self.size -= 1\r\n if self.size == 0:\r\n self.head = None\r\n return last_node\r\n\r\n def insert_node(self, new_node, previous_node):\r\n \"\"\"\r\n insert a giving node to the new list given the previous node to it\r\n \"\"\"\r\n # Time complexity: O(1)\r\n # Space complexity: O(1)\r\n next_node = previous_node.next \r\n new_node.next = next_node\r\n new_node.previous = previous_node\r\n if next_node:\r\n next_node.previous = new_node\r\n else:\r\n self.tail = new_node\r\n previous_node.next = new_node\r\n\r\n # return a list of all values of the nodes\r\n n = self.head\r\n out = [n.data]\r\n for _ in range(self.size):\r\n n = n.next\r\n out.append(n.data)\r\n return out\r\n\r\n def detete_node(self, node):\r\n \"\"\"\r\n delete a given node\r\n \"\"\"\r\n # Time complexity: O(1)\r\n # Space complexity: O(1)\r\n prev_node = node.previous # get the node before the node to delete\r\n next_node = node.next # get the node after the node to delete\r\n if (not prev_node) and (not next_node): # if there are nodes before and after the node (linked list size is 1)\r\n self.head = None # set the head to None\r\n self.tail = None # set the tail to None\r\n if prev_node: # if previous node found\r\n prev_node.next = next_node # set its next reference to the next node\r\n else:\r\n self.head = next_node # if not found, set head of the linked list to the next node\r\n if next_node:\r\n next_node.previous = prev_node\r\n else:\r\n self.tail = prev_node\r\n\r\n\r\n\r\n","sub_path":"implementation_5/doubly_linked_list.py","file_name":"doubly_linked_list.py","file_ext":"py","file_size_in_byte":4483,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"365355437","text":"\"\"\"Add fields to Team\n\nRevision ID: d2853477b461\nRevises: a23d4d63432d\nCreate Date: 2017-01-04 17:23:20.326509\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = 'd2853477b461'\ndown_revision = 'a23d4d63432d'\nbranch_labels = None\ndepends_on = None\n\ndef upgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.add_column('teams', sa.Column('created_at', sa.DateTime(timezone=True), nullable=True))\n op.add_column('teams', sa.Column('description', sa.String(), nullable=True))\n op.add_column('teams', sa.Column('owner_id', sa.Integer(), nullable=True))\n op.add_column('teams', sa.Column('updated_at', sa.DateTime(timezone=True), nullable=True))\n op.create_foreign_key(None, 'teams', 'users', ['owner_id'], ['id'])\n op.add_column('users_teams', sa.Column('created_at', sa.DateTime(timezone=True), nullable=True))\n op.add_column('users_teams', sa.Column('id', sa.Integer(), nullable=False))\n op.add_column('users_teams', sa.Column('team_ud', sa.Integer(), nullable=True))\n op.add_column('users_teams', sa.Column('updated_at', sa.DateTime(timezone=True), nullable=True))\n op.drop_constraint('users_teams_team_id_fkey', 'users_teams', type_='foreignkey')\n op.create_foreign_key(None, 'users_teams', 'teams', ['team_ud'], ['id'])\n op.drop_column('users_teams', 'team_id')\n ### end Alembic commands ###\n\n\ndef downgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.add_column('users_teams', sa.Column('team_id', sa.INTEGER(), autoincrement=False, nullable=True))\n op.drop_constraint(None, 'users_teams', type_='foreignkey')\n op.create_foreign_key('users_teams_team_id_fkey', 'users_teams', 'teams', ['team_id'], ['id'])\n op.drop_column('users_teams', 'updated_at')\n op.drop_column('users_teams', 'team_ud')\n op.drop_column('users_teams', 'id')\n op.drop_column('users_teams', 'created_at')\n op.drop_constraint(None, 'teams', type_='foreignkey')\n op.drop_column('teams', 'updated_at')\n op.drop_column('teams', 'owner_id')\n op.drop_column('teams', 'description')\n op.drop_column('teams', 'created_at')\n ### end Alembic commands ###\n","sub_path":"alembic/versions/d2853477b461_add_fields_to_team.py","file_name":"d2853477b461_add_fields_to_team.py","file_ext":"py","file_size_in_byte":2201,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"99022","text":"from bs4 import BeautifulSoup\n\n\ndef read_file(fileName):\n file = open(fileName)\n data = file.read()\n file.close()\n return data\n\n\nsoup = BeautifulSoup(read_file('intro_to_soup_html.html'), 'lxml')\nmeta = soup.meta\nprint(meta)\nprint(meta.get('charset'))\n\nbody = soup.body\nprint(body)","sub_path":"example 5.py","file_name":"example 5.py","file_ext":"py","file_size_in_byte":293,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"288455100","text":"\"\"\"\nEndpoint routers for users.\n\nEventually might need to handle auth here as well, so write code as if\nthat was an upcoming feature.\n\"\"\"\nfrom fastapi import APIRouter\nfrom models import users as models\nfrom docs import users as docs\nimport util.users as utils\n\nrouter = APIRouter()\n\n\n@router.post(\n \"/users/register\",\n response_model=models.UserRegistrationResponse,\n description=docs.registration_desc,\n summary=docs.registration_summ,\n tags=[\"Users\"],\n status_code=201,\n)\nasync def register_user(form: models.UserRegistrationForm):\n # send the form data and DB instance to util.users.register_user\n user_id = await utils.register_user(form)\n\n # return response in reponse model\n return models.UserRegistrationResponse(user_id=user_id)\n\n\n@router.delete(\n \"/users/delete\",\n description=docs.delete_user_desc,\n summary=docs.delete_user_summ,\n tags=[\"Users\"],\n status_code=204,\n)\nasync def delete_user(identifier: models.UserIdentifier):\n await utils.delete_user(identifier)\n\n\n@router.post(\"/users/find\",\n response_model=models.UserInfoQueryResponse,\n description=docs.find_user_by_identifier_desc,\n summary=docs.find_user_by_identifier_summ,\n tags=[\"Users\"],\n status_code=200)\nasync def get_user(identifier: models.UserIdentifier):\n user_data = await utils.get_user_info_by_identifier(identifier)\n return models.UserInfoQueryResponse(**user_data.dict())\n","sub_path":"routes/users.py","file_name":"users.py","file_ext":"py","file_size_in_byte":1468,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"613148416","text":"#!/usr/bin/env python\n\nfrom optparse import OptionParser\nfrom gen.run_generators import RUNGeneratorRes\nimport os\nimport sys\n\ndef parse_args():\n parser = OptionParser(\"usage: %prog [options] [files...] \"\n \"[generators...] [param=val[,val]...]\")\n\n parser.add_option('-o', '--out-dir', dest='out_dir',\n help='directory for data output',\n default=(\"%s/%s\"% (os.getcwd(), \"RESEXPS\")))\n\n return parser.parse_args()\n\ndef main():\n opts, inFolders = parse_args()\n distr = 0.8\n res_number = 3\n \n if not os.path.exists(opts.out_dir):\n os.mkdir(opts.out_dir)\n if opts.out_dir[-1] != '/':\n opts.out_dir = opts.out_dir+'/'\n \n for folder in inFolders:\n done = False\n trial = 0\n while (not done and trial < 50):\n try:\n if folder[-1] == '/':\n folder = folder[:-1]\n foldername = folder.strip().split('/')[-1]\n out_dir = opts.out_dir+foldername+\"_res=\"+str(distr)+\"/\"\n if not os.path.exists(out_dir):\n os.mkdir(out_dir)\n \n \n generator = RUNGeneratorRes()\n generator.out_dir=out_dir\n\n params = {}\n\n ts = generator._create_taskset_from_file(params, res_number, folder, distr)\n\n generator._customize(ts, params)\n generator._write_schedule(dict(params.items() + [('task_set', ts)]))\n generator._write_params(params)\n done = True\n except:\n trial += 1\n continue\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"duplicateExpWithRes.py","file_name":"duplicateExpWithRes.py","file_ext":"py","file_size_in_byte":1514,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"313627800","text":"# https://stackoverflow.com/questions/62262538/how-to-add-an-image-in-pyqt-qdock-widget\n\nfrom PyQt5 import QtWidgets, QtGui\nfrom PyQt5.QtWidgets import (QApplication, QMainWindow)\nfrom PyQt5.QtCore import Qt\n \nimport sys\n\nclass MainWindow(QMainWindow):\n def __init__(self):\n super().__init__()\n self.resize(800, 600)\n\n dockWidget = QtWidgets.QDockWidget()\n dockWidget.setWindowTitle(\"Image Viewer\")\n \n image = QtGui.QImage('pyqt5/img_1.png')\n pixmap = QtGui.QPixmap.fromImage(image) \n label = QtWidgets.QLabel('testing', self)\n label.setPixmap(pixmap)\n\n #dockWidget.setWidget(pixmap)\n dockWidget.setWidget(label)\n dockWidget.setFloating(False)\n self.addDockWidget(Qt.RightDockWidgetArea, dockWidget)\n\n\nif __name__ == '__main__':\n app = QApplication(sys.argv)\n myWidget = MainWindow()\n myWidget.show()\n\n sys.exit(app.exec_())","sub_path":"200609-01-qdock.py","file_name":"200609-01-qdock.py","file_ext":"py","file_size_in_byte":924,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"299802642","text":"from __future__ import print_function\n\nimport sys\nimport os\nimport logging\nimport yaml\nfrom fnmatch import fnmatch\n\nfrom .files import MinipipeFilesMixin\nfrom .items import MinipipeItemsMixin\n\n__all__ = [\n 'delete_modules',\n 'Minipipe',\n 'MinipipeConfig',\n 'MinipipeUtil',\n]\n\nLOG = logging.getLogger('minipipe')\n\n\ndef delete_modules(pattern, verbose=True):\n \"\"\"\n Delete all existing python modules that match\n the given pattern.\n\n Args:\n pattern: A str pattern to match against module names\n verbose: A bool, when true, prints each module that is deleted\n \"\"\"\n mods = sys.modules.keys()\n matching_mods = [mod for mod in mods if fnmatch(mod, pattern)]\n for mod in matching_mods:\n if verbose:\n print('Deleting module: {0}'.format(mod))\n del sys.modules[mod]\n\n\ndef remove_prefix(name, prefix):\n if name.startswith(prefix):\n return name[len(prefix):]\n\n\nclass MinipipeConfig(dict):\n \"\"\"\n A config dictionary for a Minipipe project.\n\n A MinipipeConfig can be loaded using `load_config`\n which accepts any path as a starting point, and then\n searches upwards until a minipipe_config.yaml file\n is found. If no path is given, starts from the\n current working directory.\n\n Can be subclassed to change how the config file is\n found, such as modifying the default anypath.\n \"\"\"\n\n @classmethod\n def from_path(cls, anypath=None):\n inst = cls()\n inst.load_config(anypath)\n return inst\n\n def __init__(self, *args, **kwargs):\n super(MinipipeConfig, self).__init__(*args, **kwargs)\n self.config_filename = 'minipipe_config.yaml'\n self._has_attempted_load = False\n\n def has_attempted_load(self):\n return self._has_attempted_load\n\n def is_valid(self):\n return 'config_filepath' in self\n\n def get_default_anypath(self):\n \"\"\"\n Return the default path to use for anypath if none is given\n \"\"\"\n return os.getcwd()\n\n def find_config_file(self, anypath=None):\n \"\"\"\n Find and return the full path to the config file.\n\n Args:\n anypath: A string path anywhere in the project. If None, will use\n the default anypath (usually the current working directory)\n \"\"\"\n if not anypath:\n anypath = self.get_default_anypath()\n\n filepath = Minipipe.find_file_upwards(\n self.config_filename, anypath)\n\n return filepath\n\n def load_config(self, anypath=None):\n \"\"\"\n Load a config file by searching for it using any path.\n\n Args:\n anypath: A string path anywhere in the project. If None, will use\n the default anypath (usually the current working directory)\n \"\"\"\n self.clear()\n self._has_attempted_load = True\n\n config_filepath = self.find_config_file(anypath)\n if not config_filepath:\n LOG.warning('Failed to find {0} from path: {1}'.format(\n self.config_filename, anypath))\n return\n\n with open(config_filepath, 'rb') as fp:\n loaded_config = yaml.load(fp)\n\n # TODO: better error handling of failed config load\n\n if loaded_config is not None:\n self.update(loaded_config)\n self.post_load_config(config_filepath)\n else:\n LOG.warning(\n \"Failed to load minipipe config: {0}\".format(config_filepath))\n\n def post_load_config(self, filepath):\n \"\"\"\n Apply any modifications to the config after it is loaded.\n Config will not be None when this is called.\n \"\"\"\n # store config filepath and dirpath\n self['config_filepath'] = filepath\n dirpath = os.path.dirname(filepath)\n self['config_dirpath'] = dirpath\n\n # evaluate full paths from '_relpath' keys\n keys = list(self.keys())\n for key in keys:\n if key.endswith('_relpath'):\n # add new key with _path suffix\n self[key[:-8] + '_path'] = Minipipe.join_path(\n dirpath, self[key])\n elif key.endswith('.relpath'):\n # add new key with _path suffix\n self[key[:-8] + '.path'] = Minipipe.join_path(\n dirpath, self[key])\n\n\nclass MinipipeUtil(object):\n \"\"\"\n The base class for any object that uses a MinipipeConfig.\n \"\"\"\n\n def __init__(self, config=None):\n \"\"\"\n Args:\n config: A MinipipeConfig or MinipipeUtil. If given a config,\n will use the config instance directly, if given a util,\n will use the config of that util when accessing.\n \"\"\"\n self._config = None\n self.parent = None\n if isinstance(config, MinipipeConfig):\n self._config = config\n elif isinstance(config, MinipipeUtil):\n self.parent = config\n\n def create_config(self):\n \"\"\"\n Create and return a new MinipipeConfig for this util.\n \"\"\"\n config = MinipipeConfig()\n self.init_config(config)\n return config\n\n def init_config(self, config):\n \"\"\"\n Initialize a newly created MinipipeConfig. This is an\n opportunity to load the config from a specific path,\n or modify other settings.\n \"\"\"\n pass\n\n def _get_config(self):\n \"\"\"\n Return the MinipipeConfig for this util.\n \"\"\"\n if self.parent:\n return self.parent._get_config()\n else:\n if not self._config:\n self._config = self.create_config()\n return self._config\n\n def get_config(self):\n \"\"\"\n Return the Minipipe instance for this util.\n\n Attempts to automatically load the config if it\n hasn't been loaded yet.\n \"\"\"\n config = self._get_config()\n if not config.has_attempted_load():\n LOG.info(\"Auto-loading MinipipeUtil config, it is recommended \"\n \"that load_config be called explicitly before use\")\n config.load_config()\n return config\n\n def load_config(self, anypath=None):\n mp = self._get_config()\n mp.load_config(anypath)\n\n @property\n def config(self):\n return self.get_config()\n\n def config_format(self, string_fmt):\n \"\"\"\n Format a string with the MinipipeConfig of this util.\n Modifies the format string to convert '{my.key}' into '{0[my.key]}'\n so that config keys with dots to not break the formatting.\n \"\"\"\n fixed_fmt = string_fmt.replace('{', '{0[').replace('}', ']}')\n return fixed_fmt.format(self.config)\n\n\nclass Minipipe(MinipipeUtil,\n MinipipeFilesMixin,\n MinipipeItemsMixin):\n \"\"\"\n A MinipipeUtil with many core methods for common functionality\n including path and file management.\n \"\"\"\n\n def titlecase(self, name):\n \"\"\"\n Convert a name to TitleCase\n\n Args:\n name: A string name\n \"\"\"\n title_keywords = self.config.get('titlecase_keywords', [])\n if name.upper() in title_keywords:\n return name.upper()\n if len(name):\n return name[0].upper() + name[1:]\n return name\n\n def titlecase_path(self, path):\n \"\"\"\n Convert an any case path to a TitleCase path.\n\n Args:\n path: A string path\n \"\"\"\n items = path.replace('\\\\', '/').split('/')\n title_items = [self.titlecase(item) for item in items]\n return self.join_path(*title_items)\n\n\nfuncs = [\n remove_prefix\n]\nfor func in funcs:\n setattr(Minipipe, func.__name__, staticmethod(func))\n","sub_path":"src/minipipe/core.py","file_name":"core.py","file_ext":"py","file_size_in_byte":7738,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"38518079","text":"class Context(dict):\n def __init__(self, content=None, parent=None):\n self.parent = parent\n if content is None:\n content = {}\n dict.__init__(self, content)\n\n def __getitem__(self, which):\n try:\n return dict.__getitem__(self, which)\n except KeyError as err:\n if self.parent is not None:\n return self.parent[which]\n raise err\n\n def child(self):\n return Context(parent=self)\n\n def get(self, what, default=None):\n result = dict.get(self, what, default)\n if result is None and self.parent is not None:\n result = self.parent.get(what, default)\n return result\n\n\ndef default_context():\n from srl import ast\n return {\n \"digit\": ast.Digit(),\n \"letter\": ast.Letter(),\n \"whitespace\": ast.Whitespace(),\n }\n","sub_path":"srl/context.py","file_name":"context.py","file_ext":"py","file_size_in_byte":868,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"138348188","text":"from pwn import *\nobjects = []\n\nobjects_num = objects.__len__()\nobjects_string = reduce(lambda x, y: x + p32(y.__len__()) + y, objects, \"\")\nobjects_total_length = objects_string.__len__()\nanimation = p32(0)\nbinary = p32(objects_num)\nbinary += p32(objects_total_length)\nbinary += animation\nbinary += objects_string\nopen(\"test_files/no_object_and_animation\", \"w\").write(binary)\n","sub_path":"test_files/for_generates/no_object_and_animation.py","file_name":"no_object_and_animation.py","file_ext":"py","file_size_in_byte":376,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"409498422","text":"from flask_cors import CORS\n\nimport helpers\nimport logging\nfrom api import settings\nfrom flask import Flask, render_template, jsonify\nfrom flask_migrate import Migrate\nfrom werkzeug.debug import get_current_traceback\nfrom models import db\nfrom request_helpers import recover_identity\nfrom api.proposals import register_endpoints as register_proposal_endpoints\nfrom api.identities import register_endpoints as register_identity_endpoints\nfrom api.sessions import register_endpoints as register_session_endpoints\nfrom api.statistics import register_endpoints as register_statistic_endpoints\nfrom api.affiliates import register_endpoints as register_affiliates_endpoints\nfrom api.mobile import register_endpoints as register_mobile_endpoints\n\nif not settings.DISABLE_LOGS:\n helpers.setup_logger()\n\napp = Flask(__name__)\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\n\n# Enable to print SQL statements using get_debug_queries.\n# app.config['SQLALCHEMY_RECORD_QUERIES'] = True\n#\n# Usage example inside endpoints.\n# for info in get_debug_queries():\n# print(info.statement, info.parameters, info.duration, sep='\\n')\n# print('\\n')\n\nCORS(app, resources=[r'/v1/affiliates'])\nregister_proposal_endpoints(app)\nregister_identity_endpoints(app)\nregister_session_endpoints(app)\nregister_statistic_endpoints(app)\nregister_affiliates_endpoints(app)\nregister_mobile_endpoints(app)\n\n\ndef _generate_database_uri(db_config):\n return 'mysql+pymysql://{}:{}@{}/{}'.format(\n db_config['user'], db_config['passwd'], db_config['host'],\n db_config['name'])\n\n\napp.config['SQLALCHEMY_DATABASE_URI'] =\\\n _generate_database_uri(settings.DB_CONFIG)\n\n\nmigrate = Migrate(app, db)\n\n\n# TODO: move to authorization.py\n@app.route('/', methods=['GET'])\ndef home():\n return render_template(\n 'api.html',\n )\n\n\n# End Point example which recovers public address from signed payload\n@app.route('/v1/me', methods=['GET'])\n@recover_identity\ndef test_signed_payload(caller_identity):\n return jsonify({\n 'identity': caller_identity\n })\n\n\n@app.errorhandler(404)\ndef method_not_found(e):\n return jsonify(error='unknown API method'), 404\n\n\n@app.errorhandler(405)\ndef method_not_allowed(e):\n return jsonify(error='method not allowed'), 405\n\n\n@app.errorhandler(Exception)\ndef handle_error(e):\n track = get_current_traceback(\n skip=1,\n show_hidden_frames=True,\n ignore_system_exceptions=False\n )\n logging.error(track.plaintext)\n return jsonify(error=str(e)), 500\n\n\ndef start_debug_app():\n init_db()\n app.run(debug=True)\n\n\ndef init_db():\n db.init_app(app)\n\n\nif __name__ == '__main__':\n start_debug_app()\n","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":2663,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"586934858","text":"\"\"\" Defines the Worker class.\r\n\"\"\"\r\n\r\nfrom ..workforce import _store\r\nfrom .exceptions import ValidationError\r\nfrom .feature_model import FeatureModel\r\nfrom ._store import *\r\nfrom ._schemas import WorkerSchema\r\n\r\n\r\nclass Worker(FeatureModel):\r\n \"\"\"\r\n Represents a worker in a Workforce Project\r\n\r\n ================== ====================================================================\r\n **Argument** **Description**\r\n ------------------ --------------------------------------------------------------------\r\n project Required :class:`~arcgis.apps.workforce.Project`. The project that\r\n the worker belongs to.\r\n ------------------ --------------------------------------------------------------------\r\n feature Optional :class:`~arcgis.features.Feature`. The feature representing\r\n the worker. Mostly intended for\r\n internal usage. If supplied, other parameters are ignored.\r\n ------------------ --------------------------------------------------------------------\r\n geometry Optional :class:`Dict`. The geometry of the worker.\r\n ------------------ --------------------------------------------------------------------\r\n contact_number Optional :class:`String`. The contact number of the worker.\r\n ------------------ --------------------------------------------------------------------\r\n name Optional :class:`String`. The name of the worker.\r\n ------------------ --------------------------------------------------------------------\r\n notes Optional :class:`String`. The notes about the worker.\r\n ------------------ --------------------------------------------------------------------\r\n status Optional :class:`String`. The status of the worker.\r\n\r\n `not_working`, `working`, `on_break`\r\n ------------------ --------------------------------------------------------------------\r\n title Optional :class:`String`. The title of the worker.\r\n ------------------ --------------------------------------------------------------------\r\n user_id Optional :class:`String`. The user id of the worker\r\n ================== ====================================================================\r\n\r\n \"\"\"\r\n\r\n def __init__(self, project, feature=None, geometry=None, contact_number=None,\r\n name=None, notes=None, status=\"not_working\", title=None, user_id=None):\r\n super().__init__(project, project.workers_layer, feature)\r\n self._schema = WorkerSchema(project.workers_layer)\r\n if not feature:\r\n self.geometry = geometry\r\n self.contact_number = contact_number\r\n self.name = name\r\n self.notes = notes\r\n self.status = status\r\n self.title = title\r\n self.user_id = user_id\r\n\r\n def __str__(self):\r\n return \"{} ({})\".format(self.name, self.user_id)\r\n\r\n def __repr__(self):\r\n return \"\".format(self.object_id)\r\n\r\n def update(self, geometry=None, contact_number=None,\r\n name=None, notes=None, status=None, title=None, user_id=None):\r\n \"\"\"\r\n Updates the worker on the server\r\n\r\n ================== ====================================================================\r\n **Argument** **Description**\r\n ------------------ --------------------------------------------------------------------\r\n geometry Optional :class:`Dict`. The geometry of the worker.\r\n ------------------ --------------------------------------------------------------------\r\n contact_number Optional :class:`String`. The contact number of the worker.\r\n ------------------ --------------------------------------------------------------------\r\n name Optional :class:`String`. The name of the worker.\r\n ------------------ --------------------------------------------------------------------\r\n notes Optional :class:`String`. The notes about the worker.\r\n ------------------ --------------------------------------------------------------------\r\n status Optional :class:`String`. The status of the worker.\r\n\r\n `not_working`, `working`, `on_break`\r\n ------------------ --------------------------------------------------------------------\r\n title Optional :class:`String`. The title of the worker.\r\n ------------------ --------------------------------------------------------------------\r\n user_id Optional :class:`String`. The user id of the worker\r\n ================== ====================================================================\r\n\r\n \"\"\"\r\n update_worker(self.project, self, geometry, contact_number, name, notes, status, title, user_id)\r\n\r\n def delete(self):\r\n \"\"\"Deletes the worker from the server\"\"\"\r\n delete_workers(self.project, [self])\r\n\r\n @property\r\n def name(self):\r\n \"\"\"Gets/Sets the name of the worker\"\"\"\r\n return self._feature.attributes.get(self._schema.name)\r\n\r\n @name.setter\r\n def name(self, value):\r\n self._feature.attributes[self._schema.name] = value\r\n\r\n @property\r\n def contact_number(self):\r\n \"\"\"Gets/Sets the contact number of the worker\"\"\"\r\n return self._feature.attributes.get(self._schema.contact_number)\r\n\r\n @contact_number.setter\r\n def contact_number(self, value):\r\n self._feature.attributes[self._schema.contact_number] = value\r\n\r\n @property\r\n def title(self):\r\n \"\"\"Gets/Sets the title of the worker\"\"\"\r\n return self._feature.attributes.get(self._schema.title)\r\n\r\n @title.setter\r\n def title(self, value):\r\n self._feature.attributes[self._schema.title] = value\r\n\r\n @property\r\n def notes(self):\r\n \"\"\"Gets/Sets the notes of the worker\"\"\"\r\n return self._feature.attributes.get(self._schema.notes)\r\n\r\n @notes.setter\r\n def notes(self, value):\r\n self._feature.attributes[self._schema.notes] = value\r\n\r\n @property\r\n def user_id(self):\r\n \"\"\"Gets/Sets the user id of the worker\"\"\"\r\n return self._feature.attributes.get(self._schema.user_id)\r\n\r\n @user_id.setter\r\n def user_id(self, value):\r\n self._feature.attributes[self._schema.user_id] = value\r\n\r\n @property\r\n def status(self):\r\n \"\"\"\r\n Gets/Sets the :class:`String` status of the worker\r\n\r\n `not_working`, `working`, `on_break`\r\n \"\"\"\r\n lut = {\r\n 0: \"not_working\",\r\n 1: \"working\",\r\n 2: \"on_break\",\r\n }\r\n if self._feature.attributes[self._schema.status] is not None:\r\n return lut[self._feature.attributes[self._schema.status]]\r\n else:\r\n return None\r\n\r\n @status.setter\r\n def status(self, value):\r\n if (isinstance(value, int) and value >= 0 and value <= 2) or value is None:\r\n self._feature.attributes[self._schema.status] = value\r\n elif isinstance(value, str):\r\n reduced_str = value.lower().replace(\" \", \"\").replace(\"_\", \"\")\r\n if reduced_str == \"notworking\":\r\n self._feature.attributes[self._schema.status] = 0\r\n elif reduced_str == \"working\":\r\n self._feature.attributes[self._schema.status] = 1\r\n elif reduced_str == \"onbreak\":\r\n self._feature.attributes[self._schema.status] = 2\r\n else:\r\n raise ValidationError(\"Invalid status\", self)\r\n else:\r\n raise ValidationError(\"Invalid status\", self)\r\n\r\n @FeatureModel.geometry.setter\r\n def geometry(self, value):\r\n self._feature.geometry = value\r\n\r\n def _validate(self, **kwargs):\r\n \"\"\"\r\n \"\"\"\r\n errors = super()._validate(**kwargs)\r\n errors += self._validate_name()\r\n errors += self._validate_status()\r\n errors += self._validate_user_id()\r\n return errors\r\n\r\n def _validate_for_add(self, **kwargs):\r\n errors = super()._validate_for_add(**kwargs)\r\n errors += self._validate_user_id_on_server()\r\n return errors\r\n\r\n def _validate_for_update(self, **kwargs):\r\n errors = super()._validate_for_update(**kwargs)\r\n errors += self._validate_user_id_on_server()\r\n return errors\r\n\r\n def _validate_for_remove(self, **kwargs):\r\n errors = super()._validate_for_remove(**kwargs)\r\n assignments = _store.query_assignments(self.project, \"{} = {}\".format(self.project._assignment_schema.worker_id, self.object_id))\r\n if assignments:\r\n errors.append(ValidationError(\"Cannot remove a Worker that has assignments\", self))\r\n return errors\r\n\r\n def _validate_name(self):\r\n errors = []\r\n if not self.name or self.name.isspace():\r\n errors.append(ValidationError(\"Worker cannot have an empty name\", self))\r\n return errors\r\n\r\n def _validate_status(self):\r\n errors = []\r\n if self.status is None:\r\n errors.append(ValidationError(\"Worker must have a status\", self))\r\n return errors\r\n\r\n def _validate_user_id(self):\r\n errors = []\r\n if not self.user_id or self.user_id.isspace():\r\n errors.append(ValidationError(\"Worker cannot have an empty user_id\", self))\r\n return errors\r\n\r\n def _validate_user_id_on_server(self):\r\n errors = []\r\n user = self.project.gis.users.get(self.user_id)\r\n if user is None:\r\n message = \"The Worker user_id must match an accessible named user id\"\r\n errors.append(ValidationError(message, self))\r\n\r\n workers = [w for w in self.project._cached_workers.values() if w.user_id == self.user_id]\r\n duplicate_workers = [w for w in workers if w.object_id != self.object_id]\r\n if duplicate_workers:\r\n message = \"There cannot be multiple Workers with the same user_id\"\r\n errors.append(ValidationError(message, self))\r\n return errors\r\n","sub_path":"arcpyenv/arcgispro-py3-clone/Lib/site-packages/arcgis/apps/workforce/worker.py","file_name":"worker.py","file_ext":"py","file_size_in_byte":10416,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"112691110","text":"#!/usr/bin/python\n\nimport numpy as np\nimport matplotlib as mpl\nmpl.use('Agg')\nimport matplotlib.pyplot as plt\n\nsample_data = 1\nnsamp = 10000*sample_data\ndt = 1e-3\ndtSamp = dt*nsamp\n\n# Mean square displacment\nindata = np.loadtxt('msd_post.txt',dtype=float)\n\nd = indata.shape[1] # dimensions\nN = indata.shape[0] # number of data points\n\nt = np.zeros(N)\nfor i in np.arange(N):\n t[i] = indata[i][0]\n\nv = np.zeros(N)\nfor i in np.arange(N):\n v[i] = indata[i][1]/((2*20/np.pi)*(2*20/np.pi))\n\nplt.figure(figsize=(12,10))\n\nfig = plt.subplot(211)\nplt.plot(t,v)\nplt.ylabel('$<\\\\Delta r^{2}>/4R^2$',fontsize=30)\nplt.title('Mean square displacement',fontsize=35,fontweight='bold')\nfig.tick_params(axis='both', which='major', labelsize=20)\n\nfig = plt.subplot(212)\nplt.loglog(t,v)\nplt.xlabel('$\\\\Delta t$',fontsize=30)\nplt.ylabel('$<\\\\Delta r^{2}>/4R^2$',fontsize=30)\nfig.tick_params(axis='both', which='major', labelsize=20)\n\nplt.savefig('msd_post.png',bbox_inches='tight')\n\n\n# Plot exponent coefficients\nindata = np.loadtxt('msd_post.txt',dtype=float)\n\nd = indata.shape[1] # dimensions\nN = indata.shape[0] # number of data points\n\nt = np.zeros(N)\nfor i in np.arange(N):\n t[i] = indata[i][0]\n\nv = np.zeros(N)\nfor i in np.arange(N):\n v[i] = indata[i][1]\n\nt_log = np.log(t[1:np.size(t)])\nmsd_log = np.log(v[1:np.size(v)])\negim_offset = 10\nN = 90\negim = np.zeros(N)\nt_egim = np.zeros(N)\ni = 0\nfor i in np.arange(N):\n coef = np.polyfit(t_log[i:i+egim_offset],msd_log[i:i+egim_offset],1)\n egim[i] = coef[0]\n t_egim[i] = dtSamp*i\n\nplt.figure(figsize=(12,10))\nfig = plt.subplot(111)\nplt.plot(t_egim,egim,'o')\nplt.title('Exponent fits within a window of %s' %(egim_offset), fontsize=20,fontweight='bold')\nplt.xlabel('t',fontsize=20)\nplt.ylabel('$\\\\alpha$',fontsize=20)\nfig.tick_params(axis='both', which='major', labelsize=20)\n\nplt.savefig('exponents.png',bbox_inches='tight')\n\n","sub_path":"OldCellAnalysis/New_found/plot_sampling.py","file_name":"plot_sampling.py","file_ext":"py","file_size_in_byte":1877,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"203313135","text":"from geometry_msgs.msg import Twist\n\nclass Quadrotor:\n \"\"\"\n Holds all information about a quadrotor and can control the quadrotor\n completely\n \"\"\"\n def __init__(self):\n self.pos = [0, 0, 0]\n self.vel = [0, 0, 0]\n self.acc = [0, 0, 0]\n\n self.ang = [0, 0, 0]\n self.ang_vel = [0, 0, 0]\n self.ang_acc = [0, 0, 0]\n\n self.publisher = rospy.Publisher('cmd_vel', Twist, queue_size=10)\n\n def send(self):\n \"\"\"\n Makes a Twist message and sends it\n \"\"\"\n msg = Twist()\n\n msg.linear.x = self.vel[0]\n msg.linear.y = self.vel[1]\n msg.linear.z = self.vel[2]\n msg.angular.x = self.ang_vel[0]\n msg.angular.y = self.ang_vel[1]\n msg.angular.z = self.ang_vel[2]\n\n self.publisher.publish(msg)","sub_path":"hector_keyboard_controller/scripts/quadrotor/Quadrotor.py","file_name":"Quadrotor.py","file_ext":"py","file_size_in_byte":813,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"221643476","text":"# -*- coding: utf-8 -*-\r\n\r\n\"\"\"\r\n\r\nSpyder Editor\r\n\r\n\r\n\r\nThis is a temporary script file.\r\n\r\n\"\"\"\r\n\r\nimport numpy as np\r\n\r\nimport pylab as plt\r\n\r\nfrom PIL import Image\r\n\r\nimport math\r\nI = Image.open( \"..//images/161062.jpg\")\r\n#I = Image.open( \"C://Users/tianzixie/Desktop/1610621.jpg\")\r\n#imgsrc = \"https://www2.eecs.berkeley.edu/Research/Projects/CS/vision/bsds/BSDS300/html/images/plain/normal/gray/161062.jpg\"\r\n#from skimage import io\r\n#I = io.imread(imgsrc)\r\n\r\nimgArr = np.array(I)\r\n\r\nimgArrr=np.array(I)\r\n\r\nImage.fromarray(imgArr).show()\r\n\r\nhis = np.zeros(256)\r\n\r\nprob = np.zeros(256)\r\n\r\n#create array\r\n\r\nfor i in range(imgArr.shape[0]):\r\n\r\n for j in range(imgArr.shape[1]):\r\n\r\n his[imgArr[i, j]] = his[imgArr[i, j]] + 1\r\n\r\nfor i in range(256):\r\n\r\n prob[i]=his[i]/(imgArr.shape[0]*imgArr.shape[1])\r\n\r\nplt.plot(prob)\r\n\r\nplt.show()\r\n\r\n#show a picture of probability before operation\r\n\r\nplt.plot(his)\r\n\r\nplt.show()\r\n\r\n#show a histogram before operation\r\n\r\nE=0\r\n\r\nC=0\r\n\r\nfor j in range(256):\r\n\r\n A=0\r\n\r\n B=0\r\n\r\n \r\n\r\n for i in range(j):\r\n\r\n if prob[i]>0:\r\n\r\n A=A-prob[i]*math.log(prob[i])\r\n\r\n for i in range(256-j):\r\n\r\n if prob[i]>0:\r\n\r\n B=B-prob[i]*math.log(prob[i])\r\n\r\n if C<(A+B):\r\n\r\n C=A+B\r\n\r\n D=j\r\n\r\ncu=np.zeros(256)\r\n\r\nfor i in range(256):\r\n\r\n if i==0:\r\n\r\n cu[i]=prob[i]*255\r\n\r\n else:\r\n\r\n cu[i]=cu[i-1]+prob[i]*255 \r\n\r\nP=0\r\n\r\nfor i in range(imgArrr.shape[0]):\r\n\r\n for j in range(imgArrr.shape[1]):\r\n\r\n a= cu[imgArrr[i,j]]+0.5 \r\n\r\n imgArrr[i,j]=int(a)\r\n\r\n#operation\r\n\r\nImage.fromarray(imgArrr).show() \r\n\r\nhis2 = np.zeros(256)\r\n\r\nprob = np.zeros(256) \r\n\r\nfor i in range(imgArr.shape[0]):\r\n\r\n for j in range(imgArr.shape[1]):\r\n\r\n his2[imgArrr[i, j]] = his2[imgArrr[i, j]] + 1 \r\n\r\nfor i in range(256):\r\n\r\n prob[i]=his2[i]/(imgArr.shape[0]*imgArr.shape[1])\r\n\r\nplt.plot(prob)\r\n\r\nplt.show()\r\n\r\n#show a picture of probability after operation\r\n\r\nplt.plot(his2)\r\n\r\nplt.show()\r\n\r\n#show a histogram after operation","sub_path":"Answers/code/Ansewer5_1.py","file_name":"Ansewer5_1.py","file_ext":"py","file_size_in_byte":2064,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"542177195","text":"\"\"\"\nDescription\n++++++++++++++++++++++\nAddition losses module defines classses which are commonly used particularly in segmentation and are not part of standard pytorch library.\n\nUsage\n++++++++++++++++++++++\nImport the package and Instantiate any loss class you want to you::\n\n from nn_common_modules import losses as additional_losses\n loss = additional_losses.DiceLoss()\n\n Note: If you use DiceLoss, insert Softmax layer in the architecture. In case of combined loss, do not put softmax as it is in-built\n\nMembers\n++++++++++++++++++++++\n\"\"\"\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torch.nn.modules.loss import _Loss, _WeightedLoss\nimport numpy as np\nfrom torch.autograd import Variable\n\n\nclass DiceLoss(_WeightedLoss):\n \"\"\"\n Dice Loss for a batch of samples\n \"\"\"\n\n def forward(self, output, target, weights=None, ignore_index=None, binary=False):\n \"\"\"\n Forward pass\n\n :param output: NxCxHxW logits\n :param target: NxHxW LongTensor\n :param weights: C FloatTensor\n :param ignore_index: int index to ignore from loss\n :param binary: bool for binarized one chaneel(C=1) input\n :return: torch.tensor\n \"\"\"\n output = F.softmax(output, dim=1)\n if binary:\n return self._dice_loss_binary(output, target)\n return self._dice_loss_multichannel(output, target, weights, ignore_index)\n\n @staticmethod\n def _dice_loss_binary(output, target):\n \"\"\"\n Dice loss for one channel binarized input\n\n :param output: Nx1xHxW logits\n :param target: NxHxW LongTensor\n :return:\n \"\"\"\n eps = 0.0001\n\n intersection = output * target\n numerator = 2 * intersection.sum(0).sum(1).sum(1)\n denominator = output + target\n denominator = denominator.sum(0).sum(1).sum(1) + eps\n loss_per_channel = 1 - (numerator / denominator)\n\n return loss_per_channel.sum() / output.size(1)\n\n @staticmethod\n def _dice_loss_multichannel(output, target, weights=None, ignore_index=None):\n \"\"\"\n Forward pass\n\n :param output: NxCxHxW Variable\n :param target: NxHxW LongTensor\n :param weights: C FloatTensor\n :param ignore_index: int index to ignore from loss\n :param binary: bool for binarized one chaneel(C=1) input\n :return:\n \"\"\"\n eps = 0.0001\n encoded_target = output.detach() * 0\n\n if ignore_index is not None:\n mask = target == ignore_index\n target = target.clone()\n target[mask] = 0\n encoded_target.scatter_(1, target.unsqueeze(1), 1)\n mask = mask.unsqueeze(1).expand_as(encoded_target)\n encoded_target[mask] = 0\n else:\n encoded_target.scatter_(1, target.unsqueeze(1), 1)\n\n if weights is None:\n weights = 1\n\n intersection = output * encoded_target\n numerator = 2 * intersection.sum(0).sum(1).sum(1)\n denominator = output + encoded_target\n\n if ignore_index is not None:\n denominator[mask] = 0\n denominator = denominator.sum(0).sum(1).sum(1) + eps\n loss_per_channel = weights * (1 - (numerator / denominator))\n\n return loss_per_channel.sum() / output.size(1)\n\n\nclass IoULoss(_WeightedLoss):\n \"\"\"\n IoU Loss for a batch of samples\n \"\"\"\n\n def forward(self, output, target, weights=None, ignore_index=None):\n \"\"\"Forward pass\n \n :param output: shape = NxCxHxW\n :type output: torch.tensor [FloatTensor]\n :param target: shape = NxHxW\n :type target: torch.tensor [LongTensor]\n :param weights: shape = C, defaults to None\n :type weights: torch.tensor [FloatTensor], optional\n :param ignore_index: index to ignore from loss, defaults to None\n :type ignore_index: int, optional\n :return: loss value\n :rtype: torch.tensor\n \"\"\"\n\n output = F.softmax(output, dim=1)\n\n eps = 0.0001\n encoded_target = output.detach() * 0\n\n if ignore_index is not None:\n mask = target == ignore_index\n target = target.clone()\n target[mask] = 0\n encoded_target.scatter_(1, target.unsqueeze(1), 1)\n mask = mask.unsqueeze(1).expand_as(encoded_target)\n encoded_target[mask] = 0\n else:\n encoded_target.scatter_(1, target.unsqueeze(1), 1)\n\n if weights is None:\n weights = 1\n\n intersection = output * encoded_target\n numerator = intersection.sum(0).sum(1).sum(1)\n denominator = (output + encoded_target) - (output * encoded_target)\n\n if ignore_index is not None:\n denominator[mask] = 0\n denominator = denominator.sum(0).sum(1).sum(1) + eps\n loss_per_channel = weights * (1 - (numerator / denominator))\n\n return loss_per_channel.sum() / output.size(1)\n\n\nclass CrossEntropyLoss2d(_WeightedLoss):\n \"\"\"\n Standard pytorch weighted nn.CrossEntropyLoss\n \"\"\"\n\n def __init__(self, weight=None):\n super(CrossEntropyLoss2d, self).__init__()\n self.nll_loss = nn.CrossEntropyLoss(weight)\n\n def forward(self, inputs, targets):\n \"\"\"\n Forward pass\n\n :param inputs: torch.tensor (NxC)\n :param targets: torch.tensor (N)\n :return: scalar\n \"\"\"\n return self.nll_loss(inputs, targets)\n\n\nclass CombinedLoss(_Loss):\n \"\"\"\n A combination of dice and cross entropy loss\n \"\"\"\n\n def __init__(self):\n super(CombinedLoss, self).__init__()\n self.cross_entropy_loss = CrossEntropyLoss2d()\n self.dice_loss = DiceLoss()\n self.focal_loss = FocalLoss()\n self.l2_loss = nn.MSELoss()\n\n def forward(self, input, target, weight=None):\n \"\"\"\n Forward pass\n\n :param input: torch.tensor (NxCxHxW)\n :param target: torch.tensor (NxHxW)\n :param weight: torch.tensor (NxHxW)\n :return: scalar\n \"\"\"\n # input_soft = F.softmax(input, dim=1)\n target = target.type(torch.long)\n y_2 = torch.mean(self.dice_loss(input, target))\n if weight is None:\n y_1 = torch.mean(self.cross_entropy_loss.forward(input, target))\n else:\n y_1 = torch.mean(\n torch.mul(self.cross_entropy_loss.forward(input, target), weight))\n return y_1 + y_2\n\n\nclass CombinedLoss_KLdiv(_Loss):\n \"\"\"\n A combination of dice and cross entropy loss\n \"\"\"\n\n def __init__(self):\n super(CombinedLoss_KLdiv, self).__init__()\n self.cross_entropy_loss = CrossEntropyLoss2d()\n self.dice_loss = DiceLoss()\n\n def forward(self, input, target, weight=None):\n \"\"\"\n Forward pass\n\n \"\"\"\n input, kl_div_loss = input\n # input_soft = F.softmax(input, dim=1)\n y_2 = torch.mean(self.dice_loss(input, target))\n if weight is None:\n y_1 = torch.mean(self.cross_entropy_loss.forward(input, target))\n else:\n y_1 = torch.mean(\n torch.mul(self.cross_entropy_loss.forward(input, target), weight))\n return y_1, y_2, kl_div_loss\n\n\n# Credit to https://github.com/clcarwin/focal_loss_pytorch\nclass FocalLoss(nn.Module):\n \"\"\"\n Focal Loss for Dense Object Detection\n \"\"\"\n\n def __init__(self, gamma=2, alpha=None, size_average=True):\n\n super(FocalLoss, self).__init__()\n self.gamma = gamma\n self.alpha = alpha\n if isinstance(alpha, (float, int)):\n self.alpha = torch.Tensor([alpha, 1 - alpha])\n if isinstance(alpha, list):\n self.alpha = torch.Tensor(alpha)\n self.size_average = size_average\n\n def forward(self, input, target):\n \"\"\"Forward pass\n\n :param input: shape = NxCxHxW\n :type input: torch.tensor\n :param target: shape = NxHxW\n :type target: torch.tensor\n :return: loss value\n :rtype: torch.tensor\n \"\"\"\n\n if input.dim() > 2:\n # N,C,H,W => N,C,H*W\n input = input.view(input.size(0), input.size(1), -1)\n input = input.transpose(1, 2) # N,C,H*W => N,H*W,C\n input = input.contiguous().view(-1, input.size(2)) # N,H*W,C => N*H*W,C\n target = target.view(-1, 1)\n\n logpt = F.log_softmax(input, dim=1)\n logpt = logpt.gather(1, target)\n logpt = logpt.view(-1)\n pt = Variable(logpt.data.exp())\n\n if self.alpha is not None:\n if self.alpha.type() != input.data.type():\n self.alpha = self.alpha.type_as(input.data)\n at = self.alpha.gather(0, target.data.view(-1))\n logpt = logpt * Variable(at)\n\n loss = -1 * (1 - pt) ** self.gamma * logpt\n if self.size_average:\n return loss.mean()\n else:\n return loss.sum()\n\n\nclass KLDCECombinedLoss(nn.Module):\n \"\"\"\n Combined loss of KL-Divergence and CrossEntropy.\n \"\"\"\n\n def __init__(self, gamma_value=1, beta_value=1.1):\n super(KLDCECombinedLoss, self).__init__()\n self.cross_entropy_loss = CrossEntropyLoss2d()\n self.dice_loss = DiceLoss()\n self.beta_value = beta_value\n self.gamma_value = gamma_value\n\n def forward(self, inp, target, weight=(None, None)):\n \"\"\"\n\n :param inp: tuple with (prior, posterior, predicted_y), prior, posterior can be dict for multi-layer KLDiv.\n :param target: Tensor (Ground truth)\n :param weight: Tuple, (None, None) | (False, False) | (weights, class_weights) and any mix\n :return: dice_loss, CE_loss, KL_div_loss, total_loss\n \"\"\"\n prior, posterior, y_p = inp\n if target is not None:\n target = target.type(torch.long)\n\n dice_loss = torch.tensor([0]).type(torch.FloatTensor)\n cross_entropy_loss = torch.tensor([0]).type(torch.FloatTensor)\n kl_div_loss = torch.tensor([0]).type(torch.FloatTensor)\n criterion = nn.KLDivLoss(reduction='batchmean')\n w, cw = weight\n if w is None:\n dice_loss = torch.mean(self.dice_loss(y_p, target))\n elif w is not False:\n dice_loss = torch.mean(torch.mul(self.dice_loss(y_p, target), w))\n\n if cw is None:\n cross_entropy_loss = torch.mean(self.cross_entropy_loss.forward(y_p, target))\n elif cw is not False:\n cross_entropy_loss = torch.mean(\n torch.mul(self.cross_entropy_loss.forward(y_p, target), cw))\n\n if prior is not None and posterior is not None:\n if type(prior) is dict and type(posterior) is dict:\n for i, j in zip(prior, posterior):\n # kl_div_loss += criterion(F.log_softmax(posterior[j].type(torch.FloatTensor), dim=0),\n # F.softmax(prior[i].type(torch.FloatTensor), dim=0))\n kl_div_loss += self.loss_to_normal(posterior[j]) + self.loss_to_normal(prior[i])\n else:\n\n # kl_div_loss = criterion(F.log_softmax(posterior.type(torch.FloatTensor), dim=0),\n # F.softmax(prior.type(torch.FloatTensor), dim=0))\n kl_div_loss += self.loss_to_normal(posterior) + self.loss_to_normal(prior)\n\n if posterior is not None and prior is None:\n kl_div_loss = posterior\n\n dice_loss = dice_loss.cuda(0)\n cross_entropy_loss = cross_entropy_loss.cuda(0)\n kl_div_loss = kl_div_loss.cuda(0)\n\n cumulative_loss = dice_loss + cross_entropy_loss + kl_div_loss\n\n cumulative_loss = cumulative_loss.cuda(0)\n\n return dice_loss, cross_entropy_loss, kl_div_loss, cumulative_loss\n\n def loss_to_normal(self, tup):\n mu, logvar = tup\n mu, logvar = mu.type(torch.FloatTensor), logvar.type(torch.FloatTensor)\n KLD_ = -0.5 * torch.sum(1 + logvar - mu.pow(2) - logvar.exp())\n return KLD_\n \nclass KLDivLossFunc(nn.Module):\n\n def __init__(self, beta_value=1):\n super(KLDivLossFunc, self).__init__()\n self.beta_value = beta_value\n\n def forward(self, inp, target):\n \"\"\"Forward pass\n :param inp:\n :type inp: input data tensor\n :param target: shape = NxHxW\n :type target: torch.tensor\n :return: combined loss value\n :rtype: torch.tensor\n \"\"\"\n criterion = nn.KLDivLoss(reduction='batchmean')\n kldivloss = criterion(F.log_softmax(target.type(torch.FloatTensor), dim=0),\n F.softmax(inp.type(torch.FloatTensor), dim=0)).cuda()\n\n return kldivloss\n\n @staticmethod\n def loss_to_normal(z_mu, z_var):\n kl_loss = 0.5 * torch.sum(torch.exp(z_var) + z_mu ** 2 - 1. - z_var)\n return kl_loss\n\n","sub_path":"build/lib/nn_common_modules/losses.py","file_name":"losses.py","file_ext":"py","file_size_in_byte":12797,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"19533145","text":"\"\"\"\n本题思路比较简单,首先要满足的条件是所有汽油的量要大于等于路上消耗的量,否则不可能跑完全程\n然后就是开始找一个起点,当某个点的汽油量足以支撑到下一个点时,认为可以开始.然后向后走,看能否走完全程,走不完就找到下一个可以开始的点继续.\n但是这么暴力的遍历的话,时间复杂度是n^2,可以用合适的贪心算法去优化.\n可以稍微想一想:如果从A走到B失败了,其实从A到A+1资源剩余是大于等于0的,那么从AB之间的任一点出发到B的资源更为紧张,所以从AB之间任一点出发都不可能到达B了,\n接下来选定B作为起始点即可.这样大大减少了起始点的数量,也省了不少时间\n\"\"\"\nclass Solution(object):\n def canCompleteCircuit(self, gas, cost):\n \"\"\"\n :type gas: List[int]\n :type cost: List[int]\n :rtype: int\n \"\"\"\n if sum(gas) - sum(cost) < 0:\n return -1\n l = len(gas)\n i = 0\n while i < l:\n if gas[i] - cost[i] >= 0:\n remain = 0\n j = 0\n start = i\n for j in range(i, i + l):\n remain += (gas[j % l] - cost[j % l])\n if remain < 0:\n i = j % l\n if i < start:\n return -1\n break\n if j == start + l - 1:\n return start\n i += 1\n return -1\n\n\nif __name__ == '__main__':\n sol = Solution()\n gas = [1, 2, 3, 4, 5]\n cost = [3, 4, 5, 1, 2]\n print(sol.canCompleteCircuit(gas, cost))\n","sub_path":"101-200/134.py","file_name":"134.py","file_ext":"py","file_size_in_byte":1681,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"184074934","text":"# 实现一个简单的商城购物系统\nfrom bll import handles\nfrom dal import shopping_goods_data\n\n\ndef shopping():\n\n\n prompt = \"您好,欢迎使用薯条橙子在线购物系统chipscoco,输入<>中对应的指令来使用购物系统:\\n\" \\\n \"<1>:查看所有商品\\n<2>:对商品按售价进行排序(asc表示升序,desc表示降序)\\n\" \\\n \"<3>:添加商品到购物车\\n<4>:查看购物车\\n<5>:删除购物车指定商品\\n<6>:下单结账\\n<0>:退出系统\"\n\n commands = {1: handles.show_all_goods, 2: handles.sort_goods, 3: handles.add_goods, 4: handles.show_shopping_cart,\n 5: handles.remove_goods, 6: handles.shopping_cart_paybill }\n # commands 是数字编号+函数的内存地址\n\n while True:\n print(prompt)\n command = int(input(\"输入指令:__\\b\\b\"))\n if command in commands:\n commands[command](shopping_goods_data.CHIPSCOCO)\n # 因为把 shopping_goods_data 作为模块分出去以后,这里再调用chipscoco就要\n # 加上模块名了\n elif command == 0:\n break\n else:\n print(\"您输入了非法的指令\")\n input(\"按下键盘任意键,继续使用系统......\")\n\n\nif __name__ == \"__main__\":\n shopping()\n","sub_path":"chipsprange/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":1290,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"466803598","text":"import random\r\n\r\nx = random.randint(1, 6)\r\ny = random.random()\r\n\r\nmyList = ['rock', 'paper', 'scissors']\r\nresult = random.choice(myList)\r\nprint(result)\r\ncards = [1,2,3,4,5,6,7,8,9, \"J\", \"Q\", \"K\", \"A\"]\r\n\r\nrandom.shuffle(cards)\r\n\r\nprint(cards)","sub_path":"Random.py","file_name":"Random.py","file_ext":"py","file_size_in_byte":241,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"407926796","text":"import sys\nimport argparse\nimport time\n\nsys.path.append(\"/home/hadoop/hiro_tests/\")\nimport hiroStatAndMatchDefault as hiro\n\nparser = argparse.ArgumentParser(description=\"Outputs default Hiro stat and match tests from input file\")\nparser.add_argument(\"--s3loc\", help=\"Match file. Should be only one column\")\nparser.add_argument(\"--ticket\", help=\"The relevant JIRA ticket. To be used for saving match_test outputs\")\nparser.add_argument(\"--test\", default=\"all\", choices=[\"stat\",\"match\",\"all\"], help=\"Type of test desired. Choices: [stat,match,all]. Default is all\")\nparser.add_argument(\"--htype\", default=\"md5\", choices=[\"md5\",\"sha1\",\"sha2\"], help=\"Type of hem in the original client file\")\nargs = parser.parse_args()\n\nif args.ticket:\n ticket = args.ticket\nelse:\n timeId = str(int(time.time()))\n ticket = 'noTicket_%s' %(timeId)\n\nif args.htype:\n htype = args.htype\nelse:\n htype = 'md5'\n\nclientTest = hiro.defaultTest(clientFile=args.s3loc, ticket=ticket, hem=htype)\n\nif args.test == 'stat':\n clientTest.stat_test()\nelif args.test == 'match':\n clientTest.match_test()\nelse:\n clientTest.stat_test()\n clientTest.match_test()","sub_path":"emr_scripts/hiro_tests/hiroTerminal.py","file_name":"hiroTerminal.py","file_ext":"py","file_size_in_byte":1145,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"254049345","text":"import mpids.MPInumpy as mpi_np\nimport numpy as np\nfrom mpi4py import MPI\nfrom operations import _max, _mean, _sum, _std\n\nif __name__ == '__main__':\n comm = MPI.COMM_WORLD\n rank = comm.Get_rank()\n n_procs = comm.Get_size()\n size = 2**25\n iters = 1\n mpi_np_arr = mpi_np.arange(size, dtype=np.float64)\n\n max_time = _max(mpi_np_arr, iters=iters)\n mean_time = _mean(mpi_np_arr, iters=iters)\n sum_time = _sum(mpi_np_arr, iters=iters)\n std_time = _std(mpi_np_arr, iters=iters)\n\n if rank == 0:\n print(\"mpi_np,max,%d,%d,%.9f\" %(n_procs, size, max_time))\n print(\"mpi_np,mean,%d,%d,%.9f\" %(n_procs, size, mean_time))\n print(\"mpi_np,sum,%d,%d,%.9f\" %(n_procs, size, sum_time))\n print(\"mpi_np,std,%d,%d,%.9f\" %(n_procs, size, std_time))\n","sub_path":"MPInumpy/Strong/reductions_mpi_np.py","file_name":"reductions_mpi_np.py","file_ext":"py","file_size_in_byte":786,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"113151606","text":"from PyQt5.QtCore import Qt, pyqtSlot\n\nfrom resources.ui.python.ParticipantWidget_ui import Ui_frmParticipant\nfrom libopenimu.models.Participant import Participant\nfrom libopenimu.qt.DataEditor import DataEditor\n\n\nclass ParticipantWindow(DataEditor):\n\n participant = Participant()\n dbMan = None\n\n def __init__(self, dbManager, participant=None, parent=None, default_group = None):\n super().__init__(parent=parent)\n self.UI = Ui_frmParticipant()\n self.UI.setupUi(self)\n\n self.participant = participant\n self.dbMan = dbManager\n self.data_type = \"participant\"\n\n # Signals / Slots connections\n self.UI.btnCancel.clicked.connect(self.cancel_clicked)\n self.UI.btnSave.clicked.connect(self.save_clicked)\n self.UI.txtName.textEdited.connect(self.name_edited)\n self.UI.txtDesc.textChanged.connect(self.desc_edited)\n self.UI.cmbGroups.currentIndexChanged.connect(self.group_edited)\n\n # Load groups\n groups = self.dbMan.get_all_groups()\n self.UI.cmbGroups.clear()\n self.UI.cmbGroups.addItem(\"Aucun\", userData=None)\n\n for group in groups:\n self.UI.cmbGroups.addItem(group.name, userData=group.id_group)\n\n # Update data\n self.update_data()\n\n # Set default group for new participants\n if default_group is not None:\n self.UI.cmbGroups.setCurrentIndex(self.UI.cmbGroups.findData(default_group.id_group, Qt.UserRole))\n\n self.enable_buttons(False)\n\n def validate(self):\n rval = True\n if self.UI.txtName.text() == '':\n self.UI.txtName.setStyleSheet('background-color: #ffcccc;')\n rval = False\n else:\n self.UI.txtName.setStyleSheet('background-color: rgba(226, 226, 226, 90%);')\n\n if self.UI.cmbGroups.currentIndex == -1:\n rval = False\n\n return rval\n\n def update_data(self):\n if self.participant is not None:\n self.UI.txtName.setText(self.participant.name)\n self.UI.txtDesc.setPlainText(self.participant.description)\n # if self.participant.group is not None and self.participant.group.name is not None:\n # self.UI.lblGroupValue.setText(self.participant.group.name)\n # else:\n # self.UI.lblGroupValue.setText(\"Aucun\")\n self.UI.cmbGroups.setCurrentIndex(self.UI.cmbGroups.findData(self.participant.id_group))\n else:\n self.UI.txtName.setText(\"\")\n self.UI.txtDesc.setPlainText(\"\")\n self.UI.cmbGroups.setCurrentIndex(0)\n\n def enable_buttons(self, enable):\n self.UI.btnCancel.setEnabled(enable or self.participant is None)\n self.UI.btnSave.setEnabled(enable)\n\n def update_modified_status(self):\n self.enable_buttons(\n (self.participant is not None and self.UI.txtName.text() != self.participant.name) or\n (self.participant is None and self.UI.txtName.text() != \"\") or\n (self.participant is not None and self.UI.txtDesc.toPlainText() != self.participant.description) or\n (self.participant is None and self.UI.txtDesc.toPlainText() != \"\") or\n (self.participant is not None and self.UI.cmbGroups.currentData() != self.participant.id_group)\n )\n @pyqtSlot()\n def save_clicked(self):\n if self.validate():\n if self.participant is None:\n self.participant = Participant()\n self.participant.name = self.UI.txtName.text()\n self.participant.description = self.UI.txtDesc.toPlainText()\n self.participant.id_group = self.UI.cmbGroups.currentData()\n self.participant = self.dbMan.update_participant(self.participant)\n self.enable_buttons(False)\n self.dataSaved.emit()\n\n @pyqtSlot()\n def cancel_clicked(self):\n self.update_data()\n self.dataCancelled.emit()\n\n @pyqtSlot(str)\n def name_edited(self, new_value):\n self.update_modified_status()\n\n @pyqtSlot()\n def desc_edited(self):\n self.update_modified_status()\n\n @pyqtSlot()\n def group_edited(self):\n self.update_modified_status()","sub_path":"python/libopenimu/qt/ParticipantWindow.py","file_name":"ParticipantWindow.py","file_ext":"py","file_size_in_byte":4290,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"483988336","text":"import pytest\n\nfrom .steps.get_games_by_user_uc_steps import ShouldGetTypeGamesByUserSteps\nfrom src.bp.domain import TypeGame\n\nTYPE_GAMES__RETURN_DATA = [\n {\n TypeGame(1, \"game1\", \"image1\", \"ARCADE\", \"PSICO\", \"20\", \"Puntos\"),\n TypeGame(2, \"game2\", \"image2\", \"ARCADE\", \"ING\", \"10\", \"Puntos\"),\n TypeGame(3, \"game3\", \"image3\", \"ARCADE\", \"MED\", \"30\", \"Puntos\"),\n }\n]\n\nUSER_ID_DATA = \"test_user_get_type_games\"\n\n\n@pytest.mark.parametrize(\n \"user_id, type_games\", [(USER_ID_DATA, TYPE_GAMES__RETURN_DATA)]\n)\ndef test_should_get_all_type_question(user_id, type_games):\n steps = ShouldGetTypeGamesByUserSteps()\n steps.given(user_id, type_games)\n steps.when()\n steps.then()\n","sub_path":"tests/bptest/get_type_games_by_user_use_case_test.py","file_name":"get_type_games_by_user_use_case_test.py","file_ext":"py","file_size_in_byte":707,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"449510894","text":"#!/usr/bin/env python\n\nimport serial\nimport rospy\nfrom std_msgs.msg import String\nfrom geometry_msgs.msg import Twist\nimport time\nleft_speed = \"\"\nright_speed = \"\"\nmotor_1_string = \"\"\nmotor_2_string = \"\"\nmotor_max_val = 2047; #define the maximum speed for a given motor\n\nser = serial.Serial(port = \"/dev/ttyACM0\",baudrate = 9600)\nif ser.isOpen:\n\tser.write(\"M1: \"+str(0)+\"\\r\\n\") #ensure that motors are stopped when the node is initialized\n\tser.write(\"M2: \"+str(0)+\"\\r\\n\") #ensure that motors are stopped when the node is initialized\nser.close()\nser.open()\n\ndef message_cb(data):\n# rospy.loginfo(rospy.get_caller_id() + \"I heard %s\", data.data)\n left_speed = (data.linear.x - data.angular.z)*2047;\n left_speed = min(motor_max_val,left_speed); #set to max if it goes over\n left_speed = max(-motor_max_val,left_speed); #set to max if it goes over\n ser.write(\"M1: \" + str(int(left_speed)) + \"\\r\\n\")\n # rospy.loginfo(\"left_speed: \" + str(left_speed));\n\n # rospy.loginfo(rospy.get_caller_id() + \"I heard %s\", data.data)\n right_speed = (data.linear.x + data.angular.z)*2047;\n right_speed = min(motor_max_val,right_speed); #set to max if it goes over\n right_speed = max(-motor_max_val,right_speed); #set to max if it goes over\n ser.write(\"M2: \" + str(int(right_speed)) + \"\\r\\n\")\n # rospy.loginfo(\"right_speed: \" + str(right_speed));\n\ndef listener():\n rospy.init_node('sabertooth_control', anonymous=True)\n rospy.Subscriber(\"cmd_vel\", Twist, message_cb)\n\n # spin() simply keeps python from exiting until this node is stopped\n rospy.spin()\n\nif __name__ == '__main__':\n #with serial.Serial(port = \"/dev/ttyACM0\",baudrate = 9600) as ser: \n# while True:\n listener()\n # motor_1_string = \"M1: \" + left_speed + \"\\r\\n\"\n # motor_2_string = \"M2: \" + right_speed + \"\\r\\n\"\n #ser.write(\"M1: \"+left_speed+\"\\r\\n\")\n\t #ser.write(\"M2: \"+right_speed+\"\\r\\n\")\n\t #ser.write(motor_1_string)\n #ser.write(motor_2_string)\n","sub_path":"plow_motor_control_py/scripts/sabertooth_drive.py","file_name":"sabertooth_drive.py","file_ext":"py","file_size_in_byte":1985,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"361722658","text":"import numpy as np\n\n\nclass MultiLayerPerceptron():\n def __init__(self, inputs, targets, n_hidden, beta):\n self.n_data = len(inputs)\n self.inputs = np.concatenate((inputs, -np.ones((self.n_data, 1))), axis=1)\n self.targets = targets\n self.n_in = inputs.shape[1] # Number of attributes\n self.n_out = targets.shape[1] # Number of outputs\n self.n_hidden = n_hidden # Number of hidden neurons\n self.beta = beta\n\n self.weights1 = np.random.randn(self.n_in + 1, n_hidden) * 0.1 - 0.05\n self.weights2 = np.random.randn(self.n_hidden + 1,\n self.n_out) * 0.1 - 0.05\n\n def train(self, eta, iterations):\n\n for n in range(iterations):\n\n self.outputs = self.fwd(self.inputs)\n error = 0.5 * sum((self.targets - self.outputs)**2)\n print(\"Iteration: \" + str(n) + '\\t' + \"Error: \" + str(error))\n\n deltao = (self.outputs - self.targets) * self.outputs * (1.0 - self.outputs)\n deltah = self.hidden * (1.0 - self.hidden) * \\\n np.dot(-deltao, np.transpose(self.weights2))\n\n updatew1 = np.zeros((np.shape(self.weights1)))\n updatew2 = np.zeros((np.shape(self.weights2)))\n\n updatew1 = eta * np.dot(np.transpose(self.inputs), deltah[:, :-1])\n updatew2 = eta * np.dot(np.transpose(self.hidden), deltao)\n\n self.weights1 += updatew1\n self.weights2 -= updatew2\n\n def fwd(self, inputs):\n\n # If inputs are not the same as inputs stored in the nn object add column of ones\n if not np.array_equal(inputs, self.inputs):\n inputs = np.concatenate((inputs, -np.ones((len(inputs), 1))), axis=1)\n\n self.hidden = np.dot(inputs, self.weights1)\n self.hidden = 1.0 / (1.0 + np.exp(-self.beta * self.hidden))\n self.hidden = np.concatenate((self.hidden, -np.ones((inputs.shape[0], 1))), axis=1)\n outputs = np.dot(self.hidden, self.weights2)\n outputs = 1.0 / (1.0 + np.exp(-self.beta * outputs))\n\n return outputs\n\n def confmat(self, inputs, outputs):\n \"\"\"\n UNFINISHED\n \"\"\"\n\n nn_outputs = np.where(self.fwd(inputs) > 0.5, 1, 0)\n\n return nn_outputs\n","sub_path":"src/ml_algorithms/nn.py","file_name":"nn.py","file_ext":"py","file_size_in_byte":2262,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"343454405","text":"\"\"\"\nNumber Bases\n_________________________________\n\nIt's the \"language\" that a number is written down in\n\nDouze(french) == (english)Twelve\nsame thing!!\n\n1100(binary) == 12 \nsame thing!!\n\nbase 2: binary\nbase 8: Octal(rarely used)\nbase 10: decimal (what we know from grade school)\nbase 16: hexadecimal \"hex\"\nbase 64: base 64\n\nbase 10(decimal)\n\n\n\n\n|<----- 1000's place 10^3\n||<----- 100's place 10^2\n|||<---- 10's place 10^1\n||||<---- 1's place 10^0\nabcd\n1234\n1 1000\n2 100s\n3 10s\n4 1s\n\n1234 = 1 * 1000 + 2 * 100 + 3 + 10 + 4 * 1\n ^ ^ ^ ^\n\n\nFor Hexadecimal, our digits are represented\nby 0-9, and then A-F\nand each place value is the next 16th power\n\nHere's a binary conversion\n\n|<----- 8's place 2^3\n||<----- 4's place 2^2\n|||<---- 2's place 2^1\n||||<---- 1's place 2^0\nabcd\n\n0011 binary\n\n0011 binary == 0 * 8 + 0 * 4 + 1 * 2 + 1 * 1 == 3 decimal\n\n*** FOR ANY NUMBER BASE, YOU CAN PUT ANY NUMBER OF LEADING 0'S***\n\nbinary digits == (\"bit\")\n\n8 bits == \"byte\"\n\nOne byte is the most common standard unit of memory used for our CPU\n\n4 bits == \"nybble\"\n\nThe number base only matters\nwhen you write the number down.\nOnce it's inside of a machine, the\nnumber base doesn't matter.\n\nDefault languages work in base 10 most of the time.\n\nTo specify the base in code:\nPrefix\n______\n[none] decimal\n0b binary\n0x hex\n0o octal\n\nbases only matter when you want to print it out\nall numeric values are simply numeric values\nwritten in one language or another\n\nAs the base number gets bigger, \nthe amount of digits required to represent\na numerical value goes down.\n\n4 bits(One nybble) are(is) required to store\none hex digit.\n\n\nConverting from binary to hex is very easy!\nNybbles align the digit counting between binary and hexadecimal very well.\nWhen converting, just chop up a binary number into 4's\nand then translate!\n\"\"\"\n\n# Beej's Emulator\n# Memory works like a giant array\n# Think of your RAM as a massive Array\n# In the computer, we have a RAM which contains memory\n\n\n# Index into the memory array\n# Address\n# Location\n# Pointer\n# The above all mean the same thing!!\n\n# For Tonight's assignment, do this in base 2 instead of base 10\n# The above code prints all lines of a given file\n# it'll print line by line, but we still need to make sure\n# to avoid all whitespace\n# avoid all comments\n# avoid all blank lines\n# print out errors for non-commands\n\nmemory = [0] * 256\n\n# This is a \"Data Driven\" program.\n# We have to pass in a file to this program in order to get output\n\nimport sys\n\nif len(sys.argv) !=2:\n print(\"usage: comp.py filename\")\n sys.exit(1)\n\ntry:\n address = 0\n\n with open(sys.argv[1]) as f:\n for line in f:\n t = line.split('#')\n n = t[0].strip()\n\n try:\n\n n = int(n)\n except ValueError:\n print(f\"Invalid Number {n}\")\n sys.exit(1)\n if n == '':\n continue\n\n print(repr(n))\n memory[address] = n\n address += 1\n\nexcept FileNotFoundError:\n print(f\"File not found: {sys.argv[1]}\")\n sys.exit()\n\nregister = [0] * 8 # represent r0 - r7\n\n# \"Variables\" in hardware, known as \"registers\"\n# There are a fixed number of registers\n# They have fixed names\n# On the LS8, they're called...\n# R1 R2 R3 R4 R5 R6 R7\n\n\npc = 0# Program Counter, address of the currently-executing instruction\n# Give the register for the stack pointer a symbolic name\n# So that developers know where it is\nSP = 7\nregister[SP] = 0xF4\n\ndef push_value(value):\n # Decrement SP\n register[SP] -= 1\n\n # copy the value to the SP address\n top_of_stack_addr = register[SP]\n memory[top_of_stack_addr] = value\n\n\ndef pop_value():\n \n # Get the top of stack addr\n top_of_stack_addr = register[SP]\n\n # Get the value of the top of stack\n value = memory[top_of_stack_addr]\n\n # Increment SP\n register[SP] += 1\n\n return value\n\n\n\nrunning = True\n\nwhile running:\n ir = memory[pc] # Instruction Register\n # This holds a copy of the currently executing instruction\n \n if ir == 1:\n print(\"David\")\n pc += 1\n elif ir == 2:\n running = False\n\n elif ir == 3: # Save Reg\n reg_num = memory[pc + 1]\n value = memory[pc + 2]\n register[reg_num] = value\n print(register)\n pc += 3\n\n elif ir == 4: # Print_reg\n reg_num = memory[pc + 1]\n print(register[reg_num])\n pc += 2\n\n elif ir == 5: # PUSH\n # Get the reg num to push\n reg_num = memory[pc + 1]\n\n # Get the value to push\n value = register[reg_num]\n\n push_value(value)\n\n # print(memory[0xea:0xf4])\n\n pc += 2\n\n elif ir == 6: # POP\n # Get the reg to pop into\n reg_num = memory[pc + 1]\n\n value = pop_value()\n\n #Store the value in the register\n register[reg_num] = value\n\n # Increment SP\n register[SP] += 1\n\n pc += 2\n\n print(memory[0xea:0xf4])\n\n elif ir == CALL:\n\n # Compute the return addr\n return_addr = pc + 2\n\n # Push the return addr on stack\n push_value(return_addr)\n\n # get the value from the operand reg\n reg_num = memory[pc + 1]\n value = register[reg_num]\n\n # set the pc to that value\n pc = value\n\n else:\n print(f\"Unknown Instruction {ir}\")\n\n\n# For moving the PC, use an if else statement which checks\n# the fourth bit of the instruction. if that fourth bit is\n# true, then the instruction sets the value and we can continue.\n# if that fourth bit is false, then set the pc to the new value.\n\n\n\n# inst_sets_pc = (ir >> 4) & 1 == 1:\n\n\n# if not inst_sets_pc:\n\n#_________________________________________________\n# Instruction location\n# \"POP register\"\n# copy the value from the address pointed to by the stack pointer,\n# put it at the given register\n# increment the SP(Stack works Top-Down)\n#_____________________________________________\n# Instruction location\n# \"PUSH register\"\n# Decrement the stack pointer\n# place the value at the given register\n# Stack pointer points at the item most recently pushed\n# \n\n\n### The above is a very basic emulation\n### Memory has numbers. Those numbers\n# have meaning, and we can tell the computer\n# what those meanings are.\n\n# Interrupts are a stretch goal\n\n\n#______________________________________________________________\n# Frame the Plan from Inputs to Outputs\n# Parsing, Normalize, Sanitize\n# all mean the same thing\n# Take the data, and make it into the same format\n\n\n#____________________________________________________________\n# CPU Stack notes\n\n# These are just like the stack data structure we know\n\n# Push and Pop are standard\n\n# Stack data is stored in RAM\n\n# The \"Stack Pointer\" keeps track of the address of the top of the stack.\n\n# Typically the stackgrows down from the higher memory addresses\n\n#__________________________\n# A Minimal Stack\n\n# A stack needs somewhere to store data: RAM in this case\n\n# A stack needs to keep track of where the top of the stack is: stack pointer\n\n# A Stack needs functionality to push and pop, like always. Push and pop instructions\n\n# In order to store something(PUSH)\n# We first decrement the stack pointer\n# then push the value onto the memory address which the stack pointer is pointing at\n# Decrement stack pointer -> push value to that address\n\n# In order to remove something(POP)\n# We POP it into a Register\n# POP the value at the stack pointer and copy it into Register\n# Increment the stack pointer\n# First, copy the value, POP it onto the register. Then Increment the SP\n# This doesn't remove values from the stack, but it copies them to a register\n# When we push onto the stack, it overrwrites values\n\n# The stack is typically used to store variables\n# Also used to return addresses from a subroutine\n# Storage of registers and CPU state while handling an interrupt\n# Allocation of local variables for a subroutine\n\n# If you PUSH too many items on the stack, you'll begin to \n# overwrite values \n\n# If you POP from an empty stack, you'll copy NONE onto a register\n\n# Check if a stack is empty by trying to POP from it\n\n# What information must be saved on the stack when the CPU is servicing\n# an interrupt? Why? The current state of the processor, and all of it's\n# counters, registers, and flags. This is all saved so that it can\n# handle the interruption and then pick back up where it left off\n\n\n#________________________________________________________________\n# CPU Interrupts\n\n# Interrupts are commonly generated by peripherals(keyboard, mouse)\n# who need to alert the CPU that some work needs to be done.\n\n# When an interrupt occurs, the current state of the processor is saved\n# on the stack, and execution continues at the address of the interrupt handler.\n\n# Most CPU's have a lookup table: Interrupt Vector Table.\n# This is an array of interrupts to tell the PC how to handle each one.\n# It's an array of pointers to handlers, one per interrupt.\n# Different CPU's keep the table in different areas of RAM\n\n#______________________________________________________\n# Beej's notes\n\n# Stacks are ALWAYS USED for CPU's\n\n# Stack is good for:\n\n# Temporarily storing values\n# Making subroutines possible\n# Implementing local variables\n\n\n# It's easy to implement in the CPU hardware\n\n\n# Conditionals are what we're missing from our emulator\n\n# How does the stack work?\n\n# Low level Stack concept will be used. Boil the stack down\n# to it's purest essence. In the case of a CPU, we just need\n# to be able to push and pop. memory and location is handled by RAM.\n# The stack pointer points to the top of the stack.\n# The stack pointer is a general purpose register\n\n# Stacks start from TOP-DOWN\n# if it's empty, it always points to the top(f4)\n#_________________________________________________\n# Instruction location\n# \"POP register\"\n# copy the value from the address pointed to by the stack pointer,\n# put it at the given register\n# increment the SP(Stack works Top-Down)\n#_____________________________________________\n# Instruction location\n# \"PUSH register\"\n# Decrement the stack pointer\n# place the value at the given register\n# Stack pointer points at the item most recently pushed\n#\n\n\n#____________________________________________________________________________\n#################### Subroutines #############################\n\n# Think of subroutines in a CPU as functions in higher-level languages\n\n# In assembly, we \"CALL\" a subroutine at a particular address.\n\n# Then we \"RET\"(return) from that subroutine to pick up where we left off, just like a function\n# does in a higher level language.\n\n####### Limitations with assembly-level subroutines\n# CPU's are pretty simple machines.\n# No arguments. Only takes one operand\n# No return values. \n# These can be implemented in a variety of ways (as you'll learn)\n\n###### Use the stack!!!\n\n# When we call a subroutine, we need to store the return address somewhere so we\n# know where to go when we hit the \"RET\" instruction.\n\n# CPU's use a stack for this.\n\n# Call will push the address of the instruction after it onto the stack, then move \n# the PC to the subroutine address\n\n# RET will ppo the return address off of the stack, and store it in the PC.\n\n# Use any place you'd use functions in a higher-level language\n\n## DRY principle\n\n# High-level languages eventually use CALL and RET deep down to implement functions.\n\n# The stack is used to store the return address so that we can remember to advance \n# through our PC instead of entering an infinite loop\n\n# Since stacks are first in first out, we can create local variables by pushing \n# more values onto the stack for local variable use, and popping them off one by one\n# until we are left with the final stack pop to return our subroutine\n\n# arguments could be passed to subroutines by adding them as instructions\n\n#_____________________________________________________________\n# beej's notes\n# Subroutines are functions\n# but you can't pass anything in\n# and they cant return anything\n\n\n#----------------------------\n# def foo():\n# print(\"foo 1\")\n\n# return\n\n# def bar():\n# print(\"bar 1\")\n# foo()\n# print(\"bar 2\")\n\n# return\n\n# print(\"main 1\")\n# bar()\n# print(\"main 2\")\n#_______________________________\n\n# The above is a good example of how a CPU keeps track of where they are\n# in a calling process. They assign addresses to different commands.\n\n\n#CALL:\n # push return address on stack. This is the instruction which follows the Call instruction\n # set pc to address of subroutine\n\n# RET:\n # pop the return address value from the stack\n # assign the pc to that value\n\n# When you call:\n # Allocate a stack frame\n # stack frame is the return address and the locals\n\n\n# When you return:\n # Deallocate (pop) that stack frame\n # set the pc to the return address\n\n","sub_path":"notes.py","file_name":"notes.py","file_ext":"py","file_size_in_byte":12908,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"508638845","text":"\"\"\"\nSome ASNI escape codes for use in terminal output formatting.\nSee https://www.lihaoyi.com/post/BuildyourownCommandLinewithANSIescapecodes.html\n\"\"\"\n\n# Colors\n_YELLOW = \"\\u001b[33m\"\n\n# Decorations\n_BOLD = \"\\u001b[1m\"\n\n# Aliases\nOK = _BOLD\nRESET = \"\\u001b[0m\"\nWARNING = _YELLOW + _BOLD\n","sub_path":"comparetransactionsets/terminalcolors.py","file_name":"terminalcolors.py","file_ext":"py","file_size_in_byte":287,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"559667862","text":"import netCDF4 as nc\nfrom tensorflow import keras\nimport numpy as np\nfrom keras.models import load_model\nfrom keras.callbacks import ModelCheckpoint, EarlyStopping\n\n\n# import mask to define catchment boundaries\nmask = np.load('mask.npy')\nmask_inv = np.invert(mask)\n\n\n# # # PREPROCESS LABELS, UPPER ZONE SOIL MOISTURE # # #\n\n# load label data\nfn = '/home/WUR/keppl001/MScThesis_env/data/outmaps.nc'\nlabel_data = nc.Dataset(fn)\n\n# select labels and apply mask to fill area outside catchment with nan\nsoil_moisture = label_data['ust_0_']\nsoil_moisture = np.ma.filled(soil_moisture, fill_value = np.nan)\nsoil_moisture = np.ma.getdata(soil_moisture)\nsoil_moisture[mask_inv] = np.nan\n\n# loop through dataset to fill values outside catchment with median per timestep\nfor i in np.arange(len(soil_moisture)):\n median = np.nanmedian(soil_moisture[i])\n soil_moisture[i] = np.nan_to_num(soil_moisture[i], copy=False, nan = median)\nlabels = np.expand_dims(soil_moisture, axis = 1)\n\n# define train and validation period\ntrain_window = [0, 11323]\nval_window = [11323, 13514]\n\nlag = 10\n \n# load train and validation features\nfeatures_train = 'PATH'\nfeatures_val = 'PATH' \nfeatures_train = np.load(features_train)\nfeatures_val = np.load(features_val)\n\n# reshape labels to fit shape (BATCH, HEIGHT, WIDTH, CHANNEL)\nlabels_train = labels[train_window[0] + lag:train_window[1]]\nlabels_val = labels[val_window[0] + lag:val_window[1]]\n \nlabels_train = np.reshape(labels_train, (features_train.shape[0], 91, 134, 1))\nlabels_val = np.reshape(labels_val, (features_val.shape[0], 91, 134, 1))\n\n\n# # # MODEL TRAINING # # #\n\n# load S2 model\nmodel = load_model('models/model_stacked_2')\n\n# define optimizer and lr\nopt = keras.optimizers.Adam(learning_rate=0.001)\nmodel.compile(loss = 'mean_squared_error', optimizer = opt, metrics = 'mse')\n\n# add model checkpoints to save model after improvement\nmc = ModelCheckpoint('model_stacked_2.h5', monitor='val_loss', mode='min', save_best_only=True, verbose=1)\nes = EarlyStopping(monitor='val_loss', mode='min', verbose=1, patience=100)\n\n# train model\nhistory = model.fit(x = features_train, y = labels_train, validation_data = (features_val, labels_val), epochs = 250,\n batch_size = 2, verbose = 2, callbacks=[es, mc], shuffle = False, initial_epoch=29)\n\n# save learning\narray_hist = np.array(list(history.history.values())).transpose()\nnp.save('history_stacked_2.csv', array_hist)","sub_path":"training/runfile_stacked_2.py","file_name":"runfile_stacked_2.py","file_ext":"py","file_size_in_byte":2427,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"24912075","text":"#!/usr/bin/env python\n\nimport os, sys\nimport collections\nimport numpy as np\nimport cv2\nimport math\nimport random\nimport time\n\nimport torch\nimport torch.nn as nn\nimport torch.nn.parallel\nimport torch.backends.cudnn as cudnn\nimport torch.optim\nimport torch.utils.data\nimport torchvision.transforms as transforms\nimport torchvision.datasets as datasets\nimport csv\nimport json,pickle\n\nsys.path.insert(0, \"../\")\nimport models\nfrom VideoSpatialPrediction import VideoSpatialPrediction\n\nos.environ[\"CUDA_DEVICE_ORDER\"]=\"PCI_BUS_ID\" \nos.environ[\"CUDA_VISIBLE_DEVICES\"]=\"0,1,2,3\"\n\ndef softmax(x):\n y = [math.exp(k) for k in x]\n sum_y = math.fsum(y)\n z = [k/sum_y for k in y]\n\n return z\n\ndef def_my_result(spat_prediction,layers = 2,topk = 5):\n sort_order = np.argsort(spat_prediction,axis=0)\n input_img_num_fromsingmp4 = np.argsort(spat_prediction,axis=0).shape[1] #int 250\n pre_result = sort_order[-layers:,:]\n finalpredict = np.reshape(pre_result, (1, input_img_num_fromsingmp4*layers))[0].tolist()\n count = np.zeros(90)\n for i,label in enumerate(finalpredict):\n count[label]=count[label]+1\n final_num = np.sort(count)[-topk:]\n final_label = np.argsort(count)[-topk:]\n return final_label ,final_num\n\ndef write_json(mp4_name,label,score,class_list):\n single_result = []\n single_result.append(mp4_name)\n temp_single_result = []\n for i in range(len(label)):\n temp_single_result.append({\"label\": class_list[label[-i-1]][:-1], \"score\": float('%.6f' % score[-i-1])})\n final_result['results'][mp4_name] = temp_single_result\n\n with open(\"./result.json\", \"w\") as file:\n json.dump(final_result, file)\n file.close()\n\n\n\ndef main():\n\n model_path ='/home/thl/Desktop/challeng/checkpoints/model_best.pth.tar'\n class_name_file = '/home/thl/Desktop/challeng/datasets/settings/class_name.txt'\n class_list = []\n for line in open(class_name_file, \"r\"):\n class_list.append(line)\n\n start_frame = 0\n num_categories = 90\n\n model_start_time = time.time()\n params = torch.load(model_path)\n\n spatial_net = models.rgb_vgg16(pretrained=False, num_classes=90)\n if torch.cuda.is_available():\n spatial_net = torch.nn.DataParallel(spatial_net)\n spatial_net.load_state_dict(params['state_dict'])\n spatial_net.cuda()\n spatial_net.eval()\n model_end_time = time.time()\n model_time = model_end_time - model_start_time\n print(\"Action recognition model is loaded in %4.4f seconds.\" % (model_time))\n\n val_file_dir = '/home/thl/Desktop/challeng/datasets/settings/test_set.txt'\n val_list = []\n for line in open(val_file_dir, \"r\"):\n val_list.append(line)\n\n print(\"we got %d test videos\" % len(val_list))\n\n line_id = 1\n\n result_list = []\n for line in val_list:\n clip_path ='/home/thl/Desktop/challeng/datasets/frame_and_flow/test/'+line[:-1]\n spatial_prediction = VideoSpatialPrediction(\n clip_path,\n spatial_net,\n num_categories,\n start_frame)\n\n final_lab,final_num= def_my_result(spatial_prediction, layers=2)\n # avg_spatial_pred_fc8 = np.mean(spatial_prediction, axis=1)\n final_softmax = softmax(final_num/sum(final_num))\n write_json(line[:-1], final_lab, final_softmax,class_list)\n # result_list.append(avg_spatial_pred_fc8)\n\n # pred_index = np.argmax(avg_spatial_pred_fc8)\n\n # print(final_lab,\" \",final_softmax)\n print_score = [float('%.2f' % final_softmax[0]),float('%.2f' % final_softmax[1]),float('%.2f' % final_softmax[2]),\n float('%.2f' % final_softmax[3]),float('%.2f' % final_softmax[4])]\n\n print(final_lab,print_score, ' ',line_id ,' / ',len(val_list),' video ')\n line_id += 1\n print(len(val_list))\n\n\nif __name__ == \"__main__\":\n final_result = {}\n final_result['results'] = {}\n main()\n\n\n\n\n \n","sub_path":"script/HISTORY/testset_eval.py","file_name":"testset_eval.py","file_ext":"py","file_size_in_byte":3914,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"490265033","text":"#!/usr/bin/env python\nimport grequests\nfrom urllib.request import urlopen, Request\nfrom bs4 import BeautifulSoup\nfrom gamelogs import Gamelogs\nfrom export import CSV\nfrom gamelogsurls import GamelogUrls\nfrom parse_to_csv import Parse_To_Csv\n\n\nyear = None\nyear_is_valid = False\ninitial_url = 'https://www.sports-reference.com/cbb/seasons/2019-school-stats.html'\napp_is_running = True\n\nwhile app_is_running:\n \n #simple input validation\n while year_is_valid==False:\n input_value = input('input year or press enter/return to exit: ')\n\n if(input_value==''):\n raise SystemExit\n\n elif input_value.isdigit():\n if not len(input_value)==4:\n print('invalid year')\n continue\n\n elif int(input_value)<2011:\n print('no gamelogs before 2011')\n continue\n else:\n year=int(input_value)\n year_is_valid=True\n break\n else:\n print('not a number')\n continue\n \n \n\n \n \n \n if year_is_valid:\n urls = GamelogUrls(year, initial_url)\n\n url_has_data = True\n data = []\n counter=0\n \n request = (grequests.get(link) for link in urls.get_gamelogs_urls())\n response = grequests.imap(request)\n\n input_choice_is_valid = False\n \n\n while input_choice_is_valid==False:\n input_value = input('fetch gamelogs for {} ONLY? (y/n): '.format(year))\n \n if input_value=='y':\n input_choice_is_valid = True\n print('fetching gamelogs for {} season...'.format(year))\n \n for link in response:\n if link.status_code==200:\n parse = Parse_To_Csv(link.url, year)\n parsed_gamelogs = parse.inputted_year_gamelogs()\n else:\n continue\n \n if parsed_gamelogs:\n csv = CSV(parsed_gamelogs,year)\n csv.generate_csv()\n \n else:\n year_is_valid = False\n raise SystemExit\n \n\n\n\n elif input_value=='n':\n input_choice_is_valid = True\n print('fetching gamelogs since {} season...'.format(year))\n \n for link in response:\n if link.status_code==200:\n parse = Parse_To_Csv(link.url, year)\n parsed_gamelogs = parse.recursive_year_gamelogs()\n else:\n continue\n\n if parsed_gamelogs:\n csv = CSV(parsed_gamelogs,year)\n csv.generate_years_csv()\n \n \n else:\n year_is_valid=False\n raise SystemExit\n \n \n else:\n print('y for yes, n for no... enter year again')\n year_is_valid=False\n break\n \n \n \n \n \n","sub_path":"scraper.py","file_name":"scraper.py","file_ext":"py","file_size_in_byte":3249,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"351180840","text":"\"\"\"Chunk processor engine interface\"\"\"\n\nfrom socket_messages import SocketTranscriptMessage, SocketErrorMessage\n\nclass EngineInterface():\n \"\"\"Interface for chunk processor engines\"\"\"\n def __init__(self, send_message = None):\n self.send_message = send_message\n self.accept_chunks = True\n self.is_open = True\n # Defaults\n self._sample_rate = float(16000)\n self._language = \"\" # \"de-DE\", \"en-US\", etc. (could be: \"de_DE\", \"de\", ...)\n self._asr_model_path = \"\" # model folder relative to: settings.asr_models_folder\n self._continuous_mode = False # send final result once after stop event\n self._optimize_final_result = False # use text processors to optimize final result\n\n async def process(self, chunk: bytes):\n \"\"\"Process chunk\"\"\"\n async def finish_processing(self):\n \"\"\"Block new process requests, wait for last process to finish and send result\"\"\"\n async def close(self):\n \"\"\"Close and clean up\"\"\"\n def get_options(self):\n \"\"\"Return possible options as object (optionally) with defaults\"\"\"\n return {}\n\n async def send_transcript(self,\n transcript, is_final = False, confidence = -1, features = None, alternatives = None):\n \"\"\"Send transcript result\"\"\"\n if self.send_message is not None:\n msg = SocketTranscriptMessage(\n transcript, is_final, confidence, features, alternatives)\n await self.send_message(msg)\n\n async def on_before_close(self):\n \"\"\"Run before close for any required extra action\"\"\"\n self.is_open = False\n\n async def on_error(self, error_message):\n \"\"\"Send error message\"\"\"\n self.accept_chunks = False\n if self.send_message is not None:\n await self.send_message(\n SocketErrorMessage(500, \"AsrEngineError\", error_message))\n","sub_path":"src/engine_interface.py","file_name":"engine_interface.py","file_ext":"py","file_size_in_byte":1914,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"281720830","text":"# encoding: UTF-8\n\"\"\"\nModule dataclient defines ThanfDataClient.\n\"\"\"\n\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nimport os\nimport asyncio\nimport datetime\nimport ujson\nimport pandas as pd\nimport requests\nimport logging\nfrom . import utils\n\n\nclass ThanfDataClient(object):\n\n def __init__(self, address=\"http://data.thanf.com\"):\n self._address = address\n self._bar_columns = [\n 'symbol',\n 'trade_date',\n 'open',\n 'high',\n 'low',\n 'close',\n 'volume',\n 'turnover',\n 'trade_status']\n self._index_weights_columns = [\n 'trade_date',\n 'symbol',\n 'sec_name',\n 'weight',\n 'index_code']\n self.adj_factor_columns = [\n 'symbol',\n 'trade_date',\n 'adjust_factor']\n self._index_member_columns = [\n 'in_date',\n 'out_date',\n 'symbol'\n ]\n\n @staticmethod\n def _parse_error(content):\n err = eval(content)\n return \"{0},{1}\".format(err.get('error_code'), err.get('error_msg'))\n\n @staticmethod\n def _parse(content):\n try:\n if content.startswith(b\"{'error_code'\"):\n return None, ThanfDataClient._parse_error(content)\n else:\n return ujson.loads(content), None\n except ValueError:\n with open(os.path.join(os.getcwd(), \"error.txt\"), 'wb') as f:\n f.write(content)\n raise\n\n def query_trade_dates(self, start_date, end_date):\n if start_date == \"\":\n start_date = \"20100101\"\n if end_date == \"\":\n end_date = \"{0}1231\".format(datetime.datetime.today().year)\n params = {'start_date': start_date, \"end_date\": end_date}\n r = requests.get(\"{0}/trading_dates\".format(self._address), params)\n dates, err_msg = self._parse(r.content)\n columnset = {\"istradeday\": ['T'] * len(dates), \"trade_date\": dates}\n return utils.to_dataframe(columnset), err_msg\n\n def _get_bar_url(self, symbol, start_date, end_date, ktype, atype):\n url_pattern = \"{0}/get_history_bars?{1}\"\n items = {\n \"order_book_id\": symbol,\n \"start_date\": start_date,\n \"end_date\": end_date,\n \"ktype\": ktype,\n \"atype\": utils.get_atype(atype)\n }\n return url_pattern.format(self._address, utils.dict2url(items))\n\n @staticmethod\n def _get_response_async(urls: list):\n async def get_json(url):\n json = await loop.run_in_executor(None, requests.get, url)\n responses.append(json)\n\n async def run(items: list):\n await asyncio.gather(*[get_json(x) for x in items])\n\n loop = asyncio.new_event_loop()\n responses = []\n for i in utils.chunks(urls, max(len(urls) // 1, 1)):\n loop.run_until_complete(run(i))\n return responses\n\n @staticmethod\n def _get_response(urls: list):\n logging.info('get_response begin')\n responses = []\n with requests.Session() as s:\n for i in urls:\n responses.append(s.get(i))\n logging.info('get_response end')\n return responses\n\n def _parse_bar(self, data: list):\n if len(data) == 0 or len(data[0]) == len(self._bar_columns):\n df = pd.DataFrame(data, columns=self._bar_columns)\n df['trade_status'] = df['trade_status'].apply(lambda x: '停牌' if x != '交易' else x)\n else:\n df = pd.DataFrame(data, columns=self._bar_columns[:-1])\n df['trade_status'] = '交易'\n df['vwap'] = 0\n df.loc[df['volume'] > 0, 'vwap'] = df['turnover']/df['volume']\n return df\n\n def _parse_daily_rsp(self, rsp_list: list):\n logging.info('parse_daily_rsp begin')\n rsp_list = [self._parse(x.content) for x in rsp_list]\n df = pd.DataFrame(columns=self._bar_columns)\n err_msg = None\n for data, err in rsp_list:\n if data is not None:\n df = df.append(self._parse_bar(data), sort=False)\n else:\n err_msg = err\n logging.info('parse_daily_rsp end')\n return df, err_msg\n\n def daily(self, symbol: str, start_date, end_date, adjust_mode=None):\n urls = list(map(\n lambda x: self._get_bar_url(x, start_date, end_date, \"D\", adjust_mode),\n symbol.split(',')))\n\n return self._parse_daily_rsp(self._get_response(urls))\n\n def query_inst_info(self, symbol, fields: list):\n url = '{0}/instruments?order_book_id={1}'.format(self._address, symbol)\n data, err_msg = self._parse(requests.get(url).content)\n return pd.DataFrame(data)[fields], err_msg\n\n def query_index_weights_range(self, index, start_date, end_date):\n args = utils.dict2url({\n 'order_book_id': index,\n 'start_date': start_date,\n 'end_date': end_date})\n url = '{0}/get_index_component_info?{1}'.format(self._address, args)\n data, err_msg = self._parse(requests.get(url).content)\n return pd.DataFrame(data, columns=self._index_weights_columns), err_msg\n\n def _get_adj_factor_url(self, symbol, start_date, end_date):\n url_pattern = \"{0}/get_stock_adjfactor?{1}\"\n items = {\n \"order_book_id\": symbol,\n \"start_date\": start_date,\n \"end_date\": end_date\n }\n return url_pattern.format(self._address, utils.dict2url(items))\n\n def _parse_adj_factor(self, data: list):\n return pd.DataFrame(data, columns=self.adj_factor_columns)\n\n def _parse_adj_factor_rsp(self, rsp_list: list):\n rsp_list = [self._parse(x.content) for x in rsp_list]\n df = pd.DataFrame(columns=self.adj_factor_columns)\n err_msg = None\n for data, err in rsp_list:\n if data is not None:\n df = df.append(self._parse_adj_factor(data))\n else:\n err_msg = err\n\n return df, err_msg\n\n def query_adj_factor(self, symbol, start_date, end_date):\n urls = list(map(\n lambda x: self._get_adj_factor_url(x, start_date, end_date),\n symbol.split(',')))\n return self._parse_adj_factor_rsp(self._get_response(urls))\n\n def query_index_member(self, index, start_date, end_date):\n url_pattern = \"{0}/get_index_component_transfer_info?{1}\"\n args = utils.dict2url({\n 'order_book_id': index,\n 'start_date': start_date,\n 'end_date': end_date})\n\n rsp = requests.get(url_pattern.format(self._address, args))\n data, err_msg = self._parse(rsp.content)\n return pd.DataFrame(data, columns=self._index_member_columns), err_msg\n","sub_path":"jaqs_thanf/dataclient.py","file_name":"dataclient.py","file_ext":"py","file_size_in_byte":6828,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"223180234","text":"import os\r\nimport numpy as np\r\nfrom data_set import filepaths as fp\r\nimport pandas as pd\r\n\r\nbase_path = fp.Ml_100K.ORGINAL_DIR\r\ntrain_path = os.path.join(base_path,'ua.base')\r\ntest_path = os.path.join(base_path,'ua.test')\r\nuser_path = os.path.join(base_path,'u.user')\r\nitem_path = os.path.join(base_path,'u.item')\r\noccupation_path = os.path.join(base_path,'u.occupation')\r\n\r\n\r\ndef __read_age_index():\r\n age_levels = set()\r\n with open(user_path, 'r') as f:\r\n for line in f.readlines():\r\n d = line.strip().split('|')\r\n age_level = int(d[1])//10\r\n age_levels.add(age_level)\r\n return len(age_levels)\r\n\r\ndef __read_occupation_index(begin):\r\n occupations = {}\r\n with open(occupation_path,'r') as f:\r\n names = f.read().strip().split('\\n')\r\n for name in names:\r\n occupations[name]=begin\r\n begin+=1\r\n return occupations,begin\r\n\r\ndef generate_user_df():\r\n begin = __read_age_index()\r\n gender_dict = { 'M':begin, 'F':begin+1 }\r\n begin += 2\r\n occupation_dict,begin = __read_occupation_index(begin)\r\n uids = []\r\n all_users = []\r\n\r\n with open(user_path,'r') as f:\r\n for line in f.readlines():\r\n user_indexs=[]\r\n d = line.strip().split('|')\r\n age = int(d[1])//10\r\n uids.append(d[0])\r\n user_indexs.append(age)\r\n user_indexs.append(gender_dict[d[2]])\r\n user_indexs.append(occupation_dict[d[3]])\r\n all_users.append(user_indexs)\r\n\r\n df = pd.DataFrame(all_users,index=uids,columns=['age', 'gender', 'occupation'])\r\n df.to_csv(fp.Ml_100K.USER_DF)\r\n return begin\r\n\r\ndef __get_year_index(begin):\r\n years = set()\r\n with open(item_path, 'r', encoding = 'ISO-8859-1') as f:\r\n for line in f.readlines():\r\n d = line.strip().split('|')\r\n year = d[2].split('-')\r\n if len(year)>2:\r\n years.add(int(year[2]))\r\n years.add(0)\r\n years = sorted(years)\r\n print(years)\r\n return {k:v+begin for v,k in enumerate(years)},len(years)\r\n\r\ndef generate_item_df(begin,out):\r\n items = {}\r\n years_dict, begin = __get_year_index(begin)\r\n max_n_neibour = 0\r\n all_items = []\r\n iids = []\r\n with open( item_path, 'r', encoding = 'ISO-8859-1' ) as f:\r\n for line in f.readlines():\r\n item_index = []\r\n d = line.strip().split('|')\r\n iids.append(int(d[0]))\r\n year = d[2].split('-')\r\n if len(year) > 2:\r\n item_index.append(years_dict[int(year[2])])\r\n else:\r\n item_index.append(0)\r\n\r\n subjects = d[5:]\r\n if begin == 0:\r\n begin = len(subjects)\r\n for i in range(len(subjects)):\r\n if int(subjects[i]) == 1:\r\n item_index.append( begin+i )\r\n all_items.append( item_index )\r\n if len(item_index) > max_n_neibour:\r\n max_n_neibour = len(item_index)\r\n n_all=[]\r\n for item in all_items:\r\n n_all.append( np.random.choice( item, size = max_n_neibour, replace = True ) )\r\n\r\n df = pd.DataFrame( n_all, index = iids )\r\n df.to_csv(out )\r\n\r\n #print( all_items, max_n_neibour )\r\n return items\r\n\r\ndef get1or0(r):\r\n return 1.0 if r>3 else 0.0\r\n\r\n\r\ndef __read_rating_data(path):\r\n triples=[]\r\n with open(path,'r') as f:\r\n for line in f.readlines():\r\n d=line.strip().split('\\t')\r\n triples.append([int(d[0]),int(d[1]),get1or0(int(d[2]))])\r\n return triples\r\n\r\ndef read_data_user_item_df():\r\n user_df = pd.read_csv( fp.Ml_100K.USER_DF, index_col = 0 )\r\n item_df = pd.read_csv( fp.Ml_100K.ITEM_DF_0, index_col = 0 )\r\n train_triples = __read_rating_data(train_path)\r\n test_triples= __read_rating_data(test_path)\r\n return train_triples, test_triples, user_df, item_df, max(user_df.max())+1, max(item_df.max())+1\r\n\r\n\r\ndef read_data():\r\n user_df = pd.read_csv( fp.Ml_100K.USER_DF, index_col = 0 )\r\n item_df = pd.read_csv( fp.Ml_100K.ITEM_DF, index_col = 0 )\r\n train_triples = __read_rating_data(train_path)\r\n test_triples= __read_rating_data(test_path)\r\n return train_triples, test_triples, user_df, item_df,max(item_df.max())+1\r\n\r\n\r\nif __name__ == '__main__':\r\n item_df = generate_item_df(0, fp.Ml_100K.ITEM_DF_0)\r\n #print(item_df)\r\n\r\n train_triples, test_triples, user_df, item_df,lenitems = read_data()\r\n print(user_df)\r\n print(item_df)\r\n\r\n","sub_path":"basic_sim/dataloader4ml100kIndexs.py","file_name":"dataloader4ml100kIndexs.py","file_ext":"py","file_size_in_byte":4481,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"638651199","text":"import csv\nfrom kmpp.pandora_box import PandoraBox\n\npandora_box = PandoraBox()\nmarket_contr = {}\noutput = {}\n\nfile = 'pandora_box.csv'\nwith open(file, 'r') as csv_file:\n csv_reader = csv.reader(csv_file, delimiter=',')\n total_prod = 0\n for row in csv_reader:\n prod_score = []\n for col in row:\n prod_score.append(float(col))\n pandora_box.insert_score(prod_score) \n total_prod += 1\n# pandora_box.print_box()\n\nk_product = int(input('Masukkan jumlah produk : '))\nprint('Jumlah produk: {}'.format(k_product))\ntime_start, time_end = input('Masukkan waktu awal dan akhir (dipisahkan oleh spasi): ').split()\nprint('Interval waktu: {} - {}'.format(time_start, time_end))\n\n# hitung kontribusi pasar total selama interval waktu\n# asumsi id product integer yang berurutan\nfor i in range(0, total_prod):\n market_contr[i] = pandora_box.get_score(i, int(time_start), int(time_end))\nprint('Market Contribution')\nfor key in market_contr:\n print('{} : {}'.format(key, market_contr[key]))\n\n# sort yang paling besar\nsorted_prod = sorted(market_contr, key=lambda x: (market_contr[x]), reverse=True)\nprint(sorted_prod)\n\n# keluarkan output k teratas\nfor i in range(0, k_product):\n output[sorted_prod[i]] = market_contr[sorted_prod[i]]\n\nprint(output)\n","sub_path":"app/bak/src-rsl/solution.py","file_name":"solution.py","file_ext":"py","file_size_in_byte":1284,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"256887255","text":"import os\nimport unittest\n\nfrom ogr.services.gitlab import GitlabService\nfrom ogr.persistent_storage import PersistentObjectStorage\n\nDATA_DIR = \"test_data\"\nPERSISTENT_DATA_PREFIX = os.path.join(\n os.path.dirname(os.path.realpath(__file__)), DATA_DIR\n)\n\n\nclass GitlabTests(unittest.TestCase):\n def setUp(self):\n self.token = os.environ.get(\"GITLAB_TOKEN\")\n self.user = os.environ.get(\"GITLAB_USER\")\n test_name = self.id() or \"all\"\n\n persistent_data_file = os.path.join(\n PERSISTENT_DATA_PREFIX, f\"test_gitlab_data_{test_name}.yaml\"\n )\n PersistentObjectStorage().storage_file = persistent_data_file\n\n if PersistentObjectStorage().is_write_mode and (\n not self.user or not self.token\n ):\n raise EnvironmentError(\"please set GITLAB_TOKEN GITLAB_USER env variables\")\n else:\n self.token = \"some_token\"\n\n self.service = GitlabService(\n token=self.token, instance_url=\"https://gitlab.gnome.org\", ssl_verify=True\n )\n\n self.project = self.service.get_project(\n repo=\"testing-ogr-repo\", namespace=\"lbarcziova\"\n )\n\n def tearDown(self):\n PersistentObjectStorage().dump()\n\n\nclass GenericCommands(GitlabTests):\n def test_branches(self):\n branches = self.project.get_branches()\n assert branches\n assert \"master\" in branches\n\n def test_get_file(self):\n file_content = self.project.get_file_content(\"README.md\")\n assert file_content\n assert \"This is new README for testing-ogr-repo\" in file_content\n\n def test_nonexisting_file(self):\n with self.assertRaises(FileNotFoundError):\n self.project.get_file_content(\".blablabla_nonexisting_file\")\n\n def test_username(self):\n # check just lenght, because it is based who regenerated data files\n assert len(self.service.user.get_username()) > 3\n\n def test_email(self):\n email = self.service.user.get_email()\n assert email\n assert len(email) > 3\n assert \"@\" in email\n assert \".\" in email\n\n\nclass Issues(GitlabTests):\n def test_get_issue_list(self):\n issue_list = self.project.get_issue_list()\n assert issue_list\n assert len(issue_list) >= 1\n\n def test_issue_info(self):\n issue_info = self.project.get_issue_info(issue_id=1)\n assert issue_info\n assert issue_info.title.startswith(\"My first issue\")\n assert issue_info.description.startswith(\"This is testing issue\")\n\n def test_get_all_issue_comments(self):\n comments = self.project._get_all_issue_comments(issue_id=2)\n assert comments[0].comment.startswith(\"Comment\")\n assert comments[0].author == \"lbarcziova\"\n assert len(comments) == 2\n\n def test_create_issue(self):\n issue = self.project.create_issue(\n title=\"Issue 2\", description=\"Description for issue 2\"\n )\n assert issue.title == \"Issue 2\"\n assert issue.description == \"Description for issue 2\"\n\n def test_close_issue(self):\n issue = self.project.close_issue(issue_id=1)\n assert issue.status == \"closed\"\n\n\nclass PullRequests(GitlabTests):\n def test_pr_list(self):\n pr_list = self.project.list_pull_requests()\n count = len(pr_list)\n assert pr_list\n assert count >= 1\n assert pr_list[count - 1].title == \"change\"\n\n def test_pr_info(self):\n pr_info = self.project.get_pr_info(pr_id=1)\n assert pr_info\n assert pr_info.title == \"change\"\n assert pr_info.description == \"description of mergerequest\"\n\n def test_get_all_pr_commits(self):\n commits = self.project.get_all_pr_commits(pr_id=1)\n assert commits[0] == \"0709030b613d56752725c33df36041c2b7610506\"\n assert commits[1] == \"f3881188db863e4e053f5a82422f067ac9ba2594\"\n assert len(commits) == 2\n\n def test_get_all_pr_comments(self):\n comments = self.project._get_all_pr_comments(pr_id=1)\n count = len(comments)\n assert comments[count - 1].comment == \"first comment of mergerequest\"\n assert comments[count - 1].author == \"lbarcziova\"\n assert count == 5\n\n def test_update_pr_info(self):\n pr_info = self.project.get_pr_info(pr_id=1)\n original_description = pr_info.description\n\n self.project.update_pr_info(pr_id=1, description=\"changed description\")\n pr_info = self.project.get_pr_info(pr_id=1)\n assert pr_info.description == \"changed description\"\n\n self.project.update_pr_info(pr_id=1, description=original_description)\n pr_info = self.project.get_pr_info(pr_id=1)\n assert pr_info.description == original_description\n\n\nclass Tags(GitlabTests):\n def test_get_tags(self):\n tags = self.project.get_tags()\n count = len(tags)\n assert count >= 2\n assert tags[count - 1].name == \"0.1.0\"\n assert tags[count - 1].commit_sha == \"957d267a5b0cd9e615cd081c0eb02397dce1eb73\"\n\n def test_tag_from_tag_name(self):\n tag = self.project._git_tag_from_tag_name(tag_name=\"0.1.0\")\n assert tag.commit_sha == \"957d267a5b0cd9e615cd081c0eb02397dce1eb73\"\n\n\nclass Releases(GitlabTests):\n def test_create_release(self):\n count_before = len(self.project.get_releases())\n release = self.project.create_release(\n name=\"test\", tag_name=\"0.2.0\", description=\"testing release\", ref=\"master\"\n )\n count_after = len(self.project.get_releases())\n assert release.tag_name == \"0.2.0\"\n assert release.title == \"test\"\n assert release.body == \"testing release\"\n assert count_before + 1 == count_after\n\n def test_get_releases(self):\n releases = self.project.get_releases()\n assert releases\n count = len(releases)\n assert count >= 1\n assert releases[count - 1].title == \"test\"\n assert releases[count - 1].tag_name == \"0.1.0\"\n assert releases[count - 1].body == \"testing release\"\n","sub_path":"tests/integration/test_gitlab.py","file_name":"test_gitlab.py","file_ext":"py","file_size_in_byte":6017,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"317245171","text":"import matplotlib.pyplot as plt\nimport numpy as np\n\ndir_paths = [\"model_normal_tremor\", \"model_limits_normal\", \"model_weakness_normal\"]\nfilename = \"monitor.csv\"\nplt.figure()\nfor dir_path in dir_paths:\n print(dir_path)\n reward_arr = [];\n mean_arr = []\n count = 0\n with open(dir_path + \"/\" + filename, 'r') as fin:\n line = fin.readline()\n while line:\n values = line.split(\",\")\n print(values)\n if count >= 2:\n if(dir_path == dir_paths[1]):\n if count % 2 != 0:\n reward_arr.append(float(values[0]))\n mean_arr.append(np.mean(reward_arr[-50:]))\n else:\n reward_arr.append(float(values[0]))\n mean_arr.append(np.mean(reward_arr[-50:]))\n line = fin.readline()\n count += 1\n plt.plot(mean_arr, label=dir_path)\nplt.legend()\nplt.show()\n","sub_path":"icra2021/base/plotter.py","file_name":"plotter.py","file_ext":"py","file_size_in_byte":941,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"316404966","text":"# 端口\nSERVER_PORT = '9000'\n# 开启debug模式\nDEBUG = True\n# 显示sql语句\nSQLALCHEMY_ECHO = True\n\nSQLALCHEMY_DATABASE_URI = 'mysql+pymysql://root:123456@127.0.0.1:3306/movie'\nSQLALCHEMY_TRACK_MODIFICATIONS = False\n\n# 数据库编码 utf-8\nSQLALCHEMY_ENCODING = \"utf-8\"\n","sub_path":"order/config/local_setting.py","file_name":"local_setting.py","file_ext":"py","file_size_in_byte":276,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"42526367","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nfrom ..util.html import get_content, get_location\nfrom ..util.match import matchall, match1\nfrom ..embedextractor import EmbedExtractor\n\nimport json\n\n\"\"\"\nrefer to http://open.youku.com/tools\n\"\"\"\nyouku_embed_patterns = [ 'youku\\.com/v_show/id_([a-zA-Z0-9=]+)',\n 'player\\.youku\\.com/player\\.php/sid/([a-zA-Z0-9=]+)/v\\.swf',\n 'loader\\.swf\\?VideoIDS=([a-zA-Z0-9=]+)',\n 'player\\.youku\\.com/embed/([a-zA-Z0-9=]+)',\n 'YKU.Player\\(\\'[a-zA-Z0-9]+\\',{ client_id: \\'[a-zA-Z0-9]+\\', vid: \\'([a-zA-Z0-9]+)\\'',\n 'data-youku=\\\"[a-zA-Z0-9,:]+vid:([a-zA-Z0-9=]+)\\\"'\n ]\n\n\"\"\"\nv.qq.com\n\"\"\"\nqq_embed_patterns = [ 'v\\.qq\\.com[a-zA-Z0-9\\/\\?\\.\\;]+vid=([a-zA-Z0-9]+)',\n 'TPout\\.swf[a-zA-Z0-9=\\?\\&_]+vid=([a-zA-Z0-9]+)'\n ]\n\n\n\"\"\"\ntv.sohu.com\n\"\"\"\nsohu_embed_patterns = [ 'tv\\.sohu\\.com[a-zA-Z0-9\\/\\?=]+\\&vid=([a-zA-Z0-9]+)\\&',\n 'share\\.vrs\\.sohu\\.com\\/my\\/v.swf[&+=a-zA-z0-9]+&id=([^&]+)',\n 'my\\.tv\\.sohu\\.com\\/[a-zA-Z0-9\\/]+/([^\\.]+)'\n ]\n\n\"\"\"\nKu6\n\"\"\"\nku6_embed_url = [ '(http://v.ku6vms.com/[^\\\"]+)'\n ]\n\nku6_embed_patterns = [ 'http://player.ku6.com/refer/(.*)/v.swf'\n ]\n\"\"\"\n163\n\"\"\"\nnetease_embed_patterns = [ 'v\\.163\\.com\\/[0-9a-zA-Z\\/\\?\\.]+topicid=([^&]+)&\\;vid=([^&]+)',\n 'topicid=([a-zA-Z0-9]+)&vid=([a-zA-Z0-9]+)&'\n ]\n\n\"\"\"\niqiyi\n\"\"\"\niqiyi_embed_patterns = [ 'definitionID=([^&]+)&tvId=([^&]+)'\n ]\n\n\"\"\"\nLetv Cloud\n\"\"\"\nlecloud_embed_patterns = [ '{\"uu\":\"([^\\\"]+)\",\"vu\":\"([^\\\"]+)\"',\n 'bcloud.swf\\?uu=([^&]+)&vu=([^&]+)',\n 'uu=([^&]+)&vu=([^&]+)'\n ]\n\n\"\"\"\nifeng\n\"\"\"\nifeng_embed_patterns = [ 'v\\.ifeng\\.com\\/[a-zA-Z\\=\\/\\?\\&\\.]+guid=([^\\\"]+)'\n ]\n\n\"\"\"\nweibo\n\"\"\"\nweibo_embed_patterns = [ 'http://video.weibo.com/player/1034:(\\w{32})\\w*'\n ]\n\n\"\"\"\nSina\n\"\"\"\nsina_embed_patterns = [ 'http://video.sina.com.cn/share/video/(\\d+).swf'\n ]\n\n\"\"\"\nDilidili\n\"\"\"\ndilidili_embed_patterns = [ 'vid=([^&]+)&v=([^&]+)&'\n ]\n\n\"\"\"\nBilibili\n\"\"\"\nbilibili_embed_patterns = [ 'flashvars=\"aid=(\\d+)'\n ]\n\nclass GeneralEmbed(EmbedExtractor):\n name = u\"GeneralEmbed (通用嵌入视频)\"\n\n def prepare_playlist(self):\n content = get_content(self.url)\n\n vids = matchall(content, youku_embed_patterns)\n for vid in vids:\n self.video_info_list.append(('youku',vid))\n\n vids = matchall(content, qq_embed_patterns)\n for vid in vids:\n self.video_info_list.append(('qq.video',vid))\n\n vids = matchall(content, sohu_embed_patterns)\n for vid in vids:\n self.video_info_list.append(('sohu.my',vid))\n\n urls = matchall(content, ku6_embed_url)\n for url in urls:\n html = get_content(url)\n flashvars = matchall(html, ['vid=([^&]+)', 'style=([^&]+)', 'sn=([^&]+)'])\n data = json.loads(get_content('http://v.ku6vms.com/phpvms/player/forplayer/vid/{}/style/{}/sn/{}'.format(flashvars[0], flashvars[1],flashvars[2])))\n vid = data['ku6vid']\n self.video_info_list.append(('ku6',vid))\n vids = matchall(content, ku6_embed_patterns)\n for v in vids:\n self.video_info_list.append(('ku6', v))\n vids = matchall(content, netease_embed_patterns)\n for v in vids:\n self.video_info_list.append(('netease.video', v))\n\n vids = matchall(content, iqiyi_embed_patterns)\n for v in vids:\n videoid, tvid = v\n self.video_info_list.append(('iqiyi', (tvid, videoid)))\n\n vids = matchall(content, lecloud_embed_patterns)\n for v in vids:\n uu, vu = v\n self.video_info_list.append(('le.letvcloud', (vu, uu)))\n\n vids = matchall(content, ifeng_embed_patterns)\n for v in vids:\n v = v.split('&')[0]\n self.video_info_list.append(('ifeng.news', v))\n\n vids = matchall(content, weibo_embed_patterns)\n for v in vids:\n self.video_info_list.append(('weibo','http://weibo.com/p/' + v))\n\n vids = matchall(content, sina_embed_patterns)\n for v in vids:\n v = v.split('&')[0]\n self.video_info_list.append(('sina.video', v))\n\n vids = matchall(content, bilibili_embed_patterns)\n for v in vids:\n v = \"http://www.bilibili.com/video/av{}\".format(v)\n self.video_info_list.append(('bilibili.video', v))\n\n\n vids = matchall(content, dilidili_embed_patterns)\n for v in vids:\n v,site = v\n if site == 'bilibili':\n site = 'bilibili.video'\n elif site == 'qq':\n site = 'qq.video'\n elif site =='yun':\n site = 'le.letvcloud'\n v = v.split(':')\n self.video_info_list.append((site, v))\n\n tmp = []\n for v in self.video_info_list:\n if not v in tmp:\n tmp.append(v)\n self.video_info_list = tmp\n\n parser = EmbedExtractor.parser_list\n\nsite = GeneralEmbed()\n","sub_path":"ykdl/extractors/generalembed.py","file_name":"generalembed.py","file_ext":"py","file_size_in_byte":5386,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"49475905","text":"import datetime as dt\nfrom django.test import TestCase\n\nfrom classes.forms import ClassForm, LineOfBusinessForm, REQUIRED_ERROR\nfrom classes.models import Class, LineOfBusiness\n\n\nclass ClassFormTest(TestCase):\n\n def test_class_form_has_correct_prefix(self):\n form = ClassForm()\n self.assertEqual(form.prefix, 'class')\n\n def test_start_date_input_has_placeholder_and_css_classes(self):\n form = ClassForm()\n self.assertIn(\n 'placeholder=\"Select or enter the start date\"',\n form.as_p()\n )\n self.assertIn('class=\"form-control datepicker\"', form.as_p())\n\n def test_line_of_business_input_has_placeholder_and_css_classes(self):\n form = ClassForm()\n self.assertIn(\n 'placeholder=\"Select a Line Of Business (LOB)\"',\n form.as_p()\n )\n self.assertIn('class=\"form-control dropdown\"', form.as_p())\n\n def test_shift_input_has_placeholder_and_css_classes(self):\n form = ClassForm()\n self.assertIn(\n 'placeholder=\"Class shift\"',\n form.as_p()\n )\n self.assertIn('class=\"form-control\"', form.as_p())\n\n def test_hiring_bonus_input_has_placeholder_and_css_classes(self):\n form = ClassForm()\n self.assertIn(\n 'placeholder=\"Enter Hiring Bonus (in USD)\"',\n form.as_p()\n )\n self.assertIn('class=\"form-control\"', form.as_p())\n\n def test_referral_bonus_input_has_placeholder_and_css_classes(self):\n form = ClassForm()\n self.assertIn(\n 'placeholder=\"Enter Referral Bonus (in USD)\"',\n form.as_p()\n )\n self.assertIn('class=\"form-control\"', form.as_p())\n\n def test_start_date_input_validation_for_blank_entries(self):\n lob = LineOfBusiness.objects.create(name='PyPi')\n form = ClassForm(data={\n 'start_date': '',\n 'line_of_business': lob,\n 'shift': 1,\n 'hiring_bonus': 100,\n 'referral_bonus': 100,\n })\n self.assertFalse(form.is_valid())\n self.assertEqual(\n form.errors['start_date'],\n [REQUIRED_ERROR]\n )\n\n def test_line_of_business_input_validation_for_blank_entries(self):\n form = ClassForm(data={\n 'start_date': dt.date(2017, 7, 2),\n 'line_of_business': '',\n 'shift': 1,\n 'hiring_bonus': 100,\n 'referral_bonus': 100,\n })\n self.assertFalse(form.is_valid())\n self.assertEqual(\n form.errors['line_of_business'],\n [REQUIRED_ERROR]\n )\n\n def test_hiring_bonus_input_validation_for_blank_entries(self):\n lob = LineOfBusiness.objects.create(name='PyPi')\n form = ClassForm(data={\n 'start_date': dt.date(2017, 7, 2),\n 'line_of_business': lob,\n 'shift': 1,\n 'hiring_bonus': '',\n 'referral_bonus': 100,\n })\n self.assertFalse(form.is_valid())\n self.assertEqual(\n form.errors['hiring_bonus'],\n [REQUIRED_ERROR]\n )\n\n def test_referral_bonus_input_validation_for_blank_entries(self):\n lob = LineOfBusiness.objects.create(name='PyPi')\n form = ClassForm(data={\n 'start_date': dt.date(2017, 7, 2),\n 'line_of_business': lob,\n 'shift': 1,\n 'hiring_bonus': 100,\n 'referral_bonus': '',\n })\n self.assertFalse(form.is_valid())\n self.assertEqual(\n form.errors['referral_bonus'],\n [REQUIRED_ERROR]\n )\n\n def test_form_save(self):\n lob = LineOfBusiness.objects.create(name='iPython')\n form = ClassForm(data={\n 'start_date': ['06/20/2017'],\n 'line_of_business': lob,\n 'shift': 1,\n 'hiring_bonus': 150,\n 'referral_bonus': 300,\n })\n if not form.is_valid():\n print(form.errors)\n # self.assertTrue(form.is_valid())\n new_class = form.save()\n self.assertEqual(new_class, Class.objects.all()[0])\n\n\nclass LineOfBusinessFormTest(TestCase):\n\n def test_name_input_has_correct_placeholder(self):\n form = LineOfBusinessForm()\n self.assertIn(\n 'placeholder=\"Enter Line Of Business short name\"',\n form.as_p()\n )\n \n def test_name_input_has_correct_class(self):\n form = LineOfBusinessForm()\n self.assertIn(\n 'class=\"form-control col-sm-5\"',\n form.as_p()\n )","sub_path":"classes/tests/test_forms.py","file_name":"test_forms.py","file_ext":"py","file_size_in_byte":4556,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"212730474","text":"import chainer, numpy\nfrom matplotlib import pyplot\n\nclass mychain(chainer.Chain):\n def __init__(self, i, l, o):\n super(mychain, self).__init__()\n with self.init_scope():\n self.l1 = chainer.links.Linear(i,l)\n self.l2 = chainer.links.Linear(l,o)\n \n def __call__(self, x, y=None, train=True):\n x, y = chainer.Variable(x), chainer.Variable(y)\n\n a = chainer.functions.relu(self.l1(x))\n b = chainer.functions.relu(self.l2(a))\n if train:\n return chainer.functions.mean_squared_error(b, y)\n else:\n return b.data\n\ndef main():\n mnist = chainer.datasets.get_mnist()\n train, test = mnist\n\n model = mychain(28*28, 28*28*4, 28*28)\n\n opt = chainer.optimizers.Adam()\n opt.setup(model)\n\n epoch = 500\n batch = 20\n\n #train 1 \n def rand_train(a):\n return a[numpy.random.randint(len(a))][0].reshape(-1,len(a[0][0]))\n\n import time\n\n t0 = time.time()\n\n for i in range(epoch):\n loss = 0\n for n in range(batch):\n pix = rand_train(train)\n loss += model(pix, pix)\n print('epoch: {}, loss: {}'.format(i, loss.data))\n\n model.cleargrads()\n loss.backward()\n opt.update()\n \n #print-out train time\n t1 = time.time()\n print(t1-t0)\n\n #save for train 1\n try:\n chainer.serializers.save_npz('mymodel.npz', model)\n except:\n pritn('save error')\n \n # result for train 1\n pix = train[0][0].reshape(-1,len(train[0][0]))\n #or train[0][0][:, numpy.newaxis].T\n res = model(pix, train=False)\n\n pyplot.imshow(\n numpy.reshape(train[0][0], (28, 28)),\n cmap='gray')\n pyplot.show()\n pyplot.imshow(numpy.reshape(res, (28, 28)), cmap='gray')\n pyplot.show()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1833,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"152407004","text":"import random\nfrom race import races, Race\nfrom charclass import classes, CharacterClass\nfrom armor import armors, Armor\nfrom weapon import weapons, Weapon\nimport binascii\n\n_character_id = 0\nclass Character(object):\n\tfields = 'name:str race_name:str class_name:str var_str:int var_dex:int var_con:int var_int:int armor:armor weapon:weapon'\n\tdef __init__(self, name, race_name, class_name, str = 0, dex = 0, con = 0, int = 0, armor = None, weapon = None):\n\t\tself.name = name\n\n\t\tself.race_name = race_name\n\t\tself.class_name = class_name\n\n\t\tself.var_str = str\n\t\tself.var_dex = dex\n\t\tself.var_con = con\n\t\tself.var_int = int\n\n\t\tself.per_wc_miss_chance = {}\n\n\t\tself.armor = armor\n\t\tself.weapon = weapon\n\n\trace = property(lambda self: races[self.race_name])\n\tclass_ = property(lambda self: classes[self.class_name])\n\n\tstr = property(lambda self: 1 + self.race.base_str + self.var_str)\n\tdex = property(lambda self: 1 + self.race.base_dex + self.var_dex)\n\tcon = property(lambda self: 1 + self.race.base_con + self.var_con)\n\tint = property(lambda self: 1 + self.race.base_int + self.var_int)\n\n\tmax_hp = property(lambda self: self.con * self.class_.hp_per_con)\n\tmax_sp = property(lambda self: self.int)\n\tmax_mp = property(lambda self: self.dex)\n\n\t@classmethod\n\tdef random(cls):\n\t\tglobal _character_id\n\t\tname = 'character%d' % _character_id\n\t\t_character_id += 1\n\t\trace_name = random.choice(races.keys())\n\t\tclass_name = random.choice(classes.keys())\n\t\trndstats = [random.choice(['dex', 'con', 'int', 'str']) for i in range(random.randrange(4, 6+1))]\n\t\tstr = rndstats.count('str')\n\t\tdex = rndstats.count('dex')\n\t\tcon = rndstats.count('con')\n\t\tint = rndstats.count('int')\n\t\tarmor = random.choice(armors.values())#Armor.random()\n\t\tweapon = random.choice(weapons.values())#Weapon.random()\n\t\treturn cls(name, race_name, class_name, str, dex, con, int, armor, weapon)\n\n","sub_path":"src/character.py","file_name":"character.py","file_ext":"py","file_size_in_byte":1850,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"463933037","text":"import os\nimport sys\nsys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '../varity')))\nsys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '../varity/common')))\n\nimport varity\nimport cfg\nimport random\nimport shutil\nimport socket\nimport json\n\ndef configure():\n cfg.MAX_EXPRESSION_SIZE = 5\n cfg.MAX_NESTING_LEVELS = 3\n cfg.MAX_LINES_IN_BLOCK = 3\n cfg.ARRAY_SIZE = 10\n cfg.MAX_SAME_LEVEL_BLOCKS = 2\n cfg.MATH_FUNC_ALLOWED = True\n cfg.NUM_GROUPS = 1\n cfg.TESTS_PER_GROUP = 5\n cfg.OPT_LEVELS = [(\"-O0\", 1), (\"-O0\", 0), (\"-O1\", 0), (\"-O2\", 0), (\"-O3\", 0)]\n cfg.TESTS_DIR = \"_tests\"\n cfg.INPUT_SAMPLES_PER_RUN = 5\n cfg.REAL_TYPE = \"double\"\n\n # Set machine C compiler\n cc_path = findCCompiler()\n cfg.COMPILERS = [(\"cc\", cc_path)]\n\ndef findCCompiler():\n cc_path = shutil.which('cc')\n assert cc_path != None\n #print('cc:', cc_path)\n return cc_path\n\ndef test_driver():\n configure()\n varity.generateTests()\n print('dir:', varity.dirName())\n # Directory is generated\n assert os.path.exists(\"./\"+varity.dirName())\n\n # Check C files are generated\n c_files = 0\n for dirpath, dirnames, filenames in os.walk(\".\"):\n for filename in [f for f in filenames if f.endswith(\".c\")]:\n c_files = c_files + 1\n assert c_files == cfg.TESTS_PER_GROUP*cfg.NUM_GROUPS\n\n # Check C files are compiled\n varity.compileTests(\"./\"+varity.dirName())\n compiled_files = 0\n for dirpath, dirnames, filenames in os.walk(\".\"):\n for filename in [f for f in filenames if f.endswith(\".exe\")]:\n compiled_files = compiled_files + 1\n assert compiled_files > cfg.TESTS_PER_GROUP*cfg.NUM_GROUPS\n\n # Check executables can be run\n cwd = os.getcwd()\n varity.runTests(\"./\"+varity.dirName())\n resultsFile = cwd + \"/\" + varity.dirName() + \"/results.json\"\n with open(resultsFile) as json_file:\n data = json.load(json_file)\n testName = list(data.keys())[0]\n assert testName.endswith(\".c\")\n\n # Remove dir\n shutil.rmtree(cwd + \"/\" + varity.dirName())\n assert not os.path.exists(\"./\"+varity.dirName())\n\nif __name__ == '__main__':\n test_driver()\n\n","sub_path":"tests/test_driver.py","file_name":"test_driver.py","file_ext":"py","file_size_in_byte":2202,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"84627440","text":"from django.db import models\nfrom django.core.exceptions import ValidationError\n\nfrom edc.audit.audit_trail import AuditTrail\nfrom edc.choices import YES_NO, YES_NO_DONT_KNOW\nfrom edc.device.dispatch.models import BaseDispatchSyncUuidModel\n\nfrom apps.bcpp_household.managers import HouseholdAssessmentManager\nfrom apps.bcpp_household.exceptions import AlreadyReplaced\n\nfrom ..choices import INELIGIBLE_REASON, RESIDENT_LAST_SEEN\n\nfrom .household_structure import HouseholdStructure\nfrom .plot import Plot\n\n\nclass HouseholdAssessment(BaseDispatchSyncUuidModel):\n\n household_structure = models.OneToOneField(HouseholdStructure)\n\n residency = models.CharField(\n verbose_name=('Does anyone ever stay in this household?'),\n choices=YES_NO,\n max_length=25,\n null=True,\n editable=True,\n )\n\n member_count = models.IntegerField(\n verbose_name=(\"How many people live in this household (estimate)?\"),\n null=True,\n blank=True,\n help_text=(\"Provide the number of members in this household.\"))\n\n eligibles = models.CharField(\n verbose_name=('In speaking with the individual(s) above, at '\n 'least one member of this plot is potentially eligible'),\n choices=YES_NO_DONT_KNOW,\n max_length=25,\n null=True,\n blank=True,\n editable=True,\n )\n\n ineligible_reason = models.CharField(\n verbose_name=('If no members are eligible for this study, please state '\n 'the reason for ineligility.'),\n null=True,\n max_length=25,\n choices=INELIGIBLE_REASON,\n editable=True,\n blank=True)\n\n last_seen_home = models.CharField(\n verbose_name=('When was a resident last seen in this household?'),\n choices=RESIDENT_LAST_SEEN,\n max_length=25,\n null=True,\n blank=True,\n editable=True,\n )\n\n def __unicode__(self):\n return unicode(self.household_structure)\n\n objects = HouseholdAssessmentManager()\n\n history = AuditTrail()\n\n def save(self, *args, **kwargs):\n if self.household_structure.household.replaced_by:\n raise AlreadyReplaced('Model {0}-{1} has its container replaced.'.format(\n self._meta.object_name, self.pk))\n if self.household_structure.enumerated:\n raise ValidationError('HouseholdStructure has been enumerated')\n if self.household_structure.failed_enumeration_attempts < 3:\n raise ValidationError('Three attempts are required before Household Assessment')\n super(HouseholdAssessment, self).save(*args, **kwargs)\n\n def natural_key(self):\n return self.household_structure.natural_key()\n natural_key.dependencies = ['bcpp_household.household_structure']\n\n def dispatch_container_lookup(self, using=None):\n return (Plot, 'household_structure__household__plot__plot_identifier')\n\n @property\n def vdc_househould_status(self):\n return self.last_seen_home\n\n class Meta:\n app_label = 'bcpp_household'\n verbose_name = 'Household Residency Status Assess'\n verbose_name_plural = 'Household Residency Status Assess'\n","sub_path":"apps/bcpp_household/models/household_assessment.py","file_name":"household_assessment.py","file_ext":"py","file_size_in_byte":3199,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"350199694","text":"import os\nimport functools\nfrom conans import ConanFile, CMake, tools\n\nrequired_conan_version = \">=1.33.0\"\n\n\nclass InnoextractConan(ConanFile):\n name = \"innoextract\"\n description = \"Extract contents of Inno Setup installers\"\n license = \"innoextract License\"\n topics = (\"inno-setup\", \"decompression\")\n homepage = \"https://constexpr.org/innoextract/\"\n url = \"https://github.com/conan-io/conan-center-index\"\n exports_sources = [\"CMakeLists.txt\", \"patches/*\"]\n requires = (\n \"boost/1.78.0\",\n \"xz_utils/5.2.5\",\n \"libiconv/1.16\"\n )\n generators = \"cmake\", \"cmake_find_package\"\n settings = \"os\", \"arch\", \"compiler\", \"build_type\"\n\n @property\n def _source_subfolder(self):\n return \"source_subfolder\"\n\n @property\n def _build_subfolder(self):\n return \"build_subfolder\"\n\n def source(self):\n tools.get(**self.conan_data[\"sources\"][self.version], strip_root=True,\n destination=self._source_subfolder)\n\n def build(self):\n for patch in self.conan_data.get(\"patches\", {}).get(self.version, []):\n tools.patch(**patch)\n os.remove(os.path.join(self._source_subfolder, 'cmake', 'FindLZMA.cmake'))\n os.remove(os.path.join(self._source_subfolder, 'cmake', 'Findiconv.cmake'))\n cmake = self._configure_cmake()\n cmake.build()\n\n @functools.lru_cache(1)\n def _configure_cmake(self):\n cmake = CMake(self)\n # Turn off static library detection, which is on by default on Windows.\n # This keeps the CMakeLists.txt from trying to detect static Boost\n # libraries and use Boost components for zlib and BZip2. Getting the\n # libraries via Conan does the correct thing without other assistance.\n cmake.definitions[\"USE_STATIC_LIBS\"] = False\n cmake.configure(build_folder=self._build_subfolder)\n return cmake\n\n def package(self):\n self.copy(\"LICENSE\", dst=\"licenses\", src=self._source_subfolder)\n cmake = self._configure_cmake()\n cmake.install()\n tools.rmdir(os.path.join(self.package_folder, \"share\"))\n\n def package_id(self):\n del self.info.settings.compiler\n self.info.requires.clear()\n\n def package_info(self):\n self.cpp_info.libdirs = []\n bindir = os.path.join(self.package_folder, \"bin\")\n self.output.info(\"Appending PATH environment variable: {}\"\n .format(bindir))\n self.env_info.PATH.append(bindir)\n","sub_path":"recipes/innoextract/all/conanfile.py","file_name":"conanfile.py","file_ext":"py","file_size_in_byte":2490,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"438877414","text":"#######Implementation of totogram.py#################\n####Step1:Find the total number of nodes for the given height,function num_nodes() is implemented to compute this#######\n####step2:call on function medians() to compute median and it creates a list which as all the medians#########\n####step3:function leveltree() is called to build a balanced binary tree with an exception at level two which has three nodes####\n####step4:diff() is called to find the maximum difference in built tree########\n####step5:loop is constructed which call the function optimize() to reorders the tree to reduce the maximum difference found so far####\n####step6:loop terminates if no more optimization could be made with the tree constructed#######\n####Result Analysis######\n####k=3 solution obtained 3####\n####k=4 solution obtained 5####\n####k=5 solution obtained 11####\n####k=6 solution obtained 23####\n####k=7 solution obtained 47####\n###################End################################# \nimport statistics\nimport sys\nh=int(sys.argv[1])\nmaximum=0\nclass Tree(object): ######tree nodes are of class type TREE######\n\tdef __init__(self,data):\n\t\tself.left = None\n\t\tself.right = None\n\t\tself.middle=None\n\t\tself.data = data\ndef num_nodes(h): ####### Function to return number of nodes for given height######\n\tsum=0\n\tpnode=0\n\tfor i in range(1,h+1):\n\t\tif i ==1:\n\t\t\tsum=1\n\t\telif i==2:\n\t\t\tsum+=3\n\t\t\tpnode=3\n\t\telse:\n\t\t\tsum=sum+(pnode*2)\n\t\t\tpnode=pnode*2\n\treturn sum\ntemp={}\nqueue=[]\ntnode=Tree(0)\ndef chunks(l, n):#######find the element for each level of tree######\n\tn = max(1, n)\n\treturn [l[i:i + n] for i in range(0, len(l), n)]\ndef medians(mylist,num):\n\ttemp=chunks(mylist,int(len(mylist)/num))\n\tfor i in range(0,num):\n\t\tmedian=int(statistics.median(temp[i]))\n\t\tqueue.append(median)\n\t\tmylist.remove(median)\nnode=Tree(0)\ndef leveltree(node):\n\tfor i in range (1,h+1):\n\t\tif i==1:\n\t\t\ttemp=queue[0:1]\n\t\t\tnode=maketree(node,temp,i,0)\n\t\t\tn=1\n\t\telif i==2:\n\t\t\ttemp=queue[n:n+3]\n\t\t\tmaketree(node,temp,i,0)\n\t\t\tn=n+3\n\t\t\tp=3\n\t\telse:\n\t\t\ttemp=queue[n:n+p*2]\n\t\t\tmaketree(node,temp,i,0)\n\t\t\tn=n+p*2\n\t\t\tp=p*2\n\treturn node\ndef maketree(self,temp,i,bool):######builds a level balanced binary tree with 3 nodes in second level and two nodes each###### \n\tbool=bool+1\n\tif i==1:\n\t\tself=Tree(0)\n\t\tself.data=temp.pop(0)\n\t\treturn self\n\telse :\n\t\tself.left=maketree(self.left,temp,i-1,bool)\n\t\tif(bool==1):\n\t\t\tself.middle=maketree(self.middle,temp,i-1,bool)\n\t\tself.right=maketree(self.right,temp,i-1,bool)\n\t\treturn self\n\na=[]\t\t\ndef lprintTree(tree,bool):######prints the tree in level based order on finding the result######\n\t\tif tree==None:\n\t\t\treturn\n\t\tif bool==1:\n\t\t\ta.append(tree.data)\n\t\tif tree != None:\n\t\t\tlprintTree(tree.left,bool-1)\n\t\t\tif tree.middle != None:\n\t\t\t\tlprintTree(tree.middle,bool-1)\n\t\t\tlprintTree(tree.right,bool-1)\ndef diff(self,i,bool):######Function find the difference#######\n\tglobal maximum\n\tglobal tnode\n\tbool=bool+1\n\tif self.left == None and self.right ==None:\n\t\treturn self\n\telse :\n\t\tl=self.left.data\n\t\tself.left=diff(self.left,i-1,bool)\n\t\tif(bool==1):\n\t\t\tself.middle=diff(self.middle,i-1,bool)\n\t\tr=self.right.data\n\t\tself.right=diff(self.right,i-1,bool)\n\t\tdifference=max(abs(self.data-l),abs(self.data-r))\n\t\tif maximum < difference:\n\t\t\tif( self != None):\n\t\t\t\ttnode=self\n\t\t\t\tmaximum=difference\n\treturn self\t\nmini=999\t\nmp=Tree(0)\ndef min(tnode,parent):######Helper function to reorder the tree,returns the address of parent node having child with minimum value#######\n\tglobal mini\n\tglobal mp\n\tif(tnode==None):\n\t\treturn\n\tif(mini>tnode.data):\n\t\tmini=tnode.data\n\t\tmp=parent\n\tmin(tnode.left,tnode)\n\tmin(tnode.right,tnode)\n\treturn mp\nmx=-999\t\ndef maxi(tnode,parent):######Helper function to reorder the tree, returns the address of parent node having child with maximum value######\n\tglobal mx\n\tglobal mp\n\tif(tnode==None):\n\t\treturn\n\tif(mxtemp.data):\n\t\t\tt=tnode.left.data\n\t\t\ttnode.left.data=temp.left.data\n\t\t\ttemp.left.data=t\n\t\tif(temp.right.data>temp.data):\n\t\t\tt=tnode.left.data\n\t\t\ttnode.left.data=temp.right.data\n\t\t\ttemp.right.data=t\t\nmylist=[i for i in range(1,num_nodes(h)+1)]\nfor height in range(0,h):########call to build the initial tree######\n\tif height==0:\n\t\tmedians(mylist,1)\n\telif height==1:\n\t\tmedians(mylist,3)\n\t\tparent=3\n\telif height >1:\n\t\tmedians(mylist,parent*2)\n\t\tparent=parent*2\nnode=leveltree(node)\ns=diff(node,h,0)\nmaximum1=999\nwhile(maximummaximum):\n\t\tmaximum1=maximum\n\t\ta=list()\n\t\tfor i in range(1,h+1):\n\t\t\tlprintTree(node,i)\n\tmaximum=0\t\n\tdiff(node,h,0)\t\nprint(maximum)#######returns our result i.e maximum of all the differences in the tree#######\nprint(a)######3retiurns the level based order of the tree#######\n\n ","sub_path":"totogram.py","file_name":"totogram.py","file_ext":"py","file_size_in_byte":5291,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"45946697","text":"import matplotlib.pyplot as plt\nimport numpy as np\nfrom scipy.stats import norm\nimport math\nimport matplotlib.colors as colors\n\nfrom matplotlib import cm\nfrom matplotlib import rc\n\n__author__ = 'ernesto'\n\n# if use latex or mathtext\nrc('text', usetex=True)\nrc('mathtext', fontset='cm')\n\n# auxiliar function for plot ticks of equal length in x and y axis despite its scales.\ndef convert_display_to_data_coordinates(transData, length=10):\n # create a transform which will take from display to data coordinates\n inv = transData.inverted()\n # transform from display coordinates to data coordinates in x axis\n data_coords = inv.transform([(0, 0), (length, 0)])\n # get the length of the segment in data units\n yticks_len = data_coords[1, 0] - data_coords[0, 0]\n # transform from display coordinates to data coordinates in y axis\n data_coords = inv.transform([(0, 0), (0, length)])\n # get the length of the segment in data units\n xticks_len = data_coords[1, 1] - data_coords[0, 1]\n return xticks_len, yticks_len\n\n\n#####################################\n# PARAMETERS - This can be modified #\n#####################################\n\n# normal pdf variances\nvar1 = 0.5\nvar2 = 2\nvar_std = 1\n# normal pdf mean\ntheta = 6\nepsilon = 1.5\n\n# maximum deviation from the mean where to plot each gaussian\nmax_mean_dev = 3.1 * var2\n\n#####################\n# END OF PARAMETERS #\n#####################\n\n# abscissa values\nxmin = theta - max_mean_dev\nxmax = theta + max_mean_dev\n\nx = np.linspace(xmin, xmax, 300)\n# normal distribution and density values in x\npdf_var1 = norm.pdf(x, theta, math.sqrt(var1))\npdf_var2 = norm.pdf(x, theta, math.sqrt(var2))\npdf_std = norm.pdf(x, theta, math.sqrt(var_std))\n\n\n# axis parameters\ndx = xmax / 20\nxmin_ax = xmin - dx\nxmax_ax = xmax + dx\n\nym = np.amax(pdf_var1)\nymax_ax = ym + ym / 10\nymin_ax = -ym / 10\n\n# areas to fill limits\npdf1_xinf = np.linspace(xmin, theta-epsilon, 50)\npdf1_inf = norm.pdf(pdf1_xinf, theta, math.sqrt(var1))\npdf1_xsup = np.linspace(theta+epsilon, xmax, 50)\npdf1_sup = norm.pdf(pdf1_xsup, theta, math.sqrt(var1))\npdf2_xinf = np.linspace(xmin, theta-epsilon, 50)\npdf2_inf = norm.pdf(pdf2_xinf, theta, math.sqrt(var2))\npdf2_xsup = np.linspace(theta+epsilon, xmax, 50)\npdf2_sup = norm.pdf(pdf2_xsup, theta, math.sqrt(var2))\n\nepsilon1 = epsilon / math.sqrt(var1)\nepsilon2 = epsilon / math.sqrt(var2)\npdfstd1_xinf = np.linspace(xmin, theta-epsilon1, 50)\npdfstd1_inf = norm.pdf(pdfstd1_xinf, theta, math.sqrt(var_std))\npdfstd1_xsup = np.linspace(theta+epsilon1, xmax, 50)\npdfstd1_sup = norm.pdf(pdfstd1_xsup, theta, math.sqrt(var_std))\npdfstd2_xinf = np.linspace(xmin, theta-epsilon2, 50)\npdfstd2_inf = norm.pdf(pdfstd2_xinf, theta, math.sqrt(var_std))\npdfstd2_xsup = np.linspace(theta+epsilon2, xmax, 50)\npdfstd2_sup = norm.pdf(pdfstd2_xsup, theta, math.sqrt(var_std))\n\n\n# length of the ticks for all subplot (6 pixels)\ndisplay_length = 6 # in pixels\n# x ticks labels margin\nxtm = -0.09\nytm = 0.4\n# font size\nfontsize = 14\n# colors from coolwarm\ncNorm = colors.Normalize(vmin=0, vmax=1)\nscalarMap = cm.ScalarMappable(norm=cNorm, cmap=cm.coolwarm)\ncol10 = scalarMap.to_rgba(0)\ncol20 = scalarMap.to_rgba(1)\n\nfig = plt.figure(0, figsize=(10, 6), frameon=False)\n\n# PLOT OF F(x | x < a)\nax = plt.subplot2grid((2, 8), (0, 0), rowspan=1, colspan=4)\n\nplt.xlim(xmin_ax, xmax_ax)\nplt.ylim(ymin_ax, ymax_ax)\n\n# horizontal and vertical ticks length\nxtl, ytl = convert_display_to_data_coordinates(ax.transData, length=display_length)\n\n# axis arrows\nplt.annotate(\"\", xytext=(xmin_ax, 0), xycoords='data', xy=(xmax_ax, 0), textcoords='data',\n arrowprops=dict(width=0.1, headwidth=6, headlength=8, facecolor='black', shrink=0.002))\nplt.annotate(\"\", xytext=(0, ymin_ax), xycoords='data', xy=(0, ymax_ax), textcoords='data',\n arrowprops=dict(width=0.1, headwidth=6, headlength=8, facecolor='black', shrink=0.002))\n\nplt.plot(x, pdf_var1, color='k', linewidth=2)\n\n# filled areas\nax.fill_between(pdf1_xinf, 0, pdf1_inf, color=col10)\nax.fill_between(pdf1_xsup, 0, pdf1_sup, color=col10)\n\n# xlabels and xtickslabels\nplt.plot([theta, theta], [0, xtl], 'k')\nplt.plot([theta-epsilon, theta-epsilon], [0, xtl], 'k')\nplt.plot([theta+epsilon, theta+epsilon], [0, xtl], 'k')\nplt.text(theta, xtm, '$\\\\theta$', fontsize=fontsize, ha='center', va='baseline')\nplt.text(theta-epsilon, xtm, '$\\\\theta-\\epsilon$', fontsize=fontsize, ha='center', va='baseline')\nplt.text(theta+epsilon, xtm, '$\\\\theta+\\epsilon$', fontsize=fontsize, ha='center', va='baseline')\nplt.text(xmax_ax, xtm, '$\\hat{\\\\theta}$', fontsize=fontsize, ha='right', va='baseline')\nplt.text(ytm, ymax_ax, '$p(\\hat{\\\\theta})=\\mathcal{N}(\\\\theta,\\,\\sigma^2_{\\hat{\\\\theta}})$',\n fontsize=fontsize, ha='left', va='center')\n\n\nplt.text(xmax_ax+0.4, ymax_ax, '$\\sigma^2_{\\hat{\\\\theta}}<\\sigma^2_{\\check{\\\\theta}}$',\n fontsize=fontsize, ha='center', va='center')\n\nplt.axis('off')\n\n\n##\nax = plt.subplot2grid((2, 8), (0, 4), rowspan=1, colspan=4)\n\nplt.xlim(xmin_ax, xmax_ax)\nplt.ylim(ymin_ax, ymax_ax)\n\n# axis arrows\nplt.annotate(\"\", xytext=(xmin_ax, 0), xycoords='data', xy=(xmax_ax, 0), textcoords='data',\n arrowprops=dict(width=0.1, headwidth=6, headlength=8, facecolor='black', shrink=0.002))\nplt.annotate(\"\", xytext=(theta, ymin_ax), xycoords='data', xy=(theta, ymax_ax), textcoords='data',\n arrowprops=dict(width=0.1, headwidth=6, headlength=8, facecolor='black', shrink=0.002))\n\nplt.plot(x, pdf_std, color='k', linewidth=2)\n\n# filled areas\nax.fill_between(pdfstd1_xinf, 0, pdfstd1_inf, color=col10)\nax.fill_between(pdfstd1_xsup, 0, pdfstd1_sup, color=col10)\n\nxtm2 = -0.11\n# xlabels and xtickslabels\nplt.plot([theta-epsilon1, theta-epsilon1], [0, xtl], 'k')\nplt.plot([theta+epsilon1, theta+epsilon1], [0, xtl], 'k')\n# plt.text(theta-epsilon1, xtm, '$$-\\epsilon/\\sqrt{\\\\textrm{var}(\\hat{\\\\theta})}$$',\n# fontsize=fontsize, ha='center', va='baseline')\n# plt.text(theta-epsilon1, xtm, '$$-\\\\frac{\\epsilon}{\\sigma_{\\hat{\\\\theta}}}$$', fontsize=fontsize, ha='center', va='baseline')\nplt.text(theta-epsilon1, xtm2, '$-\\epsilon/\\sigma_{\\hat{\\\\theta}}$', fontsize=fontsize, ha='center', va='baseline')\nplt.text(theta+epsilon1, xtm2, '$\\epsilon/\\sigma_{\\hat{\\\\theta}}$', fontsize=fontsize, ha='center', va='baseline')\nplt.text(xmax_ax, xtm2, '$(\\hat{\\\\theta}-\\\\theta$)/\\sigma_{\\hat{\\\\theta}}',\n fontsize=fontsize, ha='center', va='baseline')\nplt.text(theta + ytm, ymax_ax, '$p((\\hat{\\\\theta}-\\\\theta$)/\\sigma_{\\hat{\\\\theta}})=\\mathcal{N}(0,\\,1)$',\n fontsize=fontsize, ha='left', va='center')\nplt.axis('off')\n\n\n#########################\n#########################\n\nax = plt.subplot2grid((2, 8), (1, 0), rowspan=1, colspan=4)\n\nplt.xlim(xmin_ax, xmax_ax)\nplt.ylim(ymin_ax, ymax_ax)\n\n# axis arrows\nplt.annotate(\"\", xytext=(xmin_ax, 0), xycoords='data', xy=(xmax_ax, 0), textcoords='data',\n arrowprops=dict(width=0.1, headwidth=6, headlength=8, facecolor='black', shrink=0.002))\nplt.annotate(\"\", xytext=(0, ymin_ax), xycoords='data', xy=(0, ymax_ax), textcoords='data',\n arrowprops=dict(width=0.1, headwidth=6, headlength=8, facecolor='black', shrink=0.002))\n\nplt.plot(x, pdf_var2, color='k', linewidth=2)\n\n# filled areas\nax.fill_between(pdf2_xinf, 0, pdf2_inf, color=col10)\nax.fill_between(pdf2_xsup, 0, pdf2_sup, color=col10)\n\n# xlabels and xtickslabels\nplt.plot([theta, theta], [0, xtl], 'k')\nplt.plot([theta-epsilon, theta-epsilon], [0, xtl], 'k')\nplt.plot([theta+epsilon, theta+epsilon], [0, xtl], 'k')\nplt.text(theta, xtm, '$\\\\theta$', fontsize=fontsize, ha='center', va='baseline')\nplt.text(theta-epsilon, xtm, '$\\\\theta-\\epsilon$', fontsize=fontsize, ha='center', va='baseline')\nplt.text(theta+epsilon, xtm, '$\\\\theta+\\epsilon$', fontsize=fontsize, ha='center', va='baseline')\nplt.text(xmax_ax, xtm, '$\\check{\\\\theta}$', fontsize=fontsize, ha='right', va='baseline')\nplt.text(ytm, ymax_ax, '$p(\\check{\\\\theta})=\\mathcal{N}(\\\\theta,\\,\\sigma^2_{\\check{\\\\theta}})$',\n fontsize=fontsize, ha='left', va='center')\nplt.axis('off')\n\n##\nax = plt.subplot2grid((2, 8), (1, 4), rowspan=1, colspan=4)\n\nplt.xlim(xmin_ax, xmax_ax)\nplt.ylim(ymin_ax, ymax_ax)\n\n# axis arrows\nplt.annotate(\"\", xytext=(xmin_ax, 0), xycoords='data', xy=(xmax_ax, 0), textcoords='data',\n arrowprops=dict(width=0.1, headwidth=6, headlength=8, facecolor='black', shrink=0.002))\nplt.annotate(\"\", xytext=(theta, ymin_ax), xycoords='data', xy=(theta, ymax_ax), textcoords='data',\n arrowprops=dict(width=0.1, headwidth=6, headlength=8, facecolor='black', shrink=0.002))\n\nplt.plot(x, pdf_std, color='k', linewidth=2)\n\n# filled areas\nax.fill_between(pdfstd2_xinf, 0, pdfstd2_inf, color=col10)\nax.fill_between(pdfstd2_xsup, 0, pdfstd2_sup, color=col10)\n\nxtm2 = -0.11\n# xlabels and xtickslabels\nplt.plot([theta-epsilon2, theta-epsilon2], [0, xtl], 'k')\nplt.plot([theta+epsilon2, theta+epsilon2], [0, xtl], 'k')\nplt.text(theta-epsilon2, xtm2, '$-\\epsilon/\\sigma_{\\check{\\\\theta}}$', fontsize=fontsize, ha='center', va='baseline')\nplt.text(theta+epsilon2, xtm2, '$\\epsilon/\\sigma_{\\check{\\\\theta}}$', fontsize=fontsize, ha='center', va='baseline')\nplt.text(xmax_ax, xtm2, '$(\\check{\\\\theta}-\\\\theta$)/\\sigma_{\\check{\\\\theta}}',\n fontsize=fontsize, ha='center', va='baseline')\nplt.text(theta + ytm, ymax_ax, '$p((\\check{\\\\theta}-\\\\theta$)/\\sigma_{\\check{\\\\theta}})=\\mathcal{N}(0,\\,1)$',\n fontsize=fontsize, ha='left', va='center')\nplt.axis('off')\n\n\n# save as pdf image\nplt.savefig('problem_2_7.pdf', bbox_inches='tight')\n\nplt.show()\n\n","sub_path":"figuras/PycharmKayStatisticalReport/problem_2_7.py","file_name":"problem_2_7.py","file_ext":"py","file_size_in_byte":9522,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"402047582","text":"from flask import Flask, request, send_from_directory\nimport towels\napp = Flask(__name__)\n\n@app.route('/', defaults={'path': 'index.html'})\n@app.route('//')\ndef index(path):\n return send_from_directory('pages', path)\n\n@app.route('/gen/', defaults={'num': 3, 'corpus': 'full'})\n@app.route('/gen//', defaults={'corpus': 'full'})\n@app.route('/gen///')\ndef generate(num, corpus):\n if num > 42:\n return 'Sentence limit exceeded.'\n return towels.generate(num, corpus)\n\n@app.route('/gen/s/', defaults={'char': 140, 'corpus': 'full'})\n@app.route('/gen/s//', defaults={'corpus': 'full'})\n@app.route('/gen/s///')\ndef generate_sentence(char, corpus):\n if char > 420:\n return 'Character limit exceeded.'\n sentence = towels.generate_sentence(char, corpus)\n if sentence is None:\n return ''\n return sentence\n\nif __name__ == '__main__':\n app.run(host='0.0.0.0', port=22109)\n\n","sub_path":"webserv.py","file_name":"webserv.py","file_ext":"py","file_size_in_byte":983,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"594464271","text":"# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /usr/local/lib/python2.7/dist-packages/jeju/executor/editor.py\n# Compiled at: 2016-11-09 20:33:52\nimport string, ConfigParser, io, logging\n\ndef replaceable(code, kv):\n keys = kv.keys()\n for key in keys:\n nkey = '${%s}' % key\n code = string.replace(code, nkey, kv[key])\n\n logging.debug('####################' + '\\n%s' % code)\n logging.debug('####################')\n return code\n\n\ndef find_file_path(lookahead):\n if lookahead == None:\n return\n else:\n ctx = lookahead['text']\n items = ctx.split()\n if items[0] == 'edit':\n return items[1]\n return\n\n\ndef editor_text(**kwargs):\n lookahead = kwargs['lookahead']\n code = kwargs['code']\n kv = kwargs['kv']\n file_path = find_file_path(kwargs['lookahead'])\n if file_path == None:\n msg = 'Cannot find content:%s' % lookahead['text']\n logging.error(msg)\n return msg\n else:\n fp = open(file_path, 'w')\n rcode = replaceable(code, kv)\n fp.write(rcode)\n fp.close()\n return {'output': rcode}\n\n\ndef editor_ini(**kwargs):\n lookahead = kwargs['lookahead']\n code = kwargs['code']\n kv = kwargs['kv']\n added = ConfigParser.RawConfigParser(allow_no_value=True)\n rcode = replaceable(code, kv)\n added.readfp(io.BytesIO(rcode))\n file_path = find_file_path(kwargs['lookahead'])\n if file_path == None:\n msg = 'Cannot find content path: %s' % lookahead['text']\n logging.error(msg)\n return msg\n else:\n orig = ConfigParser.ConfigParser()\n orig.readfp(open(file_path))\n for section in added.sections():\n if orig.has_section(section) == False:\n msg = 'Add new section'\n logging.debug(msg)\n orig.add_section(section)\n for item, value in added.items(section):\n if item == '...':\n msg = 'abbreviation'\n else:\n orig.set(section, item, value)\n\n fp = open(file_path, 'w')\n orig.write(fp)\n new_content = orig.readfp(open(file_path))\n fp.close()\n return {'output': new_content}","sub_path":"pycfiles/jeju-0.3.6-5.linux-x86_64.tar/editor.py","file_name":"editor.py","file_ext":"py","file_size_in_byte":2336,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"581111198","text":"from environment import spark\nfrom pyspark.ml.feature import Word2Vec\n\ndocumentDF = spark.createDataFrame([\n (\"Hi I heard about Spark\".split(\" \"), ),\n (\"I wish Java could use case classes\".split(\" \"), ),\n (\"Logistic regression models are neat\".split(\" \"), )\n], [\"text\"])\n\nword2Vec = Word2Vec(vectorSize=3, minCount=0, inputCol=\"text\", outputCol=\"result\")\nmodel = word2Vec.fit(documentDF)\n\nresult = model.transform(documentDF)\nresult.show(truncate=False)\n","sub_path":"basic-mllib/Word2Vec_/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":463,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"86314463","text":"# coding: utf-8\n\n# Usage\n# reduce_jsons_4_trajectories.pyで生成されたファイルから、点群を抜き出し、ミーンシフト法によって、\n# クラスタリングをし、各クラスタの収束点を出力する。\n# 第1引数で平均を取る半径の長さをメートルで指定。第2引数で収束条件距離を指定。\n# python3 src/mean_shift/mean_shift.py 100.0 10.0\n\nimport json\nimport sys\nimport glob\nimport calendar\nimport datetime\nimport os\nimport math\n\nparam = sys.argv\ncurrent_time = datetime.datetime.today()\n\nradius_of_mean = float(param[1])\nradius_of_convergence = float(param[2])\n\ndirectory_path = 'result/mean_shift/result_' + str(current_time.year) + '_' + str(current_time.month) + '_' + str(current_time.day) + '_' + str(current_time.hour) + '_' + str(current_time.minute)\nfw_points_path = directory_path + '/points.json'\nfw_groups_path = directory_path + '/groups.json'\nfr_reduced_path = 'data/mean_shift/reduced_jsons/reduced.json'\n\ndef latlngToDistance(lat1, lng1, lat2, lng2):\n\t# 定数 ( GRS80 ( 世界測地系 ) )\n\tGRS80_R_X = 6378137.000000 # 赤道半径\n\tGRS80_R_Y = 6356752.314140 # 極半径\n\tr_x = GRS80_R_X\n\tr_y = GRS80_R_Y\n\tdif_lat = math.pi * (lat1 - lat2) / 180.0\n\tdif_lng = math.pi * (lng1 - lng2) / 180.0\n\tmean_lat = math.pi * (lat1 + lat2) / 180.0 / 2.0\n\teccentricity = math.sqrt(( r_x ** 2 - r_y ** 2 ) / ( r_x ** 2 ))\n\tw = math.sqrt(1.0 - (eccentricity ** 2) * (math.sin(mean_lat) ** 2))\n\tm = r_x * ( 1.0 - eccentricity ** 2 ) / ( w ** 3 )\n\tn = r_x / w\n\td = math.sqrt((dif_lng * m) ** 2 + (dif_lat * n * math.cos(mean_lat)) ** 2)\n\treturn d\n\ndef decideNextPoint(this_point):\n\tmean_lat = 0.0\n\tmean_lng = 0.0\n\tsum_lat = 0.0\n\tsum_lng = 0.0\n\tnum_elements = 0.0\n\tglobal is_convergent\n\tfor point in points:\n\t\tdistance_between_2_points = latlngToDistance(this_point[\"present_point\"][\"lat\"], this_point[\"present_point\"][\"lng\"], point[\"present_point\"][\"lat\"], point[\"present_point\"][\"lng\"])\n\t\tif not distance_between_2_points <= radius_of_mean:\n\t\t\tcontinue\n\t\tnum_elements += 1.0\n\t\tsum_lat += point[\"present_point\"][\"lat\"]\n\t\tsum_lng += point[\"present_point\"][\"lng\"]\n\t\tif not distance_between_2_points <= radius_of_convergence:\n\t\t\tis_convergent = False\n\tmean_lat = sum_lat / num_elements\n\tmean_lng = sum_lng / num_elements\n\tthis_point[\"next_point\"][\"lat\"] = mean_lat\n\tthis_point[\"next_point\"][\"lng\"] = mean_lng\n\treturn this_point\n\ndef updatePoint(this_point):\n\tthis_point[\"present_point\"][\"lat\"] = this_point[\"next_point\"][\"lat\"]\n\tthis_point[\"present_point\"][\"lng\"] = this_point[\"next_point\"][\"lng\"]\n\treturn this_point\n\ndef decideGroupID():\n\tglobal points\n\tcurrent_group_id = 0\n\tfor pointA in points:\n\t\tif \"group_id\" in pointA:\n\t\t\tcontinue\n\t\tpointA[\"group_id\"] = current_group_id\n\t\tcurrent_group_id += 1\n\t\tfor pointB in points:\n\t\t\tif \"group_id\" in pointB:\n\t\t\t\tcontinue\n\t\t\tdistance_between_2_points = latlngToDistance(pointA[\"present_point\"][\"lat\"], pointA[\"present_point\"][\"lng\"], pointB[\"present_point\"][\"lat\"], pointB[\"present_point\"][\"lng\"])\n\t\t\tif distance_between_2_points <= radius_of_convergence:\n\t\t\t\tpointB[\"group_id\"] = pointA[\"group_id\"]\n\tnum_group = current_group_id\n\treturn num_group\n\ndef decideMeansOfEachGroup():\n\tmeans_of_each_group = []\n\tfor i in range(num_group):\n\t\tmean_point = {\"group_id\": i ,\"lat\": 0.0, \"lng\": 0.0, \"num_points\": 0}\n\t\tfor point in points:\n\t\t\tif point[\"group_id\"] == i:\n\t\t\t\tmean_point[\"lat\"] += point[\"present_point\"][\"lat\"]\n\t\t\t\tmean_point[\"lng\"] += point[\"present_point\"][\"lng\"]\n\t\t\t\tmean_point[\"num_points\"] += 1\n\t\tmean_point[\"lat\"] = mean_point[\"lat\"] / mean_point[\"num_points\"]\n\t\tmean_point[\"lng\"] = mean_point[\"lng\"] / mean_point[\"num_points\"]\n\t\tmeans_of_each_group.append(mean_point)\n\treturn means_of_each_group\n\n\nprint(\"reading file\")\nprint(str(datetime.datetime.today()))\n\nfr_reduced = open(fr_reduced_path,'r')\npoints = []\nfor line in fr_reduced:\n\ttweet = json.loads(line)\n\tpoints.append({\"tweet_id\": tweet[\"tweet_id\"], \"present_point\": {\"lat\": float(tweet[\"coordinates\"][1]), \"lng\": float(tweet[\"coordinates\"][0])}, \"next_point\": {\"lat\": 0.0, \"lng\": 0.0}})\n\n\nprint(\"doing mean-shift\")\nprint(str(datetime.datetime.today()))\n\nis_convergent = False\nwhile(is_convergent == False):\n\tis_convergent = True\n\tfor i, point in enumerate(points):\n\t\tpoints[i] = decideNextPoint(point)\n\tfor i, point in enumerate(points):\n\t\tpoints[i] = updatePoint(point)\n\tprint(str(datetime.datetime.today()))\n\n\nprint(\"deciding the group of each point\")\nprint(str(datetime.datetime.today()))\n\t\t\nnum_group = decideGroupID()\n\n\nprint(\"deciding the mean point of each group\")\nprint(str(datetime.datetime.today()))\n\nmeans_of_each_group = decideMeansOfEachGroup()\n\n\nprint(\"outputting\")\nprint(str(datetime.datetime.today()))\n\nos.mkdir(directory_path)\n\nfr_reduced = open('data/mean_shift/reduced_jsons/reduced.json','r')\nfor i,line in enumerate(fr_reduced):\n\ttweet = json.loads(line)\n\ttweet.update({\"group_id\": points[i][\"group_id\"]})\n\tfw = open(fw_points_path, 'a')\n\tfw.write(json.dumps(tweet))\n\tfw.write('\\n')\n\tfw.close\n\nfor mean in means_of_each_group:\n\tfw = open(fw_groups_path, 'a')\n\tfw.write(json.dumps(mean))\n\tfw.write('\\n')\n\tfw.close\n\n","sub_path":"src/mean_shift/mean_shift.py","file_name":"mean_shift.py","file_ext":"py","file_size_in_byte":5113,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"206145365","text":"def decodificar(mensaje):\n linea=mensaje.split(\",\")\n palabra=\"\"\n n=0\n\n for i in linea:\n n=decimal(i)\n palabra+=chr(n)\n\n return palabra\n\ndef decimal(n):\n n=int(n,2)\n return n\n\n\nif __name__ == \"__main__\":\n mensaje=decodificar(\"01101000,01101111,01101100,01100001\")\n print(mensaje)\n","sub_path":"tema9_ej3/tema9_ej3_1563468J.py","file_name":"tema9_ej3_1563468J.py","file_ext":"py","file_size_in_byte":320,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"644081935","text":"\n\n# Iowa Road Conditions processing script\n# daryl herzmann akrherz@iastate.edu 19 Nov 2009\n# installed by Shane Searcy, ITO DMX, shane.searcy@noaa.gov\n#\n# REQUIRES: the 'mx' package installed (standard yum install mx)\n#\n# 1) Download file from website\n# 2) Compare with old file to see if they are any different\n# IF DIFFERENT\n# 3) Replace Headers with NWS mandated stuff\n# 4) Write file to hard drive\n# 5) scp file to ls1-dmx:/data/Incoming for dissemination\n\nHTTP_SRC = \"http://ia.carsprogram.org/IAcarssegment/IA_road_conditions.txt\"\n#FINAL_LOCATION = \"/data/Incoming/WAN_NWWSDSMSTOIA.dat\"\nFINAL_LOCATION = \"/tmp/LOC_DSMSTOIA.dat\"\nLOG_FILENAME = \"/tmp/STOIA_acquisition.log\"\nPREV_PRODUCT = \"/tmp/prevSTOIA.txt\"\n\nimport urllib2, logging, traceback, sys, os, tempfile, StringIO\nimport datetime\nLOGFORMAT = \"%(asctime)-15s:: %(message)s\"\nlogging.basicConfig(filename=LOG_FILENAME,level=logging.DEBUG,\n format=LOGFORMAT)\n\ndef compare_product( newdata ):\n \"\"\"\n Compares newly downloaded data with previously saved version\n @return boolean if the data is new!\n \"\"\"\n if not os.path.isfile( PREV_PRODUCT ):\n logging.debug(\"Previous datafile %s not found\" % (PREV_PRODUCT,))\n return True\n\n # Always send the 3:01 a.m. report\n now = datetime.datetime.now()\n if now.minute == 1 and now.hour == 3:\n return True\n\n # Make sure the product is complete...\n if newdata.find(\"800-762-3947\") == -1:\n return False\n\n olddata = open( PREV_PRODUCT, 'r').read()\n if olddata != newdata:\n logging.debug(\"Datafile is new!\")\n return True\n\n return False\n\ndef ship2ldad( data ):\n \"\"\"\n Writes the data to a file for LDAD to then deal with\n \"\"\"\n f = open (\"/tmp/pre_ldad_STOIA.txt\", 'w')\n f.write( data )\n f.close()\n\n\n logging.debug(\"Shipping %s product to LDAD via scp\" % (f.name,) )\n os.system(\"cp %s %s\" % (f.name, FINAL_LOCATION) )\n os.system(\"python /home/ldm/pyWWA/util/make_text_noaaportish.py %s\" % (FINAL_LOCATION,))\n os.system(\"cat %s | python /home/ldm/pyWWA/parsers/stoia_parser.py\" % (FINAL_LOCATION,))\n\ndef fix_header( data ):\n \"\"\"\n Fixes the header the file has to make NWS protocols\n @return String fixed file\n \"\"\"\n # Formulate the new header\n now = datetime.datetime.now()\n utcnow = datetime.datetime.utcnow()\n newdata = \"\"\"000\nSXUS43 KDMX %s\nSTOIA\n\nIOWA ROAD CONDITIONS\nIOWA DEPARTMENT OF PUBLIC SAFETY\nRELAYED BY THE NATIONAL WEATHER SERVICE DES MOINES IA\n%s\n\n\"\"\" % (utcnow.strftime(\"%d%H%M\"), (now.strftime(\"%-I%M %p CST %a %b %d %Y\").upper()),)\n\n # Strip off everything before the first *\n return newdata + data[data.find(\"*\"):]\n return newdata + data \n \ndef save_data( data ):\n \"\"\"\n Save the data in a file for future comparisons\n \"\"\"\n f = open( PREV_PRODUCT , 'w')\n f.write( data )\n f.close()\n logging.debug(\"Saved downloaded data to %s\" % (PREV_PRODUCT,))\n\nlogging.debug(\"_______________ Starting download\")\ntry:\n data = urllib2.urlopen( HTTP_SRC ).read()\n data = data[data.find(\"*\"):]\n\nexcept:\n logging.error(\"Download Failure!, Abort\")\n ebuf = StringIO.StringIO()\n traceback.print_exc(file=ebuf)\n ebuf.seek(0)\n logging.error( ebuf.read() )\n logging.debug(\"__ END\")\n sys.exit()\n\nlogging.debug(\"Downloaded %s bytes\" % (len(data),))\n\nisnew = compare_product( data )\n\nsave_data(data)\n\nif isnew:\n data = fix_header( data )\n ship2ldad( data )\n\nlogging.debug(\"__ END\")\n","sub_path":"scripts/roads/stoia.py","file_name":"stoia.py","file_ext":"py","file_size_in_byte":3487,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"162443644","text":"def remove_files(file_names):\n \"\"\" removes files \"\"\"\n\n import os, glob\n\n list_files= glob.glob(file_names + '*')\n\n try:\n for f in list_files:\n os.remove(f)\n print('FILES REMOVED')\n except:\n print('FAIL TO REMOVE FILES')\n\n\ndef save_it(data, file_name):\n \"\"\" save list: data, filename \"\"\"\n\n def timeStamped(fname, fmt='{fname}_%Y-%m-%d-%H%M%S.txt'):\n import datetime\n return datetime.datetime.now().strftime(fmt).format(fname=fname)\n\n try:\n with open(timeStamped(file_name), \"w\") as f:\n f.write('\\n'.join(str(line) for line in data))\n\n print(\"DATA SAVED\")\n\n except:\n print(\"FAIL TO SAVE DATA\")\n\n\ndef load_it(file_name):\n\n try:\n with open(file_name, 'r') as f:\n data = f.read()\n except:\n print(\"FAIL to READ DATA!\")\n\n data_list = [number for number in data.split(\"\\n\")]\n print(file_name, \"LOADED \")\n return(data_list)\n","sub_path":"p51/_tools.py","file_name":"_tools.py","file_ext":"py","file_size_in_byte":961,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"497406350","text":"import numpy as np\n\nimport time\nimport util as u\n\n# simulation parameters\ndt = 0.1 # time step\nT = 3000 # number of time steps\n\n# parameters of the system\ngamma = .00002 # gravitational coefficient\nm1 = 300 # mass of the first planet\nm2 = 1 # mass of the second planet\n\n# initial conditions: state of the system in the first two time steps\nx0 = np.array([0, 0, 1, 0])\nx1 = np.array([0, 0, 1, 0.005])\n\n# initializing the vector of system states\nxs = [x0, x1]\n\n# defining the function that calculates the forces acting on the planets at a time step\ndef F(x):\n r1 = x[:2] # position of the first planet\n r2 = x[2:4] # position of the second planet\n \n posdiff_vec = r2 - r1 # vector of difference of the positions\n dist = np.linalg.norm(posdiff_vec) # distance between the planets\n \n # Newton's law of universal gravitation\n F1 = gamma * m2 * posdiff_vec / (dist**2) # force acting on the first planet\n F2 = - gamma * m1 * posdiff_vec / (dist**2) # force acting on the second planet\n \n return np.concatenate([F1, F2], axis=0)\n\n\nfor i in range(T):\n # getting previous values of the iteration\n Lx = xs[-1]\n LLx = xs[-2]\n \n # approximating the second derivative with finite differences\n F_prev = F(Lx) # force acting at the previous time step\n x = dt**2*F_prev + 2*Lx - LLx # estimated position at the next time step\n \n xs.append(x)\n\n####################################################\n\nx, y = np.split(np.array(xs), 2, axis=1)\n\nu.plotAnim(x, y, T, isSaveVideo=False)\n\n# this is slow and not working properly\n#u.plotScatter(x, y)\n","sub_path":"planets/planets.py","file_name":"planets.py","file_ext":"py","file_size_in_byte":1655,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"301685895","text":"'''\r\nhttps://adventofcode.com/2020/day/7\r\n'''\r\nwith open(\"input.txt\") as f:\r\n line = f.readline()\r\n inputs = {}\r\n while line:\r\n if line.split():\r\n # set the parent bag color as the KEY and the colors it holds (VALUE) as an array\r\n parentBag = ' '.join(line.split()[:3]).replace(\"bags\", \"\")\r\n childrenBags = []\r\n innerBags = ' '.join(line.split()[4:])\r\n # we don't care HOW MANY of each bag it can hold, so we filter that out\r\n for i in innerBags.split(\", \"):\r\n # append the filtered result to childrenBags array\r\n childrenBags.append(' '.join(i.split()[1:-1]))\r\n inputs[parentBag.strip()] = childrenBags\r\n line = f.readline()\r\n#print(json.dumps(inputs, indent=2))\r\n\r\n\r\ndef numBags(color):\r\n '''\r\n I had a tough time with this one, so after hours of trying, I decided to seek help online\r\n THIS CODE LOGIC CAME FROM https://www.youtube.com/watch?v=7IOd7wvxDX0\r\n Highly recommend watching. He explains it really well :)\r\n '''\r\n containsColor = [] # keeps track of the bags that hold the specified color\r\n for k, v in inputs.items():\r\n # check to see if the colors are in the value arrays\r\n for el in v:\r\n if color == el:\r\n # if it is, append it to containsColor array\r\n containsColor.append(k)\r\n checkedColors = []\r\n if len(containsColor) == 0:\r\n return []\r\n else:\r\n for color in containsColor:\r\n checkedColors.append(color)\r\n checkedColors += numBags(color)\r\n return set(checkedColors)\r\n\r\n\r\nprint(len(numBags(\"shiny gold\")))","sub_path":"day07/lol.py","file_name":"lol.py","file_ext":"py","file_size_in_byte":1682,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"653149172","text":"from gevent import monkey\n\nmonkey.patch_all()\n\n#从gevent库里导入monkey模块\n\nimport gevent\n\nimport time\n\nimport requests\n\nfrom gevent.queue import Queue\nmonkey.patch_all()\n\n#从gevent库里导入queue模块\n\n\n#monkey.patch_all()能把程序变成协作式运行,就是可以帮助程序实现异步。\n\nstart = time.time()\n\n\n\nurl_list = [\n \"https://www.kaikeba.com/\",\n \"https://www.csdn.net/\",\n \"https://www.json.cn/\",\n \"https://cn.bing.com/\",\n \"https://www.jianshu.com/\",\n \"http://www.techweb.com.cn/\",\n \"https://www.bilibili.com/\",\n \"https://www.huxiu.com/\"\n]\n\n\n\nwork = Queue()\n\n#创建队列对象,并赋值给work\n\nfor url in url_list:\n\n#遍历url_list\n\n work.put_nowait(url)\n\n #用put_nowait()函数可以把网址都放进队列里\n\n\n\ndef crawler():\n\n while not work.empty():\n\n #当队列不是空的时候,就执行下面的程序\n\n url = work.get_nowait()\n\n #用get_nowait()函数可以把队列里的网址都取出\n\n r = requests.get(url)\n\n #用requests.get()函数抓取网址\n\n print(url,work.qsize(),r.status_code)\n\n #打印网址、队列长度、抓取请求的状态码\n\n\n\ntasks_list = [ ]\n\n#创建空的任务列表\n\nfor x in range(2):\n\n#相当于创建了2个爬虫\n\n task = gevent.spawn(crawler)\n\n #用gevent.spawn()函数创建执行crawler()函数的任务\n\n tasks_list.append(task)\n\n #往任务列表添加任务。\n\ngevent.joinall(tasks_list)\n\n#用gevent.joinall方法,执行任务列表里的所有任务,就是让爬虫开始爬取网站\n\nend = time.time()\n\nprint(end-start)","sub_path":"python_optional_class/Web crawler/gevent库和Queue模块来实现多协程.py","file_name":"gevent库和Queue模块来实现多协程.py","file_ext":"py","file_size_in_byte":1606,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"513404785","text":"### Settings for a \"cursor tracing\" visualization. ###\r\n\r\ninput_video = \"input.mp4\" # name of a video to process\r\ndir_name = \"TEMP\" # directory name in which the frames will be extracted\r\ncursor_path = \"cursor.png\" # path to a cursor png to locate in frames\r\n\r\nstart = 0 # start point (in seconds) of a video, set to 0 if not specified\r\nend = 60 # end point (in seconds) of a video, set to 0 if not specified\r\n\r\nvideo_resolution = (1920, 1080) # resolution of a input video\r\npygame_resolution = (960, 540) # resolution of a visualization in pygame\r\nfps = 60 # fps of a video\r\n\r\nconfidence = 0.45 # confidence in searchng for a cursor\r\nmax_limit = 50 # maximum amount of pixel distance from a previous location","sub_path":"config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":729,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"102770318","text":"X, Y = [], []\nfor y, cmd in enumerate(list_cmd):\n folder = 'dataset/%s/' % cmd\n for filename in os.listdir(folder):\n sr, signal = wavfile.read(folder + filename)\n signal = np.pad(signal,\n (0, 16000 - signal.shape[0]), \n 'constant', \n constant_values=(0, 0))\n assert signal.shape[0] == 16000\n X.append(preprocess.get_feature(signal))\n Y.append(y)\nX = np.array(X)\nY = np.array(Y)\nnp.save('dataset/X.npy', X)\nnp.save('dataset/Y.npy', Y)","sub_path":"py/archive/prepare.py","file_name":"prepare.py","file_ext":"py","file_size_in_byte":543,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"111063752","text":"import turtle\r\nfrom random import random, randrange\r\nimport numpy as np\r\nimport math as mt\r\n\r\n# координати паралелепіпеда\r\nxw=600\r\nyw=600\r\nst=300\r\n# розташування координат у строках: дальній чотирикутник - A B I M, ближній чотирикутник D C F E\r\nPrlpd = np.array([[0, 0, 0, 1],\r\n [st, 0, 0, 1],\r\n [st, st, 0, 1],\r\n [0, st, 0, 1],\r\n [0, 0, st, 1],\r\n [st, 0, st, 1],\r\n [st, st, st, 1],\r\n [0, st, st, 1]])\r\n# функция проекції на xy, z=0\r\ndef ProjectXY(Figure):\r\n f = np.array([ [1, 0, 0, 0], [0, 1, 0, 0], [0, 0, 0, 0], [0, 0, 0, 1] ]) # по строках\r\n ft=f.T\r\n Prxy = Figure.dot(ft)\r\n return Prxy\r\n# зміщення\r\ndef ShiftXYZ(Figure, l, m, n):\r\n f = np.array([[1, 0, 0, l],\r\n [0, 1, 0, m],\r\n [0, 0, 1, n],\r\n [1, 0, 0, 1]]) # по строках\r\n ft=f.T\r\n Prxy = Figure.dot(ft)\r\n return Prxy\r\n# обертання коло х\r\ndef insertX(Figure, TetaG):\r\n TetaR=(3/14*TetaG)/180\r\n f = np.array([[1, 0, 0, 0],\r\n [0, mt.cos(TetaR), mt.sin(TetaR), 0],\r\n [0, -mt.sin(TetaR), mt.cos(TetaR), 0],\r\n [0, 0, 0, 1]])\r\n ft=f.T\r\n Prxy = Figure.dot(ft)\r\n return Prxy\r\n# аксонометрія\r\ndef dimetri(Figure, TetaG1, TetaG2):\r\n TetaR1=(3/14*TetaG1)/180\r\n TetaR2=(3/14*TetaG2)/180\r\n f1 = np.array([[mt.cos(TetaR1), 0 , -mt.sin(TetaR1), 0],\r\n [0, 1, 0, 0],\r\n [mt.sin(TetaR1), 0, mt.cos(TetaR1), 1],\r\n [0, 0, 0, 0],])\r\n ft1 = f1.T\r\n Prxy1 = Figure.dot(ft1)\r\n f2 = np.array([[1, 0, 0, 0],\r\n [0, mt.cos(TetaR2), mt.sin(TetaR2), 0],\r\n [0, -mt.sin(TetaR2), mt.cos(TetaR2), 0],\r\n [0, 0, 0, 1]])\r\n ft2=f2.T\r\n Prxy2 = Prxy1.dot(ft2)\r\n return Prxy2\r\n\r\n\r\n# функція побудови растрового паралелепіпеда\r\ndef PrlpdWiz(Prxy3):\r\n # дальня грань - (в проекції ліва)\r\n Ax1 = Prxy3[0, 0]\r\n Ay1 = Prxy3[0, 1]\r\n Bx1 = Prxy3[1, 0]\r\n By1 = Prxy3[1, 1]\r\n Ix1 = Prxy3[2, 0]\r\n Iy1 = Prxy3[2, 1]\r\n Mx1 = Prxy3[3, 0]\r\n My1 = Prxy3[3, 1]\r\n # ближня грань - (в проекції права)\r\n Dx1 = Prxy3[4, 0]\r\n Dy1 = Prxy3[4, 1]\r\n Cx1 = Prxy3[5, 0]\r\n Cy1 = Prxy3[5, 1]\r\n Fx1 = Prxy3[6, 0]\r\n Fy1 = Prxy3[6, 1]\r\n Ex1 = Prxy3[7, 0]\r\n Ey1 = Prxy3[7, 1]\r\n\r\n # дальня грань - (в проекції ліва)\r\n turtle.up()\r\n turtle.goto(Ax1, Ay1)\r\n turtle.down()\r\n turtle.goto(Bx1, By1)\r\n turtle.goto(Ix1, Iy1)\r\n turtle.goto(Mx1, My1)\r\n turtle.goto(Ax1, Ay1)\r\n\r\n # ближча грань - (в проекції права)\r\n turtle.up()\r\n turtle.goto(Dx1, Dy1)\r\n turtle.down()\r\n turtle.goto(Cx1, Cy1)\r\n turtle.goto(Fx1, Fy1)\r\n turtle.goto(Ex1, Ey1)\r\n turtle.goto(Dx1, Dy1)\r\n\r\n # верхеня грань - (в проекції верхня)\r\n\r\n turtle.up()\r\n turtle.goto(Ax1, Ay1)\r\n turtle.down()\r\n turtle.goto(Bx1, By1)\r\n turtle.goto(Cx1, Cy1)\r\n turtle.goto(Dx1, Dy1)\r\n turtle.goto(Ax1, Ay1)\r\n\r\n # верхеня грань - (в проекції верхня)\r\n turtle.up()\r\n turtle.goto(Mx1, My1)\r\n turtle.down()\r\n turtle.goto(Ix1, Iy1)\r\n turtle.goto(Fx1, Fy1)\r\n turtle.goto(Ex1, Ey1)\r\n turtle.goto(Mx1, My1)\r\n\r\n # ліва грань - (в проекції ближня)\r\n turtle.up()\r\n turtle.goto(Ax1, Ay1)\r\n turtle.down()\r\n turtle.goto(Mx1, My1)\r\n turtle.goto(Ex1, Ey1)\r\n turtle.goto(Dx1, Dy1)\r\n turtle.goto(Ax1, Ay1)\r\n\r\n # права грань - (в проекції дальня)\r\n turtle.up()\r\n turtle.goto(Bx1, By1)\r\n turtle.down()\r\n turtle.goto(Ix1, Iy1)\r\n turtle.goto(Fx1, Fy1)\r\n turtle.goto(Cx1, Cy1)\r\n turtle.goto(Bx1, By1)\r\n\r\n return PrlpdWiz\r\n\r\n\r\n\r\nsize = 300; n = 2;\r\ndef koch_curve(size, n):\r\n if n == 0:\r\n turtle.forward(size)\r\n else:\r\n koch_curve(size / 3, n - 1)\r\n turtle.left(60)\r\n koch_curve(size / 3, n - 1)\r\n turtle.right(120)\r\n koch_curve(size / 3, n - 1)\r\n turtle.left(60)\r\n koch_curve(size / 3, n - 1)\r\n\r\ndef draw_koch_snowflake(size, n):\r\n for i in range(3):\r\n koch_curve(size, n)\r\n turtle.right(120)\r\n\r\ndraw_koch_snowflake(size, n)\r\n# --------------- багатократний фрактал КОХА (сніжинка) - як форма черепашки ---------\r\ndef koch_curve(turtle, steps, length):\r\n if steps == 0:\r\n turtle.forward(length)\r\n else:\r\n for angle in [60, -120, 60, 0]:\r\n koch_curve(turtle, steps - 1, length / 3)\r\n turtle.left(angle)\r\n\r\ndef koch_snowflake(turtle, steps, length):\r\n turtle.begin_poly()\r\n\r\n for _ in range(3):\r\n koch_curve(turtle, steps, length)\r\n turtle.right(120)\r\n\r\n turtle.end_poly()\r\n\r\n return turtle.get_poly()\r\n# ------------------------------ зміна характеристик черепахи ---------------------\r\nturtle.speed(\"fastest\")\r\nturtle.register_shape(\"snowflake\", koch_snowflake(turtle.getturtle(), 2, 100))\r\nturtle.reset()\r\nturtle.penup()\r\nturtle.shape(\"snowflake\")\r\n\r\nwidth, height = turtle.window_width() / 2, turtle.window_height() / 2\r\nwidth=int(width)\r\nheight =int(height)\r\nfor _ in range(7):\r\n turtle.color((random(), random(), random()), (random(), random(), random()))\r\n turtle.goto(randrange(-width, width), randrange(-height, height))\r\n turtle.stamp()\r\n\r\n# ------------------------------ зміна характеристик черепахи ---------------------\r\nturtle.shape(\"triangle\")\r\nturtle.stamp()\r\nturtle.forward(1)\r\n\r\n\r\nxw=600; yw=600; st=50; TetaG1=180; TetaG2=-90\r\nl=(xw/2)-st; m=(yw/2)-st; n=m\r\n#Prlpd1=ShiftXYZ (Prlpd, l, m, n)\r\nPrlpd2=dimetri (Prlpd, TetaG1, TetaG2)\r\nPrxy3=ProjectXY (Prlpd2)\r\nPrlpdWiz(Prxy3)\r\nturtle.screen.exitonclick()\r\nturtle.screen.mainloop()\r\n#--------------------------------------------------------------------------------------\r\n","sub_path":"Labaratorywork6/Koh.py","file_name":"Koh.py","file_ext":"py","file_size_in_byte":6288,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"42324157","text":"import os\nimport json\nfrom signals import S\nfrom PyQt4 import QtCore, QtNetwork\n\n#in milliseconds\nREAD_TIMEOUT=10000\nCONNECTION_TIMEOUT=10000\n\nBROADCAST_INTERVAL=1000\nSCAN_INTERVAL=3000\n\n#in bytes\nPACKET_SIZE=1024\n\n# FIXED TCP PORT\nTCP_PORT=13373\n# FIXED UDP PORT\nUDP_PORT=13373\n\nKEY='HailHydra'\n\nclass NetworkInterface():\n \"\"\" Generic class for network connections \"\"\"\n def __init__(self):\n self.connection = None\n\n S.SEND_MSG.connect(self.sendMsg)\n S.SEND_FILE.connect(self.sendFile)\n\n def readData(self):\n instream = QtCore.QDataStream(self.connection)\n instream.setVersion(QtCore.QDataStream.Qt_4_0)\n\n if self.connection.bytesAvailable() < 4:\n return\n\n # Reading blocksize\n blockSize = instream.readUInt32()\n\n # Checking if the inputstream matches the blocksize\n if self.connection.bytesAvailable() < blockSize:\n return\n\n # Detecting the data type\n header_bytes = instream.readString()\n header = json.loads(str(header_bytes, encoding='utf-8'))\n\n if(header['TYPE']=='MSG'):\n S.MSG_RECV.emit(header)\n\n elif(header['TYPE']=='FILE'):\n actual_bytes=header['SIZE']\n recv_bytes=0\n\n # Storing file in default location\n new_path = QtCore.QDir.homePath() + '/LanChat/'\n if not QtCore.QDir(new_path).exists():\n QtCore.QDir(QtCore.QDir.homePath()).mkdir('LanChat')\n\n S.FILE_INIT.emit(header)\n\n file=open(new_path+header['NAME'],\"wb\")\n\n self.connection.waitForReadyRead(-1)\n\n # Writing raw bytes to file\n while(True):\n S.PROGRESS_UPD.emit('DOWNLOAD', (recv_bytes/actual_bytes)*100)\n\n bytes_remain=self.connection.bytesAvailable()\n if (bytes_remain==0):\n break\n\n recv_bytes=recv_bytes+bytes_remain\n content=instream.readRawData(bytes_remain)\n file.write(content)\n\n if(recv_bytes==actual_bytes):\n break\n\n self.connection.waitForReadyRead(READ_TIMEOUT)\n\n file.close()\n if(actual_bytes==recv_bytes):\n S.PROGRESS_UPD.emit('DOWNLOAD', 100)\n\n header['PATH']=new_path\n S.FILE_RECV.emit(header)\n else:\n print('Error occured')\n\n def sendMsg(self,msg):\n if (msg==''):\n return\n\n # Array of bytes to hold the data\n block = QtCore.QByteArray()\n\n # Datastream\n outstream = QtCore.QDataStream(block, QtCore.QIODevice.WriteOnly)\n outstream.setVersion(QtCore.QDataStream.Qt_4_0)\n\n # Inserting space to write the block size\n outstream.writeUInt32(0)\n\n # Message\n header={'TYPE':'MSG','MSG':msg}\n header_bytes=bytes(json.dumps(header),encoding='utf-8')\n outstream.writeString(header_bytes)\n\n # Writing block size at the beginning of stream\n outstream.device().seek(0)\n outstream.writeUInt32(block.size() - 4)\n\n # Writing data to socket\n self.connection.write(block)\n\n def sendFile(self, filepath):\n if (filepath==''):\n return\n\n S.PROGRESS_UPD.emit('UPLOAD', 0)\n\n # Array of bytes to hold the data\n block = QtCore.QByteArray()\n\n # Datastream\n outstream = QtCore.QDataStream(block, QtCore.QIODevice.WriteOnly)\n outstream.setVersion(QtCore.QDataStream.Qt_4_0)\n\n # Inserting space to write the block size\n outstream.writeUInt32(0)\n\n # Header for file\n filename = os.path.split(filepath)[-1]\n\n actual_bytes=os.path.getsize(filepath)\n send_bytes=0\n\n header={'TYPE':'FILE',\n 'NAME':filename,\n 'SIZE':actual_bytes}\n\n header_bytes=bytes(json.dumps(header),encoding='utf-8')\n outstream.writeString(header_bytes)\n\n outstream.device().seek(0)\n outstream.writeUInt32(block.size() - 4)\n\n # Writing header for file\n self.connection.write(block)\n\n # Raw bytes of the file\n f = open(filepath, \"rb\")\n try:\n byte = f.read(PACKET_SIZE)\n while byte != b'':\n S.PROGRESS_UPD.emit('UPLOAD', (send_bytes/actual_bytes)*100)\n\n block.clear()\n outstream.device().seek(0)\n outstream.writeRawData(byte)\n\n self.connection.waitForBytesWritten(-1)\n self.connection.write(block)\n\n byte = f.read(PACKET_SIZE)\n send_bytes=send_bytes+PACKET_SIZE\n finally:\n f.close()\n\n S.PROGRESS_UPD.emit('UPLOAD', 100)\n S.FILE_SENT.emit()\n\nclass Server(NetworkInterface):\n \"\"\" Class definition for the TCP server \"\"\"\n def __init__(self):\n NetworkInterface.__init__(self)\n\n self.tcpServer = QtNetwork.QTcpServer()\n self.udpSocket= QtNetwork.QUdpSocket()\n\n self.timer = QtCore.QTimer()\n\n # Send a perodic broadcast for discovery\n def broadcast(self):\n self.timer.start(BROADCAST_INTERVAL)\n self.timer.timeout.connect(self.sendDatagram)\n\n # Datagram to send\n def sendDatagram(self):\n # Code for broadcasting in network\n self.udpSocket.writeDatagram(KEY,QtNetwork.QHostAddress(QtNetwork.QHostAddress.Broadcast),UDP_PORT)\n\n # Waiting for a TCP connection\n def listen(self):\n # Listening to a specific port\n if not self.tcpServer.listen(port=TCP_PORT):\n print('Error occured')\n\n # Function is called when new connection is available\n self.tcpServer.newConnection.connect(self.newconnection)\n\n def newconnection(self):\n if(self.connection):\n return\n\n # Stops the UPD broadcast\n self.timer.stop()\n\n # Saves the socket connection\n self.connection = self.tcpServer.nextPendingConnection()\n\n # Remote terminal ipaddress\n peerAddress = self.connection.peerAddress().toString()\n\n S.DEVICE_CONNECTED.emit(peerAddress)\n\n # Codes to execute when connection is dropped\n self.connection.disconnected.connect(self.close)\n self.connection.disconnected.connect(S.DISCONNECTED.emit)\n\n # Function to run when data to read is available\n self.connection.readyRead.connect(self.readData)\n\n # Closes socket and server if disconnected\n def close(self):\n if (self.connection):\n self.connection.close()\n self.connection=None\n self.tcpServer.close()\n\nclass Client(NetworkInterface):\n \"\"\" Class definition for TCP client \"\"\"\n def __init__(self):\n NetworkInterface.__init__(self)\n\n self.connection = QtNetwork.QTcpSocket()\n self.udpSocket = QtNetwork.QUdpSocket()\n\n self.hostlist = []\n\n self.timer1 = QtCore.QTimer()\n\n def findHosts(self):\n self.udpSocket.bind(UDP_PORT)\n self.udpSocket.readyRead.connect(self.recvDatagram)\n\n # Timer to emit hostlist\n self.timer1.start(SCAN_INTERVAL)\n self.timer1.timeout.connect(self.emitHosts)\n\n S.CONNECT_HOST.connect(self.connect)\n\n def recvDatagram(self):\n while self.udpSocket.hasPendingDatagrams():\n datagram, host, port = self.udpSocket.readDatagram(self.udpSocket.pendingDatagramSize())\n key=str(datagram, encoding='ascii')\n strHost=host.toString()\n if(key==KEY):\n if(strHost not in self.hostlist):\n self.hostlist.append(strHost)\n\n def emitHosts(self):\n # Sending a copy of list\n S.HOST_LIST.emit(self.hostlist.copy())\n self.hostlist.clear()\n\n def connect(self, host):\n self.timer1.stop()\n\n self.connection.connectToHost(host, TCP_PORT)\n\n # Code to run when connection is established\n if(self.connection.waitForConnected(CONNECTION_TIMEOUT)):\n S.DEVICE_CONNECTED.emit(host)\n\n # Function to run when data to read is available\n self.connection.readyRead.connect(self.readData)\n\n # Codes to execute when connection is dropped\n self.connection.disconnected.connect(self.close)\n self.connection.disconnected.connect(S.DISCONNECTED.emit)\n\n # Code to run when connection timeout\n else:\n self.close()\n S.NO_CONNECTION.emit()\n\n # Closes socket if disconnected\n def close(self):\n self.connection.close()\n\nclass NetworkThread(QtCore.QThread):\n \"\"\"Defining a thread to be used for networking\"\"\"\n def __init__(self):\n QtCore.QThread.__init__(self)\n \n self.running = False\n self.mode = None\n \n # Runs in main thread \n def startThread(self,mode):\n self.mode = mode\n self.start() \n \n # Runs in seperate thread \n def run(self):\n # Network objects are created within this thread only\n self.running = True\n\n if (self.mode == 'CLIENT'):\n client = Client()\n\n # Finds hosts in the network\n client.findHosts()\n\n elif(self.mode == 'SERVER'):\n server = Server()\n\n # Broadcast datagram to everyone\n server.broadcast()\n\n # Listen for incoming TCP connection\n server.listen()\n\n # Event loop to prevent thread from terminating\n self.exec_()\n\n def quit(self):\n self.running = False\n QtCore.QThread.quit(self)\n\nif __name__ == '__main__':\n print('Nothing to run')\nelse:\n networkThread = NetworkThread()\n","sub_path":"network.py","file_name":"network.py","file_ext":"py","file_size_in_byte":9673,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"408279814","text":"import json\nimport sys\nimport traceback\nimport asyncio\nimport datetime\nimport discord\nfrom discord.ext import commands\n\nimport data\nimport rally_api\nimport validation\nimport errors\nimport aiohttp\n\nfrom cogs import update_cog\n\nfrom constants import *\nfrom utils import pretty_print\nfrom utils.converters import TimeframeType\n\n\nclass DefaultsCommands(commands.Cog):\n def __init__(self, bot):\n self.bot = bot\n\n async def cog_after_invoke(self, ctx):\n await pretty_print(\n ctx, \"Command completed successfully!\", title=\"Success\", color=SUCCESS_COLOR\n )\n\n @errors.standard_error_handler\n async def cog_command_error(self, ctx, error):\n # All other Errors not returned come here. And we can just print the default TraceBack.\n print(\"Ignoring exception in command {}:\".format(ctx.command), file=sys.stderr)\n traceback.print_exception(\n type(error), error, error.__traceback__, file=sys.stderr\n )\n\n @staticmethod\n async def update_setting(ctx, alert, alert_nr, value, setting):\n # check if settings have been configured on the dashboard\n settings = data.get_alerts_settings(ctx.guild.id)\n if not settings:\n return await pretty_print(ctx, \"Alert settings have not been configured on the dashboard\", title='Error', color=ERROR_COLOR)\n\n settings = settings[ALERTS_SETTINGS_KEY]\n\n # check if given alert is valid\n if alert not in settings:\n return await pretty_print(ctx, \"Invalid \", title='Error', color=ERROR_COLOR)\n\n channel_object = None\n instance = None\n # check if alert_nr is a digit and check if its valid\n if alert_nr.isdigit():\n if int(alert_nr) > len(settings[alert]['instances']) or int(alert_nr) < 0:\n return await pretty_print(ctx, \"Couldn't find an entry by that alert number\", title='Error', color=ERROR_COLOR)\n\n instance = settings[alert]['instances'][int(alert_nr) - 1]\n channel_object = discord.utils.get(ctx.guild.channels, name=instance['channel'])\n\n # check if alert_nr was valid and instance and channel_object were set\n if not channel_object or not instance:\n return await pretty_print(ctx, \"Invalid \", title='Error', color=ERROR_COLOR)\n\n # update settings\n instance['settings'][setting] = value\n data.set_alerts_settings(ctx.guild.id, json.dumps(settings))\n\n return await pretty_print(ctx, \"Alert settings have been updated\", title='Success', color=SUCCESS_COLOR)\n\n @commands.command(\n name='setmin',\n help=' - Set the minimum amount for an alert'\n )\n @commands.guild_only()\n async def setmin(self, ctx, alert, alert_nr, value):\n return await self.update_setting(ctx, alert, alert_nr, value, 'minamount')\n\n @commands.command(\n name='setmax',\n help=' - Set the minimum amount for an alert'\n )\n @commands.guild_only()\n async def setmax(self, ctx, alert, alert_nr, value):\n return await self.update_setting(ctx, alert, alert_nr, value, 'maxamount')\n\n @commands.command(\n name='settimezone',\n help=' - Set timezone setting for daily stats message'\n )\n @commands.guild_only()\n async def settimezone(self, ctx, alert_nr, value):\n return await self.update_setting(ctx, 'daily_stats', alert_nr, value, 'timezone')\n\n @commands.command(\n name='allcoinstats',\n help=' - (day/week) list the following stats in the coin alerts channel based on the time given'\n )\n @commands.guild_only()\n async def allcoinstats(self, ctx, timeframe: TimeframeType):\n # delete week old data\n data.delete_week_old_events()\n\n # if default coin isn't set, send info to user about how to set it\n default_coin = data.get_default_coin(ctx.guild.id)\n if not default_coin:\n return await pretty_print(\n ctx, \"A default coin has not been set. An admin can set the default coin by typing $setdefaultcoin . Type $help for more information.\", title=\"Error\", color=ERROR_COLOR\n )\n\n # get statistics\n if timeframe == 'day':\n coin_stats = update_cog.get_day_stats(default_coin)\n else:\n coin_stats = update_cog.get_week_stats(default_coin)\n\n rewards = rally_api.get_coin_rewards(default_coin)\n coin_image_url = rally_api.get_coin_image_url(default_coin)\n\n # format message, done through dict to make keeping this and daily_stats message similar easier\n extra_str = 'Today' if timeframe == 'day' else 'This Week'\n reward_str = 'last24HourEarned' if timeframe == 'day' else 'weeklyAccumulatedReward'\n message = {\n \"description\": f\"```xl\\n\"\n f\"- {extra_str}`s purchases: {len(coin_stats['buy'])}\\n\\n\"\n f\"- {extra_str}`s donations: {len(coin_stats['donate'])}\\n\\n\"\n f\"- {extra_str}`s transfers: {len(coin_stats['transfer'])}\\n\\n\"\n f\"- {extra_str}`s conversions: {len(coin_stats['convert'])}\\n\\n\"\n f\"- {extra_str}`s redeems: {len(coin_stats['redeem'])}\\n\\n\"\n f\"- {extra_str}`s rewards earned: {round(rewards[reward_str], 3)}\\n\"\n f\"```\",\n \"color\": 0xff0000,\n \"author\": {\n \"name\": f\"{default_coin} Stats {extra_str}\",\n \"icon_url\": coin_image_url\n },\n \"timestamp\": datetime.datetime.now().isoformat()\n }\n\n # send message\n embed = discord.Embed.from_dict(message)\n return await ctx.send(embed=embed)\n\n @commands.command(\n name=\"set_default_coin\",\n help=\" Set a default coin to be used across the server\",\n )\n @validation.owner_or_permissions(administrator=True)\n async def set_default_coin(self, ctx, coin_name):\n await pretty_print(\n ctx,\n f\"Are you sure you want to set {coin_name} as default coin?\",\n caption=\"Give 👍 reaction to confirm\",\n title=\"Warning\",\n color=WARNING_COLOR,\n )\n\n def check(reaction, user):\n return user == ctx.message.author and str(reaction.emoji) == \"👍\"\n\n try:\n await self.bot.wait_for(\"reaction_add\", timeout=30.0, check=check)\n except asyncio.TimeoutError:\n await pretty_print(\n ctx, \"Set default coin timed out 👎\", title=\"Timeout\", color=ERROR_COLOR\n )\n else:\n data.add_default_coin(ctx.guild.id, coin_name)\n await pretty_print(\n ctx,\n f\"{coin_name} is now the default coin 👍\",\n title=\"Set\",\n color=GREEN_COLOR,\n )\n\n @commands.command(\n name=\"change_prefix\",\n help=\" Prefix for bot commands\",\n )\n @validation.owner_or_permissions(administrator=True)\n async def set_prefix(self, ctx, prefix):\n data.add_prefix_mapping(ctx.guild.id, prefix)\n\n @commands.command(\n name=\"change_bot_name\",\n help=\"Change the bot's name on this server\"\n )\n @commands.is_owner()\n async def set_bot_name(self, ctx, *, name=\"\"):\n try:\n await self.bot.user.edit(username=name)\n data.set_bot_name(ctx.guild.id, name)\n except Exception as e:\n return await ctx.send(f'Error: {e.text.split(\":\")[-1]}')\n\n @commands.command(\n name=\"change_bot_avatar\",\n help=\"Changes the bot's avatar\"\n )\n @commands.is_owner()\n async def set_bot_avatar(self, ctx, url=None):\n if url is None:\n url = DEFAULT_BOT_AVATAR_URL\n\n try:\n async with aiohttp.ClientSession() as session:\n async with session.get(url) as response:\n avatar = await response.read()\n\n await self.bot.user.edit(avatar=avatar)\n data.set_bot_avatar(ctx.guild.id, url)\n except:\n return await ctx.send('Error setting new bot avatar')\n\n @commands.command(\n name=\"role_call\",\n help=\" Display users who have access to a given role\",\n )\n @validation.owner_or_permissions(administrator=True)\n async def role_call(self, ctx, role: discord.Role):\n usersStr = \"\"\n for member in ctx.guild.members:\n if role in member.roles:\n usersStr += f\"{member}\\n\"\n await pretty_print(\n ctx,\n usersStr,\n title=f\"Users with {role} role\",\n color=GREEN_COLOR,\n )\n\n @commands.command(\n name=\"list_all_users\",\n help=\"Display users who have been registered\",\n )\n @validation.owner_or_permissions(administrator=True)\n async def list_all_users(self, ctx):\n usersStr = \"\"\n registered_users = data.get_all_users()\n for user in registered_users:\n member = await ctx.guild.fetch_member(user[DISCORD_ID_KEY])\n if member:\n usersStr += f\"{member}\\nRallyId: {user[RALLY_ID_KEY]}\\nDiscordId: {user[DISCORD_ID_KEY]}\\n\\n\"\n await pretty_print(\n ctx,\n usersStr or \"No registered users on this server\",\n title=f\"All registered users\",\n color=GREEN_COLOR,\n )\n","sub_path":"rallyrolebot/cogs/defaults_cog.py","file_name":"defaults_cog.py","file_ext":"py","file_size_in_byte":9503,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"525618023","text":"# -*- coding: utf-8 -*-\n#\n\nimport sys\nimport os\n\n# -- PHP highlighting configuration --------------------------------------------\n\nfrom sphinx.highlighting import lexers\nif lexers:\n\tfrom pygments.lexers.web import PhpLexer\n\tlexers['php'] = PhpLexer(startinline=True)\n\n# -- General configuration -----------------------------------------------------\n\nextensions = ['sphinx.ext.intersphinx', 'sphinx.ext.todo', 'sphinx.ext.ifconfig']\n\n# The suffix of source filenames.\nsource_suffix = '.rst'\n\n# The master toctree document.\nmaster_doc = 'index'\n\n# General information about the project.\nproject = u'typo3forum'\ncopyright = u'2016, Mittwald CM Service'\n\n# The version info for the project you're documenting, acts as replacement for\n# |version| and |release|, also used in various other places throughout the\n# built documents.\n#\n# The short X.Y version.\nversion = '1.0'\n# The full version, including alpha/beta/rc tags.\nrelease = '1.0.0'\n\n# Else, today_fmt is used as the format for a strftime call.\ntoday_fmt = '%Y-%m-%d %H:%M'\n\n# List of patterns, relative to source directory, that match files and\n# directories to ignore when looking for source files.\nexclude_patterns = ['_make']\nexclude_trees = ['_make']\n\n# If true, sectionauthor and moduleauthor directives will be shown in the\n# output. They are ignored by default.\nshow_authors = False\n\n# The name of the Pygments (syntax highlighting) style to use.\npygments_style = 'sphinx'\n\n# -- Options for HTML output ----------------------------------------------\n\n# The theme to use for HTML and HTML Help pages. See the documentation for\n# a list of builtin themes.\nhtml_theme = 'default'\n\n# Add any paths that contain custom themes here, relative to this directory.\nhtml_theme_path = []\n\n# Add any paths that contain custom static files (such as style sheets) here,\n# relative to this directory. They are copied after the builtin static files,\n# so a file named \"default.css\" will overwrite the builtin \"default.css\".\nhtml_static_path = ['../Images']\n\n# If true, \"Created using Sphinx\" is shown in the HTML footer. Default is True.\nhtml_show_sphinx = False\n\n# If true, \"(C) Copyright ...\" is shown in the HTML footer. Default is True.\nhtml_show_copyright = False\n\n# Output file base name for HTML help builder.\nhtmlhelp_basename = 'typo3_forum'\n\n\n# -- Options for LaTeX output --------------------------------------------------\n\nlatex_elements = {\n\n# The font size ('10pt', '11pt' or '12pt').\n#'pointsize': '10pt',\n}\n\n# Grouping the document tree into LaTeX files. List of tuples\n# (source start file, target name, title,\n# author, documentclass [howto, manual, or own class]).\nlatex_documents = [\n ('index', 'typo3_forum.tex', u'typo3\\_forum',\n u'Mittwald CM Service', 'manual'),\n]\n\n# -- Options for rst2pdf output ------------------------------------------------\n\n# The options element is a dictionary that lets you override\n# this config per-document.\n# For example,\n# ('index', u'MyProject', u'My Project', u'Author Name',\n# dict(pdf_compressed = True))\n# would mean that specific document would be compressed\n# regardless of the global pdf_compressed setting.\npdf_documents = [\n ('index', 'typo3_forum', u'typo3\\_forum',\n u'Mittwald CM Service', 'manual'),\n]\n\n# A comma-separated list of custom stylesheets. Example:\npdf_stylesheets = ['sphinx','kerning','a4']\n\n# A list of folders to search for stylesheets. Example:\npdf_style_path = ['.', '_styles']\n\n# How many levels deep should the table of contents be?\npdf_toc_depth = 9999\n\n# Add section number to section references\npdf_use_numbered_links = False\n\n# Background images fitting mode\npdf_fit_background_mode = 'scale'\n\n\n# -- Options for manual page output ---------------------------------------\n\n# One entry per manual page. List of tuples\n# (source start file, name, description, authors, manual section).\nman_pages = [\n ('index', 'typo3_forum', u'typo3_forum',\n [u'Mittwald CM Service'], 1)\n]\n\n# If true, show URL addresses after external links.\n#man_show_urls = False\n\n# -- Options for Texinfo output -------------------------------------------\n\n# Grouping the document tree into Texinfo files. List of tuples\n# (source start file, target name, title, author,\n# dir menu entry, description, category)\ntexinfo_documents = [\n ('index', 'typo3_forum', u'typo3_forum',\n u'Mittwald CM Service', 'typo3_forum', ' forum plugin for TYPO3',\n 'Miscellaneous'),\n]\n\n#=================================================\n#\n# TYPO3 codeblock BEGIN:\n#\n# Insert this codeblock at the end of your Sphinx\n# builder configuration file 'conf.py'.\n# This may enable TYPO3 specific features like\n# TYPO3 themes. It makes Yaml settings files work.\n#\n#-------------------------------------------------\n\nif 1 and \"TYPO3 specific\":\n\n try:\n t3DocTeam\n except NameError:\n t3DocTeam = {}\n\n try:\n import t3sphinx\n html_theme_path.insert(0, t3sphinx.themes_dir)\n html_theme = 'typo3sphinx'\n except:\n html_theme = 'default'\n\n t3DocTeam['conf_py_file'] = None\n try:\n t3DocTeam['conf_py_file'] = __file__\n except:\n import inspect\n t3DocTeam['conf_py_file'] = inspect.getfile(\n inspect.currentframe())\n\n t3DocTeam['conf_py_package_dir'] = os.path.abspath(os.path.dirname(\n t3DocTeam['conf_py_file']))\n t3DocTeam['relpath_to_master_doc'] = '..'\n t3DocTeam['relpath_to_logdir'] = '_not_versioned'\n t3DocTeam['path_to_logdir'] = os.path.join(\n t3DocTeam['conf_py_package_dir'],\n t3DocTeam['relpath_to_logdir'])\n t3DocTeam['pathToYamlSettings'] = os.path.join(\n t3DocTeam['conf_py_package_dir'],\n t3DocTeam['relpath_to_master_doc'], 'Settings.yml')\n try:\n t3DocTeam['pathToGlobalYamlSettings'] = \\\n t3sphinx.pathToGlobalYamlSettings\n except:\n t3DocTeam['pathToGlobalYamlSettings'] = None\n if not t3DocTeam['pathToGlobalYamlSettings']:\n t3DocTeam['pathToGlobalYamlSettings'] = os.path.join(\n t3DocTeam['conf_py_package_dir'], 'GlobalSettings.yml')\n try:\n __function = t3sphinx.yamlsettings.processYamlSettings\n except:\n __function = None\n if not __function:\n try:\n import yamlsettings\n __function = yamlsettings.processYamlSettings\n except:\n __function = None\n if __function:\n __function(globals(), t3DocTeam)\n\n#-------------------------------------------------\n#\n# TYPO3 codeblock END.\n#\n#=================================================\n","sub_path":"conf.py","file_name":"conf.py","file_ext":"py","file_size_in_byte":6524,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"105682365","text":"import urllib\nimport re\n\nurl = raw_input(\"Enter the web link : \")\np = urllib.urlopen(url)\na=p.read()\n\npattern = re.compile('img src=[ \"](.*?)\"' )\na = re.findall(pattern , a)\n\nf=open(\"new.txt\",\"w\")\n\nfor i in a:\n\t\n\t \tf.write(url + i +\"\\n\")\n\t\t\n\nf.close()\n","sub_path":"imgdown.py","file_name":"imgdown.py","file_ext":"py","file_size_in_byte":252,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"62428735","text":"import tensorflow as tf\nimport common.encoding\nfrom common.datasets import SequenceDataset\n\npath = '/mnt/nfs_datasets/lakh_midi_full/drums_sequence_examples/training_drum_tracks.tfrecord'\n\ninput_size = common.encoding.DrumTimeSliceEncoder().output_size\nencoder = common.encoding.OneToOneSequenceEncoder(\n\tcommon.encoding.IdentityTimeSliceEncoder(input_size)\n)\n\ndataset = SequenceDataset([path], encoder)\n\nfeatures = dataset.load_single()\n\n# Run this graph\n_features = tf.contrib.learn.run_n(features, n=1)\nprint(_features[0])","sub_path":"musicgen/common/datasets/drums/test_dataset.py","file_name":"test_dataset.py","file_ext":"py","file_size_in_byte":525,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"441965198","text":"newList = [\"eeny\", \"meeny\", \"minee\", \"moe\"]\n\ndef menu(list, question):\n for entry in list:\n print(1 + list.index(entry), end=\"\")\n print(\") \" + entry.title())\n\n question = int(input(question)) - 1\n\n return list[question].title()\n\nanswer = menu(newList, \"Which do you prefer? \")\n\nprint(\"You prefer \" + answer)\n\n","sub_path":"Training/menu.py","file_name":"menu.py","file_ext":"py","file_size_in_byte":332,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"571577438","text":"import sys\r\n\r\nfrom Samples.geocoder import get_coordinates, get_nearest_object\r\n\r\n\r\ndef main():\r\n # Забираем адресную точку из параметров запуска.\r\n address = ''\r\n try:\r\n address = \" \".join(sys.argv[1:])\r\n except:\r\n print('No data')\r\n exit(1)\r\n\r\n if not address:\r\n print('No data')\r\n exit(1)\r\n\r\n # Получаем координаты точки\r\n address_point = get_coordinates(address)\r\n\r\n # Получаем район.\r\n district_name = get_nearest_object(address_point, \"district\")\r\n print(district_name)\r\n\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n","sub_path":"2nd_year/WEB5. Работа с HTTP-API/Home/05_what_district.py","file_name":"05_what_district.py","file_ext":"py","file_size_in_byte":662,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"437852144","text":"target = int(input())\nnum = input().split(\",\")\nnumbers = [int(i) for i in num]\nlengths = []\nfor i in range(len(numbers)-1):\n current = numbers[i]\n le = 1\n for j in range(i+1, len(numbers)):\n current += numbers[j]\n le += 1\n if current >= target:\n lengths.append(le)\n break\nif len(lengths) == 0:\n print(0)\nelse:\n lengths.sort()\n print(lengths[0])","sub_path":"Code/CodeRecords/2464/60619/252822.py","file_name":"252822.py","file_ext":"py","file_size_in_byte":405,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"264856677","text":"from __future__ import unicode_literals\n\nfrom django.db import models\nfrom Courses.models import Course\nfrom People.models import Student\n\n# Create your models here.\n\nclass Question(models.Model):\n\tquestionID = models.AutoField(primary_key = True)\n\tquestionText = models.CharField('Question', max_length = 100, null = True)\n\tchoice1 = models.CharField('Choice 1', max_length = 50, null = True)\n\tchoice2 = models.CharField('Choice 2', max_length = 50, null = True)\n\tchoice3 = models.CharField('Choice 3', max_length = 50, null = True)\n\tchoice4 = models.CharField('Choice 4', max_length = 50, null = True)\n\t# questionMarks = models.IntegerField('Question Marks')\n\tcorrect = models.CharField('Correct Answer', max_length= 1, null = True)\n\nclass Test(models.Model):\n\ttestID = models.AutoField(primary_key = True)\n\ttestTitle = models.CharField('Title', max_length = 50)\n\tquestions = models.ManyToManyField(Question)\n\nclass Assignment(models.Model):\n\tassignmentID = models.AutoField(primary_key = True)\n\tassignmentTitle = models.CharField('Title', max_length = 50)\n\tassignmentText = models.CharField('Assignment', max_length = 500)\n\nclass Lecture(models.Model):\n\tlectureID = models.AutoField(primary_key = True)\n\tlectureTitle = models.CharField('Title', max_length = 50)\n\tlectureText = models.CharField('Lecture', max_length = 500)\n\tlectureWeek = models.IntegerField('Week No.', null = True)\n\nclass CourseContent(models.Model):\n\tcourseID = models.ForeignKey(Course)\n\tlectures = models.ManyToManyField(Lecture)\n\tassignments = models.ManyToManyField(Assignment)\n\ttests = models.ManyToManyField(Test)\n\nclass Evaluation(models.Model):\n\tstudentID = models.ForeignKey(Student)\n\ttestID = models.ForeignKey(Test)\n\tmarks = models.IntegerField('Marks Obtained')\n\n","sub_path":"OCMS/CourseMatter/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":1747,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"630879702","text":"import unittest\n\nfrom a.hash_table import HashTable\n\n\nclass TestHashTable(unittest.TestCase):\n\n def setUp(self) -> None:\n self.ht = HashTable(11)\n self.ht[54] = 'cat'\n self.ht[26] = 'dog'\n self.ht[93] = 'lion'\n self.ht[17] = 'tiger'\n self.ht[77] = 'bird'\n self.ht[31] = 'cow'\n self.ht[44] = 'goat'\n\n def testHashTableSize(self):\n self.assertEqual(11, self.ht.size)\n\n def testHashTableSlots(self):\n self.assertEqual([77, 44, None, None, 26, 93, 17, None, None, 31, 54], self.ht.slots)\n\n def testHashTableItems(self):\n self.assertEqual(['bird', 'goat', None, None, 'dog', 'lion', 'tiger', None, None, 'cow', 'cat'], self.ht.items)\n\n def testHashTableGetAndSet(self):\n self.assertEqual('bird', self.ht.get(77))\n\n self.ht[55] = 'pig'\n self.assertEqual('pig', self.ht.get(55))\n\n self.ht[20] = 'chicken'\n self.assertEqual('chicken', self.ht.get(20))\n self.assertEqual([77, 44, 55, 20, 26, 93, 17, None, None, 31, 54], self.ht.slots)\n\n self.ht[20] = 'duck'\n self.assertEqual('duck', self.ht.get(20))\n self.assertEqual([77, 44, 55, 20, 26, 93, 17, None, None, 31, 54], self.ht.slots)\n\n self.assertEqual(None, self.ht.get(99))\n\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"python/hash-table/a/test_hash_table.py","file_name":"test_hash_table.py","file_ext":"py","file_size_in_byte":1329,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"500789255","text":"import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\n\nfrom torch.autograd import Variable\n\n\nclass Bottleneck(nn.Module):\n expansion = 4\n\n def __init__(self, in_planes, planes, stride=1):\n super(Bottleneck, self).__init__()\n self.conv1 = nn.Conv2d(in_planes, planes, kernel_size=1, bias=False)\n self.bn1 = nn.BatchNorm2d(planes)\n self.conv2 = nn.Conv2d(planes, planes, kernel_size=3, stride=stride, padding=1, bias=False)\n self.bn2 = nn.BatchNorm2d(planes)\n self.conv3 = nn.Conv2d(planes, self.expansion*planes, kernel_size=1, bias=False)\n self.bn3 = nn.BatchNorm2d(self.expansion*planes)\n\n self.shortcut = nn.Sequential()\n if stride != 1 or in_planes != self.expansion*planes:\n self.shortcut = nn.Sequential(\n nn.Conv2d(in_planes, self.expansion*planes, kernel_size=1, stride=stride, bias=False),\n nn.BatchNorm2d(self.expansion*planes)\n )\n\n def forward(self, x):\n out = F.relu(self.bn1(self.conv1(x)))\n out = F.relu(self.bn2(self.conv2(out)))\n out = self.bn3(self.conv3(out))\n out += self.shortcut(x)\n out = F.relu(out)\n return out\n\n\nclass FPN(nn.Module):\n def __init__(self, args):\n super(FPN, self).__init__()\n self.args = args\n\n num_blocks = [2,2,2,2]\n self.in_planes = 64\n\n self.conv1 = nn.Conv2d(3, 64, kernel_size=7, stride=2, padding=3, bias=False)\n self.bn1 = nn.BatchNorm2d(64)\n\n # Bottom-up layers\n self.layer1 = self._make_layer(Bottleneck, 64, num_blocks[0], stride=1)\n self.layer2 = self._make_layer(Bottleneck, 128, num_blocks[1], stride=2)\n self.layer3 = self._make_layer(Bottleneck, 256, num_blocks[2], stride=2)\n self.layer4 = self._make_layer(Bottleneck, 512, num_blocks[3], stride=2)\n\n # Top layer\n self.toplayer = nn.Conv2d(2048, 256, kernel_size=1, stride=1, padding=0) # Reduce channels\n\n # Smooth layers\n self.smooth1 = nn.Conv2d(256, 256, kernel_size=3, stride=1, padding=1)\n self.smooth2 = nn.Conv2d(256, 256, kernel_size=3, stride=1, padding=1)\n self.smooth3 = nn.Conv2d(256, 256, kernel_size=3, stride=1, padding=1)\n\n # Lateral layers\n self.latlayer1 = nn.Conv2d(1024, 256, kernel_size=1, stride=1, padding=0)\n self.latlayer2 = nn.Conv2d( 512, 256, kernel_size=1, stride=1, padding=0)\n self.latlayer3 = nn.Conv2d( 256, 256, kernel_size=1, stride=1, padding=0)\n\n self.to(args.device)\n\n def _make_layer(self, block, planes, num_blocks, stride):\n strides = [stride] + [1]*(num_blocks-1)\n layers = []\n for stride in strides:\n layers.append(block(self.in_planes, planes, stride))\n self.in_planes = planes * block.expansion\n return nn.Sequential(*layers)\n\n def _upsample_add(self, x, y):\n '''Upsample and add two feature maps.\n Args:\n x: (Variable) top feature map to be upsampled.\n y: (Variable) lateral feature map.\n Returns:\n (Variable) added feature map.\n Note in PyTorch, when input size is odd, the upsampled feature map\n with `F.upsample(..., scale_factor=2, mode='nearest')`\n maybe not equal to the lateral feature map size.\n e.g.\n original input size: [N,_,15,15] ->\n conv2d feature map size: [N,_,8,8] ->\n upsampled feature map size: [N,_,16,16]\n So we choose bilinear upsample which supports arbitrary output sizes.\n '''\n _,_,H,W = y.size()\n return F.upsample(x, size=(H,W), mode='bilinear') + y\n\n def forward(self, x):\n # Bottom-up\n c1 = F.relu(self.bn1(self.conv1(x)))\n c1 = F.max_pool2d(c1, kernel_size=3, stride=2, padding=1)\n c2 = self.layer1(c1)\n c3 = self.layer2(c2)\n c4 = self.layer3(c3)\n c5 = self.layer4(c4)\n # Top-down\n p5 = self.toplayer(c5)\n p4 = self._upsample_add(p5, self.latlayer1(c4))\n p3 = self._upsample_add(p4, self.latlayer2(c3))\n p2 = self._upsample_add(p3, self.latlayer3(c2))\n # Smooth\n p4 = self.smooth1(p4)\n p3 = self.smooth2(p3)\n p2 = self.smooth3(p2)\n return p2, p3, p4, p5\n\n\n\nclass FocalLoss(nn.Module):\n def __init__(self, coder):\n super().__init__()\n\n self.coder = coder\n self.priors_cxcy = self.coder.center_anchor\n self.priors_xy = cxcy_to_xy(self.priors_cxcy)\n self.num_classes = self.coder.num_classes\n self.bce = nn.BCELoss(reduction='none')\n self.smooth_l1 = SmoothL1Loss()\n # self.smooth_l1 = nn.SmoothL1Loss(reduction=None)\n\n def forward(self, pred, b_boxes, b_labels):\n \"\"\"\n Forward propagation.\n :param pred (loc, cls) prediction tuple (N, 67995, 4) / (N, 67995, num_classes) or [120087] anchors\n :param labels: true object labels, a list of N tensors\n \"\"\"\n pred_loc = pred[0]\n pred_cls = pred[1]\n\n batch_size = pred_loc.size(0)\n n_priors = self.priors_xy.size(0)\n\n assert n_priors == pred_loc.size(1) == pred_cls.size(1) # 67995 --> 120087\n\n true_locs = torch.zeros((batch_size, n_priors, 4), dtype=torch.float).to(device) # (N, 67995, 4)\n true_classes = -1 * torch.ones((batch_size, n_priors, self.num_classes), dtype=torch.float).to(device) # (N, 67995, num_classes)\n depth = -1 * torch.ones((batch_size, n_priors), dtype=torch.bool).to(device) # (N, 67995)\n\n for i in range(batch_size):\n boxes = b_boxes[i] # xy coord\n labels = b_labels[i]\n\n ###################################################\n # match strategies -> make target #\n ###################################################\n iou = find_jaccard_overlap(self.priors_xy, boxes) # [67995, num_objects]\n IoU_max, IoU_argmax = iou.max(dim=1) # [67995]\n\n negative_indices = IoU_max < 0.4\n\n # ======================= make true classes ========================\n true_classes[i][negative_indices, :] = 0 # make negative\n\n depth[i][negative_indices] = 0\n\n positive_indices = IoU_max >= 0.5 # iou 가 0.5 보다 큰 아이들 - [67995]\n argmax_labels = labels[IoU_argmax] # assigned_labels\n\n # class one-hot encoding\n # 0 으로 만들고 이후에 1 을 넣어주기\n true_classes[i][positive_indices, :] = 0\n true_classes[i][positive_indices, argmax_labels[positive_indices].long()] = 1. # objects\n\n depth[i][positive_indices] = 1\n\n # =========================== make true locs ===========================\n true_locs_ = xy_to_cxcy(boxes[IoU_argmax]) # [67995, 4] 0~1 사이이다. boxes 가\n true_locs_ = self.coder.encode(true_locs_)\n true_locs[i] = true_locs_\n\n # ------------------------------------------ cls loss ------------------------------------------\n alpha = 0.25\n gamma = 2\n\n alpha_factor = torch.ones_like(true_classes).to(device) * alpha # alpha\n a_t = torch.where((true_classes == 1), alpha_factor, 1. - alpha_factor) # a_t\n p_t = torch.where(true_classes == 1, pred_cls, 1 - pred_cls) # p_t\n bce = self.bce(pred_cls, true_classes)\n cls_loss = a_t * (1 - p_t) ** gamma * bce # focal loss\n\n cls_mask = (depth >= 0).unsqueeze(-1).expand_as(cls_loss) # both fore and back ground\n num_of_pos = (depth > 0).sum().float().clamp(min=1) # only foreground (min=1)\n cls_loss = (cls_loss * cls_mask).sum() / num_of_pos # batch 의 bce loss\n # / batch 의 object 총갯수\n\n # ------------------------------------------ loc loss ------------------------------------------\n loc_mask = (depth > 0).unsqueeze(-1).expand_as(true_locs) # only foreground\n loc_loss = self.smooth_l1(pred_loc, true_locs) # (), scalar\n loc_loss = (loc_mask * loc_loss).sum() / num_of_pos\n # loc_loss *= 2 # balance values\n\n total_loss = (cls_loss + loc_loss)\n return total_loss, (loc_loss, cls_loss)","sub_path":"models/FPN/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":8700,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"164417539","text":"__author__ = \"susmote\"\n\nimport pygame\nimport time\nimport random\nfrom pygame.locals import *\n\n\nenemy_list = pygame.sprite.Group()\n\n\nclass Base(pygame.sprite.Sprite):\n def __init__(self, screen_temp, x, y, image_name):\n pygame.sprite.Sprite.__init__(self)\n self.x = x\n self.y = y\n self.screen = screen_temp\n self.image = pygame.image.load(image_name)\n self.bullet_list = []\n\n\nclass BasePlane(Base):\n # 飞机基类\n def display(self):\n self.screen.blit(self.image, (self.x, self.y))\n\n for bullet in self.bullet_list:\n bullet.display()\n bullet.move()\n if bullet.judge():\n self.bullet_list.remove(bullet)\n\n\nclass HeroPlane(BasePlane):\n # 玩家飞机类\n def __init__(self, screen_temp):\n BasePlane.__init__(self, screen_temp, 120, 420, \"./img/hero1.png\")\n self.stop = True\n self.direction = None\n self.image = pygame.transform.scale(self.image, (int(100 * 0.6), int(124 * 0.6)))\n\n def display(self):\n self.screen.blit(self.image, (self.x, self.y))\n\n for bullet in self.bullet_list:\n bullet.display()\n bullet.move()\n if bullet.judge():\n self.bullet_list.remove(bullet)\n for enemy in enemy_list:\n bullet.hit_plane(enemy)\n\n def move(self):\n if not self.stop:\n if self.direction == \"left\":\n self.x -= 3\n if self.direction == \"right\":\n self.x += 3\n if self.direction == \"up\":\n self.y -= 3\n if self.direction == \"down\":\n self.y += 3\n\n def fire(self):\n self.bullet_list.append(Bullet(self.screen, self.x, self.y))\n\n\nclass EnemyPlane(BasePlane):\n # 敌机类\n def __init__(self, screen_temp):\n BasePlane.__init__(self, screen_temp, 0, 0, \"./img/enemy0.png\")\n self.live = True\n self.image = pygame.transform.scale(self.image, (int(51*0.6), int(39*0.6)))\n self.direction = \"right\"\n\n self.images = [pygame.image.load(\"./img/enemy0_down1.png\"),\n pygame.image.load(\"./img/enemy0_down2.png\"),\n pygame.image.load(\"./img/enemy0_down3.png\"),\n pygame.image.load(\"./img/enemy0_down4.png\")]\n self.step = 0\n self.rect = self.image.get_rect()\n\n def move(self):\n\n if self.direction == \"right\":\n self.x += 3\n elif self.direction == \"left\":\n self.x -= 3\n\n if self.x > int(0.6*480 - 0.6*51):\n self.direction = \"left\"\n elif self.x < 0:\n self.direction = \"right\"\n\n def fire(self):\n\n random_num = random.randint(1, 100)\n if random_num == 25 or random_num == 50 or random_num == 75:\n self.bullet_list.append(EnemyBullet(self.screen, self.x, self.y))\n\n def explode(self):\n while self.live:\n if self.step == len(self.images):\n self.live = False\n else:\n self.image = self.images[self.step]\n self.image = pygame.transform.scale(self.image, (int(51 * 0.6), int(39 * 0.6)))\n time.sleep(0.03)\n pygame.display.update()\n self.screen.blit(self.image, (self.x, self.y))\n self.step += 1\n\n\nclass BaseBullet(Base):\n def display(self):\n self.screen.blit(self.image, (self.x, self.y))\n\n\nclass Bullet(BaseBullet):\n # 子弹类\n def __init__(self, screen_temp, x, y):\n BaseBullet.__init__(self, screen_temp, x+24, y-11, \"./img/bullet.png\")\n self.image = pygame.transform.scale(self.image, (int(22 * 0.6), int(22 * 0.6)))\n self.rect = self.image.get_rect()\n\n def move(self):\n self.y -= 5\n\n def judge(self):\n if self.y < 0:\n return True\n else:\n return False\n\n def hit_plane(self, enemy):\n if self.judge():\n # 获取敌机的坐标\n print(\"敌机的坐标 :\",enemy.x, enemy.y)\n # 获取子弹的坐标\n print(\"子弹的坐标 :\", self.x, self.y)\n # 获取敌机的实时区域\n startX = enemy.x\n endX = enemy.x+(enemy.rect.width)\n print(startX)\n print(endX)\n if self.x > startX and self.x < endX:\n print(enemy)\n enemy.explode()\n\n\n\n\n\nclass EnemyBullet(BaseBullet):\n # 敌机子弹类\n def __init__(self, screen_temp, x, y):\n BaseBullet.__init__(self, screen_temp, x+12, y+17, \"./img/bullet1.png\")\n self.image = pygame.transform.scale(self.image, (int(9 * 0.6), int(21 * 0.6)))\n\n def move(self):\n self.y += 3\n\n def judge(self):\n if self.y > int(0.6*852):\n return True\n else:\n return False\n\n\ndef main():\n\n screen = pygame.display.set_mode((int(0.6*480), int(0.6*852)), 0, 32)\n\n pygame.display.set_caption(\"打飞机游戏 by susmote\")\n\n background = pygame.image.load(\"./img/background.png\")\n background = pygame.transform.scale(background, (int(0.6*480), int(0.6*852)))\n\n hero = HeroPlane(screen)\n\n enemy_list.add(EnemyPlane(screen))\n\n\n while True:\n\n screen.blit(background, (0, 0))\n\n hero.display()\n\n hero.move()\n\n for enemy in enemy_list:\n if enemy.live:\n enemy.display()\n enemy.move()\n enemy.fire()\n if len(enemy_list) < 1:\n enemy_list.add(EnemyPlane(screen))\n\n pygame.display.update()\n\n for event in pygame.event.get():\n\n if event.type == QUIT:\n print(\"exit\")\n exit()\n\n elif event.type == KEYDOWN:\n if event.key == K_a or event.key == K_LEFT:\n print('left')\n hero.stop = False\n hero.direction = \"left\"\n elif event.key == K_d or event.key == K_RIGHT:\n print('right')\n hero.stop = False\n hero.direction = \"right\"\n elif event.key == K_w or event.key == K_UP:\n print('up')\n hero.stop = False\n hero.direction = \"up\"\n elif event.key == K_s or event.key == K_DOWN:\n print('down')\n hero.stop = False\n hero.direction = \"down\"\n elif event.key == K_SPACE:\n print('space')\n hero.fire()\n elif event.key == K_ESCAPE:\n print(\"exit\")\n exit()\n elif event.type == KEYUP:\n if event.key == K_LEFT or event.key == K_a:\n hero.stop = True\n if event.key == K_RIGHT or event.key == K_d:\n hero.stop = True\n if event.key == K_UP or event.key == K_w:\n hero.stop = True\n if event.key == K_DOWN or event.key == K_s:\n hero.stop = True\n\n time.sleep(0.01)\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"PlaneWar/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":7181,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"14340630","text":"def division1(x, y):\n\tprint(x / y)\n\treturn x / y\n\ndef division2(x, y):\n\tprint(x // y)\n\treturn x // y\n\n# division1(5, 10)\ndivision2(5, 10)\n# division1(-8, 2)\ndivision2(-8, 2)\n# division1(17, 13)\ndivision2(17, 13)\n# division1(-10, -3)\ndivision2(-10, -3)\n# division1(15, -4)\ndivision2(15, -4)\n\n\n\n\n\n","sub_path":"python/java_vs_python.py3","file_name":"java_vs_python.py3","file_ext":"py3","file_size_in_byte":295,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"372654308","text":"#from pacman3.py import *\nimport sys\nsys.path.insert(0, \".\\\\aima-python-master\")\nfrom search import *\nclass MazeGraph:\n\n def __init__(self, pac_game):\n self.pacGame=pac_game\n # self.maze_str=pac_game.strS #[str[0]][len(str)], for example, middlwMaze 18rows, 36 cols\n self.maze_height=len(self.pacGame.strS)\n self.maze_width=len(self.pacGame.strS[0])\n self.walls =pac_game.walls\n self.map = dict()\n self.graph=dict()\n self.edgeCosts={}\n self.sortedCapsulePos=[]\n\n\n def genGraph(self):\n dict={}\n\n xMax=self.maze_width\n yMax=self.maze_height\n\n for x in range(1,xMax-1):\n for y in range(1, yMax-1):\n if(self.walls[x][y]==True):\n continue;\n else:\n dict1={}\n\n if (self.walls[x][y+1] == False):\n dict1[(x, y+1)]=1 #south\n if (self.walls[x + 1][y] == False):\n dict1[(x + 1, y)] = 1 # suppose the cost is 1 east\n if (self.walls[x][y-1] == False):\n dict1[(x, y-1)] = 1 # suppose the cost is 1 North\n if (self.walls[x-1][y] == False): #West\n dict1[(x-1, y)] = 1\n dict[(x,y)]=dict1\n self.map=dict\n self.edgeCosts=dict\n self.graph=UndirectedGraph(self.map)\n self.graph.locations= {(0,0)}\n\n\n def getkey(self,item):\n return item[1]\n\n def print_map(self):\n for k, v in self.map.items():\n print(k, v)\n\n def print_keys(self):\n for k in self.map.keys():\n print (k)\n","sub_path":"CSCI4802-2020-pacmanlab2-ch3-UCS/maze_graph.py","file_name":"maze_graph.py","file_ext":"py","file_size_in_byte":1692,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"118408876","text":"import os\nimport logging\nlogging.basicConfig(format='%(asctime)s : %(levelname)s : %(message)s', level=logging.WARNING)\n\nfrom gensim import corpora, models, similarities\n\nfrom TextsDAO import TextsDAO\nfrom CorpusDAO import DictionaryDAO\nfrom CorpusDAO import CorpusDAO\n\n\n\nBASE_DIR = \".\"\nBASE_META_DIR = \".\"\nSERIALIZED_CORPUS = os.path.join(BASE_META_DIR, \"corpus.mm\")\nSERIALIZED_TFIDF_CORPUS = os.path.join(BASE_META_DIR, \"corpus_tfidf.mm\")\n\ndef main():\n\n dictionary = DictionaryDAO(BASE_META_DIR, BASE_DIR).getDictionary()\n\n if os.path.isfile(SERIALIZED_CORPUS):\n corpus = corpora.MmCorpus(SERIALIZED_CORPUS)\n else:\n corpus = CorpusDAO(BASE_META_DIR, BASE_DIR)\n corpora.MmCorpus.serialize(SERIALIZED_CORPUS, corpus)\n\n #for key, value in corpus.getDictionary().items():\n # print(\"Key:{} Value:{}\".format(key, value))\n # Confirm if its populated\n #for vector in corpus:\n # print(vector)\n if os.path.isfile(SERIALIZED_TFIDF_CORPUS):\n tfidf = models.TfidfModel(corpus)\n corpus_tfidf = corpora.MmCorpus(SERIALIZED_TFIDF_CORPUS)\n else:\n tfidf = models.TfidfModel(corpus)\n corpus_tfidf = tfidf[corpus]\n corpora.MmCorpus.serialize(SERIALIZED_TFIDF_CORPUS, corpus_tfidf)\n \n lsi = models.LsiModel(corpus_tfidf, id2word=dictionary, num_topics=20)\n corpus_lsi = lsi[corpus_tfidf]\n\n\n count = 0\n\n for doc in corpus_lsi:\n count += 1\n\n print(\"Length of corpus is \" + str(count))\n\n document = \"\"\"b''\nb\"When setting a form's opacity should I use a decimal or double?\"\nb\"
I want to use a track-bar to change a form's opacity.
\\n\\n
This is my code:
\\n\\n
decimal trans = trackBar1.Value / 5000;\\nthis.Opacity = trans;\\n
\\n\\n
When I try to build it, I get this error:
\\n\\n
\\n Given a specific DateTime value, how do I display relative time, like:
\\n\\n
\\n
2 hours ago
\\n
3 days ago
\\n
a month ago
\\n
\\n\\n
Et cetera?\"\n\"\"\"\n document = ''.join(e for e in document if e.isalnum() or e == ' ')\n new_doc_bow = dictionary.doc2bow(document.lower().split())\n #print(tfidf[new_doc_bow])\n\n new_corpus_tfidf = [tfidf[new_doc_bow]]\n lsi.add_documents(new_corpus_tfidf)\n new_corpus_lsi = lsi[new_corpus_tfidf]\n \n doc_count = 0\n topic_count = 0\n for new_doc in new_corpus_lsi:\n doc_count += 1\n print(new_doc)\n for topic_id, coorelation in new_doc:\n if coorelation > 0.05:\n lsi.print_topic(topic_id)\n topic_count += 1\n print(\"All docs: {} AND Related topics: {}\".format(doc_count, topic_count))\n \n\nif __name__ == \"__main__\":\n main()","sub_path":"module1.py","file_name":"module1.py","file_ext":"py","file_size_in_byte":2742,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"3995191","text":"import math\nfrom enum import Enum\nfrom my_math import Vector2, moment_of_inertia\n\n\nclass AaBb:\n\n def __init__(self, x1, y1, x2, y2):\n self._x1 = x1\n self._y1 = y1\n self._x2 = x2\n self._y2 = y2\n\n def __add__(self, other):\n if isinstance(other, AaBb):\n x1 = min(other.get_x1, self.get_x1)\n x2 = min(other.get_x2, self.get_x2)\n y1 = min(other.get_y1, self.get_y1)\n y2 = min(other.get_y2, self.get_y2)\n return AaBb(x1, y1, x2, y2)\n return None\n\n def __iadd__(self, other):\n if isinstance(other, AaBb):\n x1 = min(other.get_x1, self.get_x1)\n x2 = min(other.get_x2, self.get_x2)\n y1 = min(other.get_y1, self.get_y1)\n y2 = min(other.get_y2, self.get_y2)\n self.set(x1, y1, x2, y2)\n return self\n return None\n\n def __str__(self):\n return \"x1:\" + str(self.get_x1) + \"; y1:\" + str(self.get_y1) + \"; x2:\" + str(self.get_x2) + \"; y2:\" + str(\n self.get_y2) + ';'\n\n def set(self, x1, y1, x2=None, y2=None):\n if isinstance(x1, AaBb) and isinstance(x2, AaBb):\n self._x1 = min(x1.get_x1, x2.get_x1)\n self._y1 = max(x1.get_x2, x2.get_x2)\n self._x2 = min(x1.get_y1, x2.get_y1)\n self._y2 = max(x1.get_y2, x2.get_y2)\n else:\n self._x1 = min(x1, x2)\n self._y1 = min(y1, y2)\n self._x2 = max(x1, x2)\n self._y2 = max(y1, y2)\n\n def intersection_AaBb(self, x1, y1=None, x2=None, y2=None):\n if isinstance(x1, AaBb):\n if self.get_x1 > x1.get_x2 or self.get_x2 < x1.get_x1:\n return False\n elif self.get_y1 > x1.get_y2 or self.get_y2 < x1.get_y1:\n return False\n else:\n return True\n else:\n if self.get_x1 > x2 or self.get_x2 < x1:\n return False\n elif self.get_y1 > y2 or self.get_y2 < y1:\n return False\n else:\n return True\n\n @property\n def get_x1(self):\n return self._x1\n\n @property\n def get_y1(self):\n return self._y1\n\n @property\n def get_x2(self):\n return self._x2\n\n @property\n def get_y2(self):\n return self._y2\n\n\nclass Shape:\n\n def __init__(self, vertices: list):\n self._static_vertices = tuple(vertices)\n self._vertices = vertices\n self._normals = self._create_normals()\n self._center = self._get_center()\n self._aaBb = AaBb(0, 0, 0, 0)\n\n def _get_center(self) -> Vector2:\n v = Vector2(0, 0)\n for num in range(len(self._static_vertices) // 2):\n v.x += self._vertices[2 * num]\n v.y += self._vertices[2 * num + 1]\n v *= 2 / len(self._vertices)\n return v\n\n def _create_normals(self):\n v = self._vertices\n nor = []\n for n in range(len(v) // 2 - 1):\n q = Vector2(v[2 * n] - v[2 * n + 2], v[2 * n + 1] - v[2 * n + 3]).rotate90(1).nor()\n nor.append(q)\n return nor\n\n def _update_aaBb(self):\n v = self._vertices\n x1, y1, = v[0], v[1]\n x2, y2 = x1, y1\n for i in range(len(v) // 2):\n if x1 > v[2 * i]:\n x1 = v[2 * i]\n if y1 > v[2 * i + 1]:\n y1 = v[2 * i + 1]\n if x2 < v[2 * i]:\n x2 = v[2 * i]\n if y2 < v[2 * i + 1]:\n y2 = v[2 * i + 1]\n self._aaBb.set(x1, y1, x2, y2)\n\n def _update_normals(self):\n v = self._vertices\n nor = self.normals\n for n in range(len(v) // 2 - 1):\n nor[n].set(v[2 * n] - v[2 * n + 2], v[2 * n + 1] - v[2 * n + 3]).rotate90(1).nor()\n\n def update(self, angle, position):\n s_v = self._static_vertices\n v = self._vertices\n\n cos = 1\n sin = 0\n if angle != 0:\n cos = math.cos(angle)\n sin = math.sin(angle)\n\n for num in range(len(v) // 2 - 1):\n x = s_v[2 * num]\n y = s_v[2 * num + 1]\n old_x = x\n x = x * cos - y * sin\n y = old_x * sin + y * cos\n v[2 * num] = x + position.x\n v[2 * num + 1] = y + position.y\n v[-1] = v[1]\n v[-2] = v[0]\n self._update_aaBb()\n self._update_normals()\n\n @property\n def get_aaBb(self) -> AaBb:\n return self._aaBb\n\n @property\n def vertices(self):\n return self._vertices\n\n @property\n def normals(self):\n return self._normals\n\n @property\n def center(self):\n return self._center\n\n @vertices.setter\n def vertices(self, v):\n self._vertices = v\n self._static_vertices = tuple(v)\n\n\nclass Body_Type(Enum):\n Static = 0\n Dynamic = 1\n\n\nclass Body_Def:\n\n def __init__(self, pos: Vector2, t: Body_Type):\n self.type = t\n self.mass = 10.0\n self.i = 0\n self.pos = pos\n self.velocity = Vector2(0, 0)\n\n self.angle = 0.0\n self.angular_velocity = 0.0\n self.elasticity = 1.0\n self.friction = 0.0\n\n def set_mass(self, mass: float):\n self.mass = mass\n return self\n\n def set_elasticity(self, elasticity: float):\n self.elasticity = elasticity\n return self\n\n def set_friction(self, friction: float):\n self.friction = friction\n return self\n\n def sef_velocity(self, v, v_y=None):\n if v_y is None:\n self.velocity.set(v.x, v.y)\n else:\n self.velocity.set(v, v_y)\n return self\n\n def set_angular_velocity(self, a_v):\n self.angular_velocity = a_v\n return self\n\n\nclass Body:\n\n def __init__(self, shape: Shape, body_def: Body_Def):\n self.body_def = body_def\n self._shape = shape\n\n self._force = Vector2(0, 0)\n self._moment_force = 0\n\n self._user_data = None\n\n self.body_def.i = moment_of_inertia(self._shape.vertices, body_def.mass, self.position)\n self._inv_i = 1 / self.body_def.i\n self._inv_m = 1 / body_def.mass\n\n def step(self, delta):\n if self.type == Body_Type.Static:\n pass\n elif self.type == Body_Type.Dynamic:\n self.body_def.velocity.x += self._force.x * self._inv_m * delta\n self.body_def.velocity.y += self._force.y * self._inv_m * delta\n self.body_def.angle += self.body_def.angular_velocity * delta\n self.body_def.pos.add(self.body_def.velocity.x * delta, self.body_def.velocity.y * delta)\n self._shape.update(self.angle, self.position)\n self._force.set(0, 0)\n\n def add_force(self, force, force_y=None):\n if force_y is None:\n self._force += force\n else:\n self._force.x += force\n self._force.y += force_y\n\n def add_position(self, x, y=None):\n if isinstance(x, Vector2):\n self.body_def.pos.x += x.x\n self.body_def.pos.y += x.y\n else:\n self.body_def.pos.x += x\n self.body_def.pos.y += y\n\n @property\n def position(self) -> Vector2:\n return self.body_def.pos\n\n @position.setter\n def position(self, pos):\n self.body_def.pos.set(pos.x, pos.y)\n\n @property\n def velocity(self) -> Vector2:\n return self.body_def.velocity\n\n @velocity.setter\n def velocity(self, v):\n if self.type != Body_Type.Static:\n self.body_def.velocity.set(v.x, v.y)\n\n @property\n def angular_velocity(self) -> float:\n return self.body_def.angular_velocity\n\n @angular_velocity.setter\n def angular_velocity(self, w: float):\n if self.type != Body_Type.Static:\n self.body_def.angular_velocity = w\n\n @property\n def vertices(self) -> list:\n return self._shape.vertices\n\n @vertices.setter\n def vertices(self, v: list):\n self._shape.vertices = v\n\n @property\n def angle(self) -> float:\n return self.body_def.angle\n\n @angle.setter\n def angle(self, angle: float):\n self.body_def.angle = angle\n\n @property\n def user_data(self):\n return self._user_data\n\n @user_data.setter\n def user_data(self, d):\n self._user_data = d\n\n @property\n def shape(self) -> Shape:\n return self._shape\n\n @property\n def elasticity(self):\n return self.body_def.elasticity\n\n @property\n def get_aaBb(self) -> AaBb:\n return self.shape.get_aaBb\n\n @property\n def type(self) -> Body_Type:\n return self.body_def.type\n\n @property\n def mass(self) -> float:\n return self.body_def.mass\n\n @property\n def i(self):\n return self.body_def.i\n","sub_path":"body.py","file_name":"body.py","file_ext":"py","file_size_in_byte":8722,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"351611608","text":"#\r\n# @lc app=leetcode.cn id=1373 lang=python3\r\n#\r\n# [1373] 二叉搜索子树的最大键值和\r\n#\r\n# https://leetcode-cn.com/problems/maximum-sum-bst-in-binary-tree/description/\r\n#\r\n# algorithms\r\n# Hard (39.58%)\r\n# Likes: 55\r\n# Dislikes: 0\r\n# Total Accepted: 6.5K\r\n# Total Submissions: 16.4K\r\n# Testcase Example: '[1,4,3,2,4,2,5,null,null,null,null,null,null,4,6]'\r\n#\r\n# 给你一棵以 root 为根的 二叉树 ,请你返回 任意 二叉搜索子树的最大键值和。\r\n# \r\n# 二叉搜索树的定义如下:\r\n# \r\n# \r\n# 任意节点的左子树中的键值都 小于 此节点的键值。\r\n# 任意节点的右子树中的键值都 大于 此节点的键值。\r\n# 任意节点的左子树和右子树都是二叉搜索树。\r\n# \r\n# \r\n# \r\n# \r\n# 示例 1:\r\n# \r\n# \r\n# \r\n# \r\n# 输入:root = [1,4,3,2,4,2,5,null,null,null,null,null,null,4,6]\r\n# 输出:20\r\n# 解释:键值为 3 的子树是和最大的二叉搜索树。\r\n# \r\n# \r\n# 示例 2:\r\n# \r\n# \r\n# \r\n# \r\n# 输入:root = [4,3,null,1,2]\r\n# 输出:2\r\n# 解释:键值为 2 的单节点子树是和最大的二叉搜索树。\r\n# \r\n# \r\n# 示例 3:\r\n# \r\n# \r\n# 输入:root = [-4,-2,-5]\r\n# 输出:0\r\n# 解释:所有节点键值都为负数,和最大的二叉搜索树为空。\r\n# \r\n# \r\n# 示例 4:\r\n# \r\n# \r\n# 输入:root = [2,1,3]\r\n# 输出:6\r\n# \r\n# \r\n# 示例 5:\r\n# \r\n# \r\n# 输入:root = [5,4,8,3,null,6,3]\r\n# 输出:7\r\n# \r\n# \r\n# \r\n# \r\n# 提示:\r\n# \r\n# \r\n# 每棵树有 1 到 40000 个节点。\r\n# 每个节点的键值在 [-4 * 10^4 , 4 * 10^4] 之间。\r\n# \r\n# \r\n#\r\n\r\n# @lc code=start\r\n# Definition for a binary tree node.\r\n# class TreeNode:\r\n# def __init__(self, val=0, left=None, right=None):\r\n# self.val = val\r\n# self.left = left\r\n# self.right = right\r\nclass Solution:\r\n def maxSumBST(self, root: TreeNode) -> int:\r\n self.maxSum = float('-inf')\r\n def traverse(root):\r\n if root is None:\r\n # 返回一个长度为4的数组,res[0]为以root为根的二叉树是否是BST\r\n # root[1]记录以root为根二叉树所有节点的最小值\r\n # root[2]记录以root为根二叉树所有节点的最大值\r\n # root[3]记录以root为根的二叉树所有节点值之和\r\n return [1, float('inf'), float('-inf'), 0] \r\n \r\n left = traverse(root.left)\r\n right = traverse(root.right)\r\n\r\n res = [0] * 4\r\n # 判断是否当前节点为根是否是二叉树\r\n if left[0] == 1 and right[0] == 1 and root.val > left[2] and root.val < right[1]:\r\n res[0] = 1\r\n res[1] = min(left[1], root.val)\r\n res[2] = max(right[2], root.val)\r\n res[3] = left[3] + right[3] + root.val\r\n self.maxSum = max(self.maxSum, res[3])\r\n else:\r\n res[0] = 0\r\n return res\r\n \r\n return traverse(root)[-1]\r\n \r\n\r\n# @lc code=end\r\n\r\n","sub_path":"leetcode/1373.二叉搜索子树的最大键值和.py","file_name":"1373.二叉搜索子树的最大键值和.py","file_ext":"py","file_size_in_byte":3022,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"399828026","text":"\"\"\"Provide data suitable for Fava's charts. \"\"\"\nfrom datetime import date, datetime\n\nfrom beancount.core.amount import Amount\nfrom beancount.core.number import Decimal\nfrom beancount.core.position import Position\nfrom beancount.core.inventory import Inventory\nfrom beancount.core import realization\nfrom beancount.core.data import iter_entry_dates\nfrom flask.json import JSONEncoder\n\nfrom fava.core.helpers import FavaModule\nfrom fava.core.holdings import net_worth_at_dates\n\n\nclass FavaJSONEncoder(JSONEncoder):\n\n def default(self, o): # pylint: disable=E0202\n if isinstance(o, datetime):\n return o.strftime('%Y-%m-%dT%H:%M:%SZ')\n elif isinstance(o, date):\n return o.strftime('%Y-%m-%d')\n elif isinstance(o, Decimal):\n return float(o)\n elif isinstance(o, Amount):\n return str(o)\n elif isinstance(o, Position):\n return str(o)\n elif isinstance(o, (set, frozenset)):\n return list(o)\n try:\n return JSONEncoder.default(self, o)\n except TypeError:\n # workaround for #472\n try:\n return str(o)\n except TypeError:\n return ''\n\n\ndef _serialize_inventory(inventory, at_cost=False):\n \"\"\"Renders an Inventory to a currency -> amount dict.\"\"\"\n if at_cost:\n inventory = inventory.cost()\n else:\n inventory = inventory.units()\n return {p.units.currency: p.units.number for p in inventory}\n\n\ndef _real_account(account_name, entries, begin_date, end_date):\n if begin_date:\n entries = list(iter_entry_dates(entries, begin_date, end_date))\n\n return realization.get_or_create(realization.realize(entries),\n account_name)\n\n\ndef _serialize_real_account(real_account):\n return {\n 'account': real_account.account,\n 'balance_children':\n _serialize_inventory(realization.compute_balance(real_account),\n at_cost=True),\n 'balance': _serialize_inventory(real_account.balance, at_cost=True),\n 'children': [_serialize_real_account(a)\n for n, a in sorted(real_account.items())],\n }\n\n\nclass ChartModule(FavaModule):\n __slots__ = ['ledger']\n\n def _total_balance(self, names, begin_date, end_date):\n totals = [realization.compute_balance(\n _real_account(account_name, self.ledger.entries, begin_date,\n end_date))\n for account_name in names]\n return _serialize_inventory(sum(totals, Inventory()),\n at_cost=True)\n\n def events(self, event_type):\n return [{\n 'type': entry.type,\n 'date': entry.date,\n 'description': entry.description\n } for entry in self.ledger.events(event_type)]\n\n def hierarchy(self, account_name, begin_date=None, end_date=None):\n real_account = _real_account(\n account_name, self.ledger.entries, begin_date, end_date)\n return _serialize_real_account(real_account)\n\n def interval_totals(self, interval, account_name):\n \"\"\"Renders totals for account (or accounts) in the intervals.\"\"\"\n if isinstance(account_name, str):\n names = [account_name]\n else:\n names = account_name\n\n interval_tuples = self.ledger._interval_tuples(interval)\n return [{\n 'begin_date': begin_date,\n 'totals': self._total_balance(\n names,\n begin_date, end_date),\n 'budgets': self.ledger.budgets.calculate(names[0], begin_date,\n end_date),\n } for begin_date, end_date in interval_tuples]\n\n def linechart(self, account_name):\n real_account = realization.get_or_create(self.ledger.root_account,\n account_name)\n postings = realization.get_postings(real_account)\n journal = realization.iterate_with_balance(postings)\n\n return [{\n 'date': entry.date,\n # when there's no holding for a commodity, it will be missing from\n # 'balance' field but appear in 'change' field. Use 0 for those\n # commodities.\n 'balance': dict({curr: 0 for curr in list(change.currencies())},\n **_serialize_inventory(balance)),\n } for entry, _, change, balance in journal if len(change)]\n\n def net_worth_at_dates(self, interval):\n interval_tuples = self.ledger._interval_tuples(interval)\n if not interval_tuples:\n return []\n\n dates = [interval_tuples[0][0]] + [p[1] for p in interval_tuples]\n\n return net_worth_at_dates(self.ledger.entries, dates,\n self.ledger.price_map,\n self.ledger.options)\n","sub_path":"fava/core/charts.py","file_name":"charts.py","file_ext":"py","file_size_in_byte":4927,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"167227555","text":"from rest_framework.generics import CreateAPIView\nfrom .serializers import CalculateSerializer\nfrom .models import Calculate\n\n\nclass CalculateCreateView(CreateAPIView):\n queryset = Calculate.objects.all()\n serializer_class = CalculateSerializer\n\n def post(self, request, *args, **kwargs):\n r = super(CalculateCreateView, self).create(request)\n # call the calculation funcion\n self.calculate_qrisk()\n return r\n\n def calculate_qrisk(self):\n from libs import qrisk_male, qrisk_female\n\n data = self.request.data\n\n cholesterol = 0.0 if data.get('cholesterol') == '' else data.get('cholesterol')\n sytolic = 0.0 if data.get('sytolic') == '' else data.get('sytolic')\n town = 0.0\n b_AF = 0.0 if data.get('atrial_fibrillation') is None else data.get('atrial_fibrillation')\n b_ra = 0.0 if data.get('rheumatoid') is None else data.get('rheumatoid')\n b_renal = 0.0 if data.get('kidney') is None else data.get('kidney')\n b_treatedhyp = 0.0 if data.get('on_blood_pressure_treatment') is None else data.get('on_blood_pressure_treatment')\n angina = 0.0 if data.get('angina') is None else data.get('angina')\n\n if data.get('sex') == 'M':\n result = qrisk_male.cvd_male_raw(\n int(data.get('age')), b_AF,\n b_ra, b_renal,\n b_treatedhyp, 0.0, 0.0, 1.0,\n int(data.get('ethny')), angina, cholesterol,\n sytolic, int(data.get('smoke')), 0.0, town)\n print(result)\n else:\n result = qrisk_female()\n\n return result\n","sub_path":"calculate/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1622,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"26995361","text":"#!/usr/bin/python\n\nimport os\nimport logging\nimport requests\nimport copy\nimport ast\nfrom ansible.module_utils.basic import AnsibleModule\nfrom ansible.module_utils.pask_prestapi import PrestApi,\\\n OP_DELETE, OP_GET, OP_POST, OP_PUT\nfrom ansible.module_utils.pask_module import PaskModule, try_except\n\n\ninner_ip_args = dict(\n address=dict(type='str', required=True),\n broadcast=dict(type='str'),\n overlapped=dict(type='str')\n)\n\ninner_ip6_args = dict(\n address=dict(type='str', required=True),\n broadcast=dict(type='str'),\n)\n\ninner_ip6_args['adv-on-link'] = dict(type='str')\ninner_ip6_args['adv-autonomous'] = dict(type='str')\ninner_ip6_args['adv-router-addr'] = dict(type='str')\ninner_ip6_args['adv-valid-lifetime'] = dict(type='str')\ninner_ip6_args['adv-preferred-lifetime'] = dict(type='str')\n\nmodule_args = dict(\n name=dict(type='str', required=True),\n ip=dict(type='dict', options=inner_ip_args),\n ip6=dict(type='dict', options=inner_ip6_args),\n mtu=dict(type='str'),\n rpf=dict(type='str'),\n status=dict(type='str'),\n)\n\nmodule_args['adv-cur-hop-limit'] = dict(type='str')\nmodule_args['adv-default-lifetime'] = dict(type='str')\nmodule_args['adv-reachable-time'] = dict(type='str')\nmodule_args['adv-retrans-timer'] = dict(type='str')\nmodule_args['adv-send-advert'] = dict(type='str')\nmodule_args['max-rtr-adv-interval'] = dict(type='str')\nmodule_args['min-rtr-adv-interval'] = dict(type='str')\n\nname = 'interface'\n\n\nclass PaskInterface(PaskModule):\n def __init__(self, name, module_args):\n super(PaskInterface, self).__init__(name, module_args)\n\n @try_except\n def run(self):\n data = self.make_data(self.module.params, include_inner=True)\n url = os.path.join(self.url, self.module.params['name'])\n self.resp= self.put(url, data)\n\n\ndef main():\n interface = PaskInterface(name, module_args)\n interface.set_param()\n interface.run()\n interface.set_result()\n\nif __name__ == '__main__':\n main()\n","sub_path":"library/pask_interface.py","file_name":"pask_interface.py","file_ext":"py","file_size_in_byte":1975,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"191797940","text":"import networkx as nx\nimport utility\n\nmseed = utility.mGraph()\nmseed.load_from_file('eval.mgraph')\n\nm = utility.mGraph()\nm.load_from_file('../max_degree_3.mgraph')\n\ndd = nx.degree_histogram(mseed.g) ## for graph degree\ndd = [i / sum(dd) for i in dd]\ndd_bin = mseed.get_dd_bin()\ndd_bin = [i / sum(dd_bin) for i in dd_bin]\ndd2_bin = mseed.get_dd2_bin()\n\n#print(dd)\nprint(dd_bin)\nprint(dd2_bin)\n\ndd = nx.degree_histogram(m.g) ## for graph degree\ndd = [i / sum(dd) for i in dd]\ndd_bin = m.get_dd_bin()\ndd_bin = [i / sum(dd_bin) for i in dd_bin]\ndd2_bin = m.get_dd2_bin()\n\n#print(dd)\nprint(dd_bin)\nprint(dd2_bin)\n","sub_path":"marsQuery/cal_degree.py","file_name":"cal_degree.py","file_ext":"py","file_size_in_byte":608,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"600237587","text":"import math\nimport random\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n#parte 3\ndef exponential(x):\n time = np.log(np.random.uniform()) *-1\n\n return time/x\n\n\ndef strati(values_list): \n\tprint(\"rango 0 a 1:\")\n\tev = 0\n\tfor i in range(len(values_list[0:10])):\n\t\tev = ev + (i/10)*(values_list[i])\n\tprint(\"valor esperado es: \", ev)\n\n\tev = 0\n\tprint(\"rango 1 a 3:\")\n\tfor i in range(len(values_list[10:30])):\n\t\tev = ev + (i+10)*(values_list[i+10])\n\tprint(\"valor esperado es: \", ev/10)\n\n\tev = 0\n\tprint(\"rango 3 a infinito:\")\n\tfor i in range(len(values_list[30:])):\n\t\tev = ev + (i+30)*(values_list[i+30])\n\tprint(\"valor esperado es: \", ev/10000)\n\n\nx_values =[]\ny_values =[]\nfor i in range(1,10000):\n\tx_values.append(i)\n\ty_values.append(exponential(i))\n\nstrati(values_list = y_values)\n\nplt.plot(y_values)\nplt.savefig('books_read.png')\n","sub_path":"start.py","file_name":"start.py","file_ext":"py","file_size_in_byte":836,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"293273551","text":"# -*- coding: utf-8 -*-\n\nimport pandas as pd\n#import cProfile\nimport time\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport seaborn as sns\nstart = time.time()\n\n\ndef excel_mode():\n start = time.time()\n for i in range(3):\n df0 = pd.read_csv('ex.csv')\n print(i)\n end = time.time()\n print('excel_mode duration:',str(end-start))\n#print(df0.describe())\n\ndef txt_mode():\n start = time.time()\n for i in range(3):\n df1 = pd.read_table('ex.txt',sep='\\t',encoding='gb2312')\n print(i)\n end = time.time()\n print('txt_mode duration:',str(end-start))\n\nexcel_mode()\ntxt_mode()\n\n#df0 = pd.read_excel('ex.xlsx')\n#df0 = pd.read_table('ex.txt',sep='\\t',encoding='gb2312')\n#fig = plt.figure()\n##ax = fig.add_subplot(111)\n#df0.boxplot(column= 'ADC',by='SBR')\n##sns.boxplot(x='SBR',y='ADC',data=df0)\n#print(df0.groupby('SBR')['ADC'].agg([np.mean,np.std,'count']))\n#end = time.time()\n#print('test time:',str(end-start))","sub_path":"2-work/Python/python-Intv/company_plot/plot.py","file_name":"plot.py","file_ext":"py","file_size_in_byte":955,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"42364336","text":"# -*- coding: utf-8 -*-\n\"\"\" Module intis data in DB \"\"\"\nimport logging\nfrom arttest.services import (ArticlesService, AccountService,ArticleTypesService, SubscribersService)\nfrom arttest.models.models import (Article, Account, ArticleType, Subscriber)\nfrom arttest.logic.helpers import dt_to_text\n\nLOG = logging.getLogger(__name__)\n\nclass InitDB(object):\n \"\"\" docstring for InitDB \"\"\"\n def __init__(self):\n self.articles = ArticlesService()\n self.accounts = AccountService()\n self.types = ArticleTypesService()\n self.subscribers = SubscribersService()\n self.dgroups = {}\n\n def add(self):\n \"\"\" Inits DB \"\"\"\n LOG.info(\"Initializing db data\")\n try:\n if self.types.count() == 0:\n self.init_article_types()\n if self.accounts.count() == 0:\n self.init_admin()\n LOG.info(\"Successfully finished\")\n except Exception as ex:\n LOG.error(\"An error occured while Initializing db data\")\n LOG.exception(ex)\n\n def init_admin(self):\n \"\"\" Inits first admin user \"\"\"\n account = Account()\n account.login = \"admin\"\n account.password = \"admin01\"\n account.name = \"Admin\"\n account.lastname = \"Admin\"\n account.permissions = \"Admin\"\n self.accounts.add(account)\n\n def init_article_types(self):\n \"\"\"Adds article types to DB \"\"\"\n # Init first\n article_type = ArticleType()\n article_type.code = 'grey'\n article_type.name = 'Szary'\n article_type.color = '#242424'\n # Init second\n article_type2 = ArticleType()\n article_type2.code = 'red'\n article_type2.name = 'Czerwony'\n article_type2.color = '#bddbbb'\n # Init second\n article_type3 = ArticleType()\n article_type3.code = 'green'\n article_type3.name = 'Zielony'\n article_type3.color = '#18c118'\n self.types.add(article_type)\n self.types.add(article_type2)\n self.types.add(article_type3)\n\n def init_subscriber(self):\n \"\"\" inits subscriber \"\"\"\n sub = Subscriber()\n sub.name = \"aaaa\"\n sub.lastname = \"dsadsad\"\n sub.email = \"SSS@sds\"\n sub = self.subscribers.add(sub)\n LOG.debug(str(sub.id))\n\n ","sub_path":"arttest/scripts/initdb.py","file_name":"initdb.py","file_ext":"py","file_size_in_byte":2322,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"83010535","text":"\"\"\"Turtle彩虹绘制\"\"\"\nimport turtle as t\n\n\ndef init(x, y, z, speed):\n t.setup(x, y)\n t.pensize(z)\n t.speed(speed)\n\n\ndef start(start, end):\n t.penup()\n t.setx(start)\n t.sety(end)\n t.pendown()\n\n\ndef draw(color, position, radius, extend):\n t.pendown()\n t.left(position)\n t.color(color)\n t.circle(radius, extend)\n t.penup()\n\n\ninit(800, 800, 20, 10)\n\nstart(100, 0)\ndraw('red', 90, 100, 180)\nstart(120, 0)\ndraw('orange', 180, 120, 180)\nstart(140, 0)\ndraw('yellow', 180, 140, 180)\nstart(160, 0)\ndraw('green', 180, 160, 180)\nstart(180, 0)\ndraw('cyan', 180, 180, 180)\nstart(200, 0)\ndraw('blue', 180, 200, 180)\nstart(220, 0)\ndraw('purple', 180, 220, 180)\nt.done()\n","sub_path":"L2/2_Rainbow.py","file_name":"2_Rainbow.py","file_ext":"py","file_size_in_byte":693,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"126306128","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\nfrom django.utils.encoding import python_2_unicode_compatible\nfrom django.shortcuts import render\nfrom models import Question, Choice, Build\nfrom django.http import JsonResponse,HttpResponseRedirect,Http404,HttpResponse\n\nfrom django.views.decorators.csrf import csrf_exempt\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib import auth\nfrom django.core import serializers\nimport json\n\nlogin_url='/login/'\n\n@login_required(login_url=login_url)\ndef get_data_quest(request):\n data = Build.objects.all()[:170]\n requestData=[]\n for i in request.META:\n requestData.append([i,request.META[i]])\n return render(request,r'grid/GridJQuery/index.htm',{'data': data,\n 'requestData':requestData})\n\n\n@login_required(login_url=login_url)\ndef sb_admin(request):\n data = Build.objects.all()[:170]\n return render(request,r'grid/Bootstrap/index.html')\n\n@login_required(login_url=login_url)\ndef sb_tables(request):\n return render(request,r'grid/Bootstrap/tables.html')\n\n@login_required(login_url=login_url)\ndef map(request):\n return render(request,r'v1/map-element.html',{'Dashboard':'',\n 'info':'Карта присутності провадерів інтернет','map':True})\n\n@login_required(login_url=login_url)\ndef build_detal_information(request):\n id=int(request.GET['id'])\n data = Build.objects.get(id=id)\n adress=str(data)\n balans=str(data.balans)\n service=str(data.service)\n\n return JsonResponse({'adress':adress,'balans':balans,'service':service})\n\n\n@login_required(login_url=login_url)\n@csrf_exempt # декоратор вимикає перевірку csrf -токена !!!!!!!!!!!!\ndef send_claim(request):\n id=int(request.POST['id'])\n data = Build.objects.get(id=id)\n adress=str(data)\n balans=str(data.balans)\n service=str(data.service)\n return JsonResponse({'adress':adress,'balans':balans,'service':service})\n\n@login_required(login_url=login_url)\ndef test_build(request):\n data = Build.objects.all()[1:2].values_list()\n requestData=[]\n for i in data:\n requestData.append(i)\n return render(request,r'map/Object_detal_information.html',{'requestData':requestData})\n\n\ndef login(request):\n return render(request,r'v1/login-form.html')\n\n@login_required(login_url=login_url)\ndef base(request):\n return render(request,r'v1/base.html')\n\ndef tables(request):\n return render(request,r'v1/table.html')\n\n\n##@csrf_exempt # декоратор вимикає перевірку csrf -токена !!!!!!!!!!!!\ndef login_validate(request):\n html = \"
Ніхуя не вийшло \"\n if request.method == 'POST':\n username = request.POST.get('username', '')\n password = request.POST.get('password', '')\n user = auth.authenticate(username=username, password=password)\n if user is not None and user.is_active:\n # Пароль правилен и пользователь “активный”\n auth.login(request, user)\n # Переадресовать на страницу успешного входа.\n return HttpResponseRedirect('../')\n else:\n # Переадресовать на страницу ошибок\n return HttpResponse(html)\n else:\n # Переадресовать на страницу ошибок\n return HttpResponse(html)\n\ndef logout(request):\n auth.logout(request)\n # Переадресовать на страницу успешного выхода.\n return HttpResponseRedirect(\"/login/\")\n\n@login_required(login_url=login_url)\ndef get_build_table_data(request):\n data = Build.objects.all()[:1000]\n jasonDatalist=[]\n for i in data:\n jasonDatalist.append([str(i),str(i.balans),str(i.service)])\n r=json.dumps(jasonDatalist)\n return HttpResponse(r, content_type=\"application/json\")\n\n\n","sub_path":"drtm/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":3992,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"299292541","text":"\"\"\"Support functions for verification of embedded license claims.\"\"\"\n\n__id__ = \"$Id$\"\n__version__ = \"$Revision$\"\n__copyright__ = '(c) 2004, Creative Commons, Nathan R. Yergler'\n__license__ = 'licensed under the GNU GPL2'\n\nimport ccrdf\nimport ccrdf.rdfextract as rdfextract\n\nimport cctagutils.rdf\nfrom cctagutils.metadata import metadata\n\ndef parseClaim(claim):\n results = {}\n\n vtext = 'verify at '\n vloc = claim.find(vtext)\n if vloc != -1:\n results['verify at'] = claim[vloc+len(vtext):].strip()\n claim = claim[:vloc]\n\n ltext = \"licensed to the public under \"\n lloc = claim.lower().find(ltext)\n if lloc != -1:\n results['license'] = claim[lloc+len(ltext):].strip()\n claim = claim[:lloc]\n\n results['copyright'] = claim.strip()\n\n return results\n\ndef lookup(filename):\n \"\"\"Returns True of False if the embedded claim can be verified.\"\"\"\n \n if verify(filename) > 0:\n return True\n else:\n return False\n \ndef verify(filename):\n \"\"\"Extracts license claim information from a file and verifies it.\n Returns the following status codes:\n 1 Verified\n 0 No RDF\n -1 Work information not found (possible SHA1 mismatch)\n -2 Verification license does not match claim.\n \"\"\"\n\n status = 0\n \n claim = metadata(filename).getClaim()\n if claim is None:\n raise cctag.exceptions.NotLicensedException\n \n fileinfo = parseClaim(claim)\n fileinfo['sha'] = 'urn:sha1:%s' % cctag.rdf.fileHash(filename)\n\n verifyRdf = rdfextract.RdfExtractor().extractRdfText(\n rdfextract.retrieveUrl(fileinfo['verify at'])\n )\n\n # check if we found any RDF at all, and update the status code\n if len(verifyRdf) > 0:\n status = -1\n\n # check each block of RDF\n # (a verification page may also have it's own license RDF embedded)\n for block in verifyRdf:\n # parse/validate the RDF\n verifyCc = ccrdf.ccRdf()\n verifyCc.parse(block)\n\n # for each work in the RDF block...\n for work in verifyCc.works():\n \n # if the subject matches...\n if work.subject == fileinfo['sha']:\n # we found the work information;\n # only one reason left to not verify\n status = -2\n \n # we found the work, now make sure the license matches\n for license in work.licenses():\n if license == fileinfo['license']:\n return 1\n\n # either the file wasn't found, or the license didn't match\n return status\n","sub_path":"spotlight/Source/cctagutils/lookup.py","file_name":"lookup.py","file_ext":"py","file_size_in_byte":2618,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"185838878","text":"from Problem3 import mcdm\r\nfrom Problem3.graphrank import *\r\nimport requests\r\n\r\n# {'id': 1, 'distanceCityLink': 73.15100000000001, 'distancePosLaju': 39.28, 'distanceGdex': 76.682, 'distanceJnT': 66.071, 'distanceDHL': 58.685, 'shortestDistance': 'Pos Laju', 'distanceShortest': 39.28, 'customer': 1}\r\n\r\ndef dist_adapter(resp):\r\n if resp.status_code != 200:\r\n print(\"Nothing fishy here...\")\r\n return None\r\n raw = resp.json()\r\n return {'City-Link Express': raw['distanceCityLink'], 'DHL': raw['distanceDHL'], 'GDEX': raw['distanceGdex'], 'J&T': raw['distanceJnT'], 'Pos Laju': raw['distancePosLaju']}\r\n\r\ndef prob3():\r\n r = requests.get('http://algoprojq1.herokuapp.com/api/distance/1') \r\n distance = dist_adapter(r) or {'City-Link Express': 30, 'DHL': 80, 'GDEX': 55, 'J&T': 63, 'Pos Laju': 70}\r\n try:\r\n from Problem2 import prob2\r\n semantic = prob2(gimme_senti=True)\r\n except:\r\n semantic = {'City-Link Express': 8.5, 'DHL': 6.4, 'GDEX': 9.3, 'J&T': 1.87, 'Pos Laju': 4.86}\r\n \r\n # lists from dict for mcdm\r\n courier_company = list(semantic.keys())\r\n distance_list = list(distance.values())\r\n semantic_list = list(semantic.values())\r\n # multi criteria decision making\r\n mcdm.min_normalize(distance_list)\r\n mcdm.max_normalize(semantic_list)\r\n mcdm_list = mcdm.mcdm_weighted_sum(distance_list, 0.5, semantic_list, 0.5) # result\r\n mcdm_dict = dict(zip(courier_company, mcdm_list))\r\n print(\"multi criteria decision making result: \", mcdm_dict)\r\n # plot graph for most to least recommended ranking\r\n return graphrank.plot_graph_rank(mcdm_dict)\r\n","sub_path":"Problem3/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":1639,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"523448772","text":"# -*- coding:utf-8 -*-\n# @Time : 2019-08-18 11:33\n# @Author : 胡远\n# @Github : https://github.com/QuixoteHY\n# @Email : 1290482442@qq.com\n# @Describe :\n\nimport asyncio\n\nfrom aiohttp import web\n\nfrom common.logger import logger\n\nfrom data_server_interface.python3.settings import SERVER_HOST, SERVER_PORT, HEARTBEAT_INTERVAL\nfrom data_server_interface.python3.controller import Controller\n\n\ndef _heartbeat(loop, interval):\n try:\n logger.info('心跳:'+str(interval)+'s')\n except Exception as e:\n logger.info(logger.exception(e))\n loop.call_later(interval, _heartbeat, loop, interval)\n\n\ndef start_heartbeat(loop, interval):\n _heartbeat(loop, interval)\n\n\nclass MainHandler:\n def __init__(self):\n self.controller = Controller()\n\n async def get_stock_info(self, request):\n remote = request.remote\n logger.info(str(remote))\n try:\n financial_statement_type = request.match_info.get('financial_statement_type', '')\n ts_code = request.match_info.get('ts_code', '')\n if not ts_code:\n return web.json_response({'status': 'no ts_code in your request url'})\n if financial_statement_type == 'balance_sheet':\n self.controller.get_balance_sheet(ts_code)\n return web.json_response({})\n elif financial_statement_type == 'fina_indicators':\n return web.Response(body=self.controller.get_fina_indicators(ts_code).encode('utf-8'),\n content_type='text/html')\n except Exception as e:\n logger.info(logger.exception(e))\n return web.json_response({'status': 'error in server'})\n\n\nasync def init(loop):\n app = web.Application(loop=loop)\n handler = MainHandler()\n #\n # 获取某上市公司资产负债表信息\n # http://127.0.0.1:8888/get/stock/{financial_statement_type}/{ts_code}\n app.router.add_get('/get/stock/{financial_statement_type}/{ts_code}', handler.get_stock_info)\n #\n # server = await loop.create_server(app.make_handler(), SERVER_HOST, SERVER_PORT)\n # logger.info('\\n\\tServer started at http://%s:%s...' % (SERVER_HOST, SERVER_PORT))\n # return server\n runner = web.AppRunner(app)\n await runner.setup()\n site = web.TCPSite(runner, SERVER_HOST, SERVER_PORT)\n logger.info('\\n\\tServer started at http://%s:%s...' % (SERVER_HOST, SERVER_PORT))\n await site.start()\n\nif __name__ == '__main__':\n _loop = asyncio.get_event_loop()\n _loop.run_until_complete(init(_loop))\n start_heartbeat(_loop, HEARTBEAT_INTERVAL)\n _loop.run_forever()\n\n","sub_path":"data_server_interface/python3/server_stock.py","file_name":"server_stock.py","file_ext":"py","file_size_in_byte":2623,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"531210236","text":"import pygame, random\nimport pygame, sys\nfrom pygame.locals import *\nfrom input_mech import *\nfrom random import randint\n\t\nclass Sprite:\n def load(self, file, x, y, w, h, fx1, fy1, fx2, fy2):\n self.sIMG = pygame.image.load(file)\n self.sX = x\n self.sY = y\n self.sW = w\n self.sH = h\n self.sFX1 = fx1 * w\n self.sFY1 = fy1 * h\n self.sFX2 = fx2 * w\n self.sFY2 = fy2 * h\n self.sIMG.set_clip(pygame.Rect(self.sFX1, self.sFY1, self.sW, self.sH))\n self.sF1 = self.sIMG.subsurface(self.sIMG.get_clip())\n self.sIMG.set_clip(pygame.Rect(self.sFX2, self.sFY2, self.sW, self.sH))\n self.sF2 = self.sIMG.subsurface(self.sIMG.get_clip())\n self.sFC = self.sF1\n self.animationint = 0\n self.collisionrect = pygame.Rect(self.sX, self.sY, self.sW, self.sH)\n self.clicked = False\n self.hovered = False\n return self.sIMG, self.sX, self.sY, self.sW, self.sH, self.sFX1, self.sFY1, self.sFX2, self.sFY2, self.sF1, self.sF2, self.sFC, self.animationint, self.collisionrect, self.clicked, self.hovered\n\n def update(self, mouseDown, mouseX, mouseY):\n self.collisionrect = pygame.Rect(self.sX, self.sY, self.sW, self.sH)\n if mouseClick(mouseDown, mouseX, mouseY, self.collisionrect):\n self.clicked = True\n else:\n self.clicked = False\n if mouseHover(mouseX, mouseY, self.collisionrect):\n self.hovered = True\n else:\n self.hovered = False\n def draw(self, confirmation, ds):\n if confirmation:\n if self.animationint > 30:\n self.animationint = 0\n self.animationint += 1\n if self.animationint < 15:\n self.sFC = self.sF1\n if self.animationint >= 15:\n self.sFC = self.sF2\n ds.blit(self.sFC, (self.sX, self.sY))\n\nclass GUI_Button(Sprite):\n def lGui(self, file, x, y, w, h, fx1, fy1, fx2, fy2):\n Sprite.load(self, file, x, y, w, h, fx1, fy1, fx2, fy2)\n self.originX = x\n self.originY = y\n \n def uGui(self, mouseDown, mouseX, mouseY):\n ''' WIP: Hover Text\n black = pygame.Color(0, 0, 0)\n fontObj = pygame.font.Font('freesansbold.ttf', 12)\n self.gText = fontObj.render('This is a button', True, black)\n self.gTextRect = self.gText.get_rect()\n self.gTextRect.y = self.sY + 0.5 * self.sH\n self.gTextRect.x = self.sX + 0.5 * self.sW\n '''\n if self.clicked:\n self.sFC = self.sF2\n else:\n self.sFC = self.sF1\n \n '''if self.hovered:\n self.sX = self.originX + 10\n else:\n self.sX = self.originX'''\n \n Sprite.update(self, mouseDown, mouseX, mouseY)\n def dGui(self, ds):\n Sprite.draw(self ,False , ds)\n '''\n if self.hovered:\n ds.blit(self.gText, self.gTextRect)\n '''\n\n","sub_path":"Python Game Development/sprite_classes.py","file_name":"sprite_classes.py","file_ext":"py","file_size_in_byte":3006,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"547069641","text":"import json\n\"\"\"\ndic = {}\nwith open(\"sample.json\",\"w\") as fw:\n with open(\"sample.csv\",\"r\") as fr:\n ls = [i.rstrip() for i in fr]\n ls = ls[0].split(\",\")+ls[1].split(\",\")\n for i in range(int(len(ls)/2)):\n dic[ls[i]] = ls[i+5]\n fw.write(str(dic))\n\"\"\"\n\nwith open(\"sample.json\",\"r\") as fr:\n header = fr.readline().strip().split(\",\")\n val = fr.readline().strip().split(\",\")\n\nd = {}\nfor i in range(len(header)):\n k = header[i]\n v = val[i]\n d[k] = v\n\nprint(d)\n","sub_path":"week_3/home_work.d/csvTojson.py","file_name":"csvTojson.py","file_ext":"py","file_size_in_byte":504,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"102329610","text":"\"\"\"\nEntechts spider created on the top of ATSSpider\n\nscrapy crawl entechts -a mining_job_id=9999 -a iteration=1 -a extract=1 -a url=\"http://www.entechts.com/Jobs.aspx?Keyword=a\"\n\nSample URL:\n http://www.entechts.com/Jobs.aspx?Keyword=a\n\"\"\"\n\nfrom re import compile\nfrom scrapy.http import Request\nfrom scrapy.selector import Selector\nfrom urlparse import urljoin\n\nfrom brightcorp.base.atsspiders import ATSSpider\nfrom brightcorp.items import BrightcorpItemLoader\nfrom brightcorp.processors import Prefix, RemoveBadElements, UrlJoin\n\npattern = {\n 'count': compile(r'(\\d+)\\s*jobs\\s*match'),\n 'ref_id': compile(r'JOBID=(\\d+)'),\n}\n\n\nclass Entechts(ATSSpider):\n\n name = 'entechts'\n logo_url = ''\n\n def parse(self, response):\n sel = Selector(response)\n # set expected job count\n if not self.expected_job_count_set:\n expected_count = sel.xpath(\n '//ul/li/h3[contains(text(), \"jobs match\")]/text()'\n ).extract()\n if expected_count:\n match = pattern['count'].search(expected_count[0])\n if match:\n self.expected_job_count = expected_count[0]\n if not self.logo_url:\n self.logo_url = sel.xpath(\n '//div[@id=\"logo\"]/div/a/img/@src'\n ).extract()\n\n for div in sel.xpath(\n '//div/div[@class=\"job\"]'\n ):\n href = div.xpath(\n './div[@class=\"summary\"]/a/@href'\n ).extract()\n if href:\n yield Request(\n callback=self.parse_job_callback(),\n meta={\n 'company': div.xpath(\n './div[@class=\"client\"]/text()'\n ).extract(),\n 'location': div.xpath(\n './div[@class=\"location\"]/text()'\n ).extract(),\n 'jobcategory': div.xpath(\n './div[@class=\"sector\"]/text()'\n ).extract(),\n 'jobtype': div.xpath(\n './div[@class=\"type\"]/text()'\n ).extract(),\n 'baseSalary': div.xpath(\n './div[@class=\"rate\"]/text()'\n ).extract(),\n 'title': div.xpath(\n './div[@class=\"summary\"]/a/text()'\n ).extract(),\n },\n url=urljoin(response.url, href[0])\n )\n\n # pagination\n next_page = sel.xpath(\n '//div[contains(@class, \"pagination-btm\")]/a[@class=\"next-arrow right\"]/@href'\n ).extract()\n if next_page:\n yield Request(\n callback=self.parse,\n url=next_page[0]\n )\n\n def parse_job(self, response):\n \"\"\"\n Extract all required information.\n \"\"\"\n sel = Selector(response)\n\n loader = BrightcorpItemLoader(selector=sel)\n\n loader.add_value(\n 'title', response.meta.get('title')\n )\n loader.add_value(\n 'location', response.meta.get('location')\n )\n loader.add_value(\n 'company', response.meta.get('company')\n )\n loader.add_value(\n 'referencenumber',\n response.url,\n Prefix('%s-' % self.name),\n re=pattern['ref_id']\n )\n loader.add_value('url', response.url)\n loader.add_xpath(\n 'description',\n '//div/div[@class=\"details\"]',\n RemoveBadElements(['img', ])\n )\n loader.add_value(\n 'baseSalary', response.meta.get('baseSalary')\n )\n loader.add_value(\n 'jobtype', response.meta.get('jobtype')\n )\n loader.add_value(\n 'jobcategory', response.meta.get('jobcategory')\n )\n loader.add_value(\n 'logo_url',\n self.logo_url,\n UrlJoin(response.url)\n )\n loader.add_value('apply_url', response.url)\n\n yield loader.load_item()\n","sub_path":"brightcorp/brightcorp/spiders/entechts.py","file_name":"entechts.py","file_ext":"py","file_size_in_byte":4182,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"348935834","text":"from django.contrib import admin\nfrom .models import Question, Choice\n# Register your models here.\n\nclass ChoiceInline(admin.TabularInline):\n model = Choice\n extra = 3\n\nclass QuestionAdmin(admin.ModelAdmin):\n fieldsets = [\n ('Date Info', {'fields': ['published_on']}),\n (None, {'fields': ['body']}),\n ]\n inlines = [ChoiceInline]\n\nadmin.site.register(Question, QuestionAdmin)\n","sub_path":"app/polls/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":404,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"77437371","text":"#!/usr/bin/env python\n# -*- coding: utf-8\nimport glob\nimport os\n\nimport markdown\n\nfrom scriptlib import config, const\n\n\ndef read_file(filename):\n with open(filename) as fh:\n return fh.read().decode(\"utf-8\")\n\n\ndef write_file(filename, data):\n with open(filename, \"w\") as fh:\n fh.write(data)\n\n\ndef scan_dir(path):\n markdowns = []\n for filename in glob.glob(\"%s/*.md\" % path):\n markdowns.append(read_file(filename))\n return markdowns\n\n\ndef build_chapter(path):\n chapter = []\n if os.path.isfile(path):\n chapter.append(read_file(path))\n elif os.path.isdir(path):\n chapter.extend(scan_dir(path))\n return chapter\n\n\ndef filter_front_matter(chapter):\n \"\"\"\n Scan each chapter for Jekyll annotations at the beginning of each section\n and remove them.\n \"\"\"\n sections = []\n for section in chapter:\n counts = 0\n start = None\n lines = section.splitlines()\n for index, line in enumerate(lines):\n if counts == 2:\n start = index\n break\n if line == const.delimiter:\n counts += 1\n sections.append(\"\\n\".join(lines[start:]))\n return sections\n\n\ndef is_heading(key):\n if (key.startswith(\"#\") and\n not key.startswith(\"#B(\") and\n not key.startswith(\"#b(\") and\n not key.startswith(\"#(\") and\n not key.startswith(\"#Fun\")):\n return True\n return False\n\n\ndef get_anchor_name(heading):\n return heading.replace(\n '#', '').strip().replace(\n '.', '').replace(\n ' ', '_').replace(\n \"'\", '').replace(\n '`', '').lower()\n\n\ndef get_anchor(heading):\n return '' % get_anchor_name(heading)\n\n\ndef is_seen(key, seen):\n if seen.intersection([key]):\n return True\n return False\n\n\ndef remove_extra_headings(chapter, headings_only=False, include_anchors=True,\n as_string=True):\n \"\"\"\n Several chapters have their headings listed more than once (due to multiple\n markdown docs). This function removes all but the first one.\n \"\"\"\n sections = []\n seen = set()\n for section in chapter:\n filtered_section = []\n for line in section.splitlines():\n key = line.strip()\n if is_heading(key):\n if not is_seen(key, seen):\n if include_anchors:\n filtered_section.append(get_anchor(line))\n filtered_section.append(line)\n seen.add(key)\n elif not headings_only:\n filtered_section.append(line)\n if as_string:\n sections.append(\"\\n\".join(filtered_section))\n else:\n sections.extend(filtered_section)\n return sections\n\n\ndef assemble_headings(book_config):\n headings = []\n for chapter_location in book_config.chapters:\n chapter = build_chapter(chapter_location)\n headings.extend(\n remove_extra_headings(\n chapter, headings_only=True, include_anchors=False,\n as_string=False))\n return headings\n\n\ndef assemble_chapters(book_config, remove_front_matter=True):\n chapters = []\n for chapter_location in book_config.chapters:\n chapter = build_chapter(chapter_location)\n if remove_front_matter:\n chapter = filter_front_matter(chapter)\n chapter = remove_extra_headings(chapter)\n chapters.extend(chapter)\n return chapters\n\n\ndef assemble_book(book_config, remove_front_matter=True):\n chapters = [const.delimiter,\n \"layout: book\",\n \"title: %s\" % book_config.title,\n \"author: %s\" % \", \".join(book_config.authors),\n const.delimiter] + assemble_chapters(\n book_config, remove_front_matter)\n book = \"\\n\".join(chapters)\n return book.encode(\"utf-8\")\n\n\ndef generate_doc(book_config):\n markdown_data = assemble_book(book_config)\n write_file(book_config.md_file, markdown_data)\n\n\ndef generate_docs():\n for book_config in config.docs:\n generate_doc(book_config)\n","sub_path":"docs/v0.8/scriptlib/md.py","file_name":"md.py","file_ext":"py","file_size_in_byte":4119,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"275817611","text":"#!/usr/bin/env python\n\n\"\"\"\non any new /initialpose, do full rotation, then delay (to hone in amcl)\n\nfollow something ~15th pose in global path for all moves (about 0.3m away?)\n -maximum path length seems to be about 35*5 (45*5 max) for 2-3 meter path\n -(longer if more turns -- go for 15th or 20th pose, or max if less, should be OK)\n\nignore local path, except for determining if at goal or not\n\tif no recent local path, must be at goal: followpath = False, goalpose = true\n\nrequires dwa_base_controller, global path updated continuously as bot moves\n\n\"\"\"\n\n\nimport rospy, tf\nimport oculusprimesocket\nfrom nav_msgs.msg import Odometry\nimport math\nfrom nav_msgs.msg import Path\nfrom geometry_msgs.msg import PoseWithCovarianceStamped\nfrom actionlib_msgs.msg import GoalStatusArray\nfrom move_base_msgs.msg import MoveBaseActionGoal\n\nlistentime = 0.6 # 0.6 # allows odom + amcl to catch up\nnextmove = 0\nodomx = 0\nodomy = 0\nodomth = 0\ntargetx = 0\t\ntargety = 0\ntargetth = 0\nfollowpath = False\npathid = None\ngoalth = 0 \ninitialturn = False\nwaitonaboutface = 0\nminturn = math.radians(8) # (was 6) -- 0.21 minimum for pwm 255\nminlinear = 0.08 # was 0.05\nmaxlinear = 0.5\nlastpath = 0 # refers to localpath\ngoalpose = False\ngoalseek = False\nmeterspersec = 0.33 # linear speed TODO: get from java\ndegperms = 0.0857 # turnspeed TODO: get from java\ntfth = 0\nglobalpathposenum = 20 # just right\nlistener = None\n\n\ndef pathCallback(data): # local path\n\tglobal goalpose, lastpath\n\t\n\tlastpath = rospy.get_time()\n\tgoalpose = False\n\t\ndef globalPathCallback(data):\n\tglobal targetx, targety, targetth , followpath, pathid\n\t\n\tn = len(data.poses)\n\tif n < 5:\n\t\treturn\n\t\t\n\tif n-1 < globalpathposenum:\n\t\tp = data.poses[n-1] \n\telse:\n\t\tp = data.poses[globalpathposenum]\n\t\n\ttargetx = p.pose.position.x\n\ttargety = p.pose.position.y\n\tquaternion = ( p.pose.orientation.x, p.pose.orientation.y,\n\tp.pose.orientation.z, p.pose.orientation.w )\n\ttargetth = tf.transformations.euler_from_quaternion(quaternion)[2]\n\t\n\tfollowpath = True\n\tpathid = data.header.seq\n\ndef odomCallback(data):\n\tglobal odomx, odomy, odomth\n\todomx = data.pose.pose.position.x\n\todomy = data.pose.pose.position.y\n\tquaternion = ( data.pose.pose.orientation.x, data.pose.pose.orientation.y,\n\tdata.pose.pose.orientation.z, data.pose.pose.orientation.w )\n\todomth = tf.transformations.euler_from_quaternion(quaternion)[2]\n\t\n\t# determine direction (angle) on map\n\tglobal tfth, listener\t \n\ttry:\n\t\t(trans,rot) = listener.lookupTransform('/map', '/odom', rospy.Time(0))\n\t\tquaternion = (rot[0], rot[1], rot[2], rot[3])\n\t\ttfth = tf.transformations.euler_from_quaternion(quaternion)[2]\n\texcept (tf.LookupException, tf.ConnectivityException, tf.ExtrapolationException):\n\t\tpass\t\n\ndef intialPoseCallback(data):\n\tif data.pose.pose.position.x == 0 and data.pose.pose.position.y == 0:\n\t\treturn\n\t# do full rotation on pose estimate, to hone-in amcl (if not docked)\n\trospy.sleep(0.5) # let amcl settle\n\toculusprimesocket.clearIncoming() # why?\n\toculusprimesocket.sendString(\"right 360\")\n\toculusprimesocket.waitForReplySearch(\" direction stop\")\n\n\t\ndef goalCallback(d):\n\tglobal goalth, goalpose, lastpath, initialturn, followpath, nextmove\n\n\t# to prevent immediately rotating wrongly towards new goal direction \n\tlastpath = rospy.get_time()\n\tgoalpose = False\n\n\t# set goal angle\n\tdata = d.goal.target_pose\n\tquaternion = ( data.pose.orientation.x, data.pose.orientation.y,\n\tdata.pose.orientation.z, data.pose.orientation.w )\n\tgoalth = tf.transformations.euler_from_quaternion(quaternion)[2]\n\tinitialturn = True\n\tfollowpath = False\n\tnextmove = lastpath + 2 # sometimes globalpath still points at previoius goal\n\t\ndef goalStatusCallback(data):\n\tglobal goalseek\n\tgoalseek = False\n\tif len(data.status_list) == 0:\n\t\treturn\n\tstatus = data.status_list[len(data.status_list)-1] # get latest status\n\tif status.status == 1:\n\t\tgoalseek = True\n\ndef move(ox, oy, oth, tx, ty, tth, gth):\n\tglobal followpath, goalpose, tfth, pathid, initialturn, waitonaboutface\n\tglobal odomx, odomy, odomth\n\n\tcurrentpathid = pathid\n\n\t# determine xy deltas for move\n\tdistance = 0\n\tif followpath:\n\t\tdx = tx - ox\n\t\tdy = ty - oy\t\n\t\tdistance = math.sqrt( pow(dx,2) + pow(dy,2) )\n\t\n\tgoalrotate = False\n\tif distance > 0:\n\t\tth = math.acos(dx/distance)\n\t\tif dy <0:\n\t\t\tth = -th\n\telif goalpose:\n\t\tth = gth - tfth\n\t\tgoalrotate = True\n\telse:\n\t\tth = tth\n\t\n\t# determine angle delta for move\n\tdth = th - oth\n\tif dth > math.pi:\n\t\tdth = -math.pi*2 + dth\n\telif dth < -math.pi:\n\t\tdth = math.pi*2 + dth\n\t\t\n\t# force minimums\t\n\tif distance > 0 and distance < minlinear:\n\t\tdistance = minlinear\n\t\t\n\tif distance > maxlinear:\n\t\tdistance = maxlinear\n\n\t# supposed to reduce zig zagging (was 0.3)\n\tif dth < minturn*0.5 and dth > -minturn*0.5:\n\t\tdth = 0\n\telif dth >= minturn*0.5 and dth < minturn:\n\t\tdth = minturn\n\telif dth <= -minturn*0.5 and dth > -minturn:\n\t\tdth = -minturn\n\n\toculusprimesocket.clearIncoming()\n\n\t# if turning more than 120 deg, inch forward, make sure not transient obstacle (like door transfer)\n\tif abs(dth) > 2.0944 and not goalrotate and not initialturn and waitonaboutface < 1: \n\t\toculusprimesocket.sendString(\"forward 0.25\")\n\t\toculusprimesocket.waitForReplySearch(\" direction stop\")\n\t\twaitonaboutface += 1 # only do this once\n\t\trospy.sleep(1)\n\t\treturn\n\t\t\n\twaitonaboutface = 0\n\n\tif not pathid == currentpathid:\n\t\treturn\n\n\tif dth > 0:\n\t\toculusprimesocket.sendString(\"left \" + str(int(math.degrees(dth))) ) \n\t\toculusprimesocket.waitForReplySearch(\" direction stop\")\n\telif dth < 0:\n\t\toculusprimesocket.sendString(\"right \" +str(int(math.degrees(-dth))) )\n\t\toculusprimesocket.waitForReplySearch(\" direction stop\")\n\n\tif distance > 0:\n\t\toculusprimesocket.sendString(\"forward \"+str(distance))\n\t\trospy.sleep(distance/meterspersec)\n\t\tinitialturn = False\n\n\t# if goalrotate:\n\t\t# rospy.sleep(1) \n\t\t\n\t\t\t\ndef cleanup():\n\t# oculusprimesocket.sendString(\"move stop\")\n\t# oculusprimesocket.sendString(\"state delete navigationenabled\")\n\toculusprimesocket.sendString(\"log global_path_follower.py disconnecting\") \n\n\n# MAIN\n\n# rospy.init_node('dwa_base_controller', anonymous=False)\nrospy.init_node('global_path_follower', anonymous=False)\nlistener = tf.TransformListener()\noculusprimesocket.connect()\nrospy.on_shutdown(cleanup)\n\nrospy.Subscriber(\"odom\", Odometry, odomCallback)\nrospy.Subscriber(\"move_base/DWAPlannerROS/local_plan\", Path, pathCallback)\nrospy.Subscriber(\"move_base/goal\", MoveBaseActionGoal, goalCallback)\nrospy.Subscriber(\"move_base/status\", GoalStatusArray, goalStatusCallback)\nrospy.Subscriber(\"move_base/DWAPlannerROS/global_plan\", Path, globalPathCallback)\nrospy.Subscriber(\"initialpose\", PoseWithCovarianceStamped, intialPoseCallback)\n\noculusprimesocket.sendString(\"log global_path_follower.py connected\") \n# oculusprimesocket.sendString(\"state odomturndpms \"+str(degperms)) # degrees per ms \n# oculusprimesocket.sendString(\"state odomturnpwm 100\") # approx starting point smooth floor\n# oculusprimesocket.sendString(\"state odomlinearmpms \"+str(meterspersec/1000)) \n# oculusprimesocket.sendString(\"state odomlinearpwm 150\") # approx starting point\n\n# oculusprimesocket.sendString(\"speed \"+str(linearspeed) )\n\nwhile not rospy.is_shutdown():\n\tt = rospy.get_time()\n\t\n\tif t >= nextmove:\n\t\t# nextmove = t + listentime\n\t\tif goalseek and (followpath or goalpose): \n\t\t\tmove(odomx, odomy, odomth, targetx, targety, targetth, goalth) # blocking\n\t\t\tnextmove = rospy.get_time() + listentime\n\t\t\tfollowpath = False\n\t\n\tif t - lastpath > 3:\n\t\tgoalpose = True\n\t\n\trospy.sleep(0.01)\n\t\n","sub_path":"src/global_path_follower.py","file_name":"global_path_follower.py","file_ext":"py","file_size_in_byte":7481,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"314010037","text":"import tensorflow as tf\n\nclass SignalCNN(object):\n # A CNN for signal regression.\n\n def __init__(self, signal_length, num_outputs, filter_sizes, num_filters):\n\n # Placeholders for input, output and dropout\n self.input_x = tf.placeholder(tf.float32, [None, 2, signal_length, 1], name=\"input_x\")\n self.input_y = tf.placeholder(tf.float32, [None, num_outputs], name=\"input_y\")\n self.dropout_keep_prob = tf.placeholder(tf.float32, name=\"dropout_keep_prob\")\n\n pooled_outputs = []\n\n for i, filter_size in enumerate (filter_sizes):\n with tf.name_scope(\"conv-maxpool-%s\" % filter_size):\n # Convolutional Layer\n conv1 = tf.layers.conv2d (\n inputs=self.input_x,\n filters=num_filters,\n kernel_size=[2, filter_size],\n padding='VALID',\n activation=tf.nn.relu)\n print ('conv1', conv1)\n\n # pool_size = tf.int32(filter_size/2)\n # Max-pooling over the outputs\n pooled1 = tf.layers.max_pooling2d(\n inputs=conv1,\n pool_size=[1, 2],\n strides=[1, 1],\n padding='VALID')\n pooled1 = tf.nn.dropout(pooled1, self.dropout_keep_prob)\n print ('pool1', pooled1)\n\n # Convolutional Layer\n conv2 = tf.layers.conv2d (\n inputs=pooled1,\n filters=num_filters,\n kernel_size=[1, filter_size],\n padding='VALID',\n activation=tf.nn.relu)\n print ('conv2', conv2)\n\n # Max-pooling over the outputs\n pooled2 = tf.layers.max_pooling2d(\n inputs=conv2,\n pool_size=[1, 2],\n strides=[1, 1],\n padding='VALID')\n pooled2 = tf.nn.dropout(pooled2, self.dropout_keep_prob)\n print ('pool2', pooled2)\n\n pooled2 = tf.contrib.layers.flatten(pooled2)\n pooled_outputs.append(pooled2)\n\n # Combine all the pooled features\n self.h_pool_flat = tf.concat(pooled_outputs, 1)\n print ('h_pool_flat', self.h_pool_flat)\n self.h_drop = tf.nn.dropout(self.h_pool_flat, self.dropout_keep_prob)\n \n with tf.name_scope(\"fully_connected_layer1\"):\n self.fclayer1 = tf.contrib.layers.fully_connected(self.h_drop, 8192, activation_fn=tf.nn.relu)\n self.h_drop1 = tf.nn.dropout(self.fclayer1, self.dropout_keep_prob)\n\n with tf.name_scope(\"fully_connected_layer2\"):\n self.fclayer2 = tf.contrib.layers.fully_connected(self.h_drop1, 4096, activation_fn=tf.nn.relu)\n self.h_drop2 = tf.nn.dropout(self.fclayer2, self.dropout_keep_prob)\n\n with tf.name_scope(\"fully_connected_layer3\"):\n self.fclayer3 = tf.contrib.layers.fully_connected(self.h_drop2, 2048, activation_fn=tf.nn.relu)\n self.h_drop3 = tf.nn.dropout(self.fclayer3, self.dropout_keep_prob)\n\n with tf.name_scope(\"fully_connected_layer4\"):\n self.fclayer4 = tf.contrib.layers.fully_connected(self.h_drop3, 1024, activation_fn=tf.nn.relu)\n self.h_drop4 = tf.nn.dropout(self.fclayer4, self.dropout_keep_prob)\n\n #with tf.name_scope(\"fully_connected_layer5\"):\n # self.fclayer5 = tf.contrib.layers.fully_connected(self.h_drop4, 512, activation_fn=tf.nn.relu)\n # self.h_drop5 = tf.nn.dropout(self.fclayer5, self.dropout_keep_prob)\n #\n # with tf.name_scope(\"fully_connected_layer6\"):\n # self.fclayer6 = tf.contrib.layers.fully_connected(self.h_drop5, 2500, activation_fn=tf.nn.relu)\n # self.h_drop6 = tf.nn.dropout(self.fclayer6, self.dropout_keep_prob)\n #\n # with tf.name_scope(\"fully_connected_layer7\"):\n # self.fclayer7 = tf.contrib.layers.fully_connected(self.h_drop6, 1250, activation_fn=tf.nn.relu)\n # self.h_drop7 = tf.nn.dropout(self.fclayer7, self.dropout_keep_prob)\n #\n # with tf.name_scope(\"fully_connected_layer8\"):\n # self.fclayer8 = tf.contrib.layers.fully_connected(self.h_drop7, 625, activation_fn=tf.nn.relu)\n # self.h_drop8 = tf.nn.dropout(self.fclayer8, self.dropout_keep_prob)\n\n with tf.name_scope(\"output_layer\"):\n self.predictions = tf.contrib.layers.fully_connected(self.h_drop4, num_outputs, activation_fn=None)\n \n with tf.name_scope(\"RMSE\"):\n self.rmse = tf.sqrt(tf.reduce_mean(tf.square(self.predictions - self.input_y)))\n\n with tf.name_scope(\"cost\"):\n self.cost = tf.reduce_mean(tf.square(self.predictions - self.input_y))\n tf.summary.scalar(\"cost\", self.cost)\n","sub_path":"mimic_cnn_class.py","file_name":"mimic_cnn_class.py","file_ext":"py","file_size_in_byte":4790,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"110700463","text":"from django.db import models\n\n# Create your models here.\nfrom zinnia.models_bases import entry\n\n\nclass Picture(models.Model):\n title = models.CharField(max_length=50)\n image = models.ImageField(upload_to='gallery')\n\n\nclass Gallery(models.Model):\n title = models.CharField(max_length=50)\n pictures = models.ManyToManyField(Picture)\n\n\nclass EntryGallery(\n entry.CoreEntry,\n entry.ContentEntry,\n entry.DiscussionsEntry,\n entry.RelatedEntry,\n entry.ExcerptEntry,\n entry.FeaturedEntry,\n entry.AuthorsEntry,\n entry.CategoriesEntry,\n entry.TagsEntry,\n entry.LoginRequiredEntry,\n entry.PasswordRequiredEntry,\n entry.ContentTemplateEntry,\n entry.DetailTemplateEntry\n):\n # image = models.ForeignKey(Picture)\n gallery = models.ForeignKey(Gallery)\n\n def __str__(self):\n return 'EntryGallery %s' % self.title\n\n class Meta(entry.CoreEntry.Meta):\n abstract = True\n","sub_path":"blogimages/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":926,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"640979712","text":"import datetime\n\nfrom django.db import models\nfrom django.utils import timezone\nfrom django.core.paginator import Paginator, EmptyPage, PageNotAnInteger\nfrom django import forms\n\nfrom modelcluster.contrib.taggit import ClusterTaggableManager\nfrom modelcluster.fields import ParentalKey\n\nfrom taggit.models import TaggedItemBase\n\nfrom wagtail.core.models import Page, Orderable\nfrom wagtail.core.fields import RichTextField\nfrom wagtail.admin.edit_handlers import FieldPanel, InlinePanel, MultiFieldPanel\nfrom wagtail.images.edit_handlers import ImageChooserPanel\nfrom wagtail.search import index\nfrom wagtail.snippets.models import register_snippet\nfrom wagtail.snippets.edit_handlers import SnippetChooserPanel\n\n@register_snippet\nclass BlogAuthor(models.Model):\n name = models.CharField(max_length=255)\n icon = models.ForeignKey(\n \"wagtailimages.Image\", null=True, blank=True,\n on_delete=models.SET_NULL, related_name=\"+\"\n )\n\n panels = [\n FieldPanel(\"name\"),\n ImageChooserPanel(\"icon\"),\n ]\n\n def __str__(self):\n return self.name\n\n class Meta:\n verbose_name_plural = \"blog authors\"\n\n# Create your models here.\nclass EventsIndexPage(Page):\n intro = RichTextField(blank=True)\n\n def get_context(self, request):\n # Update context to include only published posts, ordered by reverse-chron\n context = super().get_context(request)\n eventpages = self.get_children().live().order_by(\"-first_published_at\")\n context[\"eventpages\"] = eventpages\n return context\n\n content_panels = Page.content_panels + [\n FieldPanel(\"intro\"),\n ]\n\n# Still need to add event image field\nclass EventPage(Page):\n parent_page_types = ['blog.EventsIndexPage']\n gig_date = models.DateField(default=timezone.now)\n gig_time = models.CharField(default=\"19:30\", max_length=5)\n gig_location = models.CharField(default=\"Location TBA\", max_length=200)\n location_link = models.CharField(max_length=2083, default=\"https://www.uptheantics.co.uk\")\n ticket_link = models.CharField(max_length=2083, default=\"https://www.uptheantics.co.uk\")\n price = models.CharField(max_length=6, default=\"5\")\n intro = models.CharField(max_length=100)\n body = RichTextField(blank=True)\n representative_image = models.ForeignKey(\n 'wagtailimages.Image',\n null=True,\n blank=True,\n on_delete=models.SET_NULL,\n related_name='+'\n )\n\n\n search_fields = Page.search_fields + [\n index.SearchField(\"intro\"),\n index.SearchField(\"body\"),\n ]\n\n content_panels = Page.content_panels + [\n FieldPanel(\"gig_date\"),\n FieldPanel(\"gig_time\"),\n FieldPanel(\"gig_location\"),\n FieldPanel(\"price\"),\n MultiFieldPanel([\n FieldPanel(\"location_link\"),\n FieldPanel(\"ticket_link\"),\n ], heading=\"Relevant Links\"),\n FieldPanel(\"intro\"),\n FieldPanel(\"body\", classname=\"full\"),\n ImageChooserPanel('representative_image'),\n ]\n\nclass BlogIndexPage(Page):\n intro = RichTextField(blank=True)\n\n def get_context(self, request):\n # Update context to include only published posts, ordered by reverse-chron\n context = super().get_context(request)\n all_posts = self.get_children().live().type(BlogPage).order_by(\"-first_published_at\")\n recent_posts = all_posts[:10]\n context[\"recent_posts\"] = recent_posts\n context[\"all_posts\"] = all_posts\n subpage_types = [\"Blog\"]\n return context\n\nclass BlogArchives(Page):\n def get_context(self, request):\n context = super(BlogArchives, self).get_context(request)\n\n # Get the full unpaginated listing of resource pages as a queryset -\n blogpages = self.get_siblings(inclusive=False).live().order_by(\"-first_published_at\")[9:]\n\n paginator = Paginator(blogpages, 10) # Show 5 resources per page\n\n page = request.GET.get('page')\n try:\n blogpages = paginator.page(page)\n except PageNotAnInteger:\n # If page is not an integer, deliver first page.\n blogpages = paginator.page(1)\n except EmptyPage:\n # If page is out of range (e.g. 9999), deliver last page of results.\n blogpages = paginator.page(paginator.num_pages)\n\n # make the variable 'resources' available on the template\n context['blogpages'] = blogpages\n\n return context\n\n def __str__(self):\n return self.name\n\nclass BlogPage(Page):\n parent_page_types = ['blog.BlogIndexPage']\n date = models.DateField(default=timezone.now)\n intro = models.CharField(max_length=250)\n body = RichTextField(blank=True)\n author = models.ForeignKey(\n \"blog.BlogAuthor\",\n null=True,\n blank=True,\n on_delete=models.SET_NULL,\n related_name='+'\n )\n representative_image = models.ForeignKey(\n 'wagtailimages.Image',\n null=True,\n blank=True,\n on_delete=models.SET_NULL,\n related_name='+'\n )\n\n def main_image(self):\n gallery_item = self.gallery_images.first()\n if gallery_item:\n return gallery_item.image\n else:\n return None\n\n search_fields = Page.search_fields + [\n index.SearchField(\"intro\"),\n index.SearchField(\"body\"),\n ]\n\n content_panels = Page.content_panels + [\n MultiFieldPanel([\n SnippetChooserPanel(\"author\"),\n ], heading=\"Blog information\"),\n FieldPanel(\"intro\"),\n FieldPanel(\"body\", classname=\"full\"),\n InlinePanel(\"gallery_images\", label=\"Gallery Images\"),\n ImageChooserPanel('representative_image'),\n ]\n\nclass BlogPageGalleryImage(Orderable):\n page = ParentalKey(BlogPage, on_delete=models.CASCADE, related_name=\"gallery_images\")\n image = models.ForeignKey(\n \"wagtailimages.Image\", on_delete=models.CASCADE, related_name=\"+\"\n )\n caption = models.CharField(blank=True, max_length=250)\n\n panels = [\n ImageChooserPanel(\"image\"),\n FieldPanel(\"caption\")\n ]\n\nclass AboutPage(Page):\n date = models.DateField(default=timezone.now)\n intro = models.CharField(max_length=250)\n body = RichTextField(blank=True)\n\n def main_image(self):\n gallery_item = self.gallery_images.first()\n if gallery_item:\n return gallery_item.image\n else:\n return None\n\n content_panels = Page.content_panels + [\n MultiFieldPanel([\n FieldPanel(\"date\"),\n ], heading=\"Page information\"),\n MultiFieldPanel([\n FieldPanel(\"intro\"),\n FieldPanel(\"body\", classname=\"full\"),\n ], heading=\"Page Content\"),\n ]\n\n def child_pages(self):\n return AboutPage.objects.live().child_of(self).order_by(\"title\")\n\n# Does this need to take in Page as a param? It's just data being handed to About\nclass Bio(Page):\n name = models.CharField(max_length=100)\n description = RichTextField(blank=True)\n representative_image = models.ForeignKey(\n 'wagtailimages.Image',\n null=True,\n blank=True,\n on_delete=models.SET_NULL,\n related_name='+'\n )\n content_panels = Page.content_panels + [\n MultiFieldPanel([\n FieldPanel(\"name\"),\n FieldPanel(\"description\", classname=\"full\"),\n ImageChooserPanel('representative_image'),\n ], heading=\"Bio\")\n ]\n\n def __str__(self):\n return self.name\n\n class Meta:\n verbose_name_plural = \"bios\"\n","sub_path":"blog/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":7526,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"8432937","text":"from app.mod_blog.models import Blog\nimport pytest\nfrom app import app\nimport string\nimport random\n\n\n@pytest.fixture\ndef client(request):\n test_client = app.test_client()\n\n return test_client\n\n\ndef id_generator(size=10, chars=string.ascii_letters + string.digits):\n return ''.join(random.choice(chars) for _ in range(size))\n\n\ndef test_index(client):\n \"\"\"\n GIVEN a Flask test client\n WHEN the '/' page is requested (GET)\n THEN check the status code is valid\n \"\"\"\n response = client.get('/')\n assert response.status_code == 200\n\n\ndef test_new_article_instance():\n \"\"\"\n GIVEN a Blog Model\n WHEN a new Blog is created\n THEN check the title and description are defined correctly\n \"\"\"\n\n blog = Blog(\"New Article\", \"Article's description\")\n assert blog.title == \"New Article\"\n assert blog.body == \"Article's description\"\n\n\ndef test_adding_article_db():\n \"\"\"\n Given a Blog Model\n WHEN a new blog is added to the db\n THEN check the title and description are defined correctly from the DB\n \"\"\"\n id = id_generator()\n added_blog = Blog(f\"New Article {id}\", \"Article's description\").save()\n \n blog = Blog.query.filter_by(title=f\"New Article {id}\").first()\n\n assert added_blog.id == blog.id\n assert added_blog.title == blog.title\n assert added_blog.body == blog.body\n assert added_blog.date_created == blog.date_created\n assert added_blog.date_modified == blog.date_modified\n\n added_blog.remove()\n\n\ndef test_removing_article_db():\n \"\"\"\n GIVEN a Blog Model\n WHEN a blog article is removed from the db\n THEN check the article is no more in the db\n \"\"\"\n\n id = id_generator()\n added_blog = Blog(f\"Article being removed {id}\", \"Article's description\")\\\n .save()\n added_blog.remove()\n\n removed_blog = Blog.query.filter_by(title=f\"Article being removed {id}\")\\\n .first()\n\n assert removed_blog is None\n\n","sub_path":"tests/test_blog.py","file_name":"test_blog.py","file_ext":"py","file_size_in_byte":1924,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"456785953","text":"IN = [x.split(\"\\n\")[0] for x in open(\"input10.txt\", \"r\").readlines()]\r\nIN = [[char for char in IN[i]] for i in range(len(IN))]\r\n\r\n\r\ndef isInSight(pos1, pos2):\r\n points = abs(pgcd(pos2[0] - pos1[0], pos2[1] - pos1[1])) - 1\r\n if (points == -1 or points == 0):\r\n return 1\r\n increX = (pos2[0] - pos1[0]) / (points + 1)\r\n increY = (pos2[1] - pos1[1]) / (points + 1)\r\n\r\n cx = pos1[0]\r\n cy = pos1[1]\r\n for i in range(points):\r\n cx += increX\r\n cy += increY\r\n if (IN[int(cy)][int(cx)] != \".\"):\r\n return 0\r\n return 1\r\n\r\n\r\ndef pgcd(a, b):\r\n if b == 0:\r\n return a\r\n else:\r\n r = a % b\r\n return pgcd(b, r)\r\n\r\n\r\ndef toString():\r\n s = \"\"\r\n for y in range(h):\r\n for x in range(w):\r\n s += str(IN[y][x]) + \" \"\r\n s += \"\\n\"\r\n print(s)\r\n\r\n\r\nh = len(IN)\r\nw = len(IN[0])\r\n\r\ntoString()\r\n\r\nfor y in range(h):\r\n for x in range(w):\r\n if IN[y][x] != \".\":\r\n sum = 0\r\n for j in range(h):\r\n for i in range(w):\r\n if IN[j][i] != \".\" and (i != x or y != j):\r\n sum += isInSight((x, y), (i, j))\r\n IN[y][x] = str(sum)\r\n\r\nmax = 0\r\nfor y in range(h):\r\n for x in range(w):\r\n if IN[y][x] != \".\" and int(IN[y][x]) > max:\r\n max = int(IN[y][x])\r\n\r\ntoString()\r\n\r\nprint(max)","sub_path":"day10_2.py","file_name":"day10_2.py","file_ext":"py","file_size_in_byte":1372,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"484719151","text":"class Solution:\n def uniquePaths(self, m, n):\n \"\"\"\n :type m: int\n :type n: int\n :rtype: int\n \"\"\"\n dp = [[1] * (m + 1) for i in range(n + 1)]\n for i in range(2, n+1):\n for j in range(2, m+1):\n dp[i][j] = dp[i-1][j] + dp[i][j-1]\n return dp[-1][-1]\n\nif __name__ == \"__main__\":\n s = Solution()\n result = s.uniquePaths(7, 3)\n print(result)","sub_path":"0-100/62_unique_path.py","file_name":"62_unique_path.py","file_ext":"py","file_size_in_byte":428,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"169976740","text":"import pandas as pd\nfrom pandas.plotting import scatter_matrix\nimport matplotlib.pyplot as plt\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.linear_model import LogisticRegression\nfrom skl2onnx import convert_sklearn\nfrom skl2onnx.common.data_types import FloatTensorType\n\ndata = pd.read_csv('https://archive.ics.uci.edu/ml/machine-learning-databases/iris/iris.data', sep=\",\",\n names=['sepal length in cm', 'sepal width in cm', 'petal length in cm', 'petal width in cm',\n 'class'])\n\ndescription = data.describe()\nprint(description)\nscatter_matrix(data)\nfigure = plt.gcf()\nfigure.set_size_inches(10, 10)\nplt.savefig(\"../images/sample.png\", dpi=100)\n\narray = data.values\nx = array[:, 0:4]\ny = array[:, 4]\n\nx_train, x_test, y_train, y_test = train_test_split(x, y, test_size=0.2, random_state=2020)\n\nlr = LogisticRegression(penalty='l2', random_state=2020, solver='liblinear')\nlr.fit(x_train, y_train)\n\nscore = lr.score(x_test, y_test)\n\nprint('score:', score)\n\ninitial_type = [('float_input', FloatTensorType([None, 4]))]\nonx = convert_sklearn(lr, initial_types=initial_type)\nwith open(\"../model/rf_iris.onnx\", \"wb\") as f:\n f.write(onx.SerializeToString())\n","sub_path":"P_Iris/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1220,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"448794565","text":"import logging\nimport os\n\nfrom . import utils_config\nfrom . import utils_module_dir\nfrom . import utils_shell\n\nLOGGER = logging.getLogger(__name__)\n\n\ndef handle_pull_code(search_in=None, **ignore):\n utils_shell.execute('python2.7 -m devenv revert-patch %s' % (search_in or ''), shell=True)\n for module_dir, _ in utils_module_dir.search(search_in):\n code_dir = os.path.join(module_dir, 'code')\n if not os.path.exists(code_dir):\n continue\n module_config = utils_config.read(os.path.join(module_dir, 'supervisord.conf'))\n code_type = module_config.get_option_or_exit('code', 'type')\n if code_type == 'svn':\n utils_shell.execute('svn up', cwd=code_dir, shell=True)\n elif code_type == 'git':\n branch = module_config.get_option_or_exit('code', 'branch')\n utils_shell.execute('git pull origin %s' % branch, cwd=code_dir, shell=True)\n else:\n LOGGER.error('not supported version control: %s' % code_type)\n\n\n","sub_path":"devenv/handle_pull_code.py","file_name":"handle_pull_code.py","file_ext":"py","file_size_in_byte":1008,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"46596195","text":"import time\nimport traceback\nimport os\nfrom chatbase import Message, MessageSet, MessageTypes, InvalidMessageTypeError\n\n\nclass ChatBasePublisher:\n\n def __init__(self):\n self.api_key=os.environ[\"CHAT_BASE_API_KEY\"] # Chatbase API key\n self.platform = 'lord_lewin_chatbot' # Chat platform name\n # message_user = 'Do you know the time, please?' # User message\n # message_bot = 'It's 12 o'clock!' # Bot response message\n self.version = '1' # Bot version, useful if you want to mark them for A/B testing or compare results across versions\n # time_stamp = int(round(time.time() * 1e3)) # Mandatory\n\n def publish(self,userQuery,botResponse,not_handled,userId,intent):\n time_stamp = int(round(time.time() * 1e3)) # Mandatory\n \n # Create an instance of MessageSet to collect all the messages\n message_set = MessageSet(api_key=str(self.api_key), platform=self.platform,\n version=self.version, user_id=userId)\n # Create an instance of Message for the user message and set values in the constructor\n msg1 = Message(api_key=self.api_key, platform=self.platform, message=userQuery,\n intent=intent, version=self.version, user_id=userId,\n type=MessageTypes.USER, not_handled=not_handled,\n time_stamp=time_stamp)\n # msg1.set_as_feedback()\n\n # Create an instance of Message for the bot response message and set values in the constructor\n msg2 = Message(api_key=self.api_key, platform=self.platform, message=botResponse,\n version=self.version, user_id=userId,\n type=MessageTypes.AGENT)\n\n # Push messages into the collection (MessageSet)\n message_set.append_message(msg1)\n message_set.append_message(msg2)\n\n \n # Send the messages\n response = message_set.send()\n print('Response code for sending Chatbase Message', response.content)\n # response.status_code will be 200 if sending worked","sub_path":"docker/keras-chat-engine/app/analytics/chatbasepublisher.py","file_name":"chatbasepublisher.py","file_ext":"py","file_size_in_byte":2045,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"389839370","text":"# coding:utf-8\nfrom django.shortcuts import render_to_response\nfrom .models import *\n\n\ndef list(request):\n statue = \"列表展示页\"\n table_list = Server.objects.all()\n return render_to_response(\"list.html\", locals())\n\n\ndef content(request):\n statue = \"服务器详情页\"\n host_data = {\n \"host_name\": \"bian-PC\",\n \"ip\": \"192.168.0.14\",\n \"mac\": \"00-00-00-00-00-00-00-E0\",\n \"cpu\": \" Intel(R) Xeon(R) CPU E5-2650 v2 @ 2.60GHz\",\n \"mem\": \"12G\",\n \"disk\": \"500G\",\n \"system\": \"windows7\",\n \"model\": \"Thinkpad E431\"\n }\n return render_to_response(\"server_content.html\", locals())\n","sub_path":"ProjOMMP/Server/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":647,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"313638451","text":"def findMaxConsecutiveOnes(nums):\n max = 0\n temp = 0\n for n in nums:\n if n == 1:\n temp += 1\n if temp > max:\n max = temp\n else:\n temp = 0\n return max\n\n\nprint(findMaxConsecutiveOnes([1, 1, 0, 1, 1, 1]))\n","sub_path":"Python/Easy/maxConsecutiveOnes.py","file_name":"maxConsecutiveOnes.py","file_ext":"py","file_size_in_byte":255,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"501401736","text":"import ee\nfrom sepal.ee.image import select_and_add_missing\n\nfrom ..image_operation import ImageOperation\n\n\ndef mask_clouds(mosaic_def, collection):\n if not mosaic_def.mask_clouds:\n reduced = collection.select('cloud') \\\n .reduce(ee.Reducer.sum()\n .combine(ee.Reducer.count(), \"\", True)\n .combine(ee.Reducer.min(), \"\", True))\n # Proportion of pixels that are cloudy\n cloud_proportion = select_and_add_missing(reduced, ['cloud_sum']) \\\n .divide(select_and_add_missing(reduced, ['cloud_count']))\n # A representative proportion of pixels that are cloudy cloudy for the neighborhood\n normal_cloud_proportion = cloud_proportion.reproject(crs='EPSG:4326', scale=10000) \\\n .max(cloud_proportion.reproject(crs='EPSG:4326', scale=20000))\n # Measure of how a locations cloud proportion differs from the general area\n cloud_proportion_diff = cloud_proportion.subtract(normal_cloud_proportion)\n only_clouds = select_and_add_missing(reduced, ['cloud_min'])\n\n # When there is higher proportion of clouds than the normally, keep the clouds.\n # It's probably something (typically buildings) misclassified as clouds.\n # Also, don't trust the cloud classification enough to completely mask area with only clouds\n # Desert sand can be classified as cloud.\n keep_clouds = cloud_proportion_diff.gt(0.4).And(normal_cloud_proportion.lt(0.3))\n keep_clouds = keep_clouds.Or(only_clouds)\n else:\n keep_clouds = False\n\n return collection.map(lambda image: _MaskClouds(image, mosaic_def).apply(keep_clouds))\n\n\nclass _MaskClouds(ImageOperation):\n def __init__(self, image, mosaic_def):\n super(_MaskClouds, self).__init__(image)\n self.mosaic_def = mosaic_def\n\n def apply(self, keep_clouds):\n cloud_free = self.toImage('!i.cloud')\n buffer_meters = self.mosaic_def.cloud_buffer\n if buffer_meters:\n cloud_free = buffer_mask(self.toImage('!i.cloud'), buffer_meters).And(cloud_free)\n to_mask = self.image.select('toMask')\n if keep_clouds:\n mask = to_mask.Not().And(cloud_free.Or(keep_clouds))\n else:\n mask = to_mask.Not().And(cloud_free)\n return self.image.updateMask(mask)\n\n\ndef buffer_mask(mask, meters):\n cloud = mask.Not()\n min_cloud_radius = 50\n\n # Clouds with radius < min_cloud_radius will not have any inner pixels, and will not get buffered\n inner_pixel = mask \\\n .fastDistanceTransform(256, 'pixels').sqrt() \\\n .multiply(ee.Image.pixelArea().sqrt()) \\\n .gt(min_cloud_radius) \\\n .And(cloud)\n\n distance_to_inner_pixel = inner_pixel \\\n .fastDistanceTransform(256, 'pixels').sqrt() \\\n .multiply(ee.Image.pixelArea().sqrt())\n\n return distance_to_inner_pixel \\\n .lt(ee.Number(meters).add(min_cloud_radius)) \\\n .Or(cloud) \\\n .Not()\n","sub_path":"modules/google-earth-engine/docker/src/sepalinternal/mosaic/clouds.py","file_name":"clouds.py","file_ext":"py","file_size_in_byte":2951,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"633822408","text":"import sys, os, argparse\n\nimport job_queue\n\n\ndef parse_args(argv):\n parser = argparse.ArgumentParser(description='Submit job scripts')\n parser.add_argument('job_script', nargs='+')\n parser.add_argument('--array', '-a')\n return parser.parse_args(argv)\n\n\ndef main(argv):\n args = parse_args(argv)\n for job_script in args.job_script:\n queue = job_queue.get_job_queue(job_script)\n work_dir = os.path.dirname(job_script)\n job_id = queue.submit_job(job_script, work_dir=work_dir, array_idx=args.array)\n print(job_id)\n\n\nif __name__ == '__main__':\n main(sys.argv[1:])\n","sub_path":"submit_job.py","file_name":"submit_job.py","file_ext":"py","file_size_in_byte":608,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"408351033","text":"#! /usr/bin/env python\n\nfrom portals.permissions import permissions_check, superuser_required\n\nfrom django.contrib.auth.decorators import login_required\nfrom django.shortcuts import HttpResponse, render\n\nfrom .forms import VendorForm\nfrom .models import Vendor\n\n@login_required\n@permissions_check()\ndef index(request):\n\n vendor_find = []\n temp_name = \"cmdb/cmdb-header.html\"\n vendors = Vendor.objects.all()\n vendor_find = Vendor.objects.all()\n \n return render(request, 'cmdb/vendor.html', locals())\n\n@login_required\n@permissions_check()\ndef vendor_add(request):\n temp_name = \"cmdb/cmdb-header.html\"\n if request.POST:\n vendor_form = VendorForm(request.POST)\n if vendor_form.is_valid():\n vendor_form.save()\n tips = '厂商增加成功'\n display_control = ''\n else:\n tips = '厂商增加失败'\n display_control = ''\n return render(request, \"cmdb/vendor_add.html\", locals())\n else:\n display_control = 'none'\n vendor_form = VendorForm()\n return render(request, 'cmdb/vendor_add.html', locals())\n\n@login_required\n@permissions_check()\ndef vendor_edit(request, ids):\n status = 0\n\n obj = Vendor.objects.filter(id=ids)\n if len(obj) == 1:\n obj = obj[0]\n else:\n obj = None\n\n if request.method == 'POST':\n af = VendorForm(request.POST, instance=obj)\n if af.is_valid():\n af.save()\n status = 1\n else:\n status = 2\n else:\n af = VendorForm(instance=obj)\n\n return render(request, 'cmdb/vendor_edit.html', locals())\n\n\n@login_required()\n@permissions_check()\ndef vendor_del(request):\n vendor_id = request.GET.get('id', '')\n if vendor_id:\n Vendor.objects.filter(id=vendor_id).delete()\n\n if request.method == 'POST':\n vendor_batch = request.GET.get('arg', '')\n vendor_id_all = str(request.POST.get('vendor_id_all', ''))\n\n if vendor_batch:\n for vendor_id in vendor_id_all.split(','):\n vendor_item = HostGroup.objects.filter(id=vendor_id)\n if len(vendor_item) == 1:\n vendor_item = vendor_item[0]\n else:\n vendor_item = None\n vendor_item.delete()\n\n return HttpResponse('厂商删除成功')","sub_path":"cmdb/vendor.py","file_name":"vendor.py","file_ext":"py","file_size_in_byte":2339,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"109222506","text":"# 217 contains-duplicate/\nclass Solution(object):\n def containsDuplicate(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: bool\n \"\"\"\n d={}\n for x in nums:\n if d.has_key(x): return True\n d[x]=''\n\n return False\n","sub_path":"217.py","file_name":"217.py","file_ext":"py","file_size_in_byte":284,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"527012938","text":"from flask_wtf import CSRFProtect, FlaskForm, RecaptchaField\nfrom wtforms.fields import TextAreaField\nfrom wtforms.fields.html5 import EmailField\nfrom wtforms.validators import DataRequired, Email\n\ncsrf = CSRFProtect()\n\nclass SubscribeForm(FlaskForm):\n email = EmailField(\n label='Subscribe for the latest and greatest!',\n validators=[DataRequired(), Email()],\n id='subscribe-email',\n render_kw={'placeholder': 'Email'}\n )\n\nclass ContactForm(FlaskForm):\n sender = EmailField(\n validators=[DataRequired(), Email()],\n id='contact-sender',\n render_kw={'placeholder': 'Email'}\n )\n body = TextAreaField(\n validators=[DataRequired()],\n id='contact-body',\n render_kw={\n 'placeholder': 'Message',\n 'rows': '5'\n }\n )\n captcha = RecaptchaField()\n","sub_path":"app/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":856,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"515365491","text":"#!/usr/bin/env python\nimport sys\n\"\"\"\nThe 0/1 Knapsack Problem\n\nGiven a set of items, each with a weight and a value, and a\nknapsack with a max weight. What is the subset of items has\na the highest value without exceding the max weight of the\nknapsack.\n\nGiven a set of all items S and a subset S' which is optimal\nsubset given the knapsack weight.S' must still be optimal if\nan item is removed from it and S.\n\n\n\"\"\"\n\ndef knapsack(items, sack):\n \"\"\"\n items = [(v,w),(v,w)]\n sack = int\n \"\"\"\n A = [[0] * len(items) for x in range(sack+1)]\n for i in range(0, len(items)):\n for j in range(0, sack+1):\n if items[i][1] > j:\n A[j][i] = A[j][i-1]\n else:\n A[j][i] = max(A[j][i-1], A[j-items[i][1]][i-1] + items[i][0])\n return A\n\ndef recursive_knapsack(items, size):\n cache = {}\n def inner(items, size, totalItems, currentItem, cache):\n if currentItem >= totalItems or size <= 0:\n return 0\n key = (totalItems - currentItem -1, size)\n if key in cache:\n return cache[key]\n elif items[currentItem][1] > size:\n maxValue = inner(items, size, totalItems, currentItem+1, cache)\n else:\n maxValue = max(items[currentItem][0] + inner(items, size-items[currentItem][1], totalItems, currentItem+1, cache),\n inner(items, size, totalItems, currentItem+1, cache))\n\n cache[key] = maxValue\n return maxValue\n return inner(items, size, len(items), 0, cache)\n\ndef reconstruct(A, items):\n result = []\n j = len(A)-1\n i = len(A[j])-1\n while j > 1:\n if A[j][i] != A[j][i-1]:\n result.append(i+1)\n j -= items[i][1]\n i -= 1\n return result\n\ndef load_items(filename):\n \"\"\" Generate graph path from text file \"\"\"\n file = open(filename, 'r')\n # Map each line of test data to a line in the data list:\n data = [ [int(y) for y in x.rstrip().split(' ')] for x in file]\n return data\n\ndef print_result(result, reconstruct_flag=False):\n for i in range(len(result)-1, -1, -1):\n print(i,result[i])\n print(' '+' '.join([str(x) for x in range(1,len(items)+1)]))\n print(\"Optimal Value: %s\" % result[-1][-1])\n if reconstruct_flag == True:\n print(\"Items Chosen: %s\" % reconstruct(result, items))\n\nif __name__ == '__main__':\n items = [(3,4)\n ,(2,3)\n ,(4,2)\n ,(4,3)\n ]\n result = knapsack(items, 6)\n print_result(result, True)\n\n","sub_path":"dynamic/01knapsack.py","file_name":"01knapsack.py","file_ext":"py","file_size_in_byte":2508,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"391976240","text":"import improvedMunkres\r\nimport time\r\nimport numpy as np\r\nimport matplotlib.pylab as plt\r\nimport ver2\r\nimport pandas as pd\r\n\r\ncolors = ['r.-', 'g.-', 'b.-', 'y.-', 'c.-', 'm.-']\r\n\r\ndef change_N(Max, k, end,n_loop):\r\n\r\n data = [[], [], []]\r\n\r\n for N in range(100, end, 100):\r\n\r\n mode2 = 0\r\n mode3 = 0\r\n\r\n for loop in range(n_loop):\r\n\r\n s = np.random.rand(N, 2) * Max\r\n e = np.random.rand(N, 2) * Max\r\n\r\n start_time = time.time()\r\n improvedMunkres.improved_munkres(s, e, Max, N, k)\r\n print(\"%d---%s seconds ---\" % (N,time.time() - start_time))\r\n mode2 += time.time() - start_time\r\n\r\n start_time = time.time()\r\n ver2.improved_munkres(s, e, Max, N, k)\r\n print(\"%d---%s seconds ---\" % (N,time.time() - start_time))\r\n mode3 += time.time() - start_time\r\n\r\n data[0].append(N)\r\n data[1].append(mode2 / n_loop)\r\n data[2].append(mode3/n_loop)\r\n\r\n dataframe = pd.DataFrame(data)\r\n dataframe.to_csv(\"data.csv\")\r\n\r\ndef change_cluster(Max,N, start, end ,jump ,n_loop):\r\n\r\n data = [[],[]]\r\n\r\n for k in range(start, end, jump):\r\n\r\n result = 0\r\n\r\n s = np.random.rand(N, 2) * Max\r\n e = np.random.rand(N, 2) * Max\r\n\r\n s_time = time.time()\r\n result = ver2.improved_munkres(s, e, Max, N, k)\r\n time_result = time.time()-s_time\r\n\r\n data[0].append(k)\r\n data[1].append(result)\r\n\r\n print(k)\r\n\r\n dataframe = pd.DataFrame(data)\r\n dataframe.to_csv(\"err_data.csv\")\r\n\r\n\r\nchange_N(1, 100, 1100, 1)\r\n# change_cluster(1,1000,10,20,10,1)","sub_path":"fu.py","file_name":"fu.py","file_ext":"py","file_size_in_byte":1639,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"85993281","text":"import os\nimport logging\n\nif os.getenv('FLASK_ENV', 'prod') != 'prod':\n config_name = 'dev'\nelse:\n config_name = 'prod'\n\n\nsettings = {}\nwith open('instance/{0}.cfg'.format(config_name)) as f:\n for line in f:\n if line == '\\n':\n continue\n (key, val) = line.split('=')\n settings[key.strip()] = val.strip()\n\n\nlog_conversion = {'DEBUG': logging.DEBUG, 'INFO': logging.INFO}\nlog_level = log_conversion[settings['LOG_LEVEL']]\n\nfetch_wait_secs = 60 * 15 # 15 minutes","sub_path":"scraper/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":500,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"145440181","text":"#!usr/bin/env python\r\n# -*- coding: utf-8 -*-\r\n\r\nimport sys\r\nfrom PyQt5.QtWidgets import QDialog, QFileDialog\r\nfrom matplotlib import pyplot as plt\r\nfrom ui_figure import *\r\nfrom ccd_plot import CCDReader\r\n\r\nclass Dialog(QDialog):\r\n\r\n def browsecache(self):\r\n\r\n cachedir = QFileDialog.getExistingDirectory(self)\r\n\r\n self.ui.browsecacheline.setText(cachedir)\r\n\r\n return None\r\n\r\n def browsejson(self):\r\n\r\n jsondir = QFileDialog.getExistingDirectory(self)\r\n\r\n self.ui.browsejsonline.setText(jsondir)\r\n\r\n return None\r\n\r\n def browseoutput(self):\r\n\r\n outputdir = QFileDialog.getExistingDirectory(self)\r\n\r\n self.ui.browseoutputline.setText(outputdir)\r\n\r\n return None\r\n\r\n def __init__(self):\r\n\r\n super(Dialog, self).__init__()\r\n\r\n # set up the user interface from Qt Designer\r\n self.ui = Ui_Form()\r\n\r\n self.ui.setupUi(self)\r\n\r\n self.ui.browsecachebutton.clicked.connect(self.browsecache)\r\n\r\n self.ui.browsejsonbutton.clicked.connect(self.browsejson)\r\n\r\n self.ui.browseoutputbutton.clicked.connect(self.browseoutput)\r\n\r\n self.ui.plotbutton.clicked.connect(self.plot)\r\n\r\n self.ui.exitbutton.clicked.connect(self.exit_plot)\r\n\r\n def plot(self):\r\n\r\n cachedir = self.ui.browsecacheline.text()\r\n\r\n jsondir = self.ui.browsejsonline.text()\r\n\r\n outputdir = self.ui.browseoutputline.text()\r\n\r\n arccoords = self.ui.arccoordsline.text()\r\n\r\n hval = self.ui.hline.text()\r\n\r\n vval = self.ui.vline.text()\r\n\r\n drawmodelfit = self.ui.radiomodelfit.isChecked()\r\n\r\n drawmaskedobs = self.ui.radiomasked.isChecked()\r\n\r\n ccd_curves = CCDReader(h=int(hval), v=int(vval), cache_dir=str(cachedir),\r\n json_dir=str(jsondir), arc_coords=str(arccoords), output_dir=str(outputdir),\r\n drawmodelfit=drawmodelfit, drawmaskedobs=drawmaskedobs)\r\n return None\r\n\r\n def exit_plot(self):\r\n\r\n self.close()\r\n\r\n plt.close(\"all\")\r\n\r\n sys.exit(0)\r\n\r\n return None\r\n","sub_path":"build_qt.py","file_name":"build_qt.py","file_ext":"py","file_size_in_byte":2112,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"241957184","text":"from easydict import EasyDict as edict\n\n# initalization\n__C_FDST = edict()\ncfg_data = __C_FDST\n__C_FDST.DATASET = 'FDST'\n\n# dataset parameters\n__C_FDST.STD_SIZE = (1080, 1920)\n__C_FDST.TRAIN_SIZE = (360, 640)\n__C_FDST.DATA_PATH = ''\n__C_FDST.MEAN_STD = ([0.484614104033, 0.455819487572, 0.432390660048], [\n 0.23891659081, 0.229008644819, 0.226914435625])\n\n# standard data parameters\n__C_FDST.LABEL_FACTOR = 1\n__C_FDST.LOG_PARA = 100.\n\n# training parameters\n__C_FDST.TRAIN_BATCH_SIZE = 1\n__C_FDST.TRAIN_DOWNRATE = 3\n\n# validation parameters\n__C_FDST.VAL_BATCH_SIZE = 1\n","sub_path":"adacrowd/datasets/baselines/FDST/setting.py","file_name":"setting.py","file_ext":"py","file_size_in_byte":588,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"9951557","text":"from datetime import datetime\nfrom eve_sqlalchemy.tests import TestBaseSQL\nfrom eve.tests.utils import DummyEvent\nfrom eve import ETAG\n\n\nclass TestDeleteSQL(TestBaseSQL):\n\n def setUp(self, settings_file=None, url_converters=None):\n super(TestDeleteSQL, self).setUp(settings_file, url_converters)\n # Etag used to delete an item (a contact)\n self.etag_headers = [('If-Match', self.item_etag)]\n\n def test_unknown_resource(self):\n url = '%s%s/' % (self.unknown_resource_url, self.item_id)\n _, status = self.delete(url)\n self.assert404(status)\n\n def test_delete_from_resource_endpoint(self):\n r, status = self.delete(self.known_resource_url)\n self.assert204(status)\n r, status = self.parse_response(self.test_client.get(\n self.known_resource_url))\n self.assert200(status)\n self.assertEqual(len(r['_items']), 0)\n\n def test_delete_from_resource_endpoint_different_resource(self):\n r, status = self.delete(self.different_resource_url)\n self.assert204(status)\n r, status = self.parse_response(self.test_client.get(\n self.different_resource_url))\n self.assert200(status)\n self.assertEqual(len(r['_items']), 0)\n\n # deletion of 'users' will still lave 'contacts' untouched (same db\n # collection)\n r, status = self.parse_response(self.test_client.get(\n self.known_resource_url))\n self.assert200(status)\n self.assertEqual(len(r['_items']), 25)\n\n def test_delete_empty_resource(self):\n url = '%s%s/' % (self.empty_resource_url, self.item_id)\n _, status = self.delete(url)\n self.assert404(status)\n\n def test_delete_readonly_resource(self):\n _, status = self.delete(self.readonly_id_url)\n self.assert405(status)\n\n def test_delete_unknown_item(self):\n url = '%s%s/' % (self.known_resource_url, self.unknown_item_id)\n _, status = self.delete(url)\n self.assert404(status)\n\n def test_delete_if_match_missing(self):\n _, status = self.delete(self.item_id_url)\n self.assert403(status)\n\n def test_delete_if_match_disabled(self):\n self.app.config['IF_MATCH'] = False\n _, status = self.delete(self.item_id_url)\n self.assert204(status)\n\n def test_delete_ifmatch_bad_etag(self):\n _, status = self.delete(self.item_id_url,\n headers=[('If-Match', 'not-quite-right')])\n self.assert412(status)\n\n def test_delete(self):\n r, status = self.delete(self.item_id_url, headers=self.etag_headers)\n self.assert204(status)\n\n r = self.test_client.get(self.item_id_url)\n self.assert404(r.status_code)\n\n def test_delete_non_existant(self):\n url = self.item_id_url[:-5] + \"00000\"\n r, status = self.delete(url, headers=self.etag_headers)\n self.assert404(status)\n\n def test_delete_different_resource(self):\n r, status = self.delete(self.user_id_url,\n headers=[('If-Match', self.user_etag)])\n self.assert204(status)\n\n r = self.test_client.get(self.user_id_url)\n self.assert404(r.status_code)\n\n def test_delete_with_post_override(self):\n # POST request with DELETE override turns into a DELETE\n headers = [('X-HTTP-Method-Override', 'DELETE'),\n ('If-Match', self.item_etag)]\n r = self.test_client.post(self.item_id_url, data={}, headers=headers)\n self.assert204(r.status_code)\n\n def test_delete_subresource(self):\n _db = self.app.data.driver\n\n # create random person\n fake_person = self.test_sql_tables.People.\\\n from_tuple(self.random_people(1)[0])\n fake_person._created = datetime.now()\n fake_person._updated = datetime.now()\n _db.session.add(fake_person)\n _db.session.commit()\n fake_person_id = fake_person._id\n fake_invoice = self.test_sql_tables.Invoices(number=4)\n fake_invoice.people_id = fake_person._id\n fake_invoice._created = datetime.now()\n fake_invoice._updated = datetime.now()\n _db.session.add(fake_invoice)\n _db.session.commit()\n\n # grab parent collection count; we will use this later to make sure we\n # didn't delete all the users in the database. We add one extra invoice\n # to make sure that the actual count will never be 1 (which would\n # invalidate the test)\n response, status = self.get('invoices')\n invoices = len(response[self.app.config['ITEMS']])\n\n # verify that the only document retrieved is referencing the correct\n # parent document\n response, status = self.get('users/%s/invoices' % fake_person_id)\n person_id = response[self.app.config['ITEMS']][1]['people']['_id']\n self.assertEqual(person_id, fake_person_id)\n\n # delete all documents at the sub-resource endpoint\n response, status = self.delete('users/%s/invoices' % fake_person_id)\n self.assert204(status)\n\n # verify that the no documents are left at the sub-resource endpoint\n response, status = self.get('users/%s/invoices' % fake_person_id)\n self.assertEqual(len(response['_items']), 0)\n\n # verify that other documents in the invoices collection have not been\n # deleted\n response, status = self.get('invoices')\n self.assertEqual(len(response['_items']), invoices - 2)\n\n def test_delete_subresource_item(self):\n _db = self.app.data.driver\n\n # create random person\n fake_person = self.test_sql_tables.People.\\\n from_tuple(self.random_people(1)[0])\n fake_person._created = datetime.now()\n fake_person._updated = datetime.now()\n _db.session.add(fake_person)\n _db.session.commit()\n fake_person_id = fake_person._id\n fake_invoice = self.test_sql_tables.Invoices(number=4)\n fake_invoice.people_id = fake_person._id\n fake_invoice._created = datetime.now()\n fake_invoice._updated = datetime.now()\n _db.session.add(fake_invoice)\n _db.session.commit()\n fake_invoice_id = fake_invoice._id\n\n # GET all invoices by new contact\n response, status = self.get('users/%s/invoices/%s' %\n (fake_person_id, fake_invoice_id))\n etag = response[ETAG]\n\n headers = [('If-Match', etag)]\n response, status = self.delete('users/%s/invoices/%s' %\n (fake_person_id, fake_invoice_id),\n headers=headers)\n self.assert204(status)\n\n def delete(self, url, headers=None):\n r = self.test_client.delete(url, headers=headers)\n return self.parse_response(r)\n\n\nclass TestDeleteEvents(TestBaseSQL):\n\n def test_on_pre_DELETE_for_item(self):\n devent = DummyEvent(self.before_delete)\n self.app.on_pre_DELETE += devent\n self.delete_item()\n self.assertEqual('people', devent.called[0])\n self.assertFalse(devent.called[1] is None)\n\n def test_on_pre_DELETE_resource_for_item(self):\n devent = DummyEvent(self.before_delete)\n self.app.on_pre_DELETE_people += devent\n self.delete_item()\n self.assertFalse(devent.called is None)\n\n def test_on_pre_DELETE_for_resource(self):\n devent = DummyEvent(self.before_delete)\n self.app.on_pre_DELETE += devent\n self.delete_resource()\n self.assertFalse(devent.called is None)\n\n def test_on_pre_DELETE_resource_for_resource(self):\n devent = DummyEvent(self.before_delete)\n self.app.on_pre_DELETE_people += devent\n self.delete_resource()\n self.assertFalse(devent.called is None)\n\n def test_on_pre_DELETE_dynamic_filter(self):\n def filter_this(resource, request, lookup):\n lookup[\"_id\"] = self.unknown_item_id\n self.app.on_pre_DELETE += filter_this\n # Would normally delete the known document; will return 404 instead.\n r, s = self.parse_response(self.delete_item())\n self.assert404(s)\n\n def test_on_post_DELETE_for_item(self):\n devent = DummyEvent(self.after_delete)\n self.app.on_post_DELETE += devent\n self.delete_item()\n self.assertFalse(devent.called is None)\n\n def test_on_post_DELETE_resource_for_item(self):\n devent = DummyEvent(self.after_delete)\n self.app.on_post_DELETE_people += devent\n self.delete_item()\n self.assertFalse(devent.called is None)\n\n def test_on_post_DELETE_for_resource(self):\n devent = DummyEvent(self.after_delete)\n self.app.on_post_DELETE += devent\n self.delete_resource()\n self.assertFalse(devent.called is None)\n\n def test_on_post_DELETE_resource_for_resource(self):\n devent = DummyEvent(self.after_delete)\n self.app.on_post_DELETE_people += devent\n self.delete_resource()\n self.assertFalse(devent.called is None)\n\n def test_on_delete_resource(self):\n devent = DummyEvent(self.before_delete)\n self.app.on_delete_resource += devent\n self.delete_resource()\n self.assertEqual(('people',), devent.called)\n\n def test_on_delete_resource_people(self):\n devent = DummyEvent(self.before_delete)\n self.app.on_delete_resource_people += devent\n self.delete_resource()\n self.assertEqual(tuple(), devent.called)\n\n def test_on_deleted_resource(self):\n devent = DummyEvent(self.after_delete)\n self.app.on_deleted_resource += devent\n self.delete_resource()\n self.assertEqual(('people',), devent.called)\n\n def test_on_deleted_resource_people(self):\n devent = DummyEvent(self.after_delete)\n self.app.on_deleted_resource_people += devent\n self.delete_resource()\n self.assertEqual(tuple(), devent.called)\n\n def test_on_delete_item(self):\n devent = DummyEvent(self.before_delete)\n self.app.on_delete_item += devent\n self.delete_item()\n self.assertEqual('people', devent.called[0])\n self.assertEqual(\n self.item_id, devent.called[1][self.app.config['ID_FIELD']])\n\n def test_on_delete_item_people(self):\n devent = DummyEvent(self.before_delete)\n self.app.on_delete_item_people += devent\n self.delete_item()\n self.assertEqual(\n self.item_id, devent.called[0][self.app.config['ID_FIELD']])\n\n def test_on_deleted_item(self):\n devent = DummyEvent(self.after_delete)\n self.app.on_deleted_item += devent\n self.delete_item()\n self.assertEqual('people', devent.called[0])\n self.assertEqual(\n self.item_id, devent.called[1][self.app.config['ID_FIELD']])\n\n def test_on_deleted_item_people(self):\n devent = DummyEvent(self.after_delete)\n self.app.on_deleted_item_people += devent\n self.delete_item()\n self.assertEqual(\n self.item_id, devent.called[0][self.app.config['ID_FIELD']])\n\n def delete_resource(self):\n self.test_client.delete(self.known_resource_url)\n\n def delete_item(self):\n return self.test_client.delete(\n self.item_id_url, headers=[('If-Match', self.item_etag)])\n\n def before_delete(self):\n db = self.connection.session\n return db.query(self.test_sql_tables.People).\\\n get(self.item_id) is not None\n\n def after_delete(self):\n return not self.before_delete()\n","sub_path":"eve_sqlalchemy/tests/delete.py","file_name":"delete.py","file_ext":"py","file_size_in_byte":11528,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"148769815","text":"import numpy as np\nimport matplotlib.pyplot as plt\n\nbasisVectors = [[1.0,0.0,0.0],[0.0,1.0,0.0],[0.0,0.0,1.0]]\nminVectorComponent = 1.0\nmaxVectorComponent = 1.0\n\ndef getNIonsInRadius(radius,sieve):\n\tnIons = 0\n\tk = int (radius+1)\n\tfor index1 in range(-k,k+1):\n\t\tfor index2 in range(-k,k+1):\n\t\t\tfor index3 in range(-k,k+1):\n\t\t\t\tx,y,z = getCoordFromBasis(index1,index2,index3)\n\t\t\t\tif getDistanceFromOrigin(x,y,z)nIonsLast:\n\t\t\tnIonsInShell.append(nIonsCurr-nIonsLast)\n\t\t\tmadelungConstant = 0\n\t\t\tfor i in range (len(nIonsInShell)):\n\t\t\t\tmadelungConstant+=pow(-1,i+1)*nIonsInShell[i]/pow(i+1,0.5)\n\t\t\tx.append(nShell)\n\t\t\ty.append(madelungConstant)\n\t\t\tif printVal:\n\t\t\t\tprint ('Shell: {} Number of Ions: {} Radius: {} Madelung Constant: {}'.format(nShell,(nIonsCurr-nIonsLast),radius,madelungConstant))\n\t\t\tnShell+=1\n\t\t\tnIonsLast = nIonsCurr\n\t\t\tnIonsCurr = 0\n\n\treturn x,y\n\t\nif __name__ == \"__main__\":\n\tx,y = getMadelungConstants(20)\n\tfig = plt.figure()\n\tax = fig.add_subplot(111)\n\tax.plot(x,y)\n\tplt.title('Madelung Constant as a Function of Number of Shells') \n\tplt.xlabel('Number of Shells')\n\tplt.ylabel('Madelung Constant')\n\tplt.show()\n","sub_path":"Week1Excercise_MadelungSeries.py","file_name":"Week1Excercise_MadelungSeries.py","file_ext":"py","file_size_in_byte":1987,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"253849082","text":"#!/usr/bin/env python\n\nimport rospy\nimport json\nfrom numpy.random import randn \n\nfrom std_msgs.msg import String\nfrom geometry_msgs.msg import PoseWithCovarianceStamped\nfrom tf.transformations import quaternion_from_euler\n\nfrom humanoid_league_msgs.msg import Position2D\nfrom simulator.vision import Vision\n\n# The major concept needed to change is that the Robot does not cache the change of the pos of the ball, since the black board already cache the ball position. And in real time, it is the blackboard, or say the DSD to publish the positionInfo message\nclass Robot:\n def __init__(self, name):\n self._name = name\n\n # WARN the robot_pos stores the ground truth of the robot position\n self.robot_pos = Position2D()\n \n # no need to cache the subscriber\n rospy.Subscriber(\"/robots_pos\", String, self.rpos_callback)\n\n # but cache the publisher\n # publish the position with noise to topic /amcl_pose\n self.pos_pub = rospy.Publisher(\"/amcl_pose\", PoseWithCovarianceStamped, queue_size = 2)\n\n self.vision = Vision()\n\n def perform(self):\n self.vision.perform(self.robot_pos)\n self.pub_pose_with_noise()\n\n def pub_pose_with_noise(self):\n new_pos = PoseWithCovarianceStamped()\n # although I think it is important to fill all the blank in new_pos, but according to world_model_capsule, just fill \n # new_pos.header, \n # new_pos.pose.pose.x, \n # new_pos.pose.pose.y, \n # new_pos,pose.pose.orientation is ok\n new_pos.header.frame_id = \"map\"\n new_pos.header.stamp = rospy.Time.now()\n new_pos.pose.pose.position.x = self.add_random(self.robot_pos.pose.x)\n new_pos.pose.pose.position.y = self.add_random(self.robot_pos.pose.y)\n new_orient = quaternion_from_euler(0, 0, self.add_random(self.robot_pos.pose.theta))\n new_pos.pose.pose.orientation.x = new_orient[0]\n new_pos.pose.pose.orientation.y = new_orient[1]\n new_pos.pose.pose.orientation.z = new_orient[2]\n new_pos.pose.pose.orientation.w = new_orient[3]\n self.pos_pub(new_pos)\n \n def rpos_callback(self, inJson):\n inJson = inJson.data\n rpos_msg = json.loads(inJson)\n self.robot_pos.header.frame_id = rpos_msg['frame_id'].encode(\"utf-8\")\n secs = rpos_msg['stamp']['secs']\n nsecs = rpos_msg['stamp']['nsecs']\n self.robot_pos.header.stamp = rospy.Time(secs, nsecs)\n self.robot_pos.pose.x = rpos_msg[self._name]['x'] \n self.robot_pos.pose.y = rpos_msg[self._name]['y'] \n self.robot_pos.pose.theta = rpos_msg[self._name]['t'] \n self.robot_pos.confidence = rpos_msg[self._name]['c']\n rospy.loginfo(\"{} received x: {}, y: {}, theta{}\".format(self._name, rpos_msg[self._name]['x'], rpos_msg[self._name]['y'], rpos_msg[self._name]['t']))\n \n \"\"\"\n def bpos_callback(self, inJson):\n inJson = inJson.data\n bpos_msg = json.loads(inJson)\n self.ball_pos.header.frame_id = bpos_msg['frame_id'].encode(\"utf-8\")\n secs = bpos_msg['stamp']['secs']\n nsecs = bpos_msg['stamp']['nsecs']\n self.ball_pos.header.stamp = rospy.Time(secs, nsecs)\n self.ball_pos.ball_relative.x = bpos_msg['ball']['x'] \n self.ball_pos.ball_relative.y = bpos_msg['ball']['y'] \n self.ball_pos.confidence = bpos_msg['ball']['c'] \n \"\"\"\n\n def add_random(self, num):\n return num+float(randn(1))\n\ndef main():\n rospy.init_node('robot1')\n robot = Robot('robot1')\n rate = rospy.Rate(10)\n while not rospy.is_shutdown():\n robot.perform()\n rate.sleep()\nif __name__ == \"__main__\":\n main()\n","sub_path":"src/simulator/robot.py","file_name":"robot.py","file_ext":"py","file_size_in_byte":3673,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"113930622","text":"# https://uva.onlinejudge.org/index.php?option=com_onlinejudge&Itemid=8&page=show_problem&problem=989\n# Using Prim together with DFS/BFS. Complexity: O(t * s * log(c)) + O(t * q * (s + c))\nimport queue\nINF = float('inf')\nclass Node:\n def __init__(self, id, dist):\n self.id = id\n self.dist = dist\n def __lt__(self, other):\n return self.dist <= other.dist\n \n\ndef prim(source): # O (s * log(c))\n pq = queue.PriorityQueue()\n pq.put(Node(source, 0))\n dist[source] = 0\n while not pq.empty():\n top = pq.get()\n u = top.id\n visited[u] = True\n for neighbor in graph[u]:\n v = neighbor.id\n w = neighbor.dist\n if not visited[v] and dist[v] > w:\n dist[v] = w\n pq.put(Node(v, w))\n path[v] = u\n # print(path)\n # print(dist)\n for p in range(len(path)):\n if path[p] != -1:\n mst[p].append(Node(path[p], dist[p]))\n mst[path[p]].append(Node(p, dist[p]))\n return mst\n\n\ndef dfs(start, target, mst): # O (s + c)\n visited = [False] * len(mst)\n dist = [-INF for i in range(len(mst))] \n stack = []\n visited[start] = True\n stack.append(Node(start, 0))\n while len(stack) > 0:\n top = stack.pop()\n u = top.id\n if u == target:\n return dist[u]\n for v in mst[u]:\n if visited[v.id] == False:\n visited[v.id] = True\n stack.append(v)\n dist[v.id] = max(dist[u], v.dist)\n if dist[target] == -INF:\n return \"no path\"\n return dist[target]\n \n \ncase_no = 1\n\nwhile True:\n c, s, q = map(int, input().split())\n if c == 0 and s == 0 and q == 0:\n break\n if case_no != 1:\n print()\n graph = [[] for i in range(c)]\n \n for i in range(s):\n c1, c2, d = map(int, input().split())\n graph[c1 - 1].append(Node(c2 - 1, d))\n graph[c2 - 1].append(Node(c1 - 1, d))\n \n queries = []\n for j in range(q):\n q1, q2 = map(int, input().split())\n queries.append((q1 - 1, q2 - 1))\n\n print(\"Case #{}\".format(case_no))\n case_no += 1\n\n mst = [[] for i in range(c)]\n dist = [INF for i in range(c)]\n visited = [False for i in range(c)]\n path = [-1 for i in range(c)]\n # sum(si * log(ci)) ~ O(s * log(c))\n for i in range(c):\n if path[i] == -1:\n prim(i)\n # O (q * (s + c))\n for q1, q2 in queries:\n result = dfs(q1, q2, mst)\n print(result)\n\n\n\n# Using dynamic programming. Complexity: O(t * q * s * log(c))\nimport queue\nINF = float('inf')\nclass Node:\n def __init__(self, id, dist):\n self.id = id\n self.dist = dist\n def __lt__(self, other):\n return self.dist <= other.dist\n\ndef dp(start, target):\n dist = [INF for i in range(c)]\n# visited = [False for i in range(c)]\n pq = queue.PriorityQueue()\n pq.put(Node(start, 0))\n dist[start] = 0\n while not pq.empty():\n top = pq.get()\n u = top.id\n# visited[u] = True\n for neighbor in graph[u]:\n v = neighbor.id\n w = neighbor.dist \n if max(w, dist[u]) < dist[v]:\n dist[v] = max(w, dist[u])\n pq.put(Node(v, w)) \n return dist[target]\n\n\ncase_no = 1\n# O(t * q * c * log(s))\nwhile True:\n c, s, q = map(int, input().split())\n if c == 0 and s == 0 and q == 0:\n break\n if case_no != 1:\n print()\n graph = [[] for i in range(c)]\n \n for i in range(s):\n c1, c2, d = map(int, input().split())\n graph[c1 - 1].append(Node(c2 - 1, d))\n graph[c2 - 1].append(Node(c1 - 1, d))\n \n queries = []\n # O(q * c * log(s))\n for j in range(q):\n q1, q2 = map(int, input().split())\n queries.append((q1 - 1, q2 - 1))\n\n print(\"Case #{}\".format(case_no))\n case_no += 1\n for q1, q2 in queries:\n result = dp(q1, q2)\n if result == INF:\n print(\"no path\")\n else:\n print(result)\n\n","sub_path":"UVa/audiophobia.py","file_name":"audiophobia.py","file_ext":"py","file_size_in_byte":3614,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"617854766","text":"from flask import Flask, render_template, redirect, request, session, flash\n\napp = Flask(__name__)\napp.secret_key = 'mysecretkey'\n\n@app.route('/')\ndef index():\n return render_template(\"index.html\")\n\n\n@app.route('/users', methods =['POST'])\ndef users():\n if len(request.form['first']):\n return redirect('/')\n\n else:\n flash(\"success\")\n\n\n\n return render_template(\"userform.html\", name = name, favlocation = favlocation, comment= comment)\napp.run(debug=True)\n","sub_path":"Python/Flask_fun/counter/server1.py","file_name":"server1.py","file_ext":"py","file_size_in_byte":485,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"84303816","text":"\n\"\"\" Class description goes here. \"\"\"\n\n\"\"\"Entry point for standalone dataClay Execution Environment server.\n\nThe main can be called easily through a\n\n python -m dclay_server\n\"\"\"\n\nimport logging\n\n__author__ = 'Alex Barcelo '\n__copyright__ = '2015 Barcelona Supercomputing Center (BSC-CNS)'\n\nlogger = logging.getLogger(__name__)\n\n\n# We create a specific function that can be also run from importing the module (testing)\ndef run_main():\n # Current execution environment since they are initialized using environment variables and cannot be concurrently started in same host.\n from dataclay import initialize\n initialize()\n from dataclay.executionenv.server.ExecutionEnvironmentSrv import ExecutionEnvironmentSrv\n exec_env_srv = ExecutionEnvironmentSrv()\n exec_env_srv.start()\n\n \nif __name__ == \"__main__\":\n run_main()\n \n","sub_path":"src/dataclay/executionenv/server/__main__.py","file_name":"__main__.py","file_ext":"py","file_size_in_byte":867,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"198851284","text":"from matplotlib import cm\nfrom matplotlib import pyplot as plt\nfrom itertools import cycle, islice\nimport pandas as pd\nimport math\nimport numpy as np\ncolor = cm.inferno_r(np.linspace(.4,.8, 30))\n# np.random.seed(100)\n\n# def createTables(K):\n# tables = np.array([list(range(1,K+1)),list(range(K+1,2*K+1))])\n# return tables\n#\ndef transitionMatrix(K):\n A = np.zeros((2,2))\n A[0,0] += 0.25\n A[0,1] += 0.75\n A[1,0] += 0.75\n A[1,1] += 0.25\n\n\n return A\n\ndef emissionMatrix(K):\n# fair_emissions = np.ones((K,6)) * 1/6\n# bias_emissions = np.zeros((K,6))\n# bias_emissions[:,4] = 0.3\n# bias_emissions[:,5] = 0.7\n# fullEm = np.concatenate([fair_emissions, bias_emissions], axis=0)\n emissions = np.zeros((2,6))\n emissions[0,:] = 1/6 # fair dice state\n emissions[1,:][4:6] = 0.5 # bias dice state\n# return fair_emissions , bias_emissions, fullEm\n# print(emissions)\n return emissions\n\ndef createObservationSeq(table_seq,K, P):\n fair = 0\n bias = 1\n outcomes = []\n for i in range(K):\n hidden = np.random.choice([True, False], p=[1-P, P])\n state = ''\n if not hidden:\n if table_seq[i] == fair:\n die = np.random.multinomial(1, [1/6.]*6)\n state = 'fair'\n elif table_seq[i] == bias:\n# die = np.random.multinomial(1, [0.5] + [0.5] + [0]*4)# .\n die = np.random.multinomial(1, [0.4] + [0.2] + [0.2] + [0.1] + [0.05]+[0.05])\n state='bias'\n die = np.where(die>0)[0][0] + 1 # select the outcome of a dice throw.\n outcomes.append((die,state)) # append to list of outcomes for a player\n else:\n # ZERO INDICATES A HIDDEN OUTCOME, since a die cannot have outcome = 0\n if table_seq[i] == fair:\n die = np.random.multinomial(1, [1/6.]*6)\n state = 'fair_hidden'\n elif table_seq[i] == bias:\n# die = np.random.multinomial(1, [0.5] + [0.5] + [0]*4)# .\n die = np.random.multinomial(1, [0.4] + [0.2] + [0.2] + [0.1] + [0.05]+ [0.05])\n state='bias_hidden'\n die = np.where(die>0)[0][0] + 1 # select the outcome of a dice throw.\n outcomes.append((die,state)) # append to list of outcomes for a player\n\n df = pd.DataFrame(outcomes)\n df.columns = ['outcome', 'die']\n return df\n\ndef simulateTableSeq(K):\n start_prob = 0.5\n table_seq = []\n t = np.random.choice(2,1, p=[start_prob , 1 - start_prob])[0] # chosing first table is random\n table_seq.append(t)\n for i in range(K-1):\n if t == 0:\n # then more proble to shift table group after each throw.\n t = np.random.choice(2,1, p=[0.25, 0.75])[0]\n elif t ==1:\n t = np.random.choice(2,1, p=[0.75, 0.25])[0]\n table_seq.append(t)\n return table_seq\n\ndef simulateDice(K,players,P, distr):\n player_observations = []\n df = pd.DataFrame([])\n for i in range(players):\n if distr == 'mix':\n table_seq_vec = simulateTableSeq(K) # mixed , fiar and biased\n elif distr == 'bias':\n table_seq_vec = [1]*K # all biased\n elif distr == 'fair':\n table_seq_vec = [0]*K # all fair\n df_outcomes = createObservationSeq(table_seq_vec,K, P)\n df_outcomes['player'] = [i+1]*len(df_outcomes)\n df = pd.concat([df, df_outcomes])\n# print(df)\n return df\n\ndef plotObs(df_obs,players, distr):\n\n my_colors = [(0.2,0.4,0.5), (0.75, 0.25, 0.55)]*2\n# distr = 'mix'\n fig, ax = plt.subplots()\n if distr == 'mix' or distr == 'bias':\n b = df_obs[df_obs['die']=='bias'].groupby('outcome')['die'].count()\n b.plot.bar( stacked=True, color=my_colors ,title=\n 'Distribution of outcomes, ('+ distr + ')', rot=0, ax = ax)\n ax.set_ylabel('Counts for each outcome on a dice (all players inlcuded)')\n patches, labels = ax.get_legend_handles_labels()\n ax.legend(patches, labels, loc='best')\n\n if distr == 'mix' or distr == 'fair':\n f = df_obs[df_obs['die']=='fair'].groupby('outcome')['die'].count()\n f.plot.bar( stacked=True, color=my_colors[::-1], title=\n 'Distribution of outcomes',rot=0, ax = ax)\n ax.set_ylabel('Counts for each outcome on a dice (all players included)')\n patches, labels = ax.get_legend_handles_labels()\n ax.legend(patches, ['bias','fair'], loc='best')\n plt.show()\n\ndef plotPlayerSum(obs,distr):\n o = obs.groupby('player')['outcome','die'].sum()\n# o = o.sort_values(by=['outcome'])\n fig, ax = plt.subplots()\n o= o.reset_index()\n# print(o)\n my_colors = [(x/10.0, x/20.0, 0.75) for x in range(len(o))] # <-- Quick\n o['outcome'].plot(kind='bar',stacked = True, color=my_colors,title=\n 'Total sum for each player, ' + distr + ' dice distribution', rot = 0)\n ax.set_ylabel('Sum')\n ax.set_xlabel('Player')\n plt.show()\n\ndef multiPlot(obs_f,obs_b): # plot distributions, all fair and all bias\n fig, ax = plt.subplots()\n obs = pd.concat([obs_f,obs_b])\n o = obs.groupby(['player','die'])['outcome'].sum()\n my_colors = 'mb'\n o.plot.bar( stacked = True, color=my_colors,title= \n 'Sum for each player when using different dice distributons', rot = 0, ax=ax)\n ax.set_ylabel(\"Sum\")\n# print(o)\n plt.show()\n\ndef plotter(obs, players):\n color = 'rbg'\n fig, axes = plt.subplots(1,players, figsize=(12,3))\n for p in range(1,players+1):\n o = obs[obs['player']==p]\n ax = o['outcome'].plot.bar( stacked = True, color = color[p-1], ax = axes[p-1])\n ax.set_ylabel(\"dice outcome\")\n ax.set_xlabel(\"table i for player n\")\n patches, labels = ax.get_legend_handles_labels()\n ax.legend(patches, labels, loc='best')\n# o['outcome'].hist( color = color[p-1], ax = axes[p-1])\n plt.show()\n\ndef plotSums(d):\n ax = d.plot(kind='bar')\n ax.set_xlabel('player')\n ax.set_ylabel('sum')\n plt.show()\n\n\ndef main():\n K=2000\n N=3\n P = 0.8\n players = N\n pi_vec = np.array([0.5, 0.5])\n plot_without_hidden_outcomes = True\n# fair_emissions , bias_emissions, fullEm = emissionMatrix(K)\n# em_mat = emissionMatrix(K)\n# A_mat = transitionMatrix(K)\n# a = np.dot(pi_vec , em_mat\n\n \n distr = ['mix','fair','bias']\n obs = simulateDice(K,players,P,distr[2])\n obs = obs.assign(player_sequence = lambda x: x.outcome)\n print(obs)\n # NOW WE ARE ALLOWED TO SE THE TOTAL SUM: BUT ONLY THE OBSERVED OUTCOMES. NOT THE HIDDEN OUTCOMES.\n if plot_without_hidden_outcomes == True:\n obs['outcome'][obs['die'] == 'fair_hidden'] = 0 # 0 represents a hidden outcome\n obs['outcome'][obs['die'] == 'bias_hidden'] = 0 # 0 represents a hidden outcome\n \n d = pd.DataFrame()\n# d['player'] = [i for i in range(1,N+1)]\n\n total_sums = []\n observed_sums = []\n sums = obs.groupby('player')['outcome'].sum()\n print('observed sums for each player')\n print(sums, '\\n')\n for player in range(1,N+1):\n p = obs[obs['player'] == player]\n print('player: ', player)\n print('Observed Sum: ', p.outcome.sum())\n print('Original Sum: ', p.player_sequence.sum())\n print('number of hidden: ', p[p['outcome']==0].count().outcome)\n print(p, '\\n')\n observed_sums.append(p.outcome.sum())\n total_sums.append(p.player_sequence.sum())\n d['observed_sum'] = observed_sums\n d['original_sum'] = total_sums\n print(d)\n \n\n############## PLOT ########################################3\n plotSums(d)\n plotObs(obs,players,distr[2])\n plotPlayerSum(obs, distr[2])\n plotter(obs,players)\n# # \n \n Sum = 4\n dices = 3\n P = [[0]*dices]*Sum\n print(P)\n# for i in range()\n\n############## PLOT SEVERAL DIFFERENT DICE DISTRIBUTIONS #########################\n\n# obs_b = simulateDice(K,players,P,distr[2])\n# obs_f = simulateDice(K,players,P,distr[1])\n# plotObs(obs_b,players,distr[2])\n# plotObs(obs_f,players,distr[1])\n \n multiPlot(obs_f, obs_b)\n\n################ DYNAMIC PROGRAMMING ###############\nmain()\n","sub_path":"new_sum_HMM.py","file_name":"new_sum_HMM.py","file_ext":"py","file_size_in_byte":8150,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"218529721","text":"#!/usr/bin/python3\nf=open(\"shape_stat.csv\",\"r\")\ncount=0\nword=\"alarm clock\"\nfor i in f.readlines():\n\tk=i.split(\",\")\n\t\n\tif k[0] None:\n with open(LOGGING_FILE, 'a') as f:\n f.write(input_string+'\\n')\n print(input_string)\n return None\n\n#@profile\ndef _print_all_gnome_shell_processes() -> None:\n ps_e_process = subprocess.Popen(\"top -b -n 1\", shell=True, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)\n ps_e_stdout_string, _ = ps_e_process.communicate()\n stdout_lines = ps_e_stdout_string.decode(\"utf-8\").split('\\n')\n _logging_print(\"number of chrome processes {}\".format(len(list(filter(lambda line: 'chrome' in line, stdout_lines)))))\n _logging_print(stdout_lines[0])\n list(map(_logging_print, filter(lambda line: 'gnome-shell' in line, ps_e_stdout_string.decode(\"utf-8\").split('\\n'))))\n return None\n\n######################\n# Async IO Utilities #\n######################\n\nUNIQUE_BOGUS_RESULT_IDENTIFIER = (lambda x: x)\n\nEVENT_LOOP = asyncio.new_event_loop()\nasyncio.set_event_loop(EVENT_LOOP)\n\n#@profile\ndef _indefinitely_attempt_task_until_success(coroutine, coroutine_args):\n result = UNIQUE_BOGUS_RESULT_IDENTIFIER\n while result == UNIQUE_BOGUS_RESULT_IDENTIFIER:\n task = coroutine(*coroutine_args)\n from datetime import datetime; _logging_print(\"_indefinitely_attempt_task_until_success attempt start time {}\".format(datetime.now()))\n _logging_print(\"All shell processes 1\")\n _print_all_gnome_shell_processes()\n try:\n results = EVENT_LOOP.run_until_complete(asyncio.gather(task))\n if isinstance(results, list) and len(results) == 1:\n result = results[0]\n except Exception as err:\n _logging_print(\"err :: {}\".format(err))\n pass\n finally:\n _logging_print(\"All shell processes 2\")\n _print_all_gnome_shell_processes()\n pending_tasks = asyncio.Task.all_tasks()\n _logging_print(\"len(pending_tasks) {}\".format(len(pending_tasks)))\n for pending_task in pending_tasks:\n _logging_print(\"pending_task {}\".format(pending_task))\n _logging_print(\"All shell processes 3\")\n _print_all_gnome_shell_processes()\n if result == UNIQUE_BOGUS_RESULT_IDENTIFIER:\n warnings.warn(\"Attempting to execute {coroutine} on {coroutine_args} failed.\".format(coroutine=coroutine, coroutine_args=coroutine_args))\n time.sleep(1)\n return result\n\n##########################\n# Web Scraping Utilities #\n##########################\n\nasync def _launch_browser_page():\n browser = await pyppeteer.launch({'headless': False})\n page = await browser.newPage()\n return page\n\nBROWSER_PAGE = _indefinitely_attempt_task_until_success(_launch_browser_page, [])\n\n#############################\n# Wikidata Search Utilities #\n#############################\n\nWIKIDATA_SEARCH_URI_TEMPLATE = 'https://www.wikidata.org/w/index.php?sort=relevance&search={encoded_string}'\n\n#@profile\ndef _normalize_string_wrt_unicode(input_string: str) -> str:\n normalized_string = unicodedata.normalize('NFKD', input_string).encode('ascii', 'ignore').decode('utf-8')\n return normalized_string\n\nPUNUCTION_REMOVING_TRANSLATION_TABLE = str.maketrans('', '', string.punctuation)\n\n#@profile\ndef _normalize_string_for_wikidata_entity_label_comparison(input_string: str) -> str:\n normalized_string = input_string\n normalized_string = _normalize_string_wrt_unicode(normalized_string)\n normalized_string = normalized_string.lower()\n normalized_string = normalized_string.translate(PUNUCTION_REMOVING_TRANSLATION_TABLE)\n return normalized_string\n\n#@profile\nasync def _most_relevant_wikidata_entities_corresponding_to_string(input_string: str) -> str:\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string All shell processes 1\")\n _print_all_gnome_shell_processes()\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 1\")\n wikidata_entities_corresponding_to_string = []\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 1.1\")\n page = BROWSER_PAGE\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 2\")\n input_string_encoded = urllib.parse.quote(input_string)\n uri = WIKIDATA_SEARCH_URI_TEMPLATE.format(encoded_string=input_string_encoded)\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string All shell processes 2\")\n _print_all_gnome_shell_processes()\n try:\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 3\")\n _logging_print(\"uri {}\".format(uri))\n try:\n await page.goto(uri)\n except Exception as err:\n _logging_print(\"page.goto(uri) err {}\".format(err))\n exit()\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 3.5\")\n await page.waitForSelector('div#mw-content-text')\n search_results_div = await page.waitForSelector('div.searchresults')\n _logging_print(\"search_results_div {}\".format(search_results_div))\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 4\")\n search_results_paragraph_elements = await search_results_div.querySelectorAll('p')\n _logging_print(\"len(search_results_paragraph_elements) {}\".format(len(search_results_paragraph_elements)))\n search_results_have_shown_up = None\n for paragraph_element in search_results_paragraph_elements:\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 5\")\n paragraph_element_classname_string = await page.evaluate('(p) => p.className', paragraph_element)\n _logging_print(\"paragraph_element_classname_string {}\".format(paragraph_element_classname_string))\n paragraph_element_classnames = paragraph_element_classname_string.split(' ')\n _logging_print(\"paragraph_element_classnames {}\".format(paragraph_element_classnames))\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 6\")\n for paragraph_element_classname in paragraph_element_classnames:\n if paragraph_element_classname == 'mw-search-nonefound':\n search_results_have_shown_up = False\n elif paragraph_element_classname == 'mw-search-pager-bottom':\n search_results_have_shown_up = True\n if search_results_have_shown_up is not None:\n break\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 7\")\n if search_results_have_shown_up is not None:\n break\n _logging_print(\"search_results_have_shown_up {}\".format(search_results_have_shown_up))\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 8\")\n if search_results_have_shown_up:\n search_results_divs = await page.querySelectorAll('div.mw-search-result-heading')\n _logging_print(\"len(search_results_divs) {}\".format(len(search_results_divs)))\n # _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 9\")\n for search_results_div in search_results_divs:\n search_results_div_text_content = await page.evaluate('(search_results_div) => search_results_div.textContent', search_results_div)\n _logging_print(\"search_results_div_text_content {}\".format(search_results_div_text_content))\n parsable_text_match = re.match(r'^.+\\(Q[0-9]+\\) +$', search_results_div_text_content)\n _logging_print(\"parsable_text_match {}\".format(parsable_text_match))\n # _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 10\")\n if parsable_text_match:\n parsable_text = parsable_text_match.group()\n parsable_text = parsable_text.replace(')','')\n parsable_text_parts = parsable_text.split('(')\n # _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 11\")\n if len(parsable_text_parts)==2:\n (label, term_id) = parsable_text_parts\n label = label.strip()\n term_id = term_id.strip()\n # _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 12\")\n if _normalize_string_for_wikidata_entity_label_comparison(label) == _normalize_string_for_wikidata_entity_label_comparison(input_string):\n wikidata_entities_corresponding_to_string.append(term_id)\n if len(wikidata_entities_corresponding_to_string)>5:\n break\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 13\")\n except pyppeteer.errors.NetworkError:\n pass\n finally:\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string All shell processes 3\")\n _print_all_gnome_shell_processes()\n # await page.close()\n # await browser.close()\n # _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string All shell processes 4\")\n # _print_all_gnome_shell_processes()\n # _logging_print(\"before communicate browser.process {}\".format(browser.process))\n # _, errs = browser.process.communicate()\n # assert errs is None\n # _logging_print(\"after communicate browser.process {}\".format(browser.process))\n # _logging_print(\"errs {}\".format(errs))\n # process_is_still_running = browser.process.poll() is None\n # _logging_print(\"process_is_still_running {}\".format(process_is_still_running))\n # assert not process_is_still_running\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string All shell processes 5\")\n _print_all_gnome_shell_processes()\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string 14\")\n _logging_print(\"_most_relevant_wikidata_entities_corresponding_to_string All shell processes 6\")\n _print_all_gnome_shell_processes()\n return wikidata_entities_corresponding_to_string\n\n#@profile\ndef _string_corresponding_commonly_known_entities(input_string: str) -> List[str]:\n _logging_print(\"\")\n _logging_print(\"_string_corresponding_commonly_known_entities\")\n _logging_print(\"input_string {}\".format(input_string))\n result = _indefinitely_attempt_task_until_success(_most_relevant_wikidata_entities_corresponding_to_string, [input_string])\n return result\n\n####################################\n# Wikidata Query Service Utilities #\n####################################\n\nTYPE_TO_ID_MAPPING = bidict.bidict({\n 'Organization': 'Q43229',\n 'Anthroponym': 'Q10856962',\n 'Work': 'Q386724',\n 'Natural Geographic Entity': 'Q27096220',\n})\n\nQUERY_TEMPLATE_FOR_ENTITY_COMMONLY_KNOWN_ISAS = '''\nSELECT ?VALID_GENLS ?TERM\nWHERE \n{{\n VALUES ?TERM {{ {space_separated_term_ids} }}.\n ?TERM wdt:P31 ?IMMEDIATE_GENLS.\n ?IMMEDIATE_GENLS \twdt:P279* ?VALID_GENLS.\n VALUES ?VALID_GENLS {{ '''+' '.join(map(lambda type_string: 'wd:'+type_string, TYPE_TO_ID_MAPPING.values()))+''' }}.\n MINUS {{\n ?TERM wdt:P31 wd:Q4167410 .\n }}\n}}\n'''\n\nWIKI_DATA_QUERY_SERVICE_URI = 'https://query.wikidata.org'\n\n#@profile\ndef _sparql_query_queried_variables(sparql_query:str) -> List[str]:\n queried_variables = []\n sparql_tokens = sparql_query.split()\n assert sparql_tokens[0].lower()=='select'\n for sparql_token in sparql_tokens[1:]:\n if sparql_token[0]=='?':\n queried_variables.append(sparql_token)\n else:\n break\n return queried_variables\n\n#@profile\nasync def _query_wikidata_via_web_scraper(sparql_query:str) -> List[dict]:\n _logging_print(\"_query_wikidata_via_web_scraper All shell processes 1\")\n _print_all_gnome_shell_processes()\n results = []\n sparql_query_encoded = urllib.parse.quote(sparql_query)\n uri = WIKI_DATA_QUERY_SERVICE_URI+'/#'+sparql_query_encoded\n page = BROWSER_PAGE\n sparql_query_queried_variables = _sparql_query_queried_variables(sparql_query)\n number_of_variables_queried = len(sparql_query_queried_variables)\n _logging_print(\"_query_wikidata_via_web_scraper All shell processes 2\")\n _print_all_gnome_shell_processes()\n try:\n await page.goto(uri)\n selector_query_for_arbitrary_text_inside_query_box = 'span.cm-variable-2'\n await page.waitForSelector(selector_query_for_arbitrary_text_inside_query_box)\n button = await page.querySelector('button#execute-button')\n await page.evaluate('(button) => button.click()', button)\n await page.waitForSelector('div.th-inner.sortable.both')\n column_header_divs = await page.querySelectorAll('div.th-inner.sortable.both')\n assert len(column_header_divs) == number_of_variables_queried\n variable_names = []\n for column_header_div in column_header_divs:\n variable_name = await page.evaluate('(column_header_div) => column_header_div.textContent', column_header_div)\n variable_names.append(variable_name)\n assert sparql_query_queried_variables == list(map(lambda variable_name: '?'+variable_name, variable_names))\n anchors = await page.querySelectorAll('a.item-link')\n result = dict()\n for anchor_index, anchor in enumerate(anchors):\n anchor_variable = variable_names[anchor_index%number_of_variables_queried]\n anchor_link = await page.evaluate('(anchor) => anchor.href', anchor)\n assert len(re.findall(r\"^http://www.wikidata.org/entity/\\w+$\", anchor_link))==1\n entity_id = anchor_link.replace('http://www.wikidata.org/entity/','')\n anchor_variable_with_question_mark_prefix = '?'+anchor_variable\n result[anchor_variable_with_question_mark_prefix] = entity_id\n if (1+anchor_index)%number_of_variables_queried==0:\n assert len(result) == number_of_variables_queried\n results.append(result)\n result = dict()\n except pyppeteer.errors.NetworkError:\n pass\n finally:\n _logging_print(\"_query_wikidata_via_web_scraper All shell processes 3\")\n _print_all_gnome_shell_processes()\n # await page.close()\n # await browser.close()\n # _logging_print(\"_query_wikidata_via_web_scraper All shell processes 4\")\n # _print_all_gnome_shell_processes()\n # _logging_print(\"before communicate browser.process {}\".format(browser.process))\n # _, errs = browser.process.communicate()\n # assert errs is None\n # _logging_print(\"after communicate browser.process {}\".format(browser.process))\n # _logging_print(\"errs {}\".format(errs))\n # process_is_still_running = browser.process.poll() is None\n # _logging_print(\"process_is_still_running {}\".format(process_is_still_running))\n # assert not process_is_still_running\n _logging_print(\"_query_wikidata_via_web_scraper All shell processes 5\")\n _print_all_gnome_shell_processes()\n return results\n\n###########################\n# Most Abstract Interface #\n###########################\n\n#@profile\ndef execute_sparql_query_via_wikidata(sparql_query:str) -> List[dict]:\n _logging_print(\"\")\n _logging_print(\"execute_sparql_query_via_wikidata\")\n _logging_print(\"sparql_query {}\".format(sparql_query))\n result = _indefinitely_attempt_task_until_success(_query_wikidata_via_web_scraper, [sparql_query])\n return result\n\n#@profile\ndef _find_commonly_known_isas(term_ids_without_item_prefix: List[str]) -> Set[Tuple[str, str]]:\n term_type_pairs = set()\n if len(term_ids_without_item_prefix) != 0:\n term_ids = map(lambda raw_term_id: 'wd:'+raw_term_id, term_ids_without_item_prefix)\n space_separated_term_ids = ' '.join(term_ids)\n sparql_query = QUERY_TEMPLATE_FOR_ENTITY_COMMONLY_KNOWN_ISAS.format(space_separated_term_ids=space_separated_term_ids)\n results = execute_sparql_query_via_wikidata(sparql_query)\n for result in results:\n term = result['?TERM']\n term_type = result['?VALID_GENLS']\n term_type_pair = (term, term_type)\n term_type_pairs.add(term_type_pair)\n return term_type_pairs\n\n#@profile\ndef string_corresponding_wikidata_term_type_pairs(input_string: str) -> Set[Tuple[str, str]]:\n _logging_print(\"string_corresponding_wikidata_term_type_pairs All shell processes 1\")\n _print_all_gnome_shell_processes()\n term_ids = _string_corresponding_commonly_known_entities(input_string)\n _logging_print(\"\")\n _logging_print(\"string_corresponding_wikidata_term_type_pairs\")\n _logging_print(\"input_string {}\".format(input_string))\n _logging_print(\"term_ids {}\".format(term_ids))\n term_type_pairs = _find_commonly_known_isas(term_ids)\n term_type_pairs = [(term, TYPE_TO_ID_MAPPING.inverse[type_id]) for term, type_id in term_type_pairs]\n _logging_print(\"string_corresponding_wikidata_term_type_pairs All shell processes 2\")\n _print_all_gnome_shell_processes()\n return term_type_pairs\n\n#@profile\ndef main():\n _logging_print(\"This module contains utilities for named entity recognition via Wikidata scraping.\")\n _logging_print(\"BROWSER {}\".format(BROWSER))\n for _ in range(10):\n answer = string_corresponding_wikidata_term_type_pairs(\"friar\")\n _logging_print(\"answer {}\".format(answer))\n _logging_print(\"success\")\n\nif __name__ == '__main__':\n main()\n","sub_path":"named_entity_recognition_via_wikidata/named_entity_recognition_via_wikidata.py","file_name":"named_entity_recognition_via_wikidata.py","file_ext":"py","file_size_in_byte":18254,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"647349776","text":"# In this Example we will see how to read encrypted Messages with asymmetric Key Encryption(rsa)\n\n#First connect to a Node\n# Required Parameters are node, seed\nnode = \"https://nodes.thetangle.org:443\"\nseed = \"YOUR9SEED9GOES9HERE\"\napi = emi.connect_to_node(node,seed)\n\n#now we need the root_address\nroot_address = 'WUGJMZ9DLMWMV9ZBIQCZZS9CUCCCBAMWKXEEEUQMUBXHRGBCSHFSXYABBBTYRMSPFXNFRLD9VSXQWFLSW'\n# Now to read the Message Stream we will use a while loop. When we reach the last Message the loop will end.\nwhile True:\n #First we have to finde the Message\n message = emi.find_message(root_address)\n \n json_file = json.loads(message)\n \n #lets print out the message and next_address\n msg = json_file[\"1\"]\n encrypted_key = json_file[\"2\"]\n print(\"message: \" + str(msg))\n print(\"encrypted key: \" + str(encrypted_key))\n \n decrypted_key = emi.decrypt_pke(encrypted_key,alice_privat_key)\n \n decrypted_message = emi.decrypt_ske(msg,decrypted_key)\n print(\"decrypted key: \" + str(decrypted_key))\n print(\"decrypted_message: \" + str(decrypted_message))\n \n json_message = json.loads(decrypted_message)\n next_address = json_message['next_address']\n signature = json_message['signature']\n message = json_message['message']\n verify = emi.verify_signature(bob_public_key, signature,message)\n print(\"hash algorithmus: \" + str(verify))\n \n # Last step is to place next_addres as root_address\n root_address = next_address\n","sub_path":"main/examples/read_emi_asymmetric.py","file_name":"read_emi_asymmetric.py","file_ext":"py","file_size_in_byte":1481,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"569981281","text":"#!/usr/bin/env python2\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Aug 20 16:22:45 2018\n\n@author: Alex Daniel\n\"\"\"\n\nimport nibabel as nib\nimport sys\nimport numpy as np\n\nfin = sys.argv[1]\nimg = nib.load(fin, strict_sort=True, permit_truncated = True)\nhdr = img.header\ntry:\n nib.save(img, fin[:-3]+'nii.gz')\nexcept:\n print('Unable to convert '+fin+' to nii.gz')\n \nbvals, bvecs = hdr.get_bvals_bvecs()\necho_spacing = (1000.0 * hdr.general_info['water_fat_shift'])/(434.215*(hdr.general_info['epi_factor']+1))\n\nnp.savetxt(fin[:-3]+'bvec',bvecs.T, fmt = '%.3f')\nnp.savetxt(fin[:-3]+'bval',np.expand_dims(bvals,1).T, fmt = '%.0f')\nf = open(fin[:-3]+'echo', 'w')\nf.write('%.6f'%echo_spacing)\nf.close()\n","sub_path":"par_to_nifti.py","file_name":"par_to_nifti.py","file_ext":"py","file_size_in_byte":702,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"289848918","text":"# Install the Python Requests library:\n# `pip install requests`\n\nimport requests\nimport json\n\ndef send_request():\n # Request weather underground\n # GET http://api.wunderground.com/api/380538e19b591277/conditions/q/TH/Bangkok.json\n\n try:\n response = requests.get(\n url=\"http://api.wunderground.com/api/380538e19b591277/conditions/q/TH/Bangkok.json\",\n )\n if response.status_code == 200:\n print('Response HTTP Status Code: {status_code}'.format(\n status_code=response.status_code))\n print('Response HTTP Response Body: {content}'.format(\n content=response.content))\n return response.content\n else:\n print('Response HTTP Status Code: {status_code}'.format(\n status_code=response.status_code))\n return None\n\n except requests.exceptions.RequestException:\n print('HTTP Request failed')\n return None\n\n\ndef main():\n print(\"Start Program\")\n result = send_request()\n output = json.loads(result)\n print(type(result))\n print(type(output))\n\nif __name__ == '__main__':\n main()\n\n","sub_path":"WeatherUG/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1134,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"605591480","text":"# -*- coding: utf-8 -*-\n\nimport base64\nfrom io import StringIO\nimport csv\nimport logging\nimport time\nfrom datetime import datetime\nfrom dateutil.relativedelta import relativedelta\nfrom tempfile import NamedTemporaryFile\nfrom odoo import models, fields, api, _\nfrom odoo.osv import osv\n#from pdf417gen import encode, render_image\n_logger = logging.getLogger(__name__)\n\nclass sale_order(models.Model):\n _inherit = \"sale.order\"\n \n @api.depends('avc_import_transaction_log_ids.skip_line')\n @api.multi\n def _is_order_is_mismatch(self):\n for order in self:\n for transaction_line in order.avc_import_transaction_log_ids :\n if transaction_line.skip_line:\n order.is_mismatch_order = True\n\n is_amazon_edi_order = fields.Boolean('is Amazon Order')\n amazon_edi_order_id = fields.Char('Amazon Order ID')\n amazon_order_ack_uploaded = fields.Boolean('Amazon order Acknowledgement Uploaded')\n amazon_order_dispatch_advice_uploaded = fields.Boolean('Amazon order Dispatch Advice uploaded')\n avc_import_transaction_log_ids = fields.One2many('avc.transaction.log.line','sale_order_id', string = 'AVC Import Transaction Log',domain=[('operation_type','=','import'),('application','=','sale_order')])\n avc_export_transaction_log_ids = fields.One2many('avc.transaction.log.line','sale_order_id', string = 'AVC Export Transaction Log',domain=[('operation_type','=','export'),('application','=','sale_order_response')])\n vendor_id = fields.Many2one('amazon.vendor.instance', string = \"Vendor\")\n requested_for_routing = fields.Boolean(string = \"Requested for Routing\")\n received_routing_info = fields.Boolean(string = \"Received Routing Information\")\n bill_of_lading_number = fields.Char(string = \"Bill of Lading Number\")\n account_type = fields.Char(string = \"Account Type\")\n is_mismatch_order = fields.Boolean('Is Mismatch Order',compute='_is_order_is_mismatch')\n mismatch_product = fields.Selection([('cancel', 'Cancel'), ('reject', 'Reject'), ('backorder', 'Backorder'), ],\n string='If Product not Found')\n \n # Messing Info details \n \n sender_id = fields.Char('Sender ID',readonly=True)\n recipient_id = fields.Char('Recipient ID',readonly=True)\n message_type = fields.Char('Type',readonly=True)\n msg_version = fields.Char('Version',readonly=True)\n buyer_id = fields.Char('Buyer ID',readonly=True)\n buyer_address = fields.Char('Buyer Address',readonly=True)\n supplier_id = fields.Char('Supplier ID',readonly=True)\n delivery_party_id = fields.Char('Delivery Party ID',readonly=True)\n country_code = fields.Char('Delivery Country',readonly=True)\n invoice_id = fields.Char('Invoice Party ID',readonly=True)\n currancy_code = fields.Char('Currency Code',readonly=True)\n order_id = fields.Char('Sale Order ID',readonly=True) \n vat_number = fields.Char('VAT Registration Number',readonly=True)\n max_delivery_date_ept = fields.Date(string = 'Max Delivery Date')\n delivery_date_ept = fields.Date(string='Delivery Date')\n \n @api.multi\n def action_confirm(self):\n res = super(sale_order,self).action_confirm()\n for order in self:\n if order.is_amazon_edi_order:\n for picking in order.picking_ids:\n carrier_type = order.vendor_id and order.vendor_id.delivery_type\n picking.write({'carrier_type' : carrier_type , 'vendor_id' : order.vendor_id.id})\n \n @api.multi\n def reimport_amazon_po_file(self):\n sale_order_line_obj = self.env['sale.order.line']\n product_obj = self.env['product.product']\n job_id = self.env['avc.file.transaction.log'].search([('sale_order_id','=',self.id)])\n if job_id :\n data = job_id.attachment_id and job_id.attachment_id.datas\n file = StringIO(base64.decodestring(data).decode())\n reader = csv.reader(file,delimiter=\"'\",quotechar='|')\n order_line_info = {}\n line_no = 1\n for segment in reader:\n for seg in segment:\n if seg.startswith('LIN+'):\n order_line_info.update({'Line_'+str(line_no):{}})\n ean = seg.split(\"+\")\n ean = ean[len(ean)-1] \n if ean.upper().find('EN',0,len(ean)) !=-1 and ean.upper().find(':',0,len(ean)) !=-1:\n ean = ean.split(\":\") and ean.split(\":\")[0] or ''\n order_line_info['Line_'+str(line_no)].update({'ean':ean})\n #UP used for Universal Product Code **code edited here**\n elif ean.upper().find('UP',0,len(ean)) !=-1 and ean.upper().find(':',0,len(ean)) !=-1:\n ean = ean.split(\":\") and ean.split(\":\")[0] or ''\n order_line_info['Line_'+str(line_no)].update({'ean':ean})\n line_no += 1\n \n elif seg.startswith('PIA+'):\n code = seg.split(\"+\") \n code = code[2][:-3] if len(code)>2 else ''\n if not order_line_info['Line_'+str(line_no-1)].get('ean',False):\n order_line_info['Line_'+str(line_no-1)].update({'default_code':code})\n \n \n elif seg.startswith('QTY+'):\n qty = seg.split(\":\") \n qty = qty[1] if len(qty)>1 else 0\n order_line_info['Line_'+str(line_no-1)].update({'qty':qty})\n \n elif seg.startswith('PRI+'):\n price = seg.split(\":\") \n price = price[1] if len(price)>1 else 0\n order_line_info['Line_'+str(line_no-1)].update({'price':price})\n \n for key,value in order_line_info.items():\n amazon_code = value.get('default_code') or value.get('ean')\n sale_order_line = sale_order_line_obj.search([('amazon_edi_line_code','=',amazon_code),('order_id','=',self.id)])\n if not sale_order_line:\n product = product_obj.search([('amazon_sku','=',amazon_code)])\n if not product:\n product = product_obj.search([('default_code','=',amazon_code)])\n amazon_edi_code = 'SKU'\n if not product :\n product = product_obj.search([('barcode','=',amazon_code)])\n amazon_edi_code = 'barcode'\n if product:\n qty = value.get('qty',0.0)\n price = value.get('price',0.0) \n line=(product,price,amazon_code,qty)\n orderlinevals,product_id, qty_code = self.prepare_order_line_vals(line,self)\n if orderlinevals:\n sale_order_line = sale_order_line_obj.create(orderlinevals)\n remark = amazon_edi_code + ':' + amazon_code\n transaction_line = self.env['avc.transaction.log.line'].search([('job_id','=',job_id.id),('remark','=',remark),('operation_type','=','import')])\n if transaction_line:\n vals = {\n 'message':'Sale Order Line Created',\n 'remark':'sale order id %s'%(self.name or ''),\n 'sale_order_id':self.id,\n 'job_id':job_id.id,\n 'picking_id':False,\n 'back_order_id':False,\n 'sale_order_line_id':sale_order_line.id,\n 'product_id':orderlinevals.get('product_id',''),\n 'package_id':False,\n 'stock_inventory_id':False,\n 'company_id':job_id.company_id.id or False,\n 'user_id':self.env.user.id,\n 'picking_state':'draft',\n 'application':'sale_order',\n 'export_qty':orderlinevals.get('product_uom_qty',''),\n 'processed_qty':orderlinevals.get('product_uom_qty',''),\n 'manually_processed':False,\n 'is_mismatch_detail':False,\n 'skip_line':False,\n 'skip_order':False,\n 'filename':job_id.attachment_id.name,\n 'create_date':datetime.now(),\n 'operation_type':'import',\n 'price':price,\n }\n transaction_line.write(vals)\n \n \n return True\n \n @api.multi\n def export_dispatch_advice(self):\n \"\"\"\n Use: To send Dispatch Advice to Amazon Vendor Central via EDI 856 file. Call manually from Sale Order Form view\n :return: Boolean\n \"\"\"\n self.sync_export_dispatch_advice(sale_order_ids = self)\n return True\n \n @api.model\n def sync_import_amazon_edi_order(self,args={},file_datas=None, ):\n \"\"\"\n Use : For import EDI 850 Purchase Order file\n This method call by cron,\n :param args: arguments pass by cron (vendor_id)\n :param file_datas: If Purchase Order create manualy then here you can send file.\n :return: Boolean\n \"\"\"\n if not args.get('vendor_id'):\n vendor_id = self.env['ir.values'].get_default('avc.config.settings', 'vendor_id')\n else:\n vendor_id = args.get('vendor_id')\n print (\"cron run vendor id : %s\"%(vendor_id))\n vendor_obj = self.env['amazon.vendor.instance'].browse(vendor_id)\n self.import_sales_from_amazon_edi(vendor_obj,file_data=file_datas )\n return True\n \n @api.multi\n def import_sales_from_amazon_edi(self, vendor_ids = None,file_data=None):\n \"\"\"\n Use: Fetch the sale orders file from FTP location,\n format the data into format required by Odoo\n and create the sale in Odoo\n :param vendor_id: Amazon Vendor Central Instance ID\n :param file_data: EDI 850 Purchase Order file\n :return: Boolean\n \"\"\"\n ctx = self._context.copy() or {}\n \n for vendor in vendor_ids:\n self.job_id = None\n self.filename = None\n self.server_filename = None\n self.export_avc_line_id = []\n self.ack_error_lines=[]\n \n filenames_dict ={}\n if file_data:\n imp_file = StringIO(base64.decodestring(file_data))\n file_write = open('/tmp/order_data.txt','wb')\n file_write.writelines(imp_file.getvalue())\n file_write.close()\n file_read = open('/tmp/order_data.txt', \"rU\")\n dialect = csv.Sniffer().sniff(file_read.readline())\n file_read.seek(0)\n reader = csv.reader(file_read,delimiter=\"'\",quotechar='|')\n file_read.seek(0)\n self.process_file_and_prapare_order(file_read)\n else:\n file_to_delete = []\n connection_id = False\n if vendor.is_production_environment:\n ftp_server_id = vendor.production_ftp_connection\n directory_id = vendor.production_po_directory_id\n else :\n ftp_server_id = vendor.test_ftp_connection\n directory_id = vendor.test_po_directory_id\n \n with vendor.get_edi_receive_interface(ftp_server_id,directory_id) \\\n as edi_interface:\n # `filenames` contains a list of filenames to be imported \n filenames_dict = edi_interface.pull_from_ftp(vendor.po_file_import_prefix) \n \n for server_filename, filename in filenames_dict.items():\n \n with open(filename) as file:\n self.job_id = None\n self.filename = filename\n self.server_filename = server_filename\n ctx.update({'filename':server_filename})\n self.process_file_and_prapare_order(vendor,file)\n file_to_delete.append(server_filename) # : Ekta\n \n if self.job_id:\n binary_package = open(filename).read().encode()\n attachment_vals = {\n 'name':server_filename,\n 'datas':base64.encodestring(binary_package),\n 'datas_fname':server_filename,\n 'type':'binary',\n 'res_model': 'avc.file.transaction.log',\n 'res_id':self.job_id.id,\n }\n \n attachment=self.env['ir.attachment'].create(attachment_vals)\n self.job_id.write({'attachment_id' : attachment.id})\n self.job_id.message_post(body=_(\"PO Import File\"),attachment_ids=attachment.ids)\n self.job_id.message_post(body=_((\"Sale Order created %s\"%(self.order_id.name or '') if self.order_id else \"Information Mismatch\")))\n if vendor.auto_confirm_sale_order and vendor.auto_generate_po_ack and self.order_id:\n self.auto_send_poa(sale_order_id = self.order_id)\n\n if file_to_delete:\n with vendor.get_edi_receive_interface(ftp_server_id,directory_id) \\\n as edi_interface:\n edi_interface.sftp_client.chdir(edi_interface.download_dir)\n for filename in file_to_delete:\n edi_interface.delete_from_ftp(filename)\n return True\n\n @api.multi\n def auto_send_poa(self,sale_order_id = None):\n \"\"\"\n USE: This method will call export_po_ack() of stock.picking,\n :param sale_order_id:\n :return: stock.picking's export_po_ack()\n \"\"\"\n if self.instance_id.auto_confirm_sale_order:\n res = sale_order_id.action_confirm()\n if res:\n picking_id = sale_order_id.mapped('picking_ids')\n if picking_id:\n return picking_id[0].export_po_ack()\n else:\n message = \"First of all set Sale Order Auto Confirm as True from Amazon Vendor Central >> Configuration >> Vendor >> Purchase Order Acknowledgement.\"\n _logger.info(message)\n raise osv.except_osv(_('Purchase Order Auto Acknowledgement send error'),_(message))\n\n @api.multi\n def process_file_and_prapare_order(self,vendor,file):\n \"\"\"\n Use: Decode Amazon EDI 850 Purchase Order file and create sale order, sale order lines and required log entries.\n :param file: EDI 850 Purchase Order file\n :return: Boolean\n \"\"\"\n #declaration\n country_obj = self.env['res.country']\n partner_obj = self.env['res.partner'] \n sale_order_obj = self.env['sale.order'] \n sale_order_line_obj = self.env['sale.order.line']\n product_product_obj = self.env['product.product']\n \n delivery_address = {}\n order_line_info = {}\n inv_address_data = {}\n order_info = {}\n message_info = {} \n line_no = 1\n order_line = 0\n total_segment = 0\n self.order_type = ''\n \n #read and seprate file in diffrent part\n for segment in csv.reader(file,delimiter=\"'\",quotechar='|'):\n for seg in segment:\n if seg.startswith('UNB+UNOA') or seg.startswith('UNB+UNOC'):\n header = seg.split(\"+\")\n message_info.update({'sender_id' : header[2][:-3],'recipient_id' : header[3][:-3]})\n total_segment +=1\n \n elif seg.startswith('UNH'):\n msg_type = seg.split(\"+\")\n msg_type = msg_type[2].split(\":\")[0] if len(msg_type)>2 else ''\n message_info.update({'message_type' : msg_type})\n total_segment +=1\n \n elif seg.startswith('BGM+'):\n order_name = seg.split(\"+\")\n order_name = order_name[2] if len(order_name) >= 3 else ''\n order_info.update({'order_name':order_name})\n total_segment +=1\n \n elif seg.startswith('DTM+137'):\n date_seg = seg.split(\":\")\n date_order = datetime.strptime(date_seg[1], '%Y%m%d')\n order_info.update({'date_order':date_order})\n total_segment +=1\n \n elif seg.startswith('DTM+63'):\n date_seg = seg.split(\":\")\n delivery_date = datetime.strptime(date_seg[1], '%Y%m%d')\n order_info.update({'delivery_date':delivery_date})\n message_info.update({'max_delivery_date_ept':delivery_date})\n total_segment +=1\n \n elif seg.startswith('DTM+64'):\n date_seg = seg.split(\":\")\n earliest_date = datetime.strptime(date_seg[1], '%Y%m%d')\n message_info.update({'delivery_date_ept' : earliest_date})\n total_segment +=1\n \n elif seg.startswith('RFF+ADE'):\n order = seg.split(\":\")\n self.order_type = order[1]\n total_segment +=1\n\n elif seg.startswith('RFF+PD'):\n total_segment +=1\n \n elif seg.startswith('NAD+BY'):\n buyer_id = seg.split(\":\")\n buyer_address = buyer_id[0][7:]+':'+buyer_id[2]\n buyer_id = buyer_id and buyer_id[0][7:]\n message_info.update({'buyer_id':buyer_id,'buyer_address':buyer_address})\n total_segment +=1\n continue\n \n elif seg.startswith('NAD+SU'):\n supplier_id = seg.split(\":\")\n supplier_id = supplier_id and supplier_id[0][7:]\n message_info.update({'supplier_id':supplier_id})\n total_segment +=1\n continue\n \n elif seg.startswith('NAD+DP'):\n delivery = seg.split(\"+\")\n delivery_party_id = delivery[2][:-3]\n country_code = delivery[len(delivery)-1]\n country_id = country_obj.search([('code', 'ilike', country_code)])\n message_info.update({'delivery_party_id':delivery_party_id,'country_code':country_code})\n delivery_address = {'name': delivery[4],\n 'street':delivery[5],\n 'city':delivery[6],\n 'zip':delivery[8],\n 'country_id':country_id.id,\n }\n\n total_segment +=1\n continue\n #vendors information get from this part\n elif seg.startswith('NAD+IV'):\n invoice_seg = seg.split(\"+\")\n invoice_id = invoice_seg and invoice_seg[2][:-3]\n message_info.update({'invoice_id':invoice_id})\n if invoice_seg[4].find(\":\") >= 0 :\n customer = invoice_seg[4].split(\":\")\n elif invoice_seg[4].find(\",\") >=0:\n customer = invoice_seg[4].split(\",\")\n country_id = country_obj.search([('code', 'ilike', invoice_seg[9])])\n\n #partner_id = partner_obj.search([('name', '=', customer[0])],)\n partner_id = vendor.so_customer_id\n# if not partner_id:\n# partner_vals = {'name':customer[0],'opt_out':True}\n# partner_id = partner_obj.create(partner_vals)\n inv_address_data = {\n 'type':'invoice',\n 'name': \"%s\" %(customer[0]),\n 'street': customer[1],\n 'street2': invoice_seg[5],\n 'city': invoice_seg[6],\n 'zip': invoice_seg[8],\n 'country_id': country_id[0].id if country_id else False,\n 'parent_id': partner_id.id,\n }\n if delivery_address:\n delivery_address.update({'parent_id': partner_id.id,'type':'delivery'})\n order_info.update({'delivery_address': delivery_address})\n order_info.update({'inv_address_data':inv_address_data}) \n# customer_info.append(inv_address_data)\n total_segment +=1 \n continue \n \n elif seg.startswith('RFF+VA'):\n vat_number = seg.split(\":\")\n message_info.update({'vat_number':vat_number[1]})\n total_segment +=1\n continue\n \n elif seg.startswith('CUX+2'):\n currancy = seg.split(\":\")\n currancy_code = currancy[1]\n currency_id = self.env['res.currency'].search([('name','=',currancy_code)])\n currency_id = currency_id and currency_id[0] or False\n pricelist_id = vendor and vendor.pricelist_id or False\n pricelist_id = pricelist_id and pricelist_id.id or False\n order_info.update({'currency_id':currency_id.id,'pricelist_id':pricelist_id})\n message_info.update({'currancy_code':currancy_code})\n total_segment +=1\n continue\n #sale order line data saprate here\n elif seg.startswith('LIN+'):\n order_line_info.update({'Line_'+str(line_no):{}})\n ean = seg.split(\"+\")\n ean = ean[len(ean)-1] \n if ean.upper().find('EN',0,len(ean)) !=-1 and ean.upper().find(':',0,len(ean)) !=-1:\n ean = ean.split(\":\") and ean.split(\":\")[0] or ''\n order_line_info['Line_'+str(line_no)].update({'ean':ean})\n #UP used for Universal Product Code **code edited here**\n elif ean.upper().find('UP',0,len(ean)) !=-1 and ean.upper().find(':',0,len(ean)) !=-1:\n ean = ean.split(\":\") and ean.split(\":\")[0] or ''\n order_line_info['Line_'+str(line_no)].update({'ean':ean})\n line_no += 1\n order_line +=1\n total_segment +=1\n \n elif seg.startswith('PIA+'):\n code = seg.split(\"+\") \n code = code[2][:-3] if len(code)>2 else ''\n if not order_line_info['Line_'+str(line_no-1)].get('ean',False):\n order_line_info['Line_'+str(line_no-1)].update({'default_code':code})\n total_segment +=1\n \n elif seg.startswith('QTY+'):\n qty = seg.split(\":\") \n qty = qty[1] if len(qty)>1 else 0\n order_line_info['Line_'+str(line_no-1)].update({'qty':qty})\n total_segment +=1\n \n elif seg.startswith('PRI+'):\n price = seg.split(\":\") \n price = price[1] if len(price)>1 else 0\n order_line_info['Line_'+str(line_no-1)].update({'price':price})\n total_segment +=1 \n \n elif seg.startswith('UNS+S'):\n total_segment +=1\n \n elif seg.startswith('CNT+2'):\n total_line = seg.split(\":\") \n total_line = total_line[1] if len(total_line)>1 else 0\n total_segment +=1\n \n if int(total_line) != order_line:\n raise osv.except_osv(_('Error'), _('Order Line not integrated properly, Please Check order line data in file.')) \n \n elif seg.startswith('UNT+'):\n segments = seg.split(\"+\")\n segments = segments[1]\n if int(segments) != total_segment:\n raise osv.except_osv(_('Error'), _('File not integrated properly, Please Check file data.'))\n\n if not vendor.supplier_id == message_info.get('supplier_id', ''):\n if not self.job_id:\n avc_file_process_job_vals = {\n 'message':'Mismatch Supplier Information',\n 'filename':self.server_filename,\n 'vendor_id':vendor.id,\n 'application':'sale_order',\n 'operation_type':'import',\n 'create_date':datetime.now(),\n 'company_id':vendor.company_id.id or False,\n }\n self.job_id = self.create_avc_file_process_job(avc_file_process_job_vals)\n return True\n \n if not vendor.pricelist_id.currency_id.id == order_info.get('currency_id'):\n if not self.job_id:\n avc_file_process_job_vals = {\n 'message':'Mismatch Pricelist information',\n 'filename':self.server_filename,\n 'vendor_id':vendor.id,\n 'application':'sale_order',\n 'operation_type':'import',\n 'create_date':datetime.now(),\n 'company_id':vendor.company_id.id or False,\n }\n self.job_id = self.create_avc_file_process_job(avc_file_process_job_vals)\n return True\n #checked if order exist or not\n existing_order_id = sale_order_obj.search([('amazon_edi_order_id', '=', order_info.get('order_name', ''))])\n if self.order_type == 'firstorder':\n if existing_order_id:\n return True\n\n #message_id = self.env['amazon.edi.message.info'].create(message_info)\n order_vals = self.prepare_order_vals(vendor,order_info,message_info.get('delivery_party_id',False))\n order_vals.update({'vendor_id':vendor.id,'account_type':self.order_type, 'carrier_id':vendor.amazon_edi_carrier_method.id or False})\n order_vals.update(message_info)\n if vendor.warehouse_id : \n order_vals.update({'warehouse_id':vendor.warehouse_id.id})\n order_id = sale_order_obj.create(order_vals)\n #message_id.write({'order_id':order_id.id})\n #self.order_id = order_id \n fiscal_position_id = order_vals.get('fiscal_position_id',False)\n fiscal_position = self.env['account.fiscal.position'].browse(fiscal_position_id) or False \n line_id = False\n \n #CREATE LOG IN avc.file.transaction.log\n avc_file_process_job_vals = {\n 'message': 'Order imported',\n 'filename': self.server_filename,\n 'vendor_id': vendor.id,\n 'application' : 'sale_order',\n 'operation_type' : 'import',\n 'create_date' : datetime.now(),\n 'company_id':vendor.company_id.id or False,\n 'sale_order_id':order_id.id,\n }\n self.job_id = self.create_avc_file_process_job(avc_file_process_job_vals)\n \n for key,value in order_line_info.items():\n default_code = value.get('default_code',False)\n ean = value.get('ean',False)\n product = False\n if default_code:\n product = product_product_obj.search([('amazon_sku','=',default_code)])\n if not product :\n product = product_product_obj.search([('default_code','=',default_code)]) \n code_type=\"SKU\"\n amazon_code = default_code\n amazon_edi_line_code_type = 'sku'\n if ean :\n product = product_product_obj.search([('barcode','=',ean)])\n code_type = \"barcode\"\n amazon_code = ean\n amazon_edi_line_code_type = 'barcode'\n orderlinevals ={}\n line = ()\n if product:\n qty = value.get('qty',0.0)\n price = value.get('price',0.0) \n line=(product,price,amazon_code,qty)\n \n orderlinevals,product_id, qty_code = self.prepare_order_line_vals(line,order_id)\n orderlinevals.update({'amazon_edi_line_code_type' : amazon_edi_line_code_type})\n if orderlinevals:\n sale_order_line_id = sale_order_line_obj.create(orderlinevals)\n \n avc_transaction_log_val = {\n 'message':'Sale Order Line Created',\n 'remark':'sale order id %s'%(order_id.name or ''),\n 'sale_order_id':order_id.id,\n 'job_id':self.job_id.id,\n 'picking_id':False,\n 'back_order_id':False,\n 'sale_order_line_id':sale_order_line_id.id,\n 'product_id':orderlinevals.get('product_id',''),\n 'package_id':False,\n 'stock_inventory_id':False,\n 'company_id':self.job_id.company_id.id or False,\n 'user_id':self.env.user.id,\n 'picking_state':'draft',\n 'application':'sale_order',\n 'export_qty':orderlinevals.get('product_uom_qty',''),\n 'processed_qty':orderlinevals.get('product_uom_qty',''),\n 'manually_processed':False,\n 'is_mismatch_detail':False,\n 'skip_line':False,\n 'skip_order':False,\n 'filename':self.server_filename,\n 'create_date':datetime.now(),\n 'operation_type':'import',\n 'price':price,\n }\n self.job_id.transaction_log_ids.create(avc_transaction_log_val)\n else:\n line = line + (False,qty_code)\n self.ack_error_lines.append(line)\n self.create_avc_transaction_lines(order_id,code_type='SKU',code=code,processed_qty=qty,msg='Product not found',price=price)\n else:\n qty_code = 182\n line = line + (False,qty_code)\n self.ack_error_lines.append(line)\n self.create_avc_transaction_lines(order_id,code_type='barcode',code=amazon_code,processed_qty=qty,msg='Product not found',price=price)\n return True\n\n @api.multi\n def prepare_order_vals(self,vendor,order_info,delivery_party_id):\n \"\"\"\n Use: To generate sale order's value based on given information.\n :param order_info: Sale Order required information\n :param delivery_party_id: Delivery Party ID\n :return: sale order values (dict{})\n \"\"\"\n sale_order_obj = self.env['sale.order']\n partner_obj = self.env['res.partner']\n address_data = order_info.get('inv_address_data',{})\n delivery_address = order_info.get('delivery_address',{})\n partner_id=address_data.get('parent_id',False)\n if not partner_id:\n return\n \n inv_add_domain=[]\n for address in address_data:\n inv_add_domain.append((address,'=',address_data.get(address)))\n\n inv_add_id = partner_obj.search(inv_add_domain)\n inv_add_id = inv_add_id and inv_add_id[0] \n if not inv_add_id:\n inv_add_id = partner_obj.create(address_data)\n \n delivert_add_domain=[]\n for address in delivery_address:\n delivert_add_domain.append((address,'=',delivery_address.get(address)))\n \n delivery_add_id = partner_obj.search(delivert_add_domain)\n delivery_add_id = delivery_add_id and delivery_add_id[0]\n if not delivery_add_id:\n delivery_add_id = partner_obj.create(delivery_address)\n \n delivery_id = partner_obj.search([('edi_gln_no','=',delivery_party_id),('parent_id','=',partner_id)])\n if not delivery_id:\n delivery_id = partner_obj.search([('edi_gln_no','=',delivery_party_id)])\n if delivery_id and delivery_id.id != partner_id:\n delivery_id.write({'parent_id':partner_id})\n else:\n partner_obj.browse(partner_id).write({'edi_gln_no': delivery_party_id})\n\n partner_address = partner_obj.browse(partner_id).address_get(['contact','invoice','delivery'])\n \n ordervals={\n 'company_id':vendor.company_id.id or False,\n 'partner_id' :partner_address.get('contact',False),\n }\n new_record = sale_order_obj.new(ordervals)\n new_record.onchange_partner_id()\n ordervals = sale_order_obj._convert_to_write({name: new_record[name] for name in new_record._cache})\n new_record = sale_order_obj.new(ordervals)\n new_record.onchange_partner_shipping_id()\n ordervals = sale_order_obj._convert_to_write({name: new_record[name] for name in new_record._cache})\n ordervals.update({\n 'name' : order_info.get('order_name'),\n 'amazon_edi_order_id' : order_info.get('order_name',''),\n 'picking_policy' : vendor.picking_policy or False,\n 'date_order' : order_info.get('date_order',False) and order_info['date_order'].strftime('%Y-%m-%d') or time.strftime('%Y-%m-%d'),\n 'state' : 'draft',\n #'invoice_status' : self.instance_id.amazon_edi_invoice_status or 'invoiced',\n 'mismatch_product' : vendor.mismatch_product,\n 'is_amazon_edi_order' : True,\n 'note': order_info.get('order_name',''),\n 'client_order_ref' : order_info.get('order_name',''),\n 'pricelist_id': order_info.get('pricelist_id',False),\n })\n return ordervals\n \n# @api.multi\n# def _get_product_stock_for_ack(self,vendor,location_id,company_id):\n# ## Query for get stock based on configuration\n# # location_id = vendor.warehouse_id and vendor.warehouse_id.location_id and vendor.warehouse_id.location_id.id\n# # company_id = vendor.company_id and vendor.company_id.id\n# qry = \"\"\n# stock_dict = {}\n# if vendor.picking_policy_based_on == 'qty_on_hand' :\n# qry = \"\"\"\n# select product_id,sum(quantity)-sum(reserved_quantity) as total_qty from stock_quant \n# where location_id in (%s) and company_id=(%s)\n# group by product_id\n# \"\"\"%(location_id,company_id)\n# if vendor.picking_policy_based_on == 'forecast_sale' :\n# qry = \"\"\"\n# select product_id,sum(total_qty)\n# from\n# (\n# select product_id,sum(quantity)-sum(reserved_quantity) as total_qty from stock_quant \n# where location_id in (%s) and company_id=%s\n# group by product_id\n# union all \n# select product_id,sum(product_qty) as total_qty from stock_move\n# where location_dest_id in (%s) and company_id=%s and state not in ('draft','done','cancel')\n# group by product_id\n# union all \n# select product_id,-sum(product_qty) as total_qty from stock_move\n# where location_id in (%s) and company_id=%s and state in ('waiting','confirmed')\n# group by product_id \n# )T\n# group by product_id\n# \n# \"\"\"%(location_id,company_id,location_id,company_id,location_id,company_id)\n# self._cr.execute(qry)\n# results = self._cr.fetchall()\n# for result_tuple in results:\n# stock_dict.update({result_tuple[0] : result_tuple[1]})\n# return stock_dict\n @api.multi\n def prepare_order_line_vals(self,line_info,order_id):\n \"\"\"\n Use: create sale order line values\n :param line_info: sale order line dict\n :param order_id: sale order id\n :param fiscal_position: fiscal position\n :return: sale_order_line dict{}, product_id, qty_code\n \"\"\"\n product_id = line_info[0]\n file_price = line_info[1]\n amazon_edi_code = line_info[2]\n qty = float(line_info[3])\n product_product = self.env['product.product']\n sale_order_line_obj = self.env['sale.order.line']\n qty_code = False\n \n #product_id = product_product.search([('default_code','=',default_code)])\n if not product_id:\n qty_code = 182\n return () , product_id, qty_code\n orderlinevals = {\n 'order_id' : order_id.id,\n 'product_id' : product_id.id,\n } \n new_record = sale_order_line_obj.new(orderlinevals)\n new_record.product_id_change()\n orderlinevals=new_record._convert_to_write({name: new_record[name] for name in new_record._cache}) \n orderlinevals.update({\n 'product_uom_qty' : qty,\n 'price_unit' : float(file_price),\n 'customer_lead' : product_id and product_id.product_tmpl_id.sale_delay,\n 'invoice_status' : 'invoiced',\n 'amazon_edi_line_code': amazon_edi_code,\n })\n qty_code = 12\n return orderlinevals, product_id, qty_code\n\n @api.multi\n def prepare_line(self, fiscal_position=None, product_id=None, order_id=None, qty=None, file_price=None, amazon_edi_code=None):\n \"\"\"\n Use: Prepare sale order line data with backorder value\n :param fiscal_position: Fiscal Position\n :param product_id: Product ID\n :param order_id: Sale Order ID\n :param qty: Ordered Quantity\n :param file_price: Received price in PO file\n :param amazon_edi_code: Barcode / received from PO file\n :return: sale order line dict{}, product_id, qty_code\n \"\"\"\n sale_order_line_obj = self.env['sale.order.line']\n tax_id = False\n if fiscal_position:\n tax_id = fiscal_position.map_tax(product_id.taxes_id).ids\n\n orderlinevals = {\n 'order_id':order_id.id,\n 'product_id':product_id.id,\n }\n new_record = sale_order_line_obj.new(orderlinevals)\n new_record.product_id_change()\n orderlinevals = new_record._convert_to_write({name:new_record[name] for name in new_record._cache})\n if not orderlinevals.get('tax_id', []):\n tax_id = [(6, 0, tax_id)]\n orderlinevals.update({'tax_id':tax_id})\n orderlinevals.update({\n 'product_uom_qty':qty,\n 'price_unit':float(file_price),\n 'customer_lead':product_id and product_id.product_tmpl_id.sale_delay,\n 'invoice_status':'invoiced',\n 'amazon_edi_line_code':amazon_edi_code,\n })\n return orderlinevals, product_id\n\n @api.multi\n def create_avc_file_process_job(self,vals):\n \"\"\"\n Use: To create new record in avc.file.transaction.log model.\n :param vals: Required Value for avc.file.transaction.log\n :return: avc.file.transaction.log new record ID\n \"\"\"\n avc_file_process_job_obj=self.env['avc.file.transaction.log']\n job_id = avc_file_process_job_obj.create(vals)\n return job_id\n \n @api.multi\n def create_avc_transaction_lines(self,order,code_type=None,code=None,processed_qty=None,msg=None,price=None,product_id=None):\n \"\"\"\n Use: To create avc.transaction.log.line new line.\n :param code_type: code information sku/barcode\n :param code: code data\n :param processed_qty: ordered quantity\n :param msg: message for log line\n :param price: price ordered product\n :param product_id: product id if available\n :return: avc.transaction.log.line's id\n \"\"\"\n #it make an entry in avc.transaction.log.line\n avc_transaction_log_val = {\n 'message':msg if msg else 'Product not Found',\n 'remark': '%s:%s'%(code_type,code),\n 'sale_order_id':order.id,\n 'job_id':self.job_id.id,\n 'company_id':self.job_id.company_id.id or False,\n 'user_id':self.env.user.id,\n 'application':'sale_order',\n 'export_qty':0.0,\n 'processed_qty':processed_qty,\n 'manually_processed':False,\n 'is_mismatch_detail':False if product_id else True,\n 'skip_line':True,\n 'skip_order':False,\n 'filename':self.server_filename,\n 'create_date':datetime.now(),\n 'operation_type':'import',\n 'price':price,\n 'product_id':product_id if product_id else False,\n }\n res = self.job_id.transaction_log_ids.create(avc_transaction_log_val)\n return res\n \n \n \n \n @api.multi\n def to_pdf417(self,order_id, packages):\n \"\"\"\n Use: This method called from \"report_edi_saleorder_barcode_label\" QWeb report.\n This method used for creaate PDF417 format barcode which used in GS1-128 Label.\n :param order_id: sale order id\n :param packages: packages dictionary\n :return: barcode label base64 image data\n \"\"\"\n sale_order_id = self.browse(order_id)\n text = \"AMZN\"\n text = text + ',PO: ' + sale_order_id.amazon_edi_order_id\n for package in packages:\n if package.get('barcode'):\n text = text + ',UPC: ' + str(package.get('barcode')) + ','\n else:\n text = text + ',EAN: ' + str(package.get('default_code')) + ','\n text = text + 'QTY: ' + str(package.get('product_qty'))\n\n print (text)\n\n codes = encode(text, columns=5)\n image = render_image(codes)\n buffer = StringIO()\n image.save(buffer, format=\"JPEG\")\n img_str = base64.b64encode(buffer.getvalue())\n\n return img_str\n \n @api.multi\n def get_total_qty(self):\n \"\"\"\n Use: This method called from \"report_edi_saleorder_barcode_label\" QWeb report.\n :return: total quantity of selected sale order.\n \"\"\"\n qty = 0\n for line in self.order_line:\n qty += line.product_uom_qty\n return qty\n\n @api.multi\n def get_package_information(self,order_id=None):\n \"\"\"\n Use: To get pericular sale order's Package information.\n :param order_id: sale order id\n :return: dictionary with package information.\n \"\"\"\n res = {}\n sale_order_id = self.browse(order_id)\n for ids in sale_order_id.picking_ids.pack_operation_product_ids:\n line_info = {'product_id':ids.product_id.id, 'product_qty':ids.product_qty, 'default_code':ids.product_id.default_code or '', 'barcode':ids.product_id.barcode or ''}\n if res.get(ids.result_package_id.name):\n data = res.get(ids.result_package_id.name)\n data.append(line_info)\n res.update({ids.result_package_id.name:data})\n else:\n res.update({ids.result_package_id.name:[line_info]})\n return res\n\n \n\n @api.multi\n def _prepare_invoice(self):\n \"\"\"\n USE: this method call super _prepare_invoice of sale.order after that it checks\n whether current sale order is amazon vendor central's order if yes then it will\n update journal_id which set in amazon vendor instance.\n :return: invoice_value dict\n \"\"\"\n res = super(sale_order, self)._prepare_invoice()\n if self.is_amazon_edi_order and self.vendor_id.journal_id:\n res.update({'journal_id':self.vendor_id.journal_id.id})\n return res","sub_path":"odoo_apps/amazon_vendor_central_ept/model/amazon_sale_order.py","file_name":"amazon_sale_order.py","file_ext":"py","file_size_in_byte":45906,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"507685847","text":"import pytest\nimport src.misc as misc\n\n\n@pytest.mark.parametrize(\"answer, expected_answer\",\n [('', True),\n (' ', True),\n ('yes', True),\n ('Yes', True),\n ('YES', True),\n ('Y', True),\n ('no', False),\n ('No', False),\n ('NO', False),\n ('N', False)])\ndef test_confirm(monkeypatch, answer, expected_answer):\n monkeypatch.setattr('builtins.input', lambda x: answer)\n assert misc.confirm() == expected_answer","sub_path":"tests/misc_test.py","file_name":"misc_test.py","file_ext":"py","file_size_in_byte":657,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"276573562","text":"def main():\n # escribe tu código abajo de esta línea\n msj=int(input(\"Dame el número de mensajes: \"))\n m=float(input(\"Dame el número de megas: \"))\n minutos=int(input(\"Dame el numero de minutos¨: \"))\n\n mensajes=0.80*msj\n megas=0.80*m\n min=0.80*minutos\n\n costomensual= mensajes+megas+min\n print(\"El costo mensual es:\",costomensual)\n\nif __name__ == '__main__':\n main()\n","sub_path":"assignments/09Telefono/src/exercise.py","file_name":"exercise.py","file_ext":"py","file_size_in_byte":400,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"109095026","text":"import webbrowser\r\n\r\n\r\nclass Movie():\r\n\r\n # centents are defined here\r\n def __init__(\r\n self,\r\n movie_title,\r\n movive_storyline,\r\n poster_image,\r\n trailer_youtube\r\n ):\r\n self.title = movie_title\r\n self.storyline = movive_storyline\r\n self.poster_image_url = poster_image\r\n self.trailer_youtube_url = trailer_youtube\r\n\r\n # open webbrowser when check the poster\r\n def show_trailer(self):\r\n webbrowser.open(self.trailer_youtube_url)\r\n","sub_path":"media.py","file_name":"media.py","file_ext":"py","file_size_in_byte":515,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"12549095","text":"# encoding: utf-8\nfrom pdefc.lang import TypeEnum\nfrom pdefc.generators import Generator, Templates, GeneratorCli, PrefixMapper\n\n\nENCODING = 'utf8'\nHEADER_TEMPLATE = 'header.jinja2'\nIMPL_TEMPLATE = 'impl.jinja2'\nPACKAGE_TEMPLATE = 'package.jinja2'\n\n\nclass ObjectiveCGeneratorCli(GeneratorCli):\n def build_parser(self, parser):\n self._add_prefix_args(parser)\n\n def create_generator(self, out, args):\n prefixes = self._parse_prefix_args(args)\n return ObjectiveCGenerator(out, prefixes)\n\n\nclass ObjectiveCGenerator(Generator):\n '''Objective-C code generator.'''\n\n @classmethod\n def create_cli(cls):\n return ObjectiveCGeneratorCli()\n\n def __init__(self, out, prefixes=None):\n '''Create a new generator.'''\n super(ObjectiveCGenerator, self).__init__(out)\n\n self.prefix_mapper = PrefixMapper(prefixes)\n self.filters = _ObjectiveCFilters(self.prefix_mapper)\n self.templates = Templates(__file__, filters=self.filters)\n\n def generate(self, package):\n '''Generate a package source code.'''\n for module in package.modules:\n for definition in module.definitions:\n self._generate_header(definition)\n self._generate_impl(definition)\n\n self._generate_package(package)\n\n def _generate_header(self, definition):\n '''Generate a definition header file.'''\n code = self.templates.render(HEADER_TEMPLATE, definition=definition)\n filename = '%s.h' % self.filters.objc_name(definition)\n self.write_file(filename, code)\n return code\n\n def _generate_impl(self, definition):\n '''Generate a definition implementation file.'''\n code = self.templates.render(IMPL_TEMPLATE, definition=definition)\n filename = '%s.m' % self.filters.objc_name(definition)\n self.write_file(filename, code)\n return code\n\n def _generate_package(self, package):\n '''Generate a package file which groups all headers.'''\n code = self.templates.render(PACKAGE_TEMPLATE, package=package)\n\n names = set()\n for module in package.modules:\n for definition in module.definitions:\n names.add(self.filters.objc_name(definition).lower())\n\n # Generate a unique package file name.\n name = package.name\n while name in names:\n name += '_package'\n\n # Convert it into a CamelCase string.\n name = name.title().replace('_', '')\n\n # Write the package header file.\n filename = '%s.h' % name\n self.write_file(filename, code)\n return code\n\n\nclass _ObjectiveCFilters(object):\n '''Objective-C jinja filters.'''\n def __init__(self, prefix_mapper):\n self.prefix_mapper = prefix_mapper\n\n def objc_name(self, def0):\n name = def0.name\n prefix = self.prefix_mapper.get_prefix(def0.namespace) or ''\n return prefix + name\n\n def objc_bool(self, expression):\n return 'YES' if expression else 'NO'\n\n def objc_base(self, message):\n return self.objc_name(message.base) if message.base else 'PDMessage'\n\n def objc_isprimitive(self, type0):\n pointers = TypeEnum.COLLECTION_TYPES \\\n + (TypeEnum.MESSAGE, TypeEnum.INTERFACE, TypeEnum.STRING, TypeEnum.DATETIME)\n return type0.type not in pointers\n\n def objc_type(self, type0):\n t = type0.type\n if t in NATIVE_TYPES:\n return NATIVE_TYPES[t]\n elif t == TypeEnum.ENUM_VALUE:\n return '%s_%s ' % (self.objc_name(type0.enum), type0.name)\n elif t == TypeEnum.ENUM:\n return '%s ' % self.objc_name(type0)\n elif t == TypeEnum.INTERFACE:\n return 'id<%s> ' % self.objc_name(type0)\n elif t == TypeEnum.MESSAGE:\n return '%s *' % self.objc_name(type0)\n raise ValueError('Unsupported type %r' % type0)\n\n def objc_descriptor(self, type0):\n t = type0.type\n if t in NATIVE_DESCRIPTORS:\n return NATIVE_DESCRIPTORS[t]\n elif t == TypeEnum.ENUM:\n return '%sDescriptor()' % self.objc_name(type0)\n elif t == TypeEnum.LIST:\n return '[PDDescriptors listWithElement:%s]' % self.objc_descriptor(type0.element)\n elif t == TypeEnum.SET:\n return '[PDDescriptors setWithElement:%s]' % self.objc_descriptor(type0.element)\n elif t == TypeEnum.MAP:\n return '[PDDescriptors mapWithKey:%s value:%s]' % (\n self.objc_descriptor(type0.key),\n self.objc_descriptor(type0.value))\n elif t == TypeEnum.INTERFACE:\n return '%sDescriptor()' % self.objc_name(type0)\n elif t == TypeEnum.MESSAGE:\n return '[%s typeDescriptor]' % self.objc_name(type0)\n raise ValueError('Unsupported type %r' % type0)\n\n def objc_default(self, type0):\n t = type0.type\n value = NATIVE_DEFAULTS.get(t)\n if value:\n return value\n\n if t == TypeEnum.ENUM:\n return '0'\n\n return 'nil'\n\n def objc_result(self, type0):\n if type0.is_interface:\n return 'id<%s> ' % self.objc_name(type0)\n return 'NSOperation *'\n\n\nNATIVE_TYPES = {\n TypeEnum.BOOL: 'BOOL ',\n TypeEnum.INT16: 'int16_t ',\n TypeEnum.INT32: 'int32_t ',\n TypeEnum.INT64: 'int64_t ',\n TypeEnum.FLOAT: 'float ',\n TypeEnum.DOUBLE: 'double ',\n\n TypeEnum.STRING: 'NSString *',\n TypeEnum.DATETIME: 'NSDate *',\n\n TypeEnum.VOID: 'id',\n\n TypeEnum.LIST: 'NSArray *',\n TypeEnum.SET: 'NSSet *',\n TypeEnum.MAP: 'NSDictionary *'\n}\n\nNATIVE_DESCRIPTORS = {\n TypeEnum.BOOL: '[PDDescriptors bool0]',\n TypeEnum.INT16: '[PDDescriptors int16]',\n TypeEnum.INT32: '[PDDescriptors int32]',\n TypeEnum.INT64: '[PDDescriptors int64]',\n TypeEnum.FLOAT: '[PDDescriptors float0]',\n TypeEnum.DOUBLE: '[PDDescriptors double0]',\n\n TypeEnum.STRING: '[PDDescriptors string]',\n TypeEnum.DATETIME: '[PDDescriptors datetime]',\n\n TypeEnum.VOID: '[PDDescriptors void0]',\n}\n\nNATIVE_DEFAULTS = {\n TypeEnum.BOOL: 'NO',\n TypeEnum.INT16: '0',\n TypeEnum.INT32: '0',\n TypeEnum.INT64: '0L',\n TypeEnum.FLOAT: '0.0f',\n TypeEnum.DOUBLE: '0.0'\n}\n","sub_path":"generator/pdef_objc/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":6215,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"388614422","text":"import json\nimport os\nimport pickle\nimport random\nfrom collections import defaultdict\n\nimport h5py\nimport sys\nimport numpy as np\nfrom os.path import join\n\nimport re\n\nfrom pandas import DataFrame\nfrom progressbar import ProgressBar\n\nfrom utils import array2txt, DBI\n\n# os.chdir('data_0626')\n\nfilename = 'features_x_new.hdf5'\n\nfeatures = json.load(open(join('../setup', 'feature_list.json')))\nfeature_names = sorted([f['id'] for f in features])\n\nrandom_array = np.array([6.00000000e-05, 2.20000000e-04, 6.80000000e-04, 1.90750000e-03,\n 4.72750000e-03, 1.13875000e-02, 2.59575000e-02, 5.65350000e-02,\n 1.15300000e-01, 2.16387500e-01])\n\nprecision = [random_array]\nnames = ['random']\n\n# Random Result:\n# [ 6.00000000e-05 2.20000000e-04 6.80000000e-04 1.90750000e-03\n# 4.72750000e-03 1.13875000e-02 2.59575000e-02 5.65350000e-02\n# 1.15300000e-01 2.16387500e-01]\n\ndoc2cate = {str(doc['index']): doc['category']\n for doc in DBI().articles.find({'target': True},\n {'category': 1, 'index': 1})}\n\nprint(len(doc2cate))\ncates = set(doc2cate.values())\nprint(len(cates))\n\ncate_precision = defaultdict(list)\n\ntry:\n precision = pickle.load(open('evaluate_precision.p', 'rb'))\n\n cate_precision = pickle.load(open('evaluate_precision_cate.p', 'rb'))\nexcept:\n measure_results = h5py.File(filename, 'r')\n label = h5py.File('features_y.hdf5')\n for i, feature_name in enumerate(feature_names):\n print(i, feature_name)\n hit = np.zeros((10,))\n total = np.zeros((10,))\n cate_hit = {c: np.zeros((10,)) for c in cates}\n cate_total = {c: np.zeros((10,)) for c in cates}\n keys = list(measure_results.keys())\n pbar = ProgressBar(max_value=len(keys))\n for key in keys:\n measure = measure_results[key][:, i]\n top_ids = np.argsort(measure)[-2:]\n cate = doc2cate[key]\n for _id in top_ids:\n total += np.array([1] * 10)\n cate_total[cate] += np.array([1] * 10)\n hit += label[key][_id]\n cate_hit[cate] += label[key][_id]\n pbar.update(pbar.value + 1)\n pbar.finish()\n precision.append(hit / total)\n for c in cates:\n cate_precision[c].append(cate_hit[c] / cate_total[c])\n print(precision[-1])\n\n pickle.dump(precision, open('evaluate_precision.p', 'wb'))\n pickle.dump(cate_precision, open('evaluate_precision_cate.p', 'wb'))\n\nprecision = np.array(precision)\nfor c in cate_precision:\n cate_precision[c].insert(0, random_array)\n cate_precision[c] = np.array(cate_precision[c])\n\ntext = ['content', 'title', 'summary']\n\n\ndef get_indexes(feature_names):\n regex = '(.+)_(.+)_(.+)_(.+)'\n pattern = re.compile(regex)\n result_indexes = [[[] for _i in range(0, 3)] for _ in range(0, 3)]\n result_headers = [[[] for _i in range(0, 3)] for _ in range(0, 3)]\n for i, name in enumerate(feature_names):\n find = pattern.findall(name)\n if len(find) > 0 and ((find[0][1] == '1' and find[0][0] == 'stem')\n or (find[0][1] != '1' and find[0][0] == 'stemo')):\n find = find[0]\n result_indexes[text.index(find[2])][int(find[1]) - 1].append(i + 1)\n result_headers[text.index(find[2])][int(find[1]) - 1].append(find[3])\n return result_indexes, result_headers\n\nfeatures_indexes, feature_headers = get_indexes(feature_names)\n\n\nfrom matplotlib.backends.backend_pdf import PdfPages\nfrom matplotlib import pyplot as plt\nimport seaborn as sns\nimport matplotlib.ticker as mticker\npp = PdfPages('precision_2_3.pdf')\nsns.set('paper', style=\"dark\", color_codes=True)\nf, subfigs = plt.subplots(3, 3)\nplt.subplots_adjust(hspace=0.1, wspace=0.1)\nmodel_names = ['bow', 'jaccard', 'lda', 'lsi', 'tfidf']\ntarget_index = [1, 0, 4, 3, 2]\nfor i in range(0, 3):\n for j in range(0, 3):\n ax = subfigs[j][i]\n print(text[i], j + 1)\n index = features_indexes[i][j]\n headers = feature_headers[i][j]\n precision_feature = precision[index][:, 2]\n if j != 0:\n precision_feature[2] = 0\n frame = DataFrame([[model_names[index], precision_feature[index] * 100] for index in target_index\n if precision_feature[index] > 0],\n columns=['model', 'precision'])\n sns.barplot(x='precision', y='model', data=frame, ax=ax)\n max_width = max(precision_feature) * 100\n for p in ax.patches:\n if p.get_width() == max_width:\n ax.annotate('{:.2f}'.format(p.get_width()), (p.get_width() + 1, p.get_y() + 0.5), weight='bold',\n color='r', size=8)\n else:\n ax.annotate('{:.2f}'.format(p.get_width()), (p.get_width() + 1, p.get_y() + 0.5), size=8)\n\n ax.set(xlim=(0, 50))\n ax.set_ylabel('')\n ax.set_xlabel('')\n if j == 0:\n ax.set_title(text[i])\n if i == 2:\n ax.text(52, 2 if j == 0 else 1.5, ['unigram', 'bigram', 'trigram'][j], va='center', rotation='vertical')\n if i == 1:\n ax.yaxis.set_major_locator(mticker.NullLocator())\n if i == 2:\n ax.yaxis.set_major_locator(mticker.NullLocator())\n if j < 2:\n ax.xaxis.set_major_locator(mticker.NullLocator())\n\nf.text(0.5, 0.02, 'precision (%)', ha='center')\nf.text(0.02, 0.5, 'STS model', va='center', rotation='vertical')\npp.savefig()\npp.close()\n# plt.show()\n\n'''\nf_tex = open('../report_tex.txt', 'w')\nf_simple = open('../report_simple.txt', 'w')\nfor i in range(0, 3):\n for j in range(0, 3):\n print(text[i], j + 1)\n index = features_indexes[i][j]\n headers = feature_headers[i][j]\n precision_feature = precision[index]\n\n sort = np.argsort(precision_feature[:, 2])[::-1]\n sort_headers = [headers[sort_i] for sort_i in sort]\n sort_precision = np.array([precision_feature[sort_i] for sort_i in sort])\n latex = array2txt(sort_precision, [['{}-hops'.format(i) for i in range(1, 11)], sort_headers],\n variant_axis=[0, 1], format='latex')\n f_tex.write(text[i] + '_' + str(j + 1) + '\\n')\n f_tex.write(latex + '\\n\\n\\n')\n simple = array2txt(sort_precision, [['{}-hops'.format(i) for i in range(1, 11)], sort_headers],\n variant_axis=[0, 1], format='simple')\n f_simple.write(text[i] + '_' + str(j + 1) + '\\n')\n f_simple.write(simple + '\\n\\n\\n')\n print(array2txt(sort_precision, [['{}-hops'.format(i) for i in range(1, 11)], sort_headers],\n variant_axis=[0, 1], format='grid'))\n\n best = []\n\n for c in cates:\n cate_precision_feature = cate_precision[c][index]\n cate_sort = np.argsort(cate_precision_feature[:, 2])[::-1]\n cate_sort_headers = [headers[sort_i] for sort_i in cate_sort]\n cate_sort_precision = np.array([cate_precision_feature[sort_i] for sort_i in cate_sort])\n f_tex.write(c + '\\n')\n f_tex.write(array2txt(cate_sort_precision, [['{}-hops'.format(i) for i in range(1, 11)],\n cate_sort_headers],\n variant_axis=[0, 1], format='latex') + '\\n\\n\\n')\n f_simple.write(c + '\\n')\n f_simple.write(array2txt(cate_sort_precision, [['{}-hops'.format(i) for i in range(1, 11)],\n cate_sort_headers],\n variant_axis=[0, 1], format='simple') + '\\n\\n\\n')\n\n best.append(cate_sort_precision[0].tolist() + [cate_sort_headers[0]])\n best = np.array(best)\n cate_list = list(cates)\n best_sort = np.argsort(best[:, 2])[::-1]\n best_sort_headers = [cate_list[sort_i] for sort_i in best_sort]\n best_sort_precision = np.array([best[sort_i] for sort_i in best_sort])\n print(array2txt(best_sort_precision, [['{}-hops'.format(i) for i in range(1, 11)] + ['feature name'],\n best_sort_headers],\n variant_axis=[0, 1], format='grid'))\n\n'''\n","sub_path":"source/evaluate.py","file_name":"evaluate.py","file_ext":"py","file_size_in_byte":8252,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"335485270","text":"\"\"\"\n1162. As Far from Land as Possible\n\nGiven an n x n grid containing only values 0 and 1, where 0 represents water and 1 represents land, find a water cell\nsuch that its distance to the nearest land cell is maximized, and return the distance. If no land or water exists in the grid, return -1.\n\nThe distance used in this problem is the Manhattan distance: the distance between two cells (x0, y0) and (x1, y1) is |x0 - x1| + |y0 - y1|.\n\n\nExample 1:\n\n\nInput: grid = [[1,0,1],[0,0,0],[1,0,1]]\nOutput: 2\nExplanation: The cell (1, 1) is as far as possible from all the land with distance 2.\nExample 2:\n\n\nInput: grid = [[1,0,0],[0,0,0],[0,0,0]]\nOutput: 4\nExplanation: The cell (2, 2) is as far as possible from all the land with distance 4.\n\n\nConstraints:\n\nn == grid.length\nn == grid[i].length\n1 <= n <= 100\ngrid[i][j] is 0 or 1\n\n\n\n\"\"\"\n\n\nclass MaxDistance:\n\n def doit_bfs(self, grid: list) -> int:\n from collections import deque\n q = deque([])\n m, n = len(grid), len(grid[0])\n\n for i in range(m):\n for j in range(n):\n if grid[i][j] == 1:\n q.append((i, j))\n\n if len(q) == 0 or len(q) == m * n:\n return 0\n\n d = -1\n while q:\n size = len(q)\n for c in range(size):\n i, j = q.popleft()\n for x, y in ((i-1, j), (i, j+1), (i + 1, j), (i. i-1)):\n if 0 <= x < n and 0 <= y <= n and grid[x][y] == 0:\n grid[x][y] = 1\n q.append((x, y))\n d += 1\n\n return d\n\n def doit_bfs_1(self, grid: list) -> int:\n q, vis = [], set()\n for i in range(len(grid)):\n for j in range(len(grid[0])):\n if grid[i][j] == 1:\n q.append((i, j, 0))\n vis.add((i, j))\n\n if len(q) == 0 or len(q) == len(grid) * len(grid[0]): return -1\n ans = 0\n while q:\n i, j, d = q.pop(0)\n for x, y in [(i - 1, j), (i + 1, j), (i, j - 1), (i, j + 1)]:\n if 0 <= x < len(grid) and 0 <= y < len(grid[0]):\n if (x, y) not in vis:\n vis.add((x, y))\n ans = max(ans, d + 1)\n q.append((x, y, d + 1))\n\n return ans","sub_path":"PythonLeetcode/leetcodeM/1162_AsFarFromLandAsPossible.py","file_name":"1162_AsFarFromLandAsPossible.py","file_ext":"py","file_size_in_byte":2314,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"470430075","text":"from stormpath.client import Client as StormpathClient\n\nSTORMPATH_API_KEY_ID=\"4FESWTU76DNA7UVHUCNJ6E0AB\"\nSTORMPATH_API_KEY_SECRET=\"dx5/ABWKvnjxMM5nEMwNLyCc90y0wwUXfNsKWBJWaJ4\"\nSTORMPATH_APPLICATION_NAME=\"Daily Idea\"\n\nstormpath_client = StormpathClient(\n api_key_id=STORMPATH_API_KEY_ID,\n api_key_secret=STORMPATH_API_KEY_SECRET)\nstormpath_application = stormpath_client.applications.search(STORMPATH_APPLICATION_NAME)[0]\n\n","sub_path":"playtime.py","file_name":"playtime.py","file_ext":"py","file_size_in_byte":466,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"109354341","text":"import ROOT\nROOT.gSystem.Load(\"libHiggsAnalysisCombinedLimit\")\nimport json\n\n\nclass XZZDataCardMaker:\n def __init__(self,finalstate,category,luminosity=1.0,physics=\"LJ\"):\n self.physics=physics\n self.finalstate=finalstate\n self.category=category\n self.contributions=[]\n self.systematics=[]\n self.observation=0.0\n\n self.tag=self.physics+\"_\"+finalstate+\"_\"+category\n self.luminosity=luminosity\n\n\n def addSystematic(self,name,kind,values,addPar = \"\"):\n self.systematics.append({'name':name,'kind':kind,'values':values })\n\n def addObservation(self,observation=0):\n self.observation=observation\n\n def addContribution(self,name,ID,rate):\n self.contributions.append({'name':name,'ID':ID,'rate':rate}) \n\n def makeCard(self):\n\n f = open(self.tag+'.txt','w')\n f.write('imax 1\\n')\n f.write('jmax {n}\\n'.format(n=len(self.contributions)-1))\n f.write('kmax *\\n')\n f.write('-------------------------\\n')\n f.write('bin '+self.tag+'\\n')\n f.write('observation '+str(self.observation)+'\\n')\n f.write('-------------------------\\n')\n\n # sorted contributions\n contributions = sorted(self.contributions,key=lambda x: x['ID'])\n \n # print bin\n f.write('bin\\t') \n for contrib in contributions:\n f.write(self.tag+'\\t')\n f.write('\\n')\n\n #print names\n f.write('process\\t')\n for contrib in contributions:\n f.write(contrib['name']+'\\t')\n f.write('\\n')\n \n #print IDs\n f.write('process\\t')\n for contrib in contributions:\n f.write(str(contrib['ID'])+'\\t')\n f.write('\\n')\n \n #print rate\n f.write('rate\\t')\n for contrib in contributions:\n f.write(str(contrib['rate'])+'\\t')\n f.write('\\n')\n\n f.write('-------------------------\\n')\n\n # print systematics\n for syst in self.systematics:\n if syst['kind'] == 'param':\n f.write(syst['name']+'\\t'+'param\\t' +str(syst['values'][0])+'\\t'+str(syst['values'][1])+'\\n')\n elif syst['kind'] == 'lnN': \n f.write(syst['name']+'\\t'+ 'lnN\\t' )\n for contrib in contributions:\n has=False\n for name,v in syst['values'].iteritems():\n if contrib['name']==name:\n f.write(str(v)+'\\t' )\n has=True\n break;\n if not has:\n f.write('-\\t' )\n f.write('\\n' )\n \n \n f.close()\n\n\n \n\n","sub_path":"XZZ2l2nu/python/statistics/XZZDataCardMaker.py","file_name":"XZZDataCardMaker.py","file_ext":"py","file_size_in_byte":2736,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"62887066","text":"#!/usr/bin/env python\nimport subprocess\n\nimport rospy\nfrom std_msgs.msg import String\n\n\ndef cb(msg: String):\n args = [\"espeak\", msg.data]\n subprocess.run(args)\n\n\ndef main():\n rospy.init_node(\"tts_node\")\n sub = rospy.Subscriber(\"speak\", String, cb)\n rospy.spin()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"scripts/tts_node.py","file_name":"tts_node.py","file_ext":"py","file_size_in_byte":317,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"112304685","text":"'''\nProblem Statement\nGiven a sorted array arr[] of distinct integers. Sort the array into a wave-like array and return it.\nIn other words, arrange the elements into a sequence such that a1 >= a2 <= a3 >= a4 <= a5.....\n(considering the increasing lexicographical order).\n\nInput Format\n\nFirst line of input contains n-the size of array. Next line of input contains n integers-the\nelements of array.\n\nConstraints\n\n1 ≤ n ≤ 10^6\n0 ≤ Ai ≤10^7\n\nOutput Format\n\nPrint the array which should be sorted in wave like pattern.\n'''\n\nn=int(input())\narr = list(set(map(int,input().split())))\nfor i in range(0,n-1,2):\n arr[i], arr[i+1] = arr[i+1], arr[i]\nprint(*arr)\n","sub_path":"WaveFormArray.py","file_name":"WaveFormArray.py","file_ext":"py","file_size_in_byte":662,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"426874937","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Tue Sep 8 10:37:34 2019\r\n\r\n@author: Vik Jakkula\r\n\"\"\"\r\n\r\nimport torch \r\nfrom torch.autograd import Variable\r\nimport torch.nn as nn\r\nimport torch.optim as optim\r\n\r\nx = Variable(torch.Tensor([[1],[2],[3],[4]]))\r\ny = Variable(torch.Tensor([[2],[4],[6],[8]]))\r\n\r\nprint(x)\r\n\r\nclass LinearRegressionModel(nn.Module):\r\n def __init__(self,input_size,output_size):\r\n super(LinearRegressionModel,self).__init__()\r\n self.linear = nn.Linear(input_size,output_size)\r\n \r\n def forward(self,x):\r\n y_predict = self.linear(x)\r\n return y_predict\r\n \r\nmodel = LinearRegressionModel(1,1)\r\ncriteria = nn.MSELoss()\r\n# 0.01 is learning rate\r\noptimizer = optim.SGD(model.parameters(),0.01)\r\n\r\nfor epoch in range(500):\r\n y_predict = model(x)\r\n loss = criteria(y_predict,y)\r\n optimizer.zero_grad()\r\n loss.backward()\r\n optimizer.step()\r\n print(epoch, float(loss.data[0]))\r\n\r\n\r\ntest = Variable(torch.Tensor([20]))\r\nz = model.forward(test)\r\nprint(float(z[0]))\r\n\r\n\r\n","sub_path":"regression.py","file_name":"regression.py","file_ext":"py","file_size_in_byte":1037,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"66903006","text":"#!/usr/bin/env python3\nimport os\nimport sys\n\nsys.path.append(os.path.dirname(__file__)+\"/../../../\")\n\ndef main():\n os.environ['MODE']='develop'\n from nwpc_monitor_task_scheduler.celery_server.task import sms\n\n result = sms.get_group_sms_status_task.delay()\n # result = tasks.update_dingtalk_token_task.delay()\n\n print(result)\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"test/manual/nwpc_monitor_task_scheduler/run_tasks.py","file_name":"run_tasks.py","file_ext":"py","file_size_in_byte":380,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"621750777","text":"import os\n\nOUTFILENAME = \"output.txt\"\n\nfiles = [f for f in os.listdir(\".\") if os.path.isfile(f) and f != os.path.basename(__file__) and f != OUTFILENAME]\n\nstrings = []\n\nfor f in files:\n\twith open(f) as infile:\n\t\tstrings.append(infile.read())\n\nwith open(\"output.txt\", \"w\") as outfile:\n\toutfile.write(\"\\n\\n\".join(strings))\n","sub_path":"concat.py","file_name":"concat.py","file_ext":"py","file_size_in_byte":321,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"104468538","text":"import sys\n#print(sys.path)\nimport os\n#获取项目路径下的目录\nos.chdir('D:\\\\Test_framework-master')\n#打印出项目路径下的目录\n#for file in os.listdir(os.getcwd()):\n# print(file)\nsys.path.append('D:\\\\Test_framework-master')\n\nimport time\nimport unittest\nfrom srcut.config import Config, DATA_PATH, REPORT_PATH\nfrom srcut.log import logger\nfrom srcut.file_reader import ExcelReader\nfrom srcut.HTMLTestRunner import HTMLTestRunner\n#from srcut.test_youj import Email\nfrom test.page.xm_home_page import XmHomePage\n\nclass TestXm(unittest.TestCase):\n URL = Config().get('URL')\n excel = DATA_PATH + '/test.xlsx'\n\n def sub_setUp(self):\n # 初始页面是main page,传入浏览器类型打开浏览器\n self.page = XmHomePage(browser_type='chrome').get(self.URL, maximize_window=False)\n\n def sub_tearDown(self):\n self.page.quit()\n\n def test_login(self):\n datas = ExcelReader(self.excel).data\n for d in datas:\n with self.subTest(data=d):\n self.sub_setUp()\n self.page.search(d['xxxx'])\n time.sleep(2)\n self.page = XmLoginPage(self.page) # 页面跳转到result page\n links = self.page.result_links\n for link in links:\n logger.info(link.text)\n self.sub_tearDown()\n\n\nif __name__ == '__main__':\n report = REPORT_PATH + '\\\\report.html'\n with open(report, 'wb') as f:\n runner = HTMLTestRunner(f, verbosity=2, title='测试网 mcf', description='修改html报告')\n runner.run(TestBaiDu('test_search'))\n # e = Email(title='测试网测试报告',\n # message='这是今天的测试报告,请查收!',\n # receiver='422703409@qq.com',\n # server='...',\n # sender='...',\n # password='...',\n # path=report\n # )\n # e.send()\n","sub_path":"test/case/test_baidu_6.py","file_name":"test_baidu_6.py","file_ext":"py","file_size_in_byte":1927,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"479387535","text":"import pickle\nfrom openke.utils import DeepDict\nfrom subgraphs import read_triples\n\nclass DynamicTopk:\n def __init__(self, default=10):\n self.topk_dict_head = DeepDict()\n self.topk_dict_tail = DeepDict()\n self.default_topk = default\n\n def get_dyn_topk(self, ent, rel, type_prediction):\n if type_prediction == \"head\":\n if (ent, rel) in self.topk_dict_head:\n return self.topk_dict_head[(ent, rel)]\n else:\n return self.default_topk\n elif type_prediction == \"tail\":\n if (ent, rel) in self.topk_dict_tail:\n return self.topk_dict_tail[(ent, rel)]\n else:\n return self.default_topk\n\n def populate(self, triples_file):\n triples = read_triples(triples_file)\n for triple in triples:\n if (triple[0], triple[2]) in self.topk_dict_tail:\n self.topk_dict_tail[(triple[0],triple[2])] += 1\n else:\n self.topk_dict_tail[(triple[0],triple[2])] = 1\n\n if (triple[1], triple[2]) in self.topk_dict_head:\n self.topk_dict_head[(triple[1], triple[2])] += 1\n else:\n self.topk_dict_head[(triple[1], triple[2])] = 1\n\n def load(self, dyn_topk_head_filename, dyn_topk_tail_filename):\n with open(dyn_topk_tail_filename, 'rb') as fin:\n self.topk_dict_tail = pickle.load(fin)\n\n with open(dyn_topk_head_filename, 'rb') as fin:\n self.topk_dict_head = pickle.load(fin)\n\n def save(self, dyn_topk_head_filename, dyn_topk_tail_filename):\n with open(dyn_topk_tail_filename, 'wb') as fout:\n pickle.dump(self.topk_dict_tail, fout, protocol = pickle.HIGHEST_PROTOCOL)\n\n with open(dyn_topk_head_filename, 'wb') as fout:\n pickle.dump(self.topk_dict_head, fout, protocol = pickle.HIGHEST_PROTOCOL)\n","sub_path":"dynamic_topk.py","file_name":"dynamic_topk.py","file_ext":"py","file_size_in_byte":1893,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"645019783","text":"# coding: utf-8\n\nfrom __future__ import absolute_import\n\nfrom swagger_server import util\nfrom swagger_server.models.base_model_ import Model\n\n\nclass SecurityMixin(Model):\n \"\"\"NOTE: This class is auto generated by the swagger code generator program.\n\n Do not edit the class manually.\n \"\"\"\n \n def __init__(self, identifier: str = None, securitykey: str = None, ip: str = None): # noqa: E501\n \"\"\"SecurityMixin - a model defined in Swagger\n\n :param identifier: The identifier of this SecurityMixin. # noqa: E501\n :type identifier: str\n :param securitykey: The securitykey of this SecurityMixin. # noqa: E501\n :type securitykey: str\n :param ip: The ip of this SecurityMixin. # noqa: E501\n :type ip: str\n \"\"\"\n self.swagger_types = {\n 'identifier': str,\n 'securitykey': str,\n 'ip': str\n }\n \n self.attribute_map = {\n 'identifier': 'identifier',\n 'securitykey': 'securitykey',\n 'ip': 'ip'\n }\n self._identifier = identifier\n self._securitykey = securitykey\n self._ip = ip\n \n @classmethod\n def from_dict(cls, dikt) -> 'SecurityMixin':\n \"\"\"Returns the dict as a model\n\n :param dikt: A dict.\n :type: dict\n :return: The SecurityMixin of this SecurityMixin. # noqa: E501\n :rtype: SecurityMixin\n \"\"\"\n return util.deserialize_model(dikt, cls)\n \n @property\n def identifier(self) -> str:\n \"\"\"Gets the identifier of this SecurityMixin.\n\n [] # noqa: E501\n\n :return: The identifier of this SecurityMixin.\n :rtype: str\n \"\"\"\n return self._identifier\n \n @identifier.setter\n def identifier(self, identifier: str):\n \"\"\"Sets the identifier of this SecurityMixin.\n\n [] # noqa: E501\n\n :param identifier: The identifier of this SecurityMixin.\n :type identifier: str\n \"\"\"\n if identifier is None:\n raise ValueError(\"Invalid value for `identifier`, must not be `None`\") # noqa: E501\n \n self._identifier = identifier\n \n @property\n def securitykey(self) -> str:\n \"\"\"Gets the securitykey of this SecurityMixin.\n\n [] # noqa: E501\n\n :return: The securitykey of this SecurityMixin.\n :rtype: str\n \"\"\"\n return self._securitykey\n \n @securitykey.setter\n def securitykey(self, securitykey: str):\n \"\"\"Sets the securitykey of this SecurityMixin.\n\n [] # noqa: E501\n\n :param securitykey: The securitykey of this SecurityMixin.\n :type securitykey: str\n \"\"\"\n if securitykey is None:\n raise ValueError(\"Invalid value for `securitykey`, must not be `None`\") # noqa: E501\n \n self._securitykey = securitykey\n \n @property\n def ip(self) -> str:\n \"\"\"Gets the ip of this SecurityMixin.\n\n [] # noqa: E501\n\n :return: The ip of this SecurityMixin.\n :rtype: str\n \"\"\"\n return self._ip\n \n @ip.setter\n def ip(self, ip: str):\n \"\"\"Sets the ip of this SecurityMixin.\n\n [] # noqa: E501\n\n :param ip: The ip of this SecurityMixin.\n :type ip: str\n \"\"\"\n if ip is None:\n raise ValueError(\"Invalid value for `ip`, must not be `None`\") # noqa: E501\n \n self._ip = ip\n","sub_path":"kms_api/swagger_server/models/security_mixin.py","file_name":"security_mixin.py","file_ext":"py","file_size_in_byte":3435,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"467580111","text":"import cv2\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Conv2D, Flatten, MaxPooling2D, BatchNormalization, LSTM, TimeDistributed\nfrom keras import optimizers\nfrom keras.preprocessing.sequence import pad_sequences\nfrom sklearn import preprocessing\nfrom keras.utils import to_categorical\nfrom sklearn.model_selection import train_test_split\n\nmodel = Sequential()\n\n\ndef training_model(X_input, Y_output, num_of_classes, one_hot_encoder, le):\n global model\n in_shape = X_input[0].shape\n\n # conv 1\n model.add(Conv2D(32, kernel_size=3, input_shape=in_shape, strides=(1, 1), activation='relu', padding='same', name=\"c1\"))\n model.add(MaxPooling2D(pool_size=(3, 3), strides=(2, 2)))\n model.add(BatchNormalization())\n\n # conv 2\n model.add(Conv2D(64, kernel_size=3, strides=(2, 2), activation='relu', padding='same', name=\"c2\"))\n model.add(MaxPooling2D(pool_size=(3, 3), strides=(2, 2)))\n model.add(BatchNormalization())\n\n # conv 3\n model.add(Conv2D(96, kernel_size=3, strides=(1, 1), activation='relu', padding='same', name=\"c3\"))\n # //mai\n model.add(MaxPooling2D(pool_size=(3, 3), strides=(2, 2)))\n model.add(BatchNormalization())\n # //\n # conv 4\n # model.add(Conv2D(512, kernel_size=3, strides=(1, 1), activation='relu', padding='same', name=\"c4\"))\n #\n # # conv 5\n # model.add(Conv2D(512, kernel_size=3, strides=(1, 1), activation='relu', padding='same', name=\"c5\"))\n # model.add(MaxPooling2D(pool_size=(3, 3), strides=(2, 2)))\n\n # FC6\n model.add(TimeDistributed(Flatten()))\n model.add(Dense(256))\n\n # add LSTM\n model.add(LSTM(256, return_sequences=True))\n model.add(LSTM(256, return_sequences=True))\n model.add(Flatten())\n\n # model.add((Dense(128, activation='relu')))\n model.add((Dense(num_of_classes, activation='softmax')))\n print(model.summary())\n\n sgd = optimizers.SGD(lr=0.1)\n model.compile(optimizer='adam', loss='categorical_crossentropy', metrics=['accuracy']) # lsa hshof loss eh\n\n x_train, x_test, y_train, y_test = train_test_split(X_input, Y_output, test_size=0.2, random_state=0, shuffle=True)\n\n model.fit(x_train, y_train, epochs=50, batch_size=in_shape[0], shuffle=True, verbose=2,\n validation_data=(x_test, y_test))\n\n # result = model.evaluate(x_test, y_test, verbose=2)\n # print(\"Done testing\")\n\n # print(\"Test loss =\", result[0])\n # print(\"Test accuracy evaluation=\", result * 100)\n model.save_weights(\"wild_model_weights.h5\")\n\n prediction = model.predict(x_test)\n print(\"Predicting x_test:\")\n inverse_prediction = one_hot_encoder.inverse_transform(prediction.reshape(-1, num_of_classes))\n inverse_prediction = le.inverse_transform(inverse_prediction.astype(int)) # de kalmaat\n inverse_ytest = one_hot_encoder.inverse_transform(y_test.reshape(-1, num_of_classes))\n inverse_ytest = le.inverse_transform(inverse_ytest.astype(int))\n\n correct = 0\n total = 0\n for i in range(y_test.shape[0]):\n total += 1\n if inverse_prediction[i] == inverse_ytest[i]:\n correct += 1\n print(inverse_prediction[i], \"****************\", inverse_ytest[i])\n\n print(\"#correct:\", correct)\n print(\"total\", total)\n print(\"Wild model overall accuracy\", (correct / len(inverse_ytest)) * 100, \"%\")\n\n\ndef testing(padded_total_words_test, y_labels_test_encoded, one_hot_encoder, le, num_of_classes):\n global model\n prediction = model.predict(padded_total_words_test)\n\n print(\"Predicting One video:\")\n inverse_prediction = one_hot_encoder.inverse_transform(prediction.reshape(-1, num_of_classes))\n inverse_prediction = le.inverse_transform(inverse_prediction.astype(int))\n inverse_ytest = one_hot_encoder.inverse_transform(y_labels_test_encoded.reshape(-1, num_of_classes))\n inverse_ytest = le.inverse_transform(inverse_ytest.astype(int))\n correct = 0\n for i in range(y_labels_test_encoded.shape[0]):\n if inverse_prediction[i] == inverse_ytest[i]:\n correct += 1\n print(inverse_prediction[i], \"##################\", inverse_ytest[i])\n print(\"correct\", correct)\n print(\"total\", len(inverse_ytest))\n print(\"one video test Accuracy\", (correct / len(inverse_ytest)) * 100, \"%\")\n","sub_path":"wild_model.py","file_name":"wild_model.py","file_ext":"py","file_size_in_byte":4236,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"452628898","text":"from student import Student\nfrom fag import Fag\n\nclass Studentsystem:\n def __init__(self):\n self._studentListe = []\n self._fagListe = []\n\n# Oppgave 1\n def lesFraFil(self, filnavn):\n fil = open(filnavn)\n\n # *MAT1001\n for linje in fil:\n if linje[0] == \"*\":\n fag = Fag(linje[1:-1])\n self._fagListe.append(fag)\n else:\n student = finnStudent(linje[:-1])\n\n if student == None:\n student = Student(linje[:-1])\n self._studentListe.append(student)\n\n\n fag.leggTilStudent(student)\n student.leggTilFag(fag)\n\n\n def finnStudent(self, navn):\n student = None\n for stud in self._studentListe:\n if stud.hentStudentNavn() == navn:\n student = stud\n\n return student\n\n\n def finnFag(self, emnekode):\n fag = None\n\n for f in self._fagListe:\n if f.hentFagNavn() == emnekode:\n fag = f\n\n return fag\n\n\n\n# Oppgave 2\n def skrivStudent(self):\n navn = input(\"Oppgi navnet til person du vil hente oversikt til\\n\")\n\n student = finnStudent(navn)\n\n if student == None:\n print(navn + \" finnes ikke i systemet\")\n else:\n student.skrivFagPaaStudent()\n\n\n def skrivFag(self):\n kode = input(\"Oppgi emne til faget du vil hente oversikt til\\n\")\n\n fag = finnFag(kode)\n\n if fag == None:\n print(kode + \" finnes ikke i systemet\")\n\n def hentStudentMedFlestFag(self):\n studentMedFlest = None\n antallFlest = 0\n\n for stud in self._studentListe:\n if stud.hentAntallFag() > antallFlest:\n antallFlest = stud.hentAntallFag()\n studentMedFlest = stud\n\n print(\"Student med flest fag: \" + studentMedFlest.hentStudentNavn()\n + \" med antall: \" + antallFlest)\n\n\n def hentFagMedFlestStudenter(self):\n fag = None\n antallFlest = 0\n\n for f in self._fagListe:\n if (f.hentAntallStudent() > antallFlest):\n fag = f\n antallFlest = f.hentAntallStudent()\n\n print(\"Fag med flest studenter: \" + fag.hentFagNavn()\n + \" med antall: \" + antallFlest)\n\n\n# Oppgave 3\n def settInnStudent(self, navn):\n\n student = finnStudent(navn)\n\n if (student == None):\n self._studentListe.append(student)\n print(navn + \" lagt til\")\n\n else:\n print(navn + \" finnes allerede.\")\n\n\n def settInnFag(self, navn):\n fag = finnFag(navn)\n\n if (fag == None):\n self._fagListe.append(fag)\n print(navn + \" lagt til\")\n else:\n print(navn + \" finnes allerede.\")\n\n\n\n# Oppgave 4\n def leggTilStudentIFag(self):\n navn = input(\"Hva heter studenten du vil legge til i faget.\\n\")\n\n student = finnStudent(navn)\n\n if student == None:\n print(navn + \" finnes ikke i systemet, du må legge studenten inn i systemet først.\")\n return\n\n kode = inpurt(\"Hva heter emnet du vil legget til \" + navn + \"i\\n\")\n\n fag = finnFag(kode)\n\n if fag == None:\n print(kode + \" finnes ikke i systemet, du må legge faget inn i systemet først.\")\n return\n\n\n if student.tarFag(fag):\n print(navn + \" tar allerede faget.\")\n\n else:\n student.leggTilFag(fag)\n fag.leggTilStudent(student)\n print(navn + \" er lagt til i \" + kode)\n\n\n\n def ordrelokke(self):\n inntast = \"\"\n while inntast != \"q\":\n self.skrivMeny()\n inntast = input(\"Skriv inn ditt valg: \")\n\n if inntast == \"1\":\n self.leggTilStudent()\n elif inntast == \"2\":\n self.leggTilFag()\n elif inntast == \"3\":\n self.leggTilStudentIFag()\n elif inntast == \"4\":\n self.skrivFag()\n elif inntast == \"5\":\n self.skrivStudent()\n elif inntast == \"6\":\n self.finnFagMedFlestStudenter()\n elif inntast == \"7\":\n self.finnStudentMedFlestFag()\n #elif inntast == \"8\":\n #self.fjernStudentFraFag()\n #elif inntast == \"9\":\n #self.skrivAlt()\n elif inntast != \"q\":\n print(\"Ugylig input.\\n\")\n\n print(\"Avslutter programmet\")\n\n def skrivMeny(self):\n print(\"--Meny--\")\n print(\"1 - Legg til ny student\")\n print(\"2 - Legg til nytt fag\")\n print(\"3 - Legg til student i fag\")\n print(\"4 - Skriv ut studenter ved fag\")\n print(\"5 - Skriv ut alle fag til student\")\n print(\"6 - Finn fag som blir tatt av flest\")\n print(\"7 - Finn student som tar flest fag\")\n #print(\"8 - Fjern student fra fag\")\n #print(\"9 - Fullstendig oversikt\")\n print(\"q - Avslutt\")\n\n #HVIS TID: 9 - fullstendig oversikt\n def skrivAlt(self):\n for fag in self._alleFag:\n fag.skrivStudenterVedFag()\n\n #HVIS TID:\n #11-7\n def fjernStudentFraFag(self):\n navn = input(\"Hva heter studenten du vil fjerne fra faget?\")\n stud = self.finnStudent(navn)\n if stud == None:\n print(navn + \" finnes ikke.\")\n return\n\n fagNavn = input(\"Fra hvilket fag vil du fjerne \" + navn +\"?\")\n fag = self.finnFag(fagNavn)\n if fag == None:\n print(fagNavn + \" finnes ikke.\")\n return\n\n #sjekker om studenten tar faget, hvis den ikke tar det gjoer vi ikke noe.\n if not stud.tarFag(fag):\n print(navn + \" tar ikke \" + fagNavn)\n else:\n #hvis studenten finnes, faget finnes, og studenten faktisk tar faget, saa kan vi fjerne den!\n stud.fjernFag(fag)\n fag.fjernStudent(stud)\n\n print(navn + \" fjernet fra \" + fagNavn)\n","sub_path":"uke10/studentsystem.py","file_name":"studentsystem.py","file_ext":"py","file_size_in_byte":5997,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"390944568","text":"import numpy as np\nfrom abc import ABC, abstractmethod\nfrom skimage.draw import circle, rectangle\n\n\ndef circular_mask(shape):\n mask = np.zeros(shape, dtype=np.uint8)\n rr, cc = circle(shape[0]/2, shape[1]/2, radius=shape[0] / 3, shape=shape)\n mask[rr, cc] = 1\n \n return mask\n\n\ndef striped_mask(shape):\n mask = np.zeros(shape, dtype=np.uint8)\n mask[::2] = 1\n\n return mask\n\n\ndef concentric_rectangle_mask(shape, width):\n mask = np.ones(shape, dtype=np.uint8)\n rect = np.zeros(shape, dtype=np.uint8)\n\n for i in range(1, int(shape[0] / (2 * width)), 2):\n rect = np.zeros(shape, dtype=np.uint8)\n start = (i*width, i*width)\n end = (shape[0] - i*width, shape[1] - i*width)\n rr, cc = rectangle(start=start, end=end, shape=mask.shape)\n rect[rr, cc] = 1\n mask -= rect\n\n rect = np.zeros(shape, dtype=np.uint8)\n start = ((i+1) * width, (i+1) * width)\n end = (shape[0] - (i+1) * width, shape[1] - (i+1) * width)\n rr, cc = rectangle(start=start, end=end, shape=mask.shape)\n rect[rr, cc] = 1\n mask += rect\n\n\n return mask\n\n\nclass AbstractProcessing(ABC):\n \"\"\" Base class for post-processing. \"\"\"\n @abstractmethod\n def apply(self, *args, **kwargs):\n pass\n\n\nclass Quantize(AbstractProcessing):\n \"\"\" Apply quantization to each frame of a given 3D input. \"\"\"\n def __init__(self, bins=2):\n self.bins = bins\n\n def apply(self, images):\n w = images.max() / self.bins\n\n for i in range(images.shape[0]):\n images[i, :, :] -= (images[i, :, :] - (images[i, :, :] // w) * w).astype('uint8')\n\n return images\n\n\nclass AdjustBrightness(AbstractProcessing):\n \"\"\" Gamma < 1 will decrease brightness, Gamma > 1 will increase it. \"\"\"\n def __init__(self, gamma):\n self.gamma = gamma\n \n def apply(self, images):\n # Normalize, then apply brightness correction\n images = (images / images.max()) ** (1 / self.gamma)\n # Convert back to grayscale [0, 255]\n images = ((images - images.min()) * (1 / (images.max() - images.min()) * 255)).astype('uint8')\n\n return images\n\n\nclass Mask(AbstractProcessing):\n \"\"\" Apply a binary mask to each frame of a given 3D input. \"\"\"\n def __init__(self, mask):\n self.mask = mask\n \n def apply(self, images):\n images *= self.mask\n\n return images\n\n\nclass Border(AbstractProcessing):\n def __init__(self, margin, width):\n self.margin = margin\n self.width = width\n \n def apply(self, images):\n # White border\n images[:, self.margin:self.margin + self.width, :] = 255\n images[:, -self.margin - self.width:-self.margin, :] = 255\n images[:, :, self.margin:self.margin + self.width] = 255\n images[:, :, -self.margin - self.width:-self.margin] = 255\n \n # Black margin\n images[:, 0:self.margin, :] = 0\n images[:, -self.margin:, :] = 0\n images[:, :, 0:self.margin] = 0\n images[:, :, -self.margin:] = 0\n \n return images\n\n\nclass FromFunction(AbstractProcessing):\n \"\"\" Not tested, not fully compatible yet. \"\"\"\n def __init__(self, fn=None, *args, **kwargs):\n self.fn = fn\n self.args = args\n self.kwargs = kwargs\n \n def apply(self, images):\n for i in range(images.shape[0]):\n images[i, :, :] = self.fn(images[i, :, :], *self.args, **self.kwargs)\n\n return images\n\n\nclass Pipeline():\n \"\"\" Define an AbstractProcessing pipeline object. \"\"\"\n def __init__(self, *args):\n self._processing_list = args\n\n def run(self, images):\n if not self.is_empty():\n for f in self._processing_list:\n images = f.apply(images)\n\n return images\n \n def is_empty(self):\n return len(self._processing_list) == 0\n","sub_path":"postprocessing.py","file_name":"postprocessing.py","file_ext":"py","file_size_in_byte":3863,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"242046373","text":"import numpy as np\nimport numpy.linalg\nimport math\nimport time\nfrom utils.io import *\nfrom node import *\nfrom scipy.spatial.distance import euclidean\nfrom scipy.fftpack import fftn, ifftn, ifft\nfrom scipy.special import jv\nfrom new_hp import *\n\n\ngivals = [\n 22026.5, 20368, 18840.3, 17432.5, 16134.8, 14938.4, 13834.9, 12816.8,\n 11877.4, 11010.2, 10209.4, 9469.8, 8786.47, 8154.96, 7571.17, 7031.33,\n 6531.99, 6069.98, 5642.39, 5246.52, 4879.94, 4540.36, 4225.71, 3934.08,\n 3663.7, 3412.95, 3180.34, 2964.5, 2764.16, 2578.14, 2405.39, 2244.9,\n 2095.77, 1957.14, 1828.24, 1708.36, 1596.83, 1493.05, 1396.43, 1306.47,\n 1222.68, 1144.62, 1071.87, 1004.06, 940.819, 881.837, 826.806, 775.448,\n 727.504, 682.734, 640.916, 601.845, 565.329, 531.193, 499.271, 469.412,\n 441.474, 415.327, 390.848, 367.926, 346.454, 326.336, 307.481, 289.804,\n 273.227, 257.678, 243.089, 229.396, 216.541, 204.469, 193.129, 182.475,\n 172.461, 163.047, 154.195, 145.868, 138.033, 130.659, 123.717, 117.179,\n 111.022, 105.22, 99.7524, 94.5979, 89.7372, 85.1526, 80.827, 76.7447,\n 72.891, 69.2522, 65.8152, 62.5681, 59.4994, 56.5987, 53.856, 51.2619,\n 48.8078, 46.4854, 44.2872, 42.2059, 40.2348, 38.3676, 36.5982, 34.9212,\n 33.3313, 31.8236, 30.3934, 29.0364, 27.7485, 26.526, 25.365, 24.2624,\n 23.2148, 22.2193, 21.273, 20.3733, 19.5176, 18.7037, 17.9292, 17.192,\n 16.4902, 15.822, 15.1855, 14.579, 14.0011, 13.4503, 12.9251, 12.4242,\n 11.9464, 11.4905, 11.0554, 10.6401, 10.2435, 9.86473, 9.50289, 9.15713,\n 8.82667, 8.51075, 8.20867, 7.91974, 7.64333, 7.37884, 7.12569, 6.88334,\n 6.65128, 6.42902, 6.2161, 6.01209, 5.81655, 5.62911, 5.44938, 5.27701,\n 5.11167, 4.95303, 4.80079, 4.65467, 4.51437, 4.37966, 4.25027, 4.12597,\n 4.00654, 3.89176, 3.78144, 3.67537, 3.57337, 3.47528, 3.38092, 3.29013,\n 3.20276, 3.11868, 3.03773, 2.9598, 2.88475, 2.81247, 2.74285, 2.67577,\n 2.61113, 2.54884, 2.48881, 2.43093, 2.37513, 2.32132, 2.26944, 2.21939,\n 2.17111, 2.12454, 2.07961, 2.03625, 1.99441, 1.95403, 1.91506, 1.87744,\n 1.84113, 1.80608, 1.77223, 1.73956, 1.70802, 1.67756, 1.64815, 1.61976,\n 1.59234, 1.56587, 1.54032, 1.51564, 1.49182, 1.46883, 1.44664, 1.42522,\n 1.40455, 1.3846, 1.36536, 1.3468, 1.3289, 1.31164, 1.29501, 1.27898,\n 1.26353, 1.24866, 1.23434, 1.22056, 1.2073, 1.19456, 1.18231, 1.17055,\n 1.15927, 1.14844, 1.13807, 1.12814, 1.11864, 1.10956, 1.10089, 1.09262,\n 1.08475, 1.07727, 1.07017, 1.06345, 1.05709, 1.05109, 1.04545, 1.04015,\n 1.03521, 1.0306, 1.02633, 1.02239, 1.01878, 1.0155, 1.01253, 1.00989,\n 1.00756, 1.00555, 1.00385, 1.00246, 1.00139, 1.00062, 1.00015, 1\n]\n\ndef GI(w,h,d, img, max_intensity, min_intensity):\n index = (int)((img[w][h][d] - min_intensity)/max_intensity * 255)\n if index > 255:\n index = 255\n return givals[index]\n\n\"\"\"\nExhaustive Tracing\n\n\"\"\"\ndef exhaustive_tracing(img, bimg, dt_result, timemap, size, seed, max_intensity,threshold,out_path,r_iter,coverage_ratio):\n # state 0 for FAR, state 1 for TRAIL, state 2 for ALIVE\n state = np.zeros((size[0], size[1], size[2]))\n result = []\n\n # initialize \n tbimg = np.copy(bimg)\n phi = np.empty((size[0], size[1], size[2]), dtype=np.float32)\n parent = np.empty((size[0], size[1], size[2]), dtype=np.int32)\n prev = np.empty((size[0], size[1], size[2]), dtype=np.int32)\n swc_index = np.empty((size[0], size[1], size[2]), dtype=np.int32)\n\n for i in range(size[0]):\n phi[i,:,:] = np.inf\n\n current_index = 0\n # put seed into ALIVE set\n state[seed[0],seed[1],seed[2]] = 2\n phi[seed[0],seed[1],seed[2]] = 0.0\n swc_index[seed[0],seed[1],seed[2]] = 1\n prev[seed[0],seed[1],seed[2]] = 1\n\n # trail set structure[phi,w,h,d,par_id]\n trail_set = np.asarray([[0,seed[0],seed[1],seed[2]]],dtype=np.float32)\n\n # alive set structure: [id,radius,w,h,d,1,par_id]\n alive_set = None\n starttime = time.time()\n totaltime = 0\n counter = 0\n while (trail_set.size != 0):\n counter+=1\n min_ind = trail_set[0,:]\n\n trail_set = np.delete(trail_set, (0), axis=0)\n i,j,k = min_ind[1:4]\n i = int(i)\n j = int(j)\n k = int(k)\n prev_ind = prev[i,j,k]\n parent[i,j,k] = prev_ind\n\n state[i][j][k] = 2\n swc_index[i][j][k] = current_index\n # print(alive_set.shape)\n if alive_set is None:\n alive_set = np.asarray([[current_index,3,i,j,k,1,0]],dtype=np.int32)\n else:\n alive_set = np.vstack((alive_set,[current_index,3,i,j,k,1,prev_ind]))\n # print('alive:',alive_set)\n\n tbimg[i][j][k] = 2\n current_index += 1\n\n neighbor_ind = get_neighbor_ind(img.shape,i-1,i+2,j-1,j+2,k-1,k+2)\n\n for ind in neighbor_ind:\n factor = 1\n if ind[3] == 2:\n factor = 1.414214\n\n w,h,d = ind[0:3]\n\n if (img[w,h,d] <= threshold and\n img[i,j,k] <= threshold):\n continue\n\n if (state[w][h][d] != 2):\n # min_intensity set as 0\n new_dist = phi[w][h][d] + (GI(\n w,h,d, img, max_intensity, 0.0) + GI(\n i,j,k, img, max_intensity, 0.0)) * factor * 0.5\n \n prev_ind = swc_index[i][j][k]\n\n if (state[w,h,d] == 0):\n phi[w,h,d] = new_dist\n # insert into trail set\n if trail_set.shape[0] == 0:\n trail_set = np.asarray([[new_dist,w,h,d]],dtype=np.float32)\n else:\n trail_set = np.vstack((trail_set,[new_dist,w,h,d]))\n trail_set = trail_set[np.argsort(trail_set[:,0])]\n\n prev[w][h][d] = prev_ind\n state[w][h][d] = 1\n\n elif (state[w][h][d] == 1):\n if (phi[w][h][d] > new_dist):\n phi[w][h][d] = new_dist\n temp_ind = np.argwhere((trail_set[:,1] == w) & (trail_set[:,2] == h) & (trail_set[:,3] == d))[0]\n trail_set[temp_ind][0] = new_dist\n trail_set = trail_set[np.argsort(trail_set[:,0])]\n sort_time = time.time()\n prev[w][h][d] = prev_ind\n\n print('alive size:',alive_set.shape)\n ini = alive_set.copy()\n swc_x = ini[:, 2].copy()\n swc_y = ini[:, 3].copy()\n ini[:, 2] = swc_y\n ini[:, 3] = swc_x\n saveswc(out_path + 'ini.swc',ini) \n bb = np.zeros(img.shape) \n hp_result,bb = hp(img,bimg,size,alive_set,out_path,threshold,bb,1,bimg,coverage_ratio)\n result = hp_result\n print(result.shape)\n\n if r_iter == 0:\n swc_x = result[:, 2].copy()\n swc_y = result[:, 3].copy()\n result[:, 2] = swc_y\n result[:, 3] = swc_x\n saveswc(out_path + str(r_iter) + 'result.swc',result)\n return\n\n # enhanced iteration\n far = np.argwhere(bimg == 1)\n if far.shape[0] != 0:\n no_iteration = 0\n # current_index += 1\n far_timemap = np.array([[]])\n for f in far:\n # if (bimg[f[0]][f[1]][f[2]]] == 1):\n if far_timemap.shape[1] == 0:\n far_timemap = np.asarray([[f[0],f[1],f[2],timemap[f[0]][f[1]][f[2]]]])\n else:\n far_timemap = np.vstack((far_timemap,[f[0],f[1],f[2],timemap[f[0]][f[1]][f[2]]]))\n sort_timemap = far_timemap[np.argsort(far_timemap[:,3])]\n sort_timemap = sort_timemap[::-1]\n\n alive_loc = alive_set[2:5]\n\n \n while (far.size > 0 and sort_timemap.size > 0):\n # alive_set = []\n padding_index = current_index\n\n if(no_iteration >= r_iter):\n break\n\n #UPDATE \n trail_set = np.asarray([[0,sort_timemap[0][0],sort_timemap[0][1],sort_timemap[0][2]]])\n\n new_alive = np.asarray([[]])\n\n while (trail_set.size != 0):\n min_ind = trail_set[0,:]\n\n trail_set = np.delete(trail_set, (0), axis=0)\n i = int(min_ind[1])\n j = int(min_ind[2])\n k = int(min_ind[3])\n if state[i][j][k] != 3:\n print(state[i][j][k])\n\n prev_ind = prev[i][j][k]\n parent[i][j][k] = prev_ind\n\n state[i][j][k] = 4\n swc_index[i][j][k] = current_index\n\n if(new_alive.shape[1] == 0):\n new_alive = np.asarray([[0,3,i,j,k,1,-1]])\n alive_set = np.vstack((alive_set,[current_index,3,i,j,k,1,-1]))\n else:\n p_ind = prev_ind-padding_index\n new_alive = np.vstack((new_alive,[current_index-padding_index,3,i,j,k,1,p_ind]))\n alive_set = np.vstack((alive_set,[current_index,3,i,j,k,1,prev_ind]))\n tbimg[i][j][k] = 2\n current_index += 1\n totaltime+=(time.time()-starttime)\n\n neighbor_ind = get_neighbor_ind(img.shape,i-1,i+2,j-1,j+2,k-1,k+2)\n\n for ind in neighbor_ind:\n\n w,h,d = ind[0:3]\n factor = 1\n if ind[3] == 2:\n factor = 1.414214\n\n if (img[w][h][d] <= threshold):\n continue\n\n if (state[w][h][d] == 1 or state[w][h][d] == 2):\n break\n\n if (state[w][h][d] != 4):\n # min_intensity set as 0\n new_dist = phi[w][h][d] + (GI(\n w,h,d, img, max_intensity, 0.0) + GI(\n i,j,k, img, max_intensity, 0.0)\n ) * factor * 0.5\n prev_ind = swc_index[i][j][k]\n\n if (state[w][h][d] == 0):\n phi[w][h][d] = new_dist\n # insert into trail set\n if trail_set.shape[1] == 0:\n trail_set = np.vstack((trail_set,[new_dist,w,h,d]))\n else:\n sort_time = time.time()\n trail_set = np.vstack((trail_set,[new_dist,w,h,d]))\n trail_set = trail_set[np.argsort(trail_set[:,0])]\n\n prev[w][h][d] = prev_ind\n # 3 for reinforce trail\n state[w][h][d] = 3\n\n elif (state[w][h][d] == 3):\n if (phi[w][h][d] > new_dist):\n phi[w][h][d] = new_dist\n temp_ind = np.argwhere((trail_set[:,1] == w) & (trail_set[:,2] == h) & (trail_set[:,3] == d))[0]\n trail_set[temp_ind][0] = new_dist\n trail_set = trail_set[np.argsort(trail_set[:,0])]\n sort_time = time.time()\n prev[w][h][d] = prev_ind \n if(new_alive.size == 0):\n no_iteration += 1\n continue\n # print('new_alive shape',new_alive.shape)\n new = new_alive.copy()\n swc_x = new[:, 2].copy()\n swc_y = new[:, 3].copy()\n new[:, 2] = swc_y\n new[:, 3] = swc_x\n\n hp_result,bb = hp(img,bimg,size,new_alive,out_path,threshold,bb,2,bimg,coverage_ratio)\n\n # print('no of enhanced iteration: ',no_iteration)\n if(hp_result is None or hp_result.shape[1] == 0):\n no_iteration += 1\n continue\n\n # print('padding index', padding_index)\n hp_result[:,0] += padding_index\n hp_result[:,6] += padding_index\n hp_result[:,5] = 1\n result = np.vstack((result,hp_result))\n sort_timemap = np.delete(sort_timemap,0)\n no_iteration += 1\n \n print('--Enhanced tracing finished')\n print('--Enhanced iteration: %.2f sec.' % (time.time() - starttime))\n r = alive_set\n\n swc_x = alive_set[:, 2].copy()\n swc_y = alive_set[:, 3].copy()\n alive_set[:, 2] = swc_y\n alive_set[:, 3] = swc_x\n\n swc_x = result[:, 2].copy()\n swc_y = result[:, 3].copy()\n result[:, 2] = swc_y\n result[:, 3] = swc_x\n saveswc(out_path + '_ini.swc',alive_set)\n saveswc(out_path + '_result.swc',result)\n\n return r\n\ndef get_neighbor_ind(size,i_min,i_max,j_min,j_max,k_min,k_max):\n result = []\n i = i_min + 1\n j = j_min + 1\n k = k_min + 1\n if i-1 < 0:\n i_min = i\n if i+1 >= size[0]:\n i_max = i+1\n if j-1 < 0:\n j_min = j\n if j+1 >= size[1]:\n j_max = j+1\n if k-1 < 0:\n k_min = k\n if k+1 >= size[2]:\n k_max = k+1\n\n for kk in range(k_min,k_max):\n for jj in range(j_min,j_max):\n for ii in range(i_min,i_max):\n offset = abs(k-kk) + abs(j-jj) + abs(i-ii)\n if offset <= 2 and offset != 0:\n result.append([ii,jj,kk,offset])\n\n return np.asarray(result)\n\n\ndef crop(img,spatial1,spatial2):\n \"\"\"Crop a 3D block with value > thr\"\"\"\n minx = int(np.minimum(spatial1[0],spatial2[0]))\n maxx = int(np.maximum(spatial1[0],spatial2[0]))\n miny = int(np.minimum(spatial1[1],spatial2[1]))\n maxy = int(np.maximum(spatial1[1],spatial2[1]))\n minz = int(np.minimum(spatial1[2],spatial2[2]))\n maxz = int(np.maximum(spatial1[2],spatial2[2]))\n return img[minx:maxx, miny:maxy, minz:maxz]\n\ndef enhance(img,spatial1,spatial2,img2):\n minx = int(np.minimum(spatial1[0],spatial2[0]))\n maxx = int(np.maximum(spatial1[0],spatial2[0]))\n miny = int(np.minimum(spatial1[1],spatial2[1]))\n maxy = int(np.maximum(spatial1[1],spatial2[1]))\n minz = int(np.minimum(spatial1[2],spatial2[2]))\n maxz = int(np.maximum(spatial1[2],spatial2[2]))\n img[minx:maxx, miny:maxy, minz:maxz] = img2\n return img\n\n\ndef response(img, radii,rsptype='oof'):\n eps = 1e-12\n rsp = np.zeros(img.shape)\n # bar = progressbar.ProgressBar(max_value=kwargs['radii'].size)\n # bar.update(0)\n\n W = np.zeros((img.shape[0], img.shape[1], img.shape[2], 3)) # Eigen values to save\n V = np.zeros((img.shape[0], img.shape[1], img.shape[2], 3, 3)) # Eigen vectors to save\n\n if rsptype == 'oof' :\n rsptensor = ooftensor(img, radii)\n\n # pbar = tqdm(total=len(radii))\n for i, tensorfield in enumerate(rsptensor):\n # Make the tensor from tensorfield\n f11, f12, f13, f22, f23, f33 = tensorfield\n tensor = np.stack((f11, f12, f13, f12, f22, f23, f13, f23, f33), axis=-1)\n del f11\n del f12\n del f13\n del f22\n del f23\n del f33\n tensor = tensor.reshape(img.shape[0], img.shape[1], img.shape[2], 3, 3)\n w, v = np.linalg.eigh(tensor)\n del tensor\n sume = w.sum(axis=-1)\n nvox = img.shape[0] * img.shape[1] * img.shape[2]\n sortidx = np.argsort(np.abs(w), axis=-1)\n sortidx = sortidx.reshape((nvox, 3))\n\n # Sort eigenvalues according to their abs\n w = w.reshape((nvox, 3))\n for j, (idx, value) in enumerate(zip(sortidx, w)):\n w[j,:] = value[idx]\n w = w.reshape(img.shape[0], img.shape[1], img.shape[2], 3)\n\n # Sort eigenvectors according to their abs\n v = v.reshape((nvox, 3, 3))\n for j, (idx, vec) in enumerate(zip(sortidx, v)):\n v[j,:,:] = vec[:, idx]\n del sortidx\n v = v.reshape(img.shape[0], img.shape[1], img.shape[2], 3, 3)\n\n mine = w[:,:,:, 0]\n mide = w[:,:,:, 1]\n maxe = w[:,:,:, 2]\n\n if rsptype == 'oof':\n feat = maxe\n elif rsptype == 'bg':\n feat = -mide / maxe * (mide + maxe) # Medialness measure response\n cond = sume >= 0\n feat[cond] = 0 # Filter the non-anisotropic voxels\n\n del mine\n del maxe\n del mide\n del sume\n\n cond = np.abs(feat) > np.abs(rsp)\n W[cond, :] = w[cond, :]\n V[cond, :, :] = v[cond, :, :]\n rsp[cond] = feat[cond]\n del v\n del w\n del tensorfield\n del feat\n del cond\n # pbar.update(1)\n # print('rsp value',np.max(rsp),np.min(rsp))\n\n return rsp, V, W\n\n\n\ndef ooftensor(img, radii, memory_save=True):\n '''\n type: oof, bg\n '''\n # sigma = 1 # TODO: Pixel spacing\n eps = 1e-12\n # ntype = 1 # The type of normalisation\n fimg = fftn(img, overwrite_x=True)\n shiftmat = ifftshiftedcoormatrix(fimg.shape)\n x, y, z = shiftmat\n x = x / fimg.shape[0]\n y = y / fimg.shape[1]\n z = z / fimg.shape[2]\n kernel_radius = np.sqrt(x ** 2 + y ** 2 + z ** 2) + eps # The distance from origin\n\n for r in radii:\n # Make the fourier convolutional kernel\n jvbuffer = oofftkernel(kernel_radius, r) * fimg\n\n if memory_save:\n # F11\n buffer = ifftshiftedcoordinate(img.shape, 0) ** 2 * x * x * jvbuffer\n buffer = ifft(buffer, axis=0)\n buffer = ifft(buffer, axis=1)\n buffer = ifft(buffer, axis=2)\n f11 = buffer.copy()\n\n # F12\n buffer = ifftshiftedcoordinate(img.shape, 0) * ifftshiftedcoordinate(img.shape, 1) * x * y * jvbuffer\n buffer = ifft(buffer, axis=0)\n buffer = ifft(buffer, axis=1)\n buffer = ifft(buffer, axis=2)\n f12 = buffer.copy()\n\n # F13\n buffer = ifftshiftedcoordinate(img.shape, 0) * ifftshiftedcoordinate(img.shape, 2) * x * z * jvbuffer\n buffer = ifft(buffer, axis=0)\n buffer = ifft(buffer, axis=1)\n buffer = ifft(buffer, axis=2)\n f13 = buffer.copy()\n\n # F22\n buffer = ifftshiftedcoordinate(img.shape, 1) ** 2 * y ** 2 * jvbuffer\n buffer = ifft(buffer, axis=0)\n buffer = ifft(buffer, axis=1)\n buffer = ifft(buffer, axis=2)\n f22 = buffer.copy()\n\n # F23\n buffer = ifftshiftedcoordinate(img.shape, 1) * ifftshiftedcoordinate(img.shape, 2) * y * z * jvbuffer\n buffer = ifft(buffer, axis=0)\n buffer = ifft(buffer, axis=1)\n buffer = ifft(buffer, axis=2)\n f23 = buffer.copy()\n\n # F33\n buffer = ifftshiftedcoordinate(img.shape, 2) * ifftshiftedcoordinate(img.shape, 2) * z * z * jvbuffer\n buffer = ifft(buffer, axis=0)\n buffer = ifft(buffer, axis=1)\n buffer = ifft(buffer, axis=2)\n f33 = buffer.copy()\n else:\n f11 = np.real(ifftn(x * x * jvbuffer))\n f12 = np.real(ifftn(x * y * jvbuffer))\n f13 = np.real(ifftn(x * z * jvbuffer))\n f22 = np.real(ifftn(y * y * jvbuffer))\n f23 = np.real(ifftn(y * z * jvbuffer))\n f33 = np.real(ifftn(z * z * jvbuffer))\n yield [f11, f12, f13, f22, f23, f33]\n\ndef ifftshiftedcoormatrix(shape):\n shape = np.asarray(shape)\n p = np.floor(np.asarray(shape) / 2).astype('int')\n coord = []\n for i in range(shape.size):\n a = np.hstack((np.arange(p[i], shape[i]), np.arange(0, p[i]))) - p[i] - 1.\n repmatpara = np.ones((shape.size,)).astype('int')\n repmatpara[i] = shape[i]\n A = a.reshape(repmatpara)\n repmatpara = shape.copy()\n repmatpara[i] = 1\n coord.append(np.tile(A, repmatpara))\n\n return coord\n\n\ndef ifftshiftedcoordinate(shape, axis):\n shape = np.asarray(shape)\n p = np.floor(np.asarray(shape) / 2).astype('int')\n a = (np.hstack((np.arange(p[axis], shape[axis]), np.arange(0, p[axis]))) - p[axis] - 1.).astype('float')\n a /= shape[axis].astype('float')\n reshapepara = np.ones((shape.size,)).astype('int');\n reshapepara[axis] = shape[axis];\n A = a.reshape(reshapepara);\n repmatpara = shape.copy();\n repmatpara[axis] = 1;\n return np.tile(A, repmatpara)\n\ndef oofftkernel(kernel_radius, r, sigma=1, ntype=1):\n eps = 1e-12\n normalisation = 4/3 * np.pi * r**3 / (jv(1.5, 2*np.pi*r*eps) / eps ** (3/2)) / r**2 * \\\n (r / np.sqrt(2.*r*sigma - sigma**2)) ** ntype\n jvbuffer = normalisation * np.exp( (-2 * sigma**2 * np.pi**2 * kernel_radius**2) / (kernel_radius**(3/2) ))\n return (np.sin(2 * np.pi * r * kernel_radius) / (2 * np.pi * r * kernel_radius) - np.cos(2 * np.pi * r * kernel_radius)) * \\\n jvbuffer * np.sqrt( 1./ (np.pi**2 * r *kernel_radius ))\n\n","sub_path":"exhaustive_tracing.py","file_name":"exhaustive_tracing.py","file_ext":"py","file_size_in_byte":20237,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"185326344","text":"def min_max(arr):\n min=arr[0]\n max=arr[0]\n sum=0\n l=len(arr)\n for x in range(l):\n if min > arr[x]:\n min=arr[x]\n if max < arr[x]:\n max=arr[x]\n sum=sum+arr[x]\n temp=max\n max=sum-min\n min=sum-temp\n print(max,' ',min)\nmin_max([1,2,3,4,5])","sub_path":"problem1/problem13.py","file_name":"problem13.py","file_ext":"py","file_size_in_byte":305,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"608910511","text":"# coding=utf-8\r\nimport os\r\nclass MDFileSplitter:\r\n\r\n def __init__(self, s):\r\n self.path = s\r\n self.file = open(s)\r\n self.article_summary = \"\"\r\n self.splitter()\r\n\r\n def splitter(self):\r\n url = os.path.basename(self.path)\r\n title = self.file.readline().replace(' ', '').replace('\\'', '').replace('title:', '').strip() # title\r\n time_in = map(lambda x: x.strip(), str(self.file.readline()).split(' ')) # time\r\n time_str = time_in[1] + ' ' + time_in[2]\r\n categories = self.file.readline().replace(' ', '').replace('categories:', '') # categories\r\n # skip article splitter\r\n self.file.readline()\r\n # contents\r\n content = self.file.read()\r\n\r\n # summary\r\n self.article_summary = {'url': url, 'title': title, 'TimeStr': time_str, 'Categories': categories,\r\n 'Content': content}\r\n\r\n def get_info(self):\r\n return self.article_summary\r\n","sub_path":"mdfilesplitter.py","file_name":"mdfilesplitter.py","file_ext":"py","file_size_in_byte":981,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"479738392","text":"import sys\n\nimport matplotlib\nmatplotlib.use('Agg')\n\nimport numpy as np\nimport scipy.stats as stats\nimport matplotlib.pyplot as plt\n\nargs = sys.argv\n\ntrial_file = args[1]\nscore_file = args[2]\nsave_file = args[3]\n\ntgt_scores = []\nnontgt_scores = []\n\nwith open(trial_file, 'r') as tf, open(score_file, 'r') as sf:\n tf_lines = tf.read().splitlines()\n trial_types = [x.split()[2] for x in tf_lines]\n sf_lines = sf.read().splitlines()\n scores = [float(x.split()[2]) for x in sf_lines]\n\n for ttype, score in zip(trial_types, scores):\n if ttype == \"target\":\n tgt_scores.append(score)\n else:\n nontgt_scores.append(score)\n\ntgt_scores = sorted(tgt_scores)\nnontgt_scores = sorted(nontgt_scores)\n\nfit_tgt = stats.norm.pdf(tgt_scores, np.mean(tgt_scores), np.std(tgt_scores))\nfit_nontgt = stats.norm.pdf(nontgt_scores, np.mean(nontgt_scores), np.std(nontgt_scores))\n\nplt.plot(tgt_scores, fit_tgt, '-g')\nplt.hist(tgt_scores, bins=20, normed=True)\n\nplt.plot(nontgt_scores, fit_nontgt, '-r')\nplt.hist(nontgt_scores, bins=20, normed=True)\n\nsp = save_file.split('_')\nif len(sp) > 9:\n title = \"PPG type = \"+sp[-9]\n title = title+\", Rand level = \"+sp[-7]\n title = title+\", Cross-gender =\"+sp[-5]\n title = title+\", Distance = \"+sp[-3]\n title = title+\", Proximity = \"+sp[-1]\n plt.title(title+'.')\n\nplt.savefig(save_file, dpi=300)\n\n\n","sub_path":"egs/voice_privacy/v1/local/plot/plot_trial_score_dist.py","file_name":"plot_trial_score_dist.py","file_ext":"py","file_size_in_byte":1379,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"242958784","text":"#!python3\n# coding:utf-8\n\n'''\nscreen-length 0 temporary\ndisplay interface brief main\ndisplay interface description | exclude \\.\ndisplay lldp neighbor\n'''\n\n'''\n['Interface', 'PHY', 'Protocol', 'InUti', 'OutUti', 'inErrors', 'outErrors']\n['Interface', 'PHY', 'Protocol', 'Description']\n['Local Intf', 'Neighbor Dev', 'Neighbor Intf', 'Exptime (sec)']\n'''\n\n'''\n取消屏幕显示长度\n获取 聚合组 与 物理口 的关系、获取 物理up/down\n获取 端口描述\n获取 对端实际设备信息\n'''\n\n\nfrom datetime import datetime\nimport platform\nfrom colorama import Fore, Back, Style, init\nimport pandas as pd\nfrom pathlib import Path\nimport re\n_version = '1.0'\n\n'''\n# ChangeLog\n\n### v1.0 (2021-09-30) \n\n- 根据采集信息整理输出\n'''\n\ninit(autoreset=True)\nif platform.system() == 'Windows':\n init(wrap=True)\nfile_names = Path().rglob('*.log')\ncontents = []\nfor file in file_names:\n if file == 'AutoRun.log':\n continue\n if '\\\\' in str(file):\n continue\n try:\n with open(file, encoding='utf-8') as f:\n contents.append(f.read().splitlines())\n except:\n ...\n\nrecord_list = ['Interface PHY Protocol InUti OutUti inErrors outErrors',\n 'Interface PHY Protocol Description ',\n 'Local Intf Neighbor Dev Neighbor Intf Exptime (sec)'\n ]\nflag_list = [['Interface', 'PHY', 'Protocol', 'InUti', 'OutUti', 'inErrors', 'outErrors'],\n ['Interface', 'PHY', 'Protocol', 'Description', ''],\n ['Local', 'Intf', 'Neighbor', 'Dev', 'Neighbor', 'Intf', 'Exptime', '(sec)']]\nflag_list = [['Interface', 'PHY', 'Protocol', 'InUti'],\n ['Interface', 'PHY', 'Protocol', 'Description'],\n ['Local', 'Intf', 'Neighbor', 'Dev']]\nflag_key = ['flag_interface_brief',\n 'flag_interface_desc',\n 'flag_lldp_neighbor']\n\n\ndef get_device_name(content=[]) -> str:\n '''\n 获取设备名\n '''\n device_name = ''\n len_content = len(content)\n i = 0\n while True:\n r_match_name = re.search(r'<(.+[ME60|CMNET\\-SW].+)>', content[i])\n if r_match_name != None:\n device_name = r_match_name.group(1)\n return device_name\n elif i < len_content:\n i += 1\n else:\n print('No device name was found! Abort program.')\n return ''\n\n\ndef get_device_nick_name(device_name='') -> str:\n '''\n 获取设备别名\n '''\n return re.search(r'BAS[0-9]+', device_name).group(0)\n\n\ndef get_start_index(content=[], flag=[]) -> int:\n '''\n 定位数据采集起始点\n '''\n for index, value in enumerate(content):\n _list = re.split(r' +', value)\n if _list[:4] == flag:\n return index + 1\n return False\n\n\ndef get_interfaces(content=[]) -> list:\n '''\n 获取 聚合组 端口信息、up/down、\n 当前匹配 ME60\n '''\n\n def _get_frame_slot(port):\n r = re.search(r'([0-9]+)/([0-9]+)', port)\n frame = r.group(1)\n solt = r.group(2)\n return frame, solt\n\n # 定位数据采集起始点\n _start = get_start_index(content, flag_list[0])\n if not _start:\n print('无法定位数据采集点', 'get_interfaces')\n return []\n device_name = get_device_name(content)\n interfaces = []\n for value in content[_start:]:\n # 定位数据采集结束点\n if device_name in value:\n break\n _list = re.split(r' +', value)\n if _list[0][:9] == 'Eth-Trunk':\n _trunk = _list[0]\n phy = '{}-{}'.format(_list[1], _list[2])\n interfaces.append([device_name, '-', '-', _trunk, _trunk, phy])\n elif _list[0] == '':\n port = _list[1]\n frame, solt = _get_frame_slot(port)\n phy = '{}-{}'.format(_list[2], _list[3])\n interfaces.append([device_name, frame, solt, port, _trunk, phy])\n elif 'Ethernet' in _list[0]:\n port = _list[0]\n frame, solt = _get_frame_slot(port)\n phy = '{}-{}'.format(_list[1], _list[2])\n interfaces.append([device_name, frame, solt, port, '-', phy])\n else:\n continue\n\n return interfaces\n\n\ndef get_desc(content=[]) -> list:\n '''\n 获取端口描述\n '''\n # 定位数据采集起始点\n _start = get_start_index(content, flag_list[1])\n if not _start:\n print('无法定位数据采集点', 'get_desc')\n return []\n device_name = get_device_name(content)\n desc = []\n for value in content[_start:]:\n # 定位数据采集结束点\n r = re.search(device_name, value)\n if r and r.start() < 3:\n break\n _list = re.split(r' +', value)\n port = _list[0]\n if 'GE' in port:\n if '100GE' not in port:\n port = port.replace('GE', 'GigabitEthernet')\n elif not 'Eth-Trunk' in port:\n continue\n description = ''.join(_list[3:])\n # print(device_name, 'port', Fore.GREEN+port)\n # print('description', Fore.GREEN+description)\n desc.append([port, description])\n return desc\n\n\ndef get_lldp(content=[]) -> list:\n '''\n 获取 lldp 信息\n '''\n # 定位数据采集起始点\n # _start = get_start_index(content, flag_list[2])\n _start = False\n for index, value in enumerate(content):\n _list = re.split(r' +', value)\n if len(_list) == 4 and _list[1] == 'has' and _list[3][:8] == 'neighbor':\n _start = index\n break\n if not _start:\n print('无法定位数据采集点', 'get_lldp')\n return []\n device_name = get_device_name(content)\n lldp = []\n for value in content[_start:]:\n # 定位数据采集结束点\n # if re.search(r'[<\\[][~\\*]?'+device_name+'.+', value):\n # break\n r = re.search(device_name, value)\n if r and r.start() < 3:\n break\n _list = re.split(r' +', value)\n # print(\"_list:\", Fore.BLUE+str(_list))\n # lldp.append([_list[0], _list[1], _list[2]])\n if len(_list) == 4 and _list[1] == 'has' and _list[3][:8] == 'neighbor':\n port2 = _list[0]\n # print('port2', Back.BLUE+port2)\n elif value[:9] == 'Port ID ':\n lldp_port = _list[2][1:]\n # print('Port ID',Fore.GREEN+lldp_port)\n elif value[:18] == 'Port description ':\n lldp_desc = ''.join(_list[2:])[1:]\n # print('System description',Fore.GREEN+lldp_desc)\n elif value[:13] == 'System name ':\n lldp_device = _list[2][1:]\n # print('System name',Fore.GREEN+lldp_device)\n lldp.append([port2, lldp_desc, lldp_device, lldp_port])\n elif _list[0] == 'PortId:':\n lldp_port = _list[1]\n # print('PortId:',Fore.BLUE+lldp_port)\n elif _list[0] == 'PortDesc:':\n lldp_desc = ''.join(_list[1:])\n # print('SysDesc',Fore.BLUE+lldp_desc)\n elif _list[0] == 'SysName:':\n lldp_device = _list[1]\n # print('SysName',Fore.BLUE+lldp_device)\n lldp.append([port2, lldp_desc, lldp_device, lldp_port])\n return lldp\n\n\ndatas = pd.DataFrame()\nfor content in contents:\n # try:\n interfaces = get_interfaces(content)\n # print(interfaces)\n desc = get_desc(content)\n # print(desc)\n lldp = get_lldp(content)\n # print(lldp)\n df_interface = pd.DataFrame(\n interfaces, columns=['设备名', '框', '槽', '端口', '所属聚合组', '状态'])\n df_desc = pd.DataFrame(desc, columns=['端口', '描述'])\n df_lldp = pd.DataFrame(\n lldp, columns=['port2', 'LLDP描述', 'LLDP对端设备', 'LLDP对端端口'])\n # ['列1'] 为辅助列,用于其他表格 vlookup\n df_interface['列1'] = (df_interface['设备名']+\"-\"+df_interface['端口']\n ).map(lambda port: port.replace('(10G)', '').replace('XG', 'G'))\n data = df_interface.merge(df_desc, how='left', on='端口').fillna('')\n data['port2'] = data['端口'].map(\n lambda port: port.replace('(10G)', '').replace('(100G)', ''))\n data = data.merge(df_lldp, how='left', on='port2').fillna('')\n data.drop('port2', axis=1, inplace=True)\n datas = datas.append(data, ignore_index=True)\n # except Exception as err:\n # print(Back.RED+str(err))\n\ndatas.sort_values(by=['设备名', '框', '槽'], inplace=True)\ntry:\n writer = pd.ExcelWriter('output.{}.xlsx'.format(\n datetime.now().strftime(\"%Y-%m-%d.%H_%M_%S\")))\n datas.to_excel(writer, index=False)\n writer.save()\nexcept Exception as err:\n print(Back.RED+str(err))\n","sub_path":"cmnet_interface_trunk.py","file_name":"cmnet_interface_trunk.py","file_ext":"py","file_size_in_byte":8691,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"73631202","text":"from django.urls import path\n\nfrom home import views\nurlpatterns = [\n path(\"\",views.home,name=\"home\"),\n path(\"index\",views.index,name=\"index\"),\n path(\"blog/\",views.blog,name=\"blog\"),\n path(\"blogpost/\",views.blogpost,name=\"home\"),\n path(\"contact\",views.contact,name=\"contact\"),\n path(\"search\",views.search,name=\"search\")\n \n]\n ","sub_path":"home/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":345,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"13400386","text":"\ncar_status = 'stop'\nprint('Hello, welcome to Car Game, you may input \"help\" to check out how '\n 'to play')\nwhile True:\n userinput = input('>')\n userinput = userinput.lower()\n if(userinput == 'help'):\n print('''\n Start - start the car\n Stop - stop the car\n Quit - quit this gamn\n help - show this message\n ''')\n elif(userinput == 'start'):\n if(car_status == 'stop'):\n print(' The car is ready .... set ....Go!')\n car_status = 'start'\n else:\n print(' The car had started already')\n elif(userinput == 'stop'):\n if(car_status == 'start'):\n print(' Alright, the car stops now')\n car_status = 'stop'\n else:\n print(' The car had stopped already')\n elif(userinput == 'quit'):\n print(' Bye~ ')\n break\n else:\n print(\" Sorry, I don't understand.\")\n","sub_path":"Scripts/practise/cargame.py","file_name":"cargame.py","file_ext":"py","file_size_in_byte":921,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"131182934","text":"import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\n\n__all__ = [\n \"FrozenBatchNorm2d\",\n \"L2Norm\",\n \"get_norm\",\n]\n\n\nclass FrozenBatchNorm2d(nn.Module):\n _version = 3\n\n def __init__(self, num_features, eps=1e-5):\n super().__init__()\n\n self.num_features = num_features\n self.eps = eps\n self.register_buffer(\"weight\", torch.ones(num_features))\n self.register_buffer(\"bias\", torch.zeros(num_features))\n self.register_buffer(\"running_mean\", torch.zeros(num_features))\n self.register_buffer(\"running_var\", torch.ones(num_features) - eps)\n\n def forward(self, x):\n if x.requires_grad:\n scale = self.weight * (self.running_var + self.eps).rsqrt()\n bias = self.bias - self.running_mean * scale\n scale = scale.reshape(1, -1, 1, 1)\n bias = bias.reshape(1, -1, 1, 1)\n return x * scale + bias\n else:\n return F.batch_norm(\n x,\n self.running_mean,\n self.running_var,\n self.weight,\n self.bias,\n training=False,\n eps=self.eps,\n )\n\n def _load_from_state_dict(\n self,\n state_dict,\n prefix,\n local_metadata,\n strict,\n missing_keys,\n unexpected_keys,\n error_msgs\n ):\n version = local_metadata.get(\"version\", None)\n\n if version is None or version < 2:\n if prefix + \"running_mean\" not in state_dict:\n state_dict[prefix + \"running_mean\"] = torch.zeros_like(self.running_mean)\n if prefix + \"running_var\" not in state_dict:\n state_dict[prefix + \"running_var\"] = torch.ones_like(self.running_var)\n\n if version is not None and version < 3:\n logger = logging.getLogger(__name__)\n logger.info(\"FrozenBatchNorm {} is upgraded to version 3.\".format(prefix.rstrip(\".\")))\n state_dict[prefix + \"running_var\"] -= self.eps\n\n super()._load_from_state_dict(\n state_dict,\n prefix,\n local_metadata,\n strict,\n missing_keys,\n unexpected_keys,\n error_msgs\n )\n\n def __repr__(self):\n return f\"FrozenBatchNorm2d(num_features={self.num_features}, eps={self.eps})\"\n\n @classmethod\n def convert_frozen_batchnorm(cls, module):\n bn_module = nn.modules.batchnorm\n bn_module = (bn_module.BatchNorm2d, bn_module.SyncBatchNorm)\n res = module\n if isinstance(module, bn_module):\n res = cls(module.num_features)\n if module.affine:\n res.weight.data = module.weight.data.clone().detach()\n res.bias.data = module.bias.data.clone().detach()\n res.running_mean.data = module.running_mean.data\n res.running_var.data = module.running_var.data\n res.eps = module.eps\n else:\n for name, child in module.named_children():\n new_child = cls.convert_frozen_batchnorm(child)\n if new_child is not child:\n res.add_module(name, new_child)\n return res\n\n\nclass L2Norm(nn.Module):\n def __init__(self, n_dims, scale=20.0, eps=1e-10):\n super().__init__()\n\n self.n_dims = n_dims\n self.weight = nn.Parameter(torch.Tensor(self.n_dims))\n self.eps = eps\n\n nn.init.constant_(self.weight, scale)\n\n def forward(self, x):\n x_float = x.float()\n norm = x_float.pow(2).sum(1, keepdim=True).sqrt() + self.eps\n return (self.weight[None, :, None, None].float().expand_as(x_float) * x_float / norm).type_as(x)\n\n\ndef get_norm(norm, out_channels, **kwargs):\n if isinstance(norm, str):\n if len(norm) == 0:\n return None\n eps = kwargs.get(\"eps\", 1e-5)\n momentum = kwargs.get(\"momentum\", 0.1)\n affine = kwargs.get(\"affine\", True)\n track_running_stats = kwargs.get(\"track_running_stats\", True)\n norm = {\n \"BN\": lambda x: nn.BatchNorm2d(x, eps, momentum, affine, track_running_stats),\n \"GN\": lambda x: nn.GroupNorm(32, x, eps, affine),\n \"FrozenBN\": lambda x: FrozenBatchNorm2d(x, eps),\n }[norm]\n return norm(out_channels)\n","sub_path":"tkdet/layers/batch_norm.py","file_name":"batch_norm.py","file_ext":"py","file_size_in_byte":4312,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"16078538","text":"import bisect\nimport sys\n\nhaystack=[1,4,5,6,8,12,15,20,21,23,23,26,29,30]\nneedeles=[0,1,2,5,8,10,22,23,29,30,31]\n\nrow_fmt='{0:2d} @ {1:2d} {2}{0:<2d}'\n\ndef demo(bisect_fn):\n for needle in reversed(needeles):\n position=bisect_fn(haystack,needle)\n offset=position*' |'\n print(row_fmt.format(needle,position,offset))\n \n#if __name__=='__main__':\n# if sys.argv[-1]=='left':\n# bisect_fn=bisect.bisect_left#如果相等会被放置到相等元素的前面\n# else:\n# bisect_fn=bisect.bisect#否则默认放置到相等元素后面\n# print('demo:',bisect_fn.__name__)\n# print('haystack ->',' '.join('%2d' % n for n in haystack))\n# demo(bisect_fn)\n\ndef grade(score,breakpoints=[60,70,80,90],grades='FDCBA'):\n i=bisect.bisect(breakpoints,score)#二分查找,查找之前必须保证是有序序列\n return grades[i]\n\nprint([grade(score) for score in [33,99,77,70,89,90,100]])","sub_path":"pythonproject/流畅的python/第二章--序列构成的数组/2.8.1用bitsect来搜索.py","file_name":"2.8.1用bitsect来搜索.py","file_ext":"py","file_size_in_byte":932,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"164240716","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Oct 11 15:49:31 2019\n\n@author: Arvid\n\"\"\"\nfrom scipy import *\nfrom numpy import *\nfrom matplotlib.pyplot import *\nfrom scipy.linalg import lu_factor, lu_solve\n\nfrom Problem import Problem\n\n\nclass smallRoomHeatSolver():\n \n def __init__(self, interface_dir, interface_vals, problem, room):\n # geom=(1,1), heater=40, normal_wall=15):\n #super().__init__(problem)\n \n self.interface_dir = interface_dir\n self.interface_vals = interface_vals\n #Change with global geometry\n self.x_len = problem.geometry[room][0]\n self.y_len = problem.geometry[room][1]\n self.dx = problem.dx\n self.heater = problem.heater\n self.normal_wall = problem.wall\n self.n_rows = round(self.x_len/self.dx) -1 # Defines the number of rows in the coordinate mesh.\n self.n_cols = round(self.y_len/self.dx) # Defines the number of columns in the coordinate mesh. \n self.size = (self.n_rows, self.n_cols)\n self.N_elements = self.n_rows*self.n_cols # Number of points in which to calculate u \n BC, neu_ind = self._make_boundaries()\n self.BC = BC\n self.neu_ind = neu_ind\n self.A = self._make_matrix()\n lu, piv = lu_factor(self.A)\n self.lu = lu\n self.piv = piv\n self.solution = None\n \n def _make_boundaries(self):\n \n BC_W = zeros(self.size)\n BC_E = zeros(self.size)\n BC_N = zeros(self.size)\n BC_S = zeros(self.size)\n BC_N[0,0:] = self.normal_wall\n BC_S[-1,0:] = self.normal_wall\n \n if self.interface_dir == 'west':\n BC_E[:,-1] = self.heater/(self.dx**2)\n BC_W[:,0] = self.interface_vals/self.dx\n neumann_ind = nonzero(BC_W.reshape(self.N_elements))\n elif self.interface_dir == 'east':\n BC_E[:,-1] = self.interface_vals/self.dx\n BC_W[:,0] = self.heater/self.dx**2\n neumann_ind = nonzero(BC_E.reshape(self.N_elements))\n BC_tot = BC_W + BC_E + BC_N/self.dx**2 + BC_S/self.dx**2\n \n BC_tot = BC_tot.reshape(self.size[0]*self.size[1])\n return BC_tot, neumann_ind\n \n def _update_boundaries(self, interface_vals):\n self.interface_vals = interface_vals\n BC, neu = self._make_boundaries()\n self.BC = -BC\n \n def _make_matrix(self):\n A = (diag(-4*ones(self.N_elements))\n + diag(ones(self.N_elements-1), -1)\n + diag(ones(self.N_elements-1), 1)\n + diag(ones(self.N_elements-self.n_cols), self.n_cols)\n + diag(ones(self.N_elements-self.n_cols), -self.n_cols))\n for ind in self.neu_ind:\n A[ind,ind] = -3\n \n for i in range(self.n_cols-1, self.N_elements-1, self.n_cols):\n A[i, i+1] = 0\n for i in range(self.n_cols, self.N_elements-1, self.n_cols):\n A[i,i-1] = 0\n return A*(1/self.dx**2)\n \n def solve_system(self, interface_vals):\n self._update_boundaries(interface_vals)\n u = lu_solve((self.lu, self.piv), self.BC)\n mesh_vals = u.reshape(self.n_rows,self.n_cols)\n if self.interface_dir == 'east':\n interface_vals = mesh_vals[:,-1]\n elif self.interface_dir == 'west':\n interface_vals = mesh_vals[:,0]\n self.solution = u \n return u, interface_vals\n \n def getMatrix(self):\n room = zeros((self.n_rows+2, self.n_cols+1))\n if self.interface_dir == 'east':\n room[1:-1,0:-1] = flip(self.solution.reshape(self.size)) #Might have flipped to much heh (mirror flip?)\n room[0, :] = self.normal_wall*ones(self.n_cols+1)\n room[:, -1] = self.heater*ones(self.n_rows+2)\n room[-1, :] = self.normal_wall*ones(self.n_cols+1) \n elif self.interface_dir == 'west':\n room[1:-1, 1:] = flip(self.solution.reshape(self.size)) #Might have flipped to much heh (mirror flip?)\n room[0,:] = room[-1,:] = self.normal_wall*ones(self.n_cols+1)\n room[:, 0] = self.heater*ones(self.n_rows+2)\n print('Complete room is: {}'.format(room))\n return room\n \n \nif __name__ == '__main__':\n p = Problem(1/4)\n print(p.geometry)\n interface_vals = array([20,20,20])\n s = smallRoomHeatSolver('east', interface_vals, p, 'room1')\n #BC, neumann_ind = s._make_boundaries()\n A=s._make_matrix()\n s.solve_system(interface_vals)\n print(s.getMatrix())\n ","sub_path":"Project3/Arvid/smallRoomHeatSolver.py","file_name":"smallRoomHeatSolver.py","file_ext":"py","file_size_in_byte":4542,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"150032955","text":"# -*- coding: utf-8 -*- \nfrom tika import parser\nfrom pptx import Presentation\nimport olefile\nimport docx2txt\nfrom krwordrank.hangle import normalize\nfrom krwordrank.word import KRWordRank\n\n\n# pdf to txt\ndef pdf_to_txt(path):\n pdf_path = path\n raw_pdf = parser.from_file(pdf_path) \n result = raw_pdf['content'] \n result = result.strip()\n #print(result)\n return result\n\n\n# ppt to txt ---> 리스트에 txt 박스 단위로 반환 \ndef ppt_to_txt(path):\n ppt_path = Presentation(path)\n result = []\n for slide in ppt_path.slides:\n for shape in slide.shapes:\n if not shape.has_text_frame:\n continue\n for paragraph in shape.text_frame.paragraphs:\n result.append(paragraph.text)\n #print(result)\n return result\n\n\n# hwp to txt ---> Prvtxt만 가능\n# https://luji.tistory.com/18\ndef hwp_to_txt(path):\n hwp_file = olefile.OleFileIO(path)\n hwp_txt = hwp_file.openstream('Prvtext').read()\n result = hwp_txt.decode('UTF=16')\n #print (result)\n return result\n\n\n# word to txt\ndef word_to_txt(path):\n result = docx2txt.process(path)\n #print(result)\n return result\n\n\n# keyword extraction from txt\n# https://lovit.github.io/nlp/2018/04/16/krwordrank/ \n\ndef keyword_extraction(path):\n with open(path, 'r') as f:\n list_file = []\n for line in f:\n list_file.append(line)\n\n texts = list_file\n texts = [normalize(text, english=True, number=True) for text in texts]\n\n wordrank_extractor = KRWordRank(\n min_count = 5, # 단어의 최소 출현 빈도수 (그래프 생성 시)\n max_length = 10, # 단어의 최대 길이\n verbose = True\n )\n\n beta = 0.85 # PageRank의 decaying factor beta\n max_iter = 10\n\n keywords, rank, graph = wordrank_extractor.extract(texts, beta, max_iter)\n\n for word, r in sorted(keywords.items(), key=lambda x:x[1], reverse=True)[:5]: \n print('%s' % (word))\n #print('%8s:\\t%.4f' % (word, r))","sub_path":"keyword extraction/read_file.py","file_name":"read_file.py","file_ext":"py","file_size_in_byte":2000,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"613549941","text":"import haravasto\nfrom random import randint\nimport time\n\n\ntila = {\n\t\"kentta\": None,\n\t\"nakyvakentta\": None,\n\t\"liput\": None,\n\t\"miinat\": None,\n\t\"aloitus\": None\n}\n\ndef luo_kentta():\n\t\"\"\"Luo kentän käyttäjän antamien asetusten mukaan\"\"\"\n\tleveys, korkeus, maara = kysy_asetukset()\n\tprint(\"Hiiren vasen aukaisee ruutuja, hiiren oikea asettaa lipun\\nPeli päättyy, kun kaikkien miinojen paikalle on asetettu lippu\")\n\n\tnakyvakentta = []\n\tkentta = []\n\tfor rivi in range(korkeus):\n\t\tkentta.append([])\n\t\tnakyvakentta.append([])\n\t\tfor sarake in range(leveys):\n\t\t\tkentta[-1].append(\" \")\n\t\t\tnakyvakentta[-1].append(\" \")\n\ttila[\"liput\"] = []\n\ttila[\"nakyvakentta\"] = nakyvakentta\n\ttila[\"kentta\"] = kentta\n\tmiinoita(kentta, maara)\n\tnumeroi_ruudut(kentta)\n\ndef kysy_asetukset():\n\t\"\"\"Kysyy käyttäjältä asetukset ja tarkastaa ne\"\"\"\n\tprint(\"Vaikeusasteet ovat Micro$oft Minesweeperin mukaiset ja näin ollen vaikuttavat kentän kokoon, sekä miinojen määrään\\n\")\n\twhile True:\n\t\ttry:\t\n\t\t\tprint(\"Valitse 1. jos haluat pelata helpolla vaikeusasteella\\nValitse 2. jos haluat pelata keskivaikealla vaikeusasteella\\nValitse 3. jos haluat pelata vaikeimmalla vaikeusasteella\")\n\t\t\tprint(\"Valitse 4. jos haluat päättää asetukset itse\\n\")\n\t\t\tvalinta = int(input(\"Syötä valintasi: \"))\n\t\t\tif valinta == 1:\n\t\t\t\tleveys = 8\n\t\t\t\tkorkeus = 8\n\t\t\t\tmaara = 10\n\t\t\t\treturn leveys, korkeus, maara\n\t\t\telif valinta == 2:\n\t\t\t\tleveys = 16\n\t\t\t\tkorkeus = 16\n\t\t\t\tmaara = 40\n\t\t\t\treturn leveys, korkeus, maara\n\t\t\telif valinta == 3:\n\t\t\t\tleveys = 24\n\t\t\t\tkorkeus = 24\n\t\t\t\tmaara = 99\n\t\t\t\treturn leveys, korkeus, maara\n\t\t\telif valinta == 4:\n\t\t\t\tleveys = int(input(\"Syötä kentän leveys kokonaislukuna: \"))\n\t\t\t\tkorkeus = int(input(\"Syötä kentän korkeus kokonaislukuna: \"))\n\t\t\t\tmaara = int(input(\"Syötä miinojen lukumäärä: \"))\n\t\t\t\tif leveys < 1 or korkeus < 1 or maara > leveys*korkeus:\n\t\t\t\t\tprint(\"Kenttä on liian pieni tai miinoja on enemmän kuin ruutuja.\\n\")\n\t\t\t\telse:\n\t\t\t\t\treturn leveys, korkeus, maara\n\t\t\telse:\n\t\t\t\tprint(\"Virheellinen valinta.\\n\")\n\t\texcept ValueError:\n\t\t\tprint(\"Syötä arvot kokonaislukuina\\n\")\n\ndef miinoita(kentta, maara):\n\t\"\"\"Asettaa kentällä N kpl miinoja satunnaisiin paikkoihin.\"\"\"\n\tmiinat = []\n\tfor i in range(maara):\n\t\t\tx = randint(0, len(kentta) - 1)\n\t\t\ty = randint(0, len(kentta) - 1)\n\t\t\tif kentta[x][y] != \"x\":\n\t\t\t\tkentta[x][y] = \"x\"\n\t\t\t\tmiinat.append((x, y))\n\t# Asettaa kentän tiedot kirjastoon\n\ttila[\"miinat\"] = miinat\n\ttila[\"kentta\"] = kentta\n\ndef numeroi_ruudut(kentta): \n\t\"\"\"Muuttaa ruutujen arvot vastaamaan viereisten miinojen määrää\"\"\"\n\tfor rivinro, rivi in enumerate(kentta):\n\t\tfor sarakenro, sarake in enumerate(rivi):\n\t\t\tif sarake != \"x\":\n\t\t\t\t# Ottaa naapureiden arvot\n\t\t\t\tarvot = [kentta[r][s] for r, s in etsi_naapurit(rivinro, sarakenro)]\n\t\t\t\t# Laskee kuinka monta on miinoja\n\t\t\t\tif arvot.count(\"x\") > 0:\n\t\t\t\t\tkentta[rivinro][sarakenro] = str(arvot.count(\"x\"))\n\t\t\t\telse:\n\t\t\t\t\tkentta[rivinro][sarakenro] = \"0\"\n\t# Asettaa ruutujen numeroarvot kirjastoon\n\ttila[\"kentta\"] = kentta\n\ndef etsi_naapurit(x, y):\n\t\"\"\"Etsii ruudun naapurit ja palauttaa ne\"\"\"\n\tleveys = len(tila[\"kentta\"])\n\tkorkeus = len(tila[\"kentta\"][0])\n\tnaapurit = []\n\tfor nx in range(min(max(x-1, 0), leveys), min(max(x+2, 0), leveys)):\n\t\tfor ny in range(min(max(y-1, 0), korkeus), min(max(y+2, 0), korkeus)):\n\t\t\t\tnaapurit.append((nx, ny))\n\treturn naapurit\n\ndef tulvataytto(x, y, tarkastettu=[]):\n\t\"\"\"Merkitsee kentällä olevat tuntemattomat alueet turvalliseksi siten, että täyttö aloitetaan annetusta x, y -pisteestä.\"\"\"\n\tnaapurit = etsi_naapurit(x, y)\n\tfor x, y in naapurit:\n\t\tif (x, y) not in tarkastettu:\n\t\t\ttarkastettu.append((x, y))\n\t\t\tif tila[\"kentta\"][x][y] != \"x\" and tila[\"nakyvakentta\"][x][y] != \"f\":\n\t\t\t\ttila[\"nakyvakentta\"][x][y] = tila[\"kentta\"][x][y]\n\n\t\t\tif tila[\"kentta\"][x][y] == \"0\":\n\t\t\t\ttulvataytto(x, y)\n\n\ndef tarkista_voitto(x, y):\n\t#tarkistaa onko liput samoissa paikoissa kuin miinat\n\tif set(tila[\"liput\"]) == set(tila[\"miinat\"]):\n\t\tprint(\"Voitit pelin :)\")\n\t\tprint(\"Aikaa kului: {:.2f} sekunttia\".format(lopeta_kello()))\n\t\tpiirra_kentta()\n\ndef tarkista_havio(x, y):\n\t#tarkistaa onko painetussa kohdassa miina\n\tif tila[\"kentta\"][x][y] == \"x\":\n\t\tprint(\"Hävisit pelin :(\")\n\t\tprint(\"Aikaa kului: {:.2f} sekunttia\".format(lopeta_kello()))\n\t\ttila[\"nakyvakentta\"] = tila[\"kentta\"]\n\t\tpiirra_kentta()\n\ndef avaa_ruutu(x, y):\n\ttarkista_havio(x, y)\n\t#jos on lippu, poistaa sen\n\tif (x, y) == tila[\"liput\"]:\n\t\ttila[\"liput\"].remove((x, y))\n\t\t#näytä ruutu\n\t\ttila[\"nakyvakentta\"][x][y] = tila[\"kentta\"][x][y] \n\t\tpiirra_kentta()\n\n\tif tila[\"nakyvakentta\"][x][y] == \" \":\n\t\tif int(tila[\"kentta\"][x][y]) > 0:\n\t\t\ttila[\"nakyvakentta\"][x][y] = tila[\"kentta\"][x][y] \n\t\tif tila[\"kentta\"][x][y] == \"0\":\n\t\t\ttulvataytto(x, y)\n\t\tpiirra_kentta()\n\n\ndef aseta_lippu(x, y):\n\t# Tarkistaa onko ruutu tyhjä\n\tif tila[\"nakyvakentta\"][x][y] == \" \":\n\t\ttila[\"nakyvakentta\"][x][y] = \"f\"\n\t\ttila[\"liput\"].append((x, y))\n\t\ttarkista_voitto(x, y)\n\t# Poistaa lipun\n\telif tila[\"nakyvakentta\"][x][y] == \"f\":\n\t\ttila[\"nakyvakentta\"][x][y] = \" \"\n\t\ttila[\"liput\"].remove((x, y))\n\t\tprint(tila[\"liput\"])\n\telse:\n\t\tprint(\"Ei voi asettaa lippua\")\n\n\tpiirra_kentta()\n\ndef aloita_kello():\n\t# Aloittaa pelin kulkua mittaavan sekunttikellon\n\ttila[\"aloitus\"] = time.time()\n\ndef lopeta_kello():\n\t# Lopettaa pelin kulkua mittaavan sekunttikellon\n\tloppuaika = time.time()\n\ttotal = loppuaika - tila[\"aloitus\"]\n\treturn total\n\ndef hiiri_kasittelija(x, y, nappi, muokkausnapit):\n\t\"\"\"Tätä funktiota kutsutaan kun käyttäjä klikkaa sovellusikkunaa hiirellä.\"\"\"\n\tx = int(x / 40)\n\ty = int(y / 40)\n\tif nappi == haravasto.HIIRI_VASEN:\n\t\tavaa_ruutu(x, y)\n\telif nappi == haravasto.HIIRI_OIKEA:\n\t\taseta_lippu(x, y)\n\ndef piirra_kentta(): \n\t\"\"\"Käsittelijäfunktio, joka piirtää kaksiulotteisena listana kuvatun miinakentän ruudut näkyviin peli-ikkunaan.\n\tFunktiota kutsutaan aina kun pelimoottori pyytää ruudun näkymän päivitystä.\"\"\"\n\tharavasto.tyhjaa_ikkuna()\n\tharavasto.piirra_tausta()\n\tharavasto.aloita_ruutujen_piirto()\n\tfor x in range(len(tila[\"nakyvakentta\"])):\n\t\tfor y in range(len(tila[\"nakyvakentta\"][0])):\n\t\t\t\tharavasto.lisaa_piirrettava_ruutu(tila[\"nakyvakentta\"][x][y], x * 40, y * 40)\n\tharavasto.piirra_ruudut()\n\ndef main():\n\tluo_kentta()\n\tharavasto.lataa_kuvat(\"spritet\")\n\tharavasto.luo_ikkuna(len(tila[\"nakyvakentta\"] * 40), len(tila[\"nakyvakentta\"][0] * 40))\n\tharavasto.aseta_piirto_kasittelija(piirra_kentta)\n\tharavasto.aseta_hiiri_kasittelija(hiiri_kasittelija)\n\taloita_kello()\n\tharavasto.aloita()\nif __name__ == \"__main__\":\n\tmain()","sub_path":"mh.py","file_name":"mh.py","file_ext":"py","file_size_in_byte":6498,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"563388154","text":"from aws_xray_sdk.core import xray_recorder\nfrom config import get_mongo_collection\nfrom models.user_stats import UserStats\nfrom fathomapi.utils.exceptions import InvalidSchemaException\n\n\nclass UserStatsDatastore(object):\n def __init__(self, mongo_collection='athletestats'):\n self.mongo_collection = mongo_collection\n\n @xray_recorder.capture('datastore.UserStatsDatastore.get')\n def get(self, athlete_id):\n \"\"\"\n :param athlete_id: uuid\n :return:\n \"\"\"\n return self._query_mongodb(athlete_id)\n\n def put(self, items):\n if not isinstance(items, list):\n items = [items]\n try:\n for item in items:\n self._put_mongodb(item)\n except Exception as e:\n raise e\n\n def delete(self, athlete_id=None):\n if athlete_id is None:\n raise InvalidSchemaException(\"Need to provide athlete_id to delete\")\n self._delete_mongodb(athlete_id=athlete_id)\n\n @xray_recorder.capture('datastore.UserStatsDatastore._query_mongodb')\n def _query_mongodb(self, athlete_id):\n mongo_collection = get_mongo_collection(self.mongo_collection)\n if isinstance(athlete_id, list):\n query = {'athlete_id': {'$in': athlete_id}}\n mongo_results = mongo_collection.find(query)\n athlete_stats_list = []\n for mongo_result in mongo_results:\n athlete_stats_list.append(UserStats.json_deserialise(mongo_result))\n\n return athlete_stats_list\n else:\n query = {'athlete_id': athlete_id}\n mongo_result = mongo_collection.find_one(query)\n\n if mongo_result is not None:\n return UserStats.json_deserialise(mongo_result)\n else:\n return None\n\n @xray_recorder.capture('datastore.UserStatsDatastore._put_mongodb')\n def _put_mongodb(self, item):\n item = item.json_serialise()\n\n mongo_collection = get_mongo_collection(self.mongo_collection)\n query = {'athlete_id': item['athlete_id']}\n mongo_collection.replace_one(query, item, upsert=True)\n\n @xray_recorder.capture('datastore.UserStatsDatastore._delete_mongodb')\n def _delete_mongodb(self, athlete_id):\n mongo_collection = get_mongo_collection(self.mongo_collection)\n query = {}\n if isinstance(athlete_id, list):\n query['athlete_id'] = {'$in': athlete_id}\n else:\n query['athlete_id'] = athlete_id\n if len(query) > 0:\n mongo_collection.delete_many(query)","sub_path":"apigateway/datastores/user_stats_datastore.py","file_name":"user_stats_datastore.py","file_ext":"py","file_size_in_byte":2565,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"330869600","text":"\"\"\"\nThe Hamming distance between two integers is the number of positions at which the corresponding bits are different.\nGiven two integers x and y, calculate the Hamming distance.\n\nNote:\n0 ≤ x, y < 231.\n\nExample:\nInput: x = 1, y = 4\n\nOutput: 2\nExplanation:\n1 (0 0 0 1)\n4 (0 1 0 0)\n ↑ ↑\n\nThe above arrows point to positions where the corresponding bits are different.\n\"\"\"\n\nimport os\nimport sys\n\ndef main():\n x = 1\n y = 4\n\n tmp = 0\n num1 = 0\n num2 = 0\n\n while(0 <= x and y < 2**31):\n num1 = x ^ y\n print(\"num1:\",num1)\n while(num1 != 0):\n tmp += 1\n num2 = num1 -1\n num1 &= num2\n print(\"tmp:\",tmp,\"num1:\",num1,\"num2:\",num2)\n print(\"tmp:\",tmp)\n return tmp\n \nif __name__ == '__main__':\n main()\n ","sub_path":"算法_Python/汉明距离.py","file_name":"汉明距离.py","file_ext":"py","file_size_in_byte":813,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"179836184","text":"#!/usr/bin/python3\n\nimport urllib.request\nimport json\nfrom datetime import time\n\nfrom django.db.models.query import QuerySet\nfrom django.db.models import Q\n\nfrom osg.models import Attraction, Service\nfrom platour.models import Phone, TypeOfPhone, SubtypeOfExternalReference, ExternalReference, OpeningHour\nfrom platour.functions import print_error, print_log, LOG_LEVEL_DEBUG, LOG_LEVEL_INFO, LOG_LEVEL_ERROR, LOG_LEVEL_WARNING\n\ngoogle_place_key = \"AIzaSyC6zYeRdM3nMIhCp3Cy_VadKHBkbsFQkrk\"\nlanguage_pt_br = 'pt-BR'\nregion_br = 'br'\ngoogle_place_url = \"https://maps.googleapis.com/maps/api/place/details/json?placeid={placeid}&key={key}&language={language}®ion={region}\"\ngoogle_placeid_search_url = \"https://maps.googleapis.com/maps/api/place/findplacefromtext/json?key={key}&inputtype=textquery&{input}\"\nHTML_HEADER = {'User-Agent': 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:62.0) Gecko/20100101 Firefox/62.0'}\nTIMEOUT_REQUEST=60\n\nPHONE_GOOGLE = 'Google'\nWEBSITE_GOOGLE_SUBT = 'Google'\nWEBSITE_GOOGLE_TYPE = 'Site'\n\ndef get_json(placeid, language=language_pt_br, region=region_br):\n json_url = google_place_url.format(placeid=placeid, key=google_place_key, language=language, region=region)\n print_log(LOG_LEVEL_DEBUG, \"loading JSON URL '{0}'\".format(json_url))\n try:\n request = urllib.request.Request(json_url, None, HTML_HEADER)\n with urllib.request.urlopen(request, timeout=TIMEOUT_REQUEST) as url:\n data = json.loads(url.read().decode())\n print_log(LOG_LEVEL_DEBUG, \"success loading JSON URL '{0}'\".format(json_url))\n return data\n except:\n print_error(\"failed loading JSON URL '{0}'\".format(json_url))\n return\n\ndef get_placeid(establishment, force_reload=False):\n if type(establishment) not in [Attraction, Service]:\n print_log(LOG_LEVEL_ERROR, \"unexpected establishment {0} of type '{1}'\".format(establishment, type(establishment)))\n return None\n\n if force_reload == False:\n placeid = establishment.google_placeid\n if placeid:\n return placeid\n else:\n print_log(LOG_LEVEL_WARNING, \"establishment '{0}' has no google_placeid\".format(establishment))\n\n name = establishment.name\n if not name:\n print_log(LOG_LEVEL_ERROR, \"establishment of id {0} name not found, blank or empty - how is that possible?!\".format(establishment.id))\n return None\n location = establishment.locations.first().name or \"\"\n if not location:\n print_log(LOG_LEVEL_WARNING, \"establishment '{0}' location not found\".format(establishment))\n\n place = \"{0},{1}\".format(name.replace(' ', '+'), location.replace(' ', '+'))\n input_param = urllib.parse.urlencode({'input': place}, 'utf-8')\n json_url = google_placeid_search_url.format(key=google_place_key, input=input_param)\n print_log(LOG_LEVEL_DEBUG, \"loading JSON URL '{0}'\".format(json_url))\n\n try:\n request = urllib.request.Request(json_url, None, HTML_HEADER)\n with urllib.request.urlopen(request, timeout=TIMEOUT_REQUEST) as url:\n data = json.loads(url.read().decode())\n print_log(LOG_LEVEL_DEBUG, \"success loading JSON URL '{0}'\".format(json_url))\n except:\n print_error(\"failed loading JSON URL '{0}'\".format(json_url))\n return None\n\n try:\n candidates = data.get('candidates', None)\n if candidates:\n placeid = candidates[0].get('place_id', None)\n if not placeid:\n print_error(\"failed getting JSON key 'place_id' from URL '{0}'\".format(json_url))\n return None\n print_log(LOG_LEVEL_DEBUG, \"got placeid '{0}' of establishment '{1}'\".format(placeid, establishment))\n return placeid\n except:\n print_error(\"failed getting JSON key 'place_id' from URL '{0}'\".format(json_url))\n return None\n\ndef get_address(json_data):\n return json_data['result'].get('formatted_address', None)\n\ndef get_phone(json_data):\n return json_data['result'].get('international_phone_number', None)\n\ndef get_latitude(json_data):\n return str(json_data['result']['geometry']['location']['lat'])\n\ndef get_longitude(json_data):\n return str(json_data['result']['geometry']['location']['lng'])\n\ndef get_name(json_data):\n return json_data['result'].get('name', None)\n\ndef get_rating(json_data):\n return str(json_data['result'].get('rating', None))\n\ndef get_permanently_closed(json_data):\n try:\n permanently_closed = json_data['result'].get('permanently_closed', None)\n if permanently_closed == 'true':\n return True\n else:\n return False\n except KeyError as e:\n return False\n\ndef get_website(json_data):\n return json_data['result'].get('website', None)\n\ndef get_opening_hours(json_data):\n opening_hours = json_data['result'].get('opening_hours', None)\n if opening_hours:\n return opening_hours.get('periods', None)\n else:\n return None\n\ndef update(establishment):\n if establishment == None:\n print_log(LOG_LEVEL_WARNING, \"empty establishment\")\n return\n\n if type(establishment) not in [Attraction, Service]:\n print_log(LOG_LEVEL_ERROR, \"unexpected establishment {0} of type '{1}'\".format(establishment, type(establishment)))\n return\n\n print_log(LOG_LEVEL_INFO, \"started updating establishment '{0}'\".format(establishment))\n\n has_change = False\n\n ####################\n # Updating placeid #\n ####################\n data = placeid = get_placeid(establishment)\n if data == None:\n print_log(LOG_LEVEL_ERROR, \"can't get a placeid for establishment '{0}'\".format(establishment))\n return\n else:\n data_old = establishment.google_placeid\n if data != data_old:\n print_log(LOG_LEVEL_INFO, \"establishment '{0}' has a new Google placeid: from '{1}' to '{2}'\".format(establishment, data_old, data))\n establishment.google_placeid = data\n has_change = True\n\n ################\n # Getting JSON #\n ################\n e_json = get_json(placeid)\n if e_json == None:\n print_log(LOG_LEVEL_ERROR, \"can't get JSON of establishment '{0}' from Google\".format(establishment))\n return\n elif type(e_json) != dict:\n print_log(LOG_LEVEL_ERROR, \"got unexpected JSON of establishment '{0}' from Google: type: {1} | repr: {2}\".format(establishment, type(e_json), repr(e_json)))\n return\n\n ####################\n # Updating address #\n ####################\n data = get_address(e_json)\n if data == None:\n print_log(LOG_LEVEL_ERROR, \"can't get address of establishment '{0}'\".format(establishment))\n else:\n data_old = establishment.address\n if data != data_old:\n print_log(LOG_LEVEL_INFO, \"establishment '{0}' has a new address: from '{1}' to '{2}'\".format(establishment, data_old, data))\n establishment.address = data\n has_change = True\n\n ##################\n # Updating phone #\n ##################\n data = get_phone(e_json)\n if data == None:\n print_log(LOG_LEVEL_INFO, \"can't get phone of establishment '{0}'\".format(establishment))\n else:\n try:\n gphone = TypeOfPhone.objects.select_related().get(name=PHONE_GOOGLE)\n data_old_set = establishment.phone.select_related().filter(description=gphone)\n data_old = data_old_set.first()\n for d in data_old_set[1:]:\n print_log(LOG_LEVEL_WARNING, \"deleting extra {0} phone '{1}' found in establishment {2}\".format(PHONE_GOOGLE, d, establishment))\n d.delete()\n if not data_old_set:\n print_log(LOG_LEVEL_INFO, \"establishment '{0}' has a new {1} phone: from 'None' to '{2}'\".format(establishment, PHONE_GOOGLE, data))\n if type(establishment) == Attraction:\n p = Phone(description=gphone, number=data, attraction=establishment, active=True)\n elif type(establishment) == Service:\n p = Phone(description=gphone, number=data, service=establishment, active=True)\n else:\n print_log(LOG_LEVEL_ERROR, \"unexpected establishment '{0}' of type '{1}' - and how did you get here??\".format(establishment, type(establishment)))\n p.save()\n has_change = True\n elif data != data_old.number:\n print_log(LOG_LEVEL_INFO, \"establishment '{0}' has a new {1} phone: from '{2}' to '{3}'\".format(establishment, PHONE_GOOGLE, data_old, data))\n data_old.number = data\n data_old.active = True\n data_old.save()\n has_change = True # unnecessary\n except TypeOfPhone.DoesNotExist:\n print_log(LOG_LEVEL_ERROR, \"can't get type of phone '{0}'\".format(PHONE_GOOGLE))\n except:\n print_error()\n\n #####################\n # Updating latitude #\n #####################\n data = get_latitude(e_json)\n if data == None:\n print_log(LOG_LEVEL_ERROR, \"can't get latitude of establishment '{0}'\".format(establishment))\n else:\n data_old = establishment.latitude\n if data != data_old:\n print_log(LOG_LEVEL_INFO, \"establishment '{0}' has a new latitude: from '{1}' to '{2}'\".format(establishment, data_old, data))\n establishment.latitude = data\n has_change = True\n\n ######################\n # Updating longitude #\n ######################\n data = get_longitude(e_json)\n if data == None:\n print_log(LOG_LEVEL_ERROR, \"can't get longitude of establishment '{0}'\".format(establishment))\n else:\n data_old = establishment.longitude\n if data != data_old:\n print_log(LOG_LEVEL_INFO, \"establishment '{0}' has a new longitude: from '{1}' to '{2}'\".format(establishment, data_old, data))\n establishment.longitude = data\n has_change = True\n\n #################\n # Updating name #\n #################\n data = get_name(e_json)\n if data == None:\n print_log(LOG_LEVEL_ERROR, \"can't get name of establishment '{0}'\".format(establishment))\n else:\n data_old = establishment.google_name\n if data != data_old:\n print_log(LOG_LEVEL_INFO, \"establishment '{0}' has a new Google name: from '{1}' to '{2}'\".format(establishment, data_old, data))\n establishment.google_name = data\n has_change = True\n\n ###################\n # Updating rating #\n ###################\n data = get_rating(e_json)\n if data == None:\n print_log(LOG_LEVEL_ERROR, \"can't get rating of establishment '{0}'\".format(establishment))\n else:\n data_old = establishment.google_rating\n if data != data_old:\n print_log(LOG_LEVEL_INFO, \"establishment '{0}' has a new rating: from '{1}' to '{2}'\".format(establishment, data_old, data))\n establishment.google_rating = data\n has_change = True\n\n ###################\n # Updating status #\n ###################\n data = not get_permanently_closed(e_json)\n data_old = establishment.active\n if data != data_old:\n print_log(LOG_LEVEL_INFO, \"establishment '{0}' has a new status: from '{1}' to '{2}'\".format(establishment, data_old, data))\n establishment.active = data\n has_change = True\n\n ####################\n # Updating website #\n ####################\n data = get_website(e_json)\n if data == None:\n print_log(LOG_LEVEL_INFO, \"can't get website of establishment '{0}'\".format(establishment))\n else:\n try:\n gwebsite = SubtypeOfExternalReference.objects.select_related().get(name=WEBSITE_GOOGLE_SUBT, typeOfExternalReference__name=WEBSITE_GOOGLE_TYPE)\n data_old_set = establishment.external_reference.select_related().filter(subtypeOfExternalReference=gwebsite)\n data_old = data_old_set.first()\n for d in data_old_set[1:]:\n print_log(LOG_LEVEL_WARNING, \"deleting extra {0} external reference '{1}' found in establishment {2}\".format(gwebsite, d, establishment))\n d.delete()\n if not data_old_set:\n print_log(LOG_LEVEL_INFO, \"establishment '{0}' has a new {1} external reference: from 'None' to '{2}'\".format(establishment, gwebsite, data))\n if type(establishment) == Attraction:\n e = ExternalReference(subtypeOfExternalReference=gwebsite, url=data, attraction=establishment)\n elif type(establishment) == Service:\n e = ExternalReference(subtypeOfExternalReference=gwebsite, url=data, service=establishment)\n else:\n print_log(LOG_LEVEL_ERROR, \"unexpected establishment '{0}' of type '{1}' - and how did you get here??\".format(establishment, type(establishment)))\n e.save()\n has_change = True\n elif data != data_old.url:\n print_log(LOG_LEVEL_INFO, \"establishment '{0}' has a new {1} external reference: from '{2}' to '{3}'\".format(establishment, gwebsite, data_old, data))\n data_old.url = data\n data_old.save()\n except SubtypeOfExternalReference.DoesNotExist:\n print_log(LOG_LEVEL_ERROR, \"can't get subtype of external reference '{0}' of type '{1}'\".format(WEBSITE_GOOGLE_SUBT, WEBSITE_GOOGLE_TYPE))\n except:\n print_error()\n\n ###########################\n # Updating opening hours #\n ###########################\n data_json = get_opening_hours(e_json)\n if data_json == None:\n print_log(LOG_LEVEL_INFO, \"can't get opening hour of establishment '{0}'\".format(establishment))\n else:\n try:\n if type(establishment) == Attraction:\n OpeningHour.objects.select_related().filter(attraction=establishment).delete()\n elif type(establishment) == Service:\n OpeningHour.objects.select_related().filter(service=establishment).delete()\n else:\n print_log(LOG_LEVEL_ERROR, \"unexpected establishment '{0}' of type '{1}' - and how did you get here??\".format(establishment, type(establishment)))\n print(data_json)\n for ow in data_json:\n try:\n day = ow['open']['day']\n opensAt_str = ow['open']['time']\n close = ow.get('close', None)\n closesAt_str = close['time'] if close != None else None\n except KeyError:\n print_log(LOG_LEVEL_ERROR, \"wrong format on opening hour of establishment {0}\".format(establishment))\n continue\n except:\n print_error()\n opensAt = time(int(opensAt_str[:2]), int(opensAt_str[2:]))\n closesAt = time(int(closesAt_str[:2]), int(closesAt_str[2:])) if closesAt_str else None\n if type(establishment) == Attraction:\n ow_obj = OpeningHour(attraction=establishment, day=day, opensAt=opensAt, closesAt=closesAt)\n else:\n ow_obj = OpeningHour(service=establishment, day=day, opensAt=opensAt, closesAt=closesAt)\n ow_obj.save()\n print_log(LOG_LEVEL_INFO, \"got opening hour for establishment '{0}': {1}\".format(establishment, ow_obj))\n except:\n print_error()\n\n if has_change == True:\n print_log(LOG_LEVEL_INFO, \"saving change(s) on establishment '{0}': started\".format(establishment))\n establishment.save()\n print_log(LOG_LEVEL_INFO, \"saving change(s) on establishment '{0}': finished\".format(establishment))\n else:\n print_log(LOG_LEVEL_INFO, \"none changes found on establishment '{0}'\".format(establishment))\n\ndef run(establishments):\n if type(establishments) != QuerySet:\n print_log(LOG_LEVEL_ERROR, \"unexpected establishment QuerySet {0} of type '{1}'\".format(establishments, type(establishments)))\n return None\n\n if establishments.model not in [Attraction, Service]:\n print_log(LOG_LEVEL_ERROR, \"unexpected establishment model of type '{0}'\".format(establishments.model))\n return None\n\n for e in establishments:\n try:\n update(e)\n except:\n print_error()\n\ndef main():\n establishments = Attraction.objects.select_related().all()\n run(establishments)\n establishments = Service.objects.select_related().all()\n run(establishments)\n\nif __name__ == '__main__':\n main()\n ","sub_path":"osg/establishment_update.py","file_name":"establishment_update.py","file_ext":"py","file_size_in_byte":15239,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"373554017","text":"import rospy\nimport numpy as np\nimport random\nfrom std_srvs.srv import Empty\nfrom geometry_msgs.msg import Twist\nfrom nav_msgs.msg import Odometry\nfrom gazebo_connection import GazeboConnection\nfrom tf.transformations import quaternion_from_euler\n\n'''\n---------------------------------------------------------------------------------------------------\nAdapted and simplified from OpenAI's Multiagent Particle Environment class for use with a ROS\nGazebo world with a discrete action space. Continuous action space has not been implemented.\nSteps through the Gazebo simulation, gathering observations and assigning actions and rewards\nfor each agent at each time step.\n\nAuthor: Joseph Pickens, August Soderberg\n---------------------------------------------------------------------------------------------------\n'''\nclass MultiAgentGazeboEnv():\n def __init__(self, num_agents, reset_callback=None, reward_callback=None,\n observation_callback=None, info_callback=None, done_callback=None):\n # scenario callbacks\n self.reset_callback = reset_callback\n self.reward_callback = reward_callback\n self.observation_callback = observation_callback\n self.info_callback = info_callback\n self.done_callback = done_callback\n\n self.num_agents = num_agents\n self.vel_pubs = []\n for i in range(self.num_agents):\n # robot namespaces are assumed to be 'robot1', 'robot2', ...\n self.vel_pubs.append(rospy.Publisher('/robot%d/cmd_vel' % (i+1), Twist, queue_size=1))\n \n # TODO: speed definition should be specific to the scenario from which callbacks are\n # defined, rather than be defined here in the general multiagent environment class.\n self.linear_speed = 2.0\n self.angular_speed = 2.0\n\n self.gazebo = GazeboConnection(False, 'WORLD')\n\n def step(self, action_n):\n obs_n = []\n reward_n = []\n done_n = []\n info_n = {'n': []}\n\n self.gazebo.unpause_sim()\n for i, action in enumerate(action_n):\n self._set_action(action, i)\n self.gazebo.pause_sim()\n\n # record observation, etc. for each agent\n for i, _ in enumerate(action_n):\n obs_n.append(self._get_obs(i))\n reward_n.append(self._get_reward(i))\n done_n.append(self._get_done(i))\n info_n['n'].append(self._get_info(i))\n\n return obs_n, reward_n, done_n, info_n\n\n def reset(self):\n # Resets the state of the environment and returns an initial observation.\n for i in range(self.num_agents):\n model_name = 'Robot%d' % (i + 1)\n x = random.uniform(-1.2, 1.2)\n y = random.uniform(-1.2, 1.2)\n z = 0.35\n q = quaternion_from_euler(0, 0, random.uniform(0, 6.28))\n pose = [x, y, z]\n pose.extend(q)\n self.gazebo.set_model_state(model_name, pose)\n self.gazebo.unpause_sim()\n self.gazebo.pause_sim()\n obs_n = []\n for i in range(self.num_agents):\n obs_n.append(self._get_obs(i))\n return obs_n\n\n # get info used for benchmarking\n def _get_info(self, agent):\n if self.info_callback is None:\n return {}\n return self.info_callback(agent, self.world)\n\n # get observation for a particular agent\n def _get_obs(self, agent):\n if self.observation_callback is None:\n return np.zeros(0)\n return self.observation_callback(agent)\n\n # get dones for a particular agent\n def _get_done(self, agent):\n if self.done_callback is None:\n return False\n return self.done_callback(agent)\n\n # get reward for a particular agent\n def _get_reward(self, agent):\n if self.reward_callback is None:\n return 0.0\n return self.reward_callback(agent)\n\n # set env discrete action for a particular agent\n # action must be a list of 4 binary variables: [forward, backward, left, right]\n def _set_action(self, action, agent):\n t = Twist()\n t.linear.x = (action[0] - action[1]) * self.linear_speed\n t.angular.z = (action[2] - action[3]) * self.angular_speed\n self.vel_pubs[agent].publish(t)\n","sub_path":"src/multiagent_env.py","file_name":"multiagent_env.py","file_ext":"py","file_size_in_byte":4246,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"192815089","text":"# 0920\n# 가장 큰 정사각형 찾기\n# idea: 다이나믹 프로그래밍\ndef solution(board):\n n = len(board)\n m = len(board[0])\n\n dp = [[0] * m for _ in range(n)]\n dp[0] = board[0]\n for i in range(1, n):\n dp[i][0] = board[i][0]\n\n for i in range(1, n):\n for j in range(1, m):\n if board[i][j] == 1:\n dp[i][j] = min(dp[i - 1][j - 1], dp[i - 1][j], dp[i][j - 1]) + 1\n\n answer = 0\n for i in range(n):\n temp = max(dp[i])\n answer = max(answer, temp)\n print(dp)\n return answer ** 2\n\nprint(solution([[0,1,1,1],[1,1,1,1],[1,1,1,1],[0,0,1,0]]))","sub_path":"Level 2/가장 큰 정사각형 찾기.py","file_name":"가장 큰 정사각형 찾기.py","file_ext":"py","file_size_in_byte":622,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"277970247","text":"#!/usr/bin/env python3\nimport sys\n\n\ndef main():\n with open('21_input.txt', 'r') as rules_file:\n lines = rules_file.read().strip().split('\\n')\n\n rules = expand_rules(lines)\n art = to_pixel('.#./..#/###')\n for i in range(18):\n size = len(art[0])\n d = (2 if size % 2 == 0 else 3)\n partitions = partition(art, d)\n art = apply_rules(art, partitions, rules, d)\n\n print('Answer:', light_on_count(art))\n\n\ndef expand_rules(lines):\n expanded_rules = {} \n for line in lines:\n lhs, rhs = map(to_pixel, line.split('=>'))\n for r in range(4):\n expanded_rules[lhs] = rhs\n expanded_rules[flip(lhs)] = rhs\n lhs = rotate(lhs)\n return expanded_rules\n\n\ndef to_pixel(text):\n bits = {'.': 0, '#': 1}\n pixels = []\n for token in text.split('/'):\n pixels.append(tuple(bits[p] for p in token.strip()))\n return tuple(pixels)\n\n\ndef rotate(text):\n return tuple(zip(*reversed(text)))\n\n\ndef flip(text):\n return tuple(tuple(reversed(row)) for row in text)\n\n\ndef partition(art, d):\n res = []\n for outer_row in range(0, len(art), d):\n for inner_col in range(0, len(art[0]), d):\n tmp = [] \n for inner_row in range(outer_row, outer_row+d):\n tmp.append(tuple(art[inner_row][inner_col:inner_col+d]))\n res.append(tuple(tmp))\n return tuple(res)\n\n\ndef apply_rules(art, partitions, rules, d):\n output = tuple(rules[partition] for partition in partitions)\n height_in_boxes = len(art) // d\n expand_map = {2:3, 3:4}\n new_art = []\n for rows in range(0, len(partitions), height_in_boxes):\n for i in range(expand_map[d]):\n tmp = []\n for j in range(rows, rows+height_in_boxes):\n tmp.extend(output[j][i])\n new_art.append(tuple(tmp))\n return tuple(new_art)\n\n\ndef light_on_count(art):\n count = 0\n for row in art:\n count += row.count(1) \n return count\n\n\ndef print_art(art):\n for row in art:\n print(row)\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"2017/21.py","file_name":"21.py","file_ext":"py","file_size_in_byte":1909,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"246617163","text":"def Set():\n\t'''\n\t\tWhy sets?\n\t\tSet operations have a variety of common uses, some more practical than mathematical.\n\t\tFor example, because items are stored only once in a set, sets can be used to filter\n\t\tduplicates out of other collections, though items may be reordered in the process because\n\t\tsets are unordered in general. Simply convert the collection to a set, and then\n\t\tconvert it back again (sets work in the list call here because they are iterable, another\n\t\ttechnical artifact that we’ll unearth later):\n\t'''\n\tX = set('spam')\n\t{'m', 'a', 'p', 's'}\n\n\tY = {'h', 'a', 'm'} # Make a set with set literals in 3.X and 2.7\n\t{'m', 'a', 'h'}\n\ndef Set_Operations():\n\tx = set('abcde')\n\ty = set('bdxyz')\n\n\tx - y \t\t\t\t\t\t\t\t\t\t\t\t# Difference\n\tset(['a', 'c', 'e'])\n\n\tx | y \t\t\t\t\t\t\t\t\t\t\t\t# Union\n\tset(['a', 'c', 'b', 'e', 'd', 'y', 'x', 'z'])\n\n\tx & y \t\t\t\t\t\t\t\t\t\t\t\t# Intersection\n\tset(['b', 'd'])\n\n\tx ^ y \t\t\t\t\t\t\t\t\t\t\t\t# Symmetric difference (XOR)\n\tset(['a', 'c', 'e', 'y', 'x', 'z'])\n\n\tx > y, x < y \t\t\t\t\t\t\t\t\t\t# Superset, subset\n\t(False, False)\n\n\t'e' in x\t\t\t\t\t\t\t\t\t\t\t# Membership (sets)\n\tTrue\n\ndef Set_comprehensions():\n\ta=set('spam')\n\ta={x for x in 'spam'} # Same as: set('spam')\n\t{'m', 's', 'p', 'a'}\n\n\t{c * 4 for c in 'spam'} # Set of collected expression results\n\t{'pppp', 'aaaa', 'ssss', 'mmmm'}\n\n\ndef Set_Functions():\n\tx = set('abcde')\n\ty = set('bdxyz')\n\n\tz = x.intersection(y)# Same as x & y\n\t# >>> z = set(['b', 'd'])\n\n\tz.add('SPAM') # to add element to set \t\t\t\t\t# Insert one item\n\t# >>> z= set(['b', 'd', 'SPAM'])\n\n\tz.update(set(['X', 'Y'])) # to add set to another\t\t# Merge: in-place union\n\t# >>> z= set(['Y', 'X', 'b', 'd', 'SPAM'])\n\n\tz.remove('b') # Delete one item\n\t# >>> set(['Y', 'X', 'd', 'SPAM'])\n\n\tx,y={1,2,3},{1,2,3,4,5}\n\tx.issubset(y) #>>> True\n\n\ty.union(x) # # Same as x | y , y+x\n\t#>>> {'a','b','c','d','e','z','y','x'}\n\n\nif __name__=='__main__':\n\t()\n\n","sub_path":"Learning_python/07.Set.py","file_name":"07.Set.py","file_ext":"py","file_size_in_byte":1872,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"576281528","text":"from scipy.misc import imread\nimport numpy as np\nimport sys\nimport os\n\nif sys.version_info[0] == 2:\n import xml.etree.cElementTree as ET\nelse:\n import xml.etree.ElementTree as ET\n\nVOC_CLASSES = [\n 'aeroplane', 'bicycle', 'bird', 'boat',\n 'bottle', 'bus', 'car', 'cat', 'chair',\n 'cow', 'diningtable', 'dog', 'horse',\n 'motorbike', 'person', 'pottedplant',\n 'sheep', 'sofa', 'train', 'tvmonitor']\n\n\nclass VOCLoader:\n def __init__(self, root, image_sets, prop_method):\n if prop_method == 'ss':\n prop_dir = os.path.join('../data', 'voc07_proposals', 'selective_search')\n elif prop_method == 'eb':\n prop_dir = os.path.join('../data', 'voc07_proposals', 'edge_boxes_70')\n elif prop_method == 'mcg':\n prop_dir = os.path.join('../data', 'voc07_proposals', 'MCG2015')\n else:\n raise Exception('Undefined proposal name')\n self.items = []\n self.num_classes = 0\n self.name_to_index = dict(zip(VOC_CLASSES, range(len(VOC_CLASSES))))\n print('dataset loading...' + repr(image_sets))\n for (year, name) in image_sets:\n rootpath = os.path.join(root, 'VOC' + year)\n for line in open(os.path.join(rootpath, 'ImageSets', 'Main', name + '.txt')):\n data = {}\n id = line.strip()\n target = ET.parse(os.path.join(rootpath, 'Annotations', line.strip() + '.xml'))\n\n box_set = []\n category_set = []\n for obj in target.iter('object'):\n cls_name = obj.find('name').text.strip().lower()\n bbox = obj.find('bndbox')\n\n xmin = int(bbox.find('xmin').text) - 1\n ymin = int(bbox.find('ymin').text) - 1\n xmax = int(bbox.find('xmax').text) - 1\n ymax = int(bbox.find('ymax').text) - 1\n\n category = self.name_to_index[cls_name]\n box_set.append(np.array([xmin, ymin, xmax, ymax], np.float32))\n category_set.append(category)\n\n data['id'] = id\n data['boxes'] = np.array(box_set)\n data['categories'] = np.array(category_set, np.long)\n data['img_full_path'] = os.path.join(rootpath, 'JPEGImages', line.strip() + '.jpg')\n data['prop_path'] = os.path.join(prop_dir, 'mat', id[:4], '%s.mat' % id)\n self.items.append(data)\n\n print('dataset loading complete')\n\n def __len__(self):\n return len(self.items)\n\n\nclass VOCLoaderFewShot:\n def __init__(self, root, image_sets, K):\n self.items = []\n self.num_classes = 0\n self.name_to_index = dict(zip(VOC_CLASSES, range(len(VOC_CLASSES))))\n\n dupl_check = {}\n print('dataset loading...' + repr(image_sets))\n for (year, name) in image_sets:\n rootpath = os.path.join(root, 'VOC' + year)\n for cls, cls_name in enumerate(VOC_CLASSES):\n anno_file = open(os.path.join(rootpath, 'ImageSets', 'Main', cls_name + '_trainval.txt')).readlines()\n k = 0\n for idx in np.random.permutation(len(anno_file)):\n line, exist = anno_file[idx].split()\n if exist == '-1':\n continue\n if line.strip() in dupl_check:\n print('dupl')\n continue\n dupl_check[line.strip()] = True\n data = {}\n id = 'VOC' + year + '_' + line.strip()\n target = ET.parse(os.path.join(rootpath, 'Annotations', line.strip() + '.xml'))\n\n box_set = []\n category_set = []\n for obj in target.iter('object'):\n cls_name = obj.find('name').text.strip().lower()\n bbox = obj.find('bndbox')\n\n xmin = int(bbox.find('xmin').text) - 1\n ymin = int(bbox.find('ymin').text) - 1\n xmax = int(bbox.find('xmax').text) - 1\n ymax = int(bbox.find('ymax').text) - 1\n\n category = self.name_to_index[cls_name]\n box_set.append(np.array([xmin, ymin, xmax, ymax], np.float32))\n category_set.append(category)\n\n data['id'] = id\n data['boxes'] = np.array(box_set)\n data['categories'] = np.array(category_set, np.long)\n data['img_full_path'] = os.path.join(rootpath, 'JPEGImages', line.strip() + '.jpg')\n self.items.append(data)\n k += 1\n if k == K:\n break\n\n print('dataset loading complete')\n\n def __len__(self):\n return len(self.items)","sub_path":"lib/datasets/voc_loader.py","file_name":"voc_loader.py","file_ext":"py","file_size_in_byte":4909,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"259798218","text":"from ctypes import *\nfrom ctypes import util\nfrom decimal import Decimal\nfrom enum import Enum\nimport functools\nimport math\nimport unittest\n\ntry:\n import platform\n OSX_VERSION = tuple(int(v) for v in platform.mac_ver()[0].split('.')[:2])\nexcept Exception:\n OSX_VERSION = None\n\nimport faulthandler\nfaulthandler.enable()\n\nfrom rubicon.objc import (\n ObjCInstance, ObjCClass, ObjCMetaClass,\n NSObject, SEL,\n objc, objc_method, objc_classmethod, objc_property,\n NSUInteger, NSRange, NSEdgeInsets, NSEdgeInsetsMake,\n send_message, objc_const, ObjCBlock\n)\nfrom rubicon.objc import core_foundation, types\nfrom rubicon.objc.objc import ObjCBoundMethod, objc_block, objc_id, Class, Block\n\n# Load the test harness library\nrubiconharness_name = util.find_library('rubiconharness')\nif rubiconharness_name is None:\n raise RuntimeError(\"Couldn't load Rubicon test harness library. Have you set DYLD_LIBRARY_PATH?\")\nrubiconharness = CDLL(rubiconharness_name)\n\n\nclass RubiconTest(unittest.TestCase):\n def test_sel_by_name(self):\n self.assertEqual(SEL(b\"foobar\").name, b\"foobar\")\n\n def test_sel_null(self):\n with self.assertRaises(ValueError):\n SEL(None).name\n\n def test_class_by_name(self):\n \"\"\"An Objective-C class can be looked up by name.\"\"\"\n\n Example = ObjCClass(\"Example\")\n self.assertEqual(Example.name, \"Example\")\n\n def test_objcclass_caching(self):\n \"\"\"ObjCClass instances are cached.\"\"\"\n\n Example1 = ObjCClass(\"Example\")\n Example2 = ObjCClass(\"Example\")\n\n self.assertIs(Example1, Example2)\n\n def test_class_by_pointer(self):\n \"\"\"An Objective-C class can be created from a pointer.\"\"\"\n\n example_ptr = objc.objc_getClass(b\"Example\")\n Example = ObjCClass(example_ptr)\n self.assertEqual(Example, ObjCClass(\"Example\"))\n\n def test_nonexistant_class(self):\n \"\"\"A NameError is raised if a class doesn't exist.\"\"\"\n\n with self.assertRaises(NameError):\n ObjCClass('DoesNotExist')\n\n def test_metaclass_by_name(self):\n \"\"\"An Objective-C metaclass can be looked up by name.\"\"\"\n\n Example = ObjCClass(\"Example\")\n ExampleMeta = ObjCMetaClass(\"Example\")\n\n self.assertEqual(ExampleMeta.name, \"Example\")\n self.assertEqual(ExampleMeta, Example.objc_class)\n\n def test_objcmetaclass_caching(self):\n \"\"\"ObjCMetaClass instances are cached.\"\"\"\n\n ExampleMeta1 = ObjCMetaClass(\"Example\")\n ExampleMeta2 = ObjCMetaClass(\"Example\")\n\n self.assertIs(ExampleMeta1, ExampleMeta2)\n\n def test_metaclass_by_pointer(self):\n \"\"\"An Objective-C metaclass can be created from a pointer.\"\"\"\n\n examplemeta_ptr = objc.objc_getMetaClass(b\"Example\")\n ExampleMeta = ObjCMetaClass(examplemeta_ptr)\n self.assertEqual(ExampleMeta, ObjCMetaClass(\"Example\"))\n\n def test_nonexistant_metaclass(self):\n \"\"\"A NameError is raised if a metaclass doesn't exist.\"\"\"\n\n with self.assertRaises(NameError):\n ObjCMetaClass('DoesNotExist')\n\n def test_metametaclass(self):\n \"\"\"The class of a metaclass can be looked up.\"\"\"\n\n ExampleMeta = ObjCMetaClass(\"Example\")\n ExampleMetaMeta = ExampleMeta.objc_class\n\n self.assertIsInstance(ExampleMetaMeta, ObjCMetaClass)\n self.assertEqual(ExampleMetaMeta, NSObject.objc_class)\n\n def test_objcinstance_can_produce_objcclass(self):\n \"\"\"Creating an ObjCInstance for a class pointer gives an ObjCClass.\"\"\"\n\n example_ptr = objc.objc_getClass(b\"Example\")\n Example = ObjCInstance(example_ptr)\n self.assertEqual(Example, ObjCClass(\"Example\"))\n self.assertIsInstance(Example, ObjCClass)\n\n def test_objcinstance_can_produce_objcmetaclass(self):\n \"\"\"Creating an ObjCInstance for a metaclass pointer gives an ObjCMetaClass.\"\"\"\n\n examplemeta_ptr = objc.objc_getMetaClass(b\"Example\")\n ExampleMeta = ObjCInstance(examplemeta_ptr)\n self.assertEqual(ExampleMeta, ObjCMetaClass(\"Example\"))\n self.assertIsInstance(ExampleMeta, ObjCMetaClass)\n\n def test_objcclass_can_produce_objcmetaclass(self):\n \"\"\"Creating an ObjCClass for a metaclass pointer gives an ObjCMetaclass.\"\"\"\n\n examplemeta_ptr = objc.objc_getMetaClass(b\"Example\")\n ExampleMeta = ObjCClass(examplemeta_ptr)\n self.assertEqual(ExampleMeta, ObjCMetaClass(\"Example\"))\n self.assertIsInstance(ExampleMeta, ObjCMetaClass)\n\n def test_objcclass_requires_class(self):\n \"\"\"ObjCClass only accepts class pointers.\"\"\"\n\n random_obj = NSObject.alloc().init()\n with self.assertRaises(ValueError):\n ObjCClass(random_obj.ptr)\n random_obj.release()\n\n def test_objcmetaclass_requires_metaclass(self):\n \"\"\"ObjCMetaClass only accepts metaclass pointers.\"\"\"\n\n random_obj = NSObject.alloc().init()\n with self.assertRaises(ValueError):\n ObjCMetaClass(random_obj.ptr)\n random_obj.release()\n\n with self.assertRaises(ValueError):\n ObjCMetaClass(NSObject.ptr)\n\n def test_objcclass_superclass(self):\n Example = ObjCClass(\"Example\")\n BaseExample = ObjCClass(\"BaseExample\")\n\n self.assertEqual(Example.superclass, BaseExample)\n self.assertEqual(BaseExample.superclass, NSObject)\n self.assertIsNone(NSObject.superclass)\n\n def test_objcmetaclass_superclass(self):\n Example = ObjCClass(\"Example\")\n BaseExample = ObjCClass(\"BaseExample\")\n\n self.assertEqual(Example.objc_class.superclass, BaseExample.objc_class)\n self.assertEqual(BaseExample.objc_class.superclass, NSObject.objc_class)\n self.assertEqual(NSObject.objc_class.superclass, NSObject)\n\n def test_field(self):\n \"A field on an instance can be accessed and mutated\"\n\n Example = ObjCClass('Example')\n\n obj = Example.alloc().init()\n\n self.assertEqual(obj.baseIntField, 22)\n self.assertEqual(obj.intField, 33)\n\n obj.baseIntField = 8888\n obj.intField = 9999\n\n self.assertEqual(obj.baseIntField, 8888)\n self.assertEqual(obj.intField, 9999)\n\n def test_method(self):\n \"An instance method can be invoked.\"\n Example = ObjCClass('Example')\n\n obj = Example.alloc().init()\n\n self.assertEqual(obj.accessBaseIntField(), 22)\n self.assertEqual(obj.accessIntField(), 33)\n\n obj.mutateBaseIntFieldWithValue_(8888)\n obj.mutateIntFieldWithValue_(9999)\n\n self.assertEqual(obj.accessBaseIntField(), 8888)\n self.assertEqual(obj.accessIntField(), 9999)\n\n def test_method_send(self):\n \"An instance method can be invoked with send_message.\"\n Example = ObjCClass('Example')\n\n obj = Example.alloc().init()\n\n self.assertEqual(send_message(obj, \"accessBaseIntField\", restype=c_int), 22)\n self.assertEqual(send_message(obj, \"accessIntField\", restype=c_int), 33)\n\n send_message(obj, \"mutateBaseIntFieldWithValue:\", 8888, restype=None, argtypes=[c_int])\n send_message(obj, \"mutateIntFieldWithValue:\", 9999, restype=None, argtypes=[c_int])\n\n self.assertEqual(send_message(obj, \"accessBaseIntField\", restype=c_int), 8888)\n self.assertEqual(send_message(obj, \"accessIntField\", restype=c_int), 9999)\n\n def test_static_field(self):\n \"A static field on a class can be accessed and mutated\"\n Example = ObjCClass('Example')\n\n Example.mutateStaticBaseIntFieldWithValue_(1)\n Example.mutateStaticIntFieldWithValue_(11)\n\n self.assertEqual(Example.staticBaseIntField, 1)\n self.assertEqual(Example.staticIntField, 11)\n\n Example.staticBaseIntField = 1188\n Example.staticIntField = 1199\n\n self.assertEqual(Example.staticBaseIntField, 1188)\n self.assertEqual(Example.staticIntField, 1199)\n\n def test_static_method(self):\n \"A static method on a class can be invoked.\"\n Example = ObjCClass('Example')\n\n Example.mutateStaticBaseIntFieldWithValue_(2288)\n Example.mutateStaticIntFieldWithValue_(2299)\n\n self.assertEqual(Example.accessStaticBaseIntField(), 2288)\n self.assertEqual(Example.accessStaticIntField(), 2299)\n\n def test_mutator_like_method(self):\n \"A method that looks like a mutator doesn't confuse issues.\"\n Example = ObjCClass('Example')\n\n obj1 = Example.alloc().init()\n\n # setSpecialValue: looks like it might be a mutator\n # for a specialValue property, but this property doesn't exist.\n\n # We can invoke the method directly...\n obj1.setSpecialValue_(42)\n\n # ... but retrieving like a property is an error\n with self.assertRaises(AttributeError):\n obj1.specialValue\n\n # ...until you set it explicitly...\n obj1.specialValue = 37\n\n # ...at which point it's fair game to be retrieved.\n self.assertEqual(obj1.specialValue, 37)\n\n def test_property_forcing(self):\n \"An instance or property method can be explicitly declared as a property.\"\n Example = ObjCClass('Example')\n Example.declare_class_property('classMethod')\n Example.declare_class_property('classAmbiguous')\n Example.declare_property('instanceMethod')\n Example.declare_property('instanceAmbiguous')\n\n # A class method can be turned into a property\n self.assertEqual(Example.classMethod, 37)\n\n # An actual class property can be accessed as a property\n self.assertEqual(Example.classAmbiguous, 37)\n\n # An instance property can be accessed\n obj1 = Example.alloc().init()\n\n # An instance method can be turned into a property\n self.assertEqual(obj1.instanceMethod, 42)\n\n # An actual property can be accessed as a property\n self.assertEqual(obj1.instanceAmbiguous, 42)\n\n # Practical example: In Sierra, mainBundle was turned into a class property.\n # Previously, it was a method.\n NSBundle = ObjCClass('NSBundle')\n NSBundle.declare_class_property('mainBundle')\n self.assertFalse(type(NSBundle.mainBundle) == ObjCBoundMethod, 'NSBundle.mainBundle should not be a method')\n\n def test_non_existent_field(self):\n \"An attribute error is raised if you invoke a non-existent field.\"\n Example = ObjCClass('Example')\n\n obj1 = Example.alloc().init()\n\n # Non-existent fields raise an error.\n with self.assertRaises(AttributeError):\n obj1.field_doesnt_exist\n\n # Cache warming doesn't affect anything.\n with self.assertRaises(AttributeError):\n obj1.field_doesnt_exist\n\n def test_non_existent_method(self):\n \"An attribute error is raised if you invoke a non-existent method.\"\n Example = ObjCClass('Example')\n\n obj1 = Example.alloc().init()\n\n # Non-existent methods raise an error.\n with self.assertRaises(AttributeError):\n obj1.method_doesnt_exist()\n\n # Cache warming doesn't affect anything.\n with self.assertRaises(AttributeError):\n obj1.method_doesnt_exist()\n\n def test_non_existent_static_field(self):\n \"An attribute error is raised if you invoke a non-existent static field.\"\n Example = ObjCClass('Example')\n\n # Non-existent fields raise an error.\n with self.assertRaises(AttributeError):\n Example.static_field_doesnt_exist\n\n # Cache warming doesn't affect anything.\n with self.assertRaises(AttributeError):\n Example.static_field_doesnt_exist\n\n def test_non_existent_static_method(self):\n \"An attribute error is raised if you invoke a non-existent static method.\"\n Example = ObjCClass('Example')\n\n # Non-existent methods raise an error.\n with self.assertRaises(AttributeError):\n Example.static_method_doesnt_exist()\n\n # Cache warming doesn't affect anything.\n with self.assertRaises(AttributeError):\n Example.static_method_doesnt_exist()\n\n def test_polymorphic_constructor(self):\n \"Check that the right constructor is activated based on arguments used\"\n Example = ObjCClass('Example')\n\n obj1 = Example.alloc().init()\n obj2 = Example.alloc().initWithIntValue_(2242)\n obj3 = Example.alloc().initWithBaseIntValue_intValue_(3342, 3337)\n\n self.assertEqual(obj1.baseIntField, 22)\n self.assertEqual(obj1.intField, 33)\n\n self.assertEqual(obj2.baseIntField, 44)\n self.assertEqual(obj2.intField, 2242)\n\n self.assertEqual(obj3.baseIntField, 3342)\n self.assertEqual(obj3.intField, 3337)\n\n # Protected constructors can't be invoked\n with self.assertRaises(AttributeError):\n Example.alloc().initWithString_(\"Hello\")\n\n def test_static_access_non_static(self):\n \"An instance field/method cannot be accessed from the static context\"\n Example = ObjCClass('Example')\n\n obj = Example.alloc().init()\n\n with self.assertRaises(AttributeError):\n obj.staticIntField\n\n with self.assertRaises(AttributeError):\n obj.get_staticIntField()\n\n def test_non_static_access_static(self):\n \"A static field/method cannot be accessed from an instance context\"\n Example = ObjCClass('Example')\n\n with self.assertRaises(AttributeError):\n Example.intField\n\n with self.assertRaises(AttributeError):\n Example.accessIntField()\n\n def test_string_argument(self):\n \"A method with a string argument can be passed.\"\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n self.assertEqual(example.duplicateString_(\"Wagga\"), \"WaggaWagga\")\n\n def test_enum_argument(self):\n \"An enumerated type can be used as an argument.\"\n Example = ObjCClass('Example')\n\n obj = Example.alloc().init()\n\n self.assertEqual(obj.accessBaseIntField(), 22)\n self.assertEqual(obj.accessIntField(), 33)\n\n class MyEnum(Enum):\n value1 = 8888\n value2 = 9999\n value3 = 3333\n value4 = 4444\n\n obj.mutateBaseIntFieldWithValue_(MyEnum.value1)\n obj.mutateIntFieldWithValue_(MyEnum.value2)\n\n self.assertEqual(obj.accessBaseIntField(), MyEnum.value1.value)\n self.assertEqual(obj.accessIntField(), MyEnum.value2.value)\n\n obj.baseIntField = MyEnum.value3\n obj.intField = MyEnum.value4\n\n self.assertEqual(obj.accessBaseIntField(), MyEnum.value3.value)\n self.assertEqual(obj.accessIntField(), MyEnum.value4.value)\n\n def test_string_return(self):\n \"If a method or field returns a string, you get a Python string back\"\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n self.assertEqual(example.toString(), \"This is an ObjC Example object\")\n\n def test_constant_string_return(self):\n \"If a method or field returns a *constant* string, you get a Python string back\"\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n self.assertEqual(example.smiley(), \"%-)\")\n\n def test_number_return(self):\n \"If a method or field returns a NSNumber, it is converted back to native types\"\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n\n self.assertEqual(example.theAnswer(), 42)\n self.assertAlmostEqual(example.twopi(), 2.0 * math.pi, 5)\n\n def test_float_method(self):\n \"A method with a float argument can be handled.\"\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n self.assertEqual(example.areaOfSquare_(1.5), 2.25)\n\n def test_float_method_send(self):\n \"A method with a float argument can be handled by send_message.\"\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n self.assertEqual(send_message(example, \"areaOfSquare:\", 1.5, restype=c_float, argtypes=[c_float]), 2.25)\n\n def test_double_method(self):\n \"A method with a double argument can be handled.\"\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n self.assertAlmostEqual(example.areaOfCircle_(1.5), 1.5 * math.pi, 5)\n\n def test_double_method_send(self):\n \"A method with a double argument can be handled by send_message.\"\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n self.assertAlmostEqual(send_message(example, \"areaOfCircle:\", 1.5, restype=c_double, argtypes=[c_double]), 1.5 * math.pi, 5)\n\n @unittest.skipIf(OSX_VERSION and OSX_VERSION < (10, 10),\n \"Property handling doesn't work on OS X 10.9 (Mavericks) and earlier\")\n def test_decimal_method(self):\n \"A method with a NSDecimalNumber arguments can be handled.\"\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n\n result = example.areaOfTriangleWithWidth_andHeight_(Decimal('3.0'), Decimal('4.0'))\n self.assertEqual(result, Decimal('6.0'))\n self.assertIsInstance(result, Decimal, 'Result should be a Decimal')\n \n def test_auto_struct_creation(self):\n \"Structs from method signatures are created automatically.\"\n Example = ObjCClass('Example')\n \n types.unregister_encoding_all(b'{simple=ii}')\n types.unregister_encoding_all(b'{simple}')\n types.unregister_encoding_all(b'{complex=[4s]^?{simple=ii}^{complex}b8b16b8}')\n types.unregister_encoding_all(b'{complex}')\n \n # Look up the method, so the return/argument types are decoded and the structs are registered.\n Example.doStuffWithStruct_\n \n struct_simple = types.ctype_for_encoding(b'{simple=ii}')\n self.assertEqual(struct_simple, types.ctype_for_encoding(b'{simple}'))\n \n simple = struct_simple(123, 456)\n ret = Example.doStuffWithStruct_(simple)\n struct_complex = types.ctype_for_encoding(b'{complex=[4s]^?{simple=ii}^{complex}b8b16b8}')\n self.assertIsInstance(ret, struct_complex)\n self.assertEqual(struct_complex, types.ctype_for_encoding(b'{complex}'))\n self.assertEqual(list(ret.field_0), [1, 2, 3, 4])\n self.assertEqual(ret.field_1.value, None)\n self.assertEqual(ret.field_2.field_0, 123)\n self.assertEqual(ret.field_2.field_1, 456)\n self.assertEqual(cast(ret.field_3, c_void_p).value, None)\n self.assertEqual(ret.field_4, 0)\n self.assertEqual(ret.field_5, 1)\n self.assertEqual(ret.field_6, 2)\n\n def test_sequence_arg_to_struct(self):\n \"Sequence arguments are converted to structures.\"\n Example = ObjCClass('Example')\n \n ret = Example.extractSimpleStruct(([9, 8, 7, 6], None, (987, 654), None, 0, 0, 0))\n struct_simple = types.ctype_for_encoding(b'{simple=ii}')\n self.assertIsInstance(ret, struct_simple)\n self.assertEqual(ret.field_0, 987)\n self.assertEqual(ret.field_1, 654)\n\n def test_struct_return(self):\n \"Methods returning structs of different sizes by value can be handled.\"\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n\n class struct_int_sized(Structure):\n _fields_ = [(\"x\", c_char * 4)]\n types.register_encoding(b'{int_sized=[4c]}', struct_int_sized)\n\n self.assertEqual(example.intSizedStruct().x, b\"abc\")\n class struct_oddly_sized(Structure):\n _fields_ = [(\"x\", c_char * 5)]\n\n types.register_encoding(b'{oddly_sized=[5c]}', struct_oddly_sized)\n self.assertEqual(example.oddlySizedStruct().x, b\"abcd\")\n\n class struct_large(Structure):\n _fields_ = [(\"x\", c_char * 17)]\n\n types.register_encoding(b'{large=[17c]}', struct_large)\n self.assertEqual(example.largeStruct().x, b\"abcdefghijklmnop\")\n\n def test_struct_return_send(self):\n \"Methods returning structs of different sizes by value can be handled when using send_message.\"\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n\n class struct_int_sized(Structure):\n _fields_ = [(\"x\", c_char * 4)]\n\n self.assertEqual(send_message(example, \"intSizedStruct\", restype=struct_int_sized).x, b\"abc\")\n\n\n class struct_oddly_sized(Structure):\n _fields_ = [(\"x\", c_char * 5)]\n\n self.assertEqual(send_message(example, \"oddlySizedStruct\", restype=struct_oddly_sized).x, b\"abcd\")\n\n class struct_large(Structure):\n _fields_ = [(\"x\", c_char * 17)]\n\n self.assertEqual(send_message(example, \"largeStruct\", restype=struct_large).x, b\"abcdefghijklmnop\")\n\n def test_object_return(self):\n \"If a method or field returns an object, you get an instance of that type returned\"\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n\n Thing = ObjCClass('Thing')\n thing = Thing.alloc().initWithName_value_('This is thing', 2)\n\n example.thing = thing\n\n the_thing = example.thing\n self.assertEqual(the_thing.toString(), \"This is thing 2\")\n\n def test_no_convert_return(self):\n Example = ObjCClass(\"Example\")\n example = Example.alloc().init()\n\n res = example.toString(convert_result=False)\n self.assertNotIsInstance(res, ObjCInstance)\n self.assertEqual(str(ObjCInstance(res)), \"This is an ObjC Example object\")\n\n def test_partial_method_no_args(self):\n Example = ObjCClass(\"Example\")\n self.assertEqual(Example.overloaded(), 0)\n\n def test_partial_method_one_arg(self):\n Example = ObjCClass(\"Example\")\n self.assertEqual(Example.overloaded(42), 42)\n\n def test_partial_method_two_args(self):\n Example = ObjCClass(\"Example\")\n self.assertEqual(Example.overloaded(12, extraArg=34), 12+34)\n\n def test_partial_method_lots_of_args(self):\n pystring = \"Uñîçö∂€\"\n pybytestring = pystring.encode(\"utf-8\")\n nsstring = core_foundation.at(pystring)\n buf = create_string_buffer(len(pybytestring) + 1)\n usedLength = NSUInteger()\n remaining = NSRange(0, 0)\n nsstring.getBytes(\n buf,\n maxLength=32,\n usedLength=byref(usedLength),\n encoding=4, # NSUTF8StringEncoding\n options=0,\n range=NSRange(0, 7),\n remainingRange=byref(remaining),\n )\n self.assertEqual(buf.value.decode(\"utf-8\"), pystring)\n\n def test_duplicate_class_registration(self):\n \"If you define a class name twice in the same runtime, you get an error.\"\n\n NSObject = ObjCClass('NSObject')\n\n # First definition should work.\n class MyClass(NSObject):\n pass\n\n # Second definition will raise an error.\n # Without protection, this is a segfault.\n with self.assertRaises(RuntimeError):\n class MyClass(NSObject):\n pass\n\n def test_interface(self):\n \"An ObjC protocol implementation can be defined in Python.\"\n\n results = {}\n\n NSObject = ObjCClass('NSObject')\n\n class Handler(NSObject):\n @objc_method\n def initWithValue_(self, value: int):\n self.value = value\n return self\n\n @objc_method\n def peek_withValue_(self, example, value: int) -> None:\n results['string'] = example.toString() + \" peeked\"\n results['int'] = value + self.value\n\n @objc_method\n def poke_withValue_(self, example, value: int) -> None:\n results['string'] = example.toString() + \" poked\"\n results['int'] = value + self.value\n\n @objc_method\n def reverse_(self, input):\n return ''.join(reversed(input))\n\n @objc_method\n def message(self):\n return \"Alea iacta est.\"\n\n @objc_classmethod\n def fiddle_(cls, value: int) -> None:\n results['string'] = \"Fiddled with it\"\n results['int'] = value\n\n # Create two handler instances so we can check the right one\n # is being invoked.\n handler1 = Handler.alloc().initWithValue_(5)\n handler2 = Handler.alloc().initWithValue_(10)\n\n # Create an Example object, and register a handler with it.\n Example = ObjCClass('Example')\n example = Example.alloc().init()\n example.callback = handler2\n\n # Check some Python-side attributes\n self.assertEqual(handler1.value, 5)\n self.assertEqual(handler2.value, 10)\n\n # Invoke the callback; check that the results have been peeked as expected\n example.testPeek_(42)\n\n self.assertEqual(results['string'], 'This is an ObjC Example object peeked')\n self.assertEqual(results['int'], 52)\n\n example.testPoke_(37)\n\n self.assertEqual(results['string'], 'This is an ObjC Example object poked')\n self.assertEqual(results['int'], 47)\n\n self.assertEqual(example.getMessage(), 'Alea iacta est.')\n\n self.assertEqual(example.reverseIt_('Alea iacta est.'), '.tse atcai aelA')\n\n Handler.fiddle_(99)\n\n self.assertEqual(results['string'], 'Fiddled with it')\n self.assertEqual(results['int'], 99)\n\n def test_class_properties(self):\n \"A Python class can have ObjC properties with synthesized getters and setters.\"\n\n NSObject = ObjCClass('NSObject')\n NSURL = ObjCClass('NSURL')\n\n class URLBox(NSObject):\n\n # takes no type: All properties are pointers\n url = objc_property()\n\n @objc_method\n def getSchemeIfPresent(self):\n if self.url is not None:\n return self.url.scheme\n return None\n\n box = URLBox.alloc().init()\n\n # Default property value is None\n self.assertIsNone(box.url)\n\n # Assign an object via synthesized property setter and call method that uses synthesized property getter\n url = NSURL.alloc().initWithString_('https://www.google.com')\n box.url = url\n self.assertEqual(box.getSchemeIfPresent(), 'https')\n\n # Assign None to dealloc property and see if method returns expected None\n box.url = None\n self.assertIsNone(box.getSchemeIfPresent())\n\n # Try composing URLs using constructors\n base = NSURL.URLWithString('https://pybee.org')\n full = NSURL.URLWithString('contributing/', relativeToURL=base)\n\n self.assertEqual(\n \"Visit %s for details\" % full.absoluteURL,\n \"Visit https://pybee.org/contributing/ for details\"\n )\n\n def test_class_with_wrapped_methods(self):\n \"\"\"An ObjCClass can have wrapped methods.\"\"\"\n\n def deco(f):\n @functools.wraps(f)\n def _wrapper(*args, **kwargs):\n return f(*args, **kwargs)\n return _wrapper\n\n class SimpleMath(NSObject):\n @objc_method\n @deco\n def addOne_(self, num: c_int) -> c_int:\n return num + 1\n\n @objc_classmethod\n @deco\n def subtractOne_(cls, num: c_int) -> c_int:\n return num - 1\n\n simplemath = SimpleMath.alloc().init()\n self.assertEqual(simplemath.addOne_(254), 255)\n self.assertEqual(SimpleMath.subtractOne_(75), 74)\n\n def test_function_NSEdgeInsetsMake(self):\n \"Python can invoke NSEdgeInsetsMake to create NSEdgeInsets.\"\n\n insets = NSEdgeInsets(0.0, 1.1, 2.2, 3.3)\n other_insets = NSEdgeInsetsMake(0.0, 1.1, 2.2, 3.3)\n\n # structs are NOT equal\n self.assertNotEqual(insets, other_insets)\n\n # but their values are\n self.assertEqual(insets.top, other_insets.top)\n self.assertEqual(insets.left, other_insets.left)\n self.assertEqual(insets.bottom, other_insets.bottom)\n self.assertEqual(insets.right, other_insets.right)\n\n def test_cfstring_to_str(self):\n \"CFString/NSString instances can be converted to Python str.\"\n\n self.assertEqual(str(core_foundation.at(\"abcdef\")), \"abcdef\")\n\n def test_objc_const(self):\n \"objc_const works.\"\n \n string_const = objc_const(rubiconharness, \"SomeGlobalStringConstant\")\n self.assertEqual(str(string_const), \"Some global string constant\")\n\n\nclass NSArrayMixinTest(unittest.TestCase):\n nsarray = ObjCClass('NSArray')\n nsmutablearray = ObjCClass('NSMutableArray')\n\n py_list = ['one', 'two', 'three']\n\n def make_array(self, contents=None):\n a = self.nsmutablearray.alloc().init()\n if contents is not None:\n for value in contents:\n a.addObject(value)\n\n return self.nsarray.arrayWithArray(a)\n\n def test_getitem(self):\n a = self.make_array(self.py_list)\n\n for pos, value in enumerate(self.py_list):\n self.assertEqual(a[pos], value)\n\n with self.assertRaises(IndexError):\n a[len(self.py_list) + 10]\n\n def test_len(self):\n a = self.make_array(self.py_list)\n\n self.assertEqual(len(a), len(self.py_list))\n\n def test_iter(self):\n a = self.make_array(self.py_list)\n\n keys = list(self.py_list)\n for k in a:\n self.assertTrue(k in keys)\n keys.remove(k)\n\n self.assertTrue(len(keys) == 0)\n\n def test_contains(self):\n a = self.make_array(self.py_list)\n for value in self.py_list:\n self.assertTrue(value in a)\n\n def test_index(self):\n a = self.make_array(self.py_list)\n self.assertEqual(a.index('two'), 1)\n with self.assertRaises(ValueError):\n a.index('umpteen')\n\n def test_count(self):\n a = self.make_array(self.py_list)\n self.assertEqual(a.count('one'), 1)\n\n def test_copy(self):\n a = self.make_array(self.py_list)\n b = a.copy()\n self.assertEqual(b, a)\n self.assertEqual(b, self.py_list)\n\n with self.assertRaises(AttributeError):\n b.append('four')\n\n def test_equivalence(self):\n a = self.make_array(self.py_list)\n b = self.make_array(self.py_list)\n\n self.assertEqual(a, self.py_list)\n self.assertEqual(b, self.py_list)\n self.assertEqual(a, b)\n self.assertEqual(self.py_list, a)\n self.assertEqual(self.py_list, b)\n self.assertEqual(b, a)\n\n def test_slice_access(self):\n a = self.make_array(self.py_list * 2)\n self.assertEqual(a[1:4], ['two', 'three', 'one'])\n self.assertEqual(a[:-2], ['one', 'two', 'three', 'one'])\n self.assertEqual(a[4:], ['two', 'three'])\n self.assertEqual(a[1:5:2], ['two', 'one'])\n\n\nclass NSMutableArrayMixinTest(NSArrayMixinTest):\n def make_array(self, contents=None):\n a = self.nsmutablearray.alloc().init()\n if contents is not None:\n for value in contents:\n a.addObject(value)\n\n return a\n\n def test_setitem(self):\n a = self.make_array(self.py_list)\n\n a[2] = 'four'\n self.assertEqual(a[2], 'four')\n\n def test_del(self):\n a = self.make_array(self.py_list)\n del a[0]\n self.assertEqual(len(a), 2)\n self.assertEqual(a[0], 'two')\n\n def test_append(self):\n a = self.make_array()\n a.append('an item')\n self.assertTrue('an item' in a)\n\n def test_extend(self):\n a = self.make_array()\n a.extend(['an item', 'another item'])\n self.assertTrue('an item' in a)\n self.assertTrue('another item' in a)\n\n def test_clear(self):\n a = self.make_array(self.py_list)\n a.clear()\n self.assertEqual(len(a), 0)\n\n def test_count(self):\n a = self.make_array(self.py_list)\n self.assertEqual(a.count('one'), 1)\n\n a.append('one')\n self.assertEqual(a.count('one'), 2)\n\n def test_copy(self):\n a = self.make_array(self.py_list)\n b = a.copy()\n self.assertEqual(b, a)\n self.assertEqual(b, self.py_list)\n\n b.append('four')\n\n def test_insert(self):\n a = self.make_array(self.py_list)\n a.insert(1, 'four')\n self.assertEqual(a[0], 'one')\n self.assertEqual(a[1], 'four')\n self.assertEqual(a[2], 'two')\n\n def test_pop(self):\n a = self.make_array(self.py_list)\n self.assertEqual(a.pop(), 'three')\n self.assertEqual(a.pop(0), 'one')\n self.assertEqual(len(a), 1)\n self.assertEqual(a[0], 'two')\n\n def test_remove(self):\n a = self.make_array(self.py_list)\n a.remove('three')\n self.assertEqual(len(a), 2)\n self.assertEqual(a[-1], 'two')\n with self.assertRaises(ValueError):\n a.remove('umpteen')\n\n def test_slice_assignment1(self):\n a = self.make_array(self.py_list * 2)\n a[2:4] = ['four', 'five']\n self.assertEqual(a, ['one', 'two', 'four', 'five', 'two', 'three'])\n\n def test_slice_assignment2(self):\n a = self.make_array(self.py_list * 2)\n a[::2] = ['four', 'five', 'six']\n self.assertEqual(a, ['four', 'two', 'five', 'one', 'six', 'three'])\n\n def test_slice_assignment3(self):\n a = self.make_array(self.py_list * 2)\n a[2:4] = ['four']\n self.assertEqual(a, ['one', 'two', 'four', 'two', 'three'])\n\n def test_bad_slice_assignment1(self):\n a = self.make_array(self.py_list * 2)\n\n with self.assertRaises(TypeError):\n a[2:4] = 4\n\n def test_bad_slice_assignment2(self):\n a = self.make_array(self.py_list * 2)\n\n with self.assertRaises(ValueError):\n a[::2] = [4]\n\n def test_del_slice1(self):\n a = self.make_array(self.py_list * 2)\n del a[-2:]\n self.assertEqual(len(a), 4)\n self.assertEqual(a[0], 'one')\n self.assertEqual(a[-1], 'one')\n\n def test_del_slice2(self):\n a = self.make_array(self.py_list * 2)\n del a[::2]\n self.assertEqual(len(a), 3)\n self.assertEqual(a[0], 'two')\n self.assertEqual(a[1], 'one')\n self.assertEqual(a[2], 'three')\n\n def test_del_slice3(self):\n a = self.make_array(self.py_list * 2)\n del a[::-2]\n self.assertEqual(len(a), 3)\n self.assertEqual(a[0], 'one')\n self.assertEqual(a[1], 'three')\n self.assertEqual(a[2], 'two')\n\n def test_reverse(self):\n a = self.make_array(self.py_list)\n a.reverse()\n\n for pos, value in enumerate(reversed(self.py_list)):\n self.assertEqual(a[pos], value)\n\n\nclass NSDictionaryMixinTest(unittest.TestCase):\n nsdict = ObjCClass('NSDictionary')\n nsmutabledict = ObjCClass('NSMutableDictionary')\n\n py_dict = {\n 'one': 'ONE',\n 'two': 'TWO',\n 'three': 'THREE',\n }\n\n def make_dictionary(self, contents=None):\n d = self.nsmutabledict.alloc().init()\n if contents is not None:\n for key, value in contents.items():\n d.setObject_forKey_(value, key)\n\n return self.nsdict.dictionaryWithDictionary(d)\n\n def test_getitem(self):\n d = self.make_dictionary(self.py_dict)\n\n for key, value in self.py_dict.items():\n self.assertEqual(d[key], value)\n\n with self.assertRaises(KeyError):\n d['NO SUCH KEY']\n\n def test_iter(self):\n d = self.make_dictionary(self.py_dict)\n\n keys = set(self.py_dict)\n for k in d:\n self.assertTrue(k in keys)\n keys.remove(k)\n\n self.assertTrue(len(keys) == 0)\n\n def test_len(self):\n d = self.make_dictionary(self.py_dict)\n self.assertEqual(len(d), len(self.py_dict))\n\n def test_get(self):\n d = self.make_dictionary(self.py_dict)\n\n self.assertEqual(d.get('one'), 'ONE')\n self.assertEqual(d.get('two', None), 'TWO')\n self.assertEqual(d.get('four', None), None)\n self.assertEqual(d.get('five', 5), 5)\n self.assertEqual(d.get('six', None), None)\n\n def test_contains(self):\n d = self.make_dictionary(self.py_dict)\n for key in self.py_dict:\n self.assertTrue(key in d)\n\n def test_copy(self):\n d = self.make_dictionary(self.py_dict)\n e = d.copy()\n self.assertEqual(e, d)\n self.assertEqual(e, self.py_dict)\n\n with self.assertRaises(TypeError):\n e['four'] = 'FOUR'\n\n def test_keys(self):\n a = self.make_dictionary(self.py_dict)\n for k1, k2 in zip(sorted(a.keys()), sorted(self.py_dict.keys())):\n self.assertEqual(k1, k2)\n\n def test_values(self):\n a = self.make_dictionary(self.py_dict)\n for v1, v2 in zip(sorted(a.values()), sorted(self.py_dict.values())):\n self.assertEqual(v1, v2)\n\n def test_items(self):\n d = self.make_dictionary(self.py_dict)\n for i1, i2 in zip(sorted(d.items()), sorted(self.py_dict.items())):\n self.assertEqual(i1[0], i2[0])\n self.assertEqual(i1[1], i2[1])\n\nclass NSMutableDictionaryMixinTest(NSDictionaryMixinTest):\n def make_dictionary(self, contents=None):\n d = self.nsmutabledict.alloc().init()\n if contents is not None:\n for key, value in contents.items():\n d.setObject_forKey_(value, key)\n\n return d\n\n def test_setitem(self):\n d = self.make_dictionary()\n for key, value in self.py_dict.items():\n d[key] = value\n\n for key, value in self.py_dict.items():\n self.assertEqual(d[key], value)\n\n def test_del(self):\n d = self.make_dictionary(self.py_dict)\n del d['one']\n self.assertEqual(len(d), 2)\n with self.assertRaises(KeyError):\n d['one']\n\n def test_clear(self):\n d = self.make_dictionary(self.py_dict)\n d.clear()\n self.assertEqual(len(d), 0)\n\n def test_copy(self):\n d = self.make_dictionary(self.py_dict)\n e = d.copy()\n self.assertEqual(e, d)\n self.assertEqual(e, self.py_dict)\n\n e['four'] = 'FOUR'\n\n def test_pop1(self):\n d = self.make_dictionary(self.py_dict)\n\n self.assertEqual(d.pop('one'), 'ONE')\n self.assertEqual(len(d), 2)\n with self.assertRaises(KeyError):\n d['one']\n\n def test_pop2(self):\n d = self.make_dictionary(self.py_dict)\n\n with self.assertRaises(KeyError):\n d.pop('four')\n\n def test_pop3(self):\n d = self.make_dictionary(self.py_dict)\n\n self.assertEqual(d.pop('four', 4), 4)\n\n def test_popitem(self):\n d = self.make_dictionary(self.py_dict)\n\n keys = set(self.py_dict)\n\n while len(d) > 0:\n key, value = d.popitem()\n self.assertTrue(key in keys)\n self.assertEqual(value, self.py_dict[key])\n self.assertTrue(key not in d)\n\n def test_setdefault1(self):\n d = self.make_dictionary(self.py_dict)\n\n self.assertEqual(d.setdefault('one', 'default'), 'ONE')\n self.assertEqual(len(d), len(self.py_dict))\n\n def test_setdefault2(self):\n d = self.make_dictionary(self.py_dict)\n\n self.assertTrue('four' not in d)\n self.assertEqual(d.setdefault('four', 'FOUR'), 'FOUR')\n self.assertEqual(len(d), len(self.py_dict) + 1)\n self.assertEqual(d['four'], 'FOUR')\n\n def test_setdefault3(self):\n d = self.make_dictionary(self.py_dict)\n\n self.assertTrue('four' not in d)\n self.assertEqual(d.setdefault('four'), None)\n self.assertEqual(len(d), len(self.py_dict))\n with self.assertRaises(KeyError):\n d['four']\n\n def test_update1(self):\n d = self.make_dictionary(self.py_dict)\n\n self.assertEqual(d, self.py_dict)\n d.update({'one': 'two', 'three': 'four', 'four': 'FIVE'})\n self.assertNotEqual(d, self.py_dict)\n self.assertEqual(d['one'], 'two')\n self.assertEqual(d['two'], 'TWO')\n self.assertEqual(d['three'], 'four')\n self.assertEqual(d['four'], 'FIVE')\n self.assertEqual(len(d), len(self.py_dict) + 1)\n\n def test_update2(self):\n d = self.make_dictionary(self.py_dict)\n\n self.assertEqual(d, self.py_dict)\n d.update([('one', 'two'), ('three', 'four'), ('four', 'FIVE')])\n self.assertNotEqual(d, self.py_dict)\n self.assertEqual(d['one'], 'two')\n self.assertEqual(d['two'], 'TWO')\n self.assertEqual(d['three'], 'four')\n self.assertEqual(len(d), len(self.py_dict) + 1)\n\n def test_update3(self):\n d = self.make_dictionary(self.py_dict)\n\n self.assertEqual(d, self.py_dict)\n d.update(one='two', three='four', four='FIVE')\n self.assertNotEqual(d, self.py_dict)\n self.assertEqual(d['one'], 'two')\n self.assertEqual(d['two'], 'TWO')\n self.assertEqual(d['three'], 'four')\n self.assertEqual(d['four'], 'FIVE')\n self.assertEqual(len(d), len(self.py_dict) + 1)\n\n\nclass BlockTests(unittest.TestCase):\n def test_block_property_ctypes(self):\n BlockPropertyExample = ObjCClass(\"BlockPropertyExample\")\n instance = BlockPropertyExample.alloc().init()\n result = ObjCBlock(instance.blockProperty, c_int, c_int, c_int)(1, 2)\n self.assertEqual(result, 3)\n\n def test_block_property_pytypes(self):\n BlockPropertyExample = ObjCClass(\"BlockPropertyExample\")\n instance = BlockPropertyExample.alloc().init()\n result = ObjCBlock(instance.blockProperty, int, int, int)(1, 2)\n self.assertEqual(result, 3)\n\n def test_block_delegate_method_manual_ctypes(self):\n class DelegateManualC(NSObject):\n @objc_method\n def exampleMethod_(self, block):\n ObjCBlock(block, c_void_p, c_int, c_int)(2, 3)\n BlockObjectExample = ObjCClass(\"BlockObjectExample\")\n delegate = DelegateManualC.alloc().init()\n instance = BlockObjectExample.alloc().initWithDelegate_(delegate)\n result = instance.blockExample()\n self.assertEqual(result, 5)\n\n def test_block_delegate_method_manual_pytypes(self):\n class DelegateManualPY(NSObject):\n @objc_method\n def exampleMethod_(self, block):\n ObjCBlock(block, None, int, int)(2, 3)\n BlockObjectExample = ObjCClass(\"BlockObjectExample\")\n delegate = DelegateManualPY.alloc().init()\n instance = BlockObjectExample.alloc().initWithDelegate_(delegate)\n result = instance.blockExample()\n self.assertEqual(result, 5)\n\n def test_block_delegate_auto(self):\n class DelegateAuto(NSObject):\n @objc_method\n def exampleMethod_(self, block: objc_block):\n block(4, 5)\n BlockObjectExample = ObjCClass(\"BlockObjectExample\")\n delegate = DelegateAuto.alloc().init()\n instance = BlockObjectExample.alloc().initWithDelegate_(delegate)\n result = instance.blockExample()\n self.assertEqual(result, 9)\n\n def test_block_delegate_auto_struct(self):\n class BlockStruct(Structure):\n _fields_ = [\n ('a', c_int),\n ('b', c_int),\n ]\n class DelegateAutoStruct(NSObject):\n @objc_method\n def structBlockMethod_(self, block: objc_block) -> int:\n return block(BlockStruct(42, 43))\n BlockObjectExample = ObjCClass(\"BlockObjectExample\")\n delegate = DelegateAutoStruct.alloc().init()\n instance = BlockObjectExample.alloc().initWithDelegate_(delegate)\n result = instance.structBlockExample()\n self.assertEqual(result, 85)\n\n def test_block_receiver(self):\n BlockReceiverExample = ObjCClass(\"BlockReceiverExample\")\n instance = BlockReceiverExample.alloc().init()\n\n values = []\n\n def block(a: int, b: int) -> None:\n values.append(a + b)\n instance.receiverMethod_(block)\n\n self.assertEqual(values, [27])\n\n def test_block_receiver_unannotated(self):\n BlockReceiverExample = ObjCClass(\"BlockReceiverExample\")\n instance = BlockReceiverExample.alloc().init()\n\n def block(a, b):\n return a + b\n with self.assertRaises(ValueError):\n instance.receiverMethod_(block)\n\n def test_block_receiver_lambda(self):\n BlockReceiverExample = ObjCClass(\"BlockReceiverExample\")\n instance = BlockReceiverExample.alloc().init()\n with self.assertRaises(ValueError):\n instance.receiverMethod_(lambda a, b: a + b)\n\n def test_block_receiver_explicit(self):\n BlockReceiverExample = ObjCClass(\"BlockReceiverExample\")\n instance = BlockReceiverExample.alloc().init()\n\n values = []\n\n block = Block(lambda a, b: values.append(a + b), None, int, int)\n instance.receiverMethod_(block)\n\n self.assertEqual(values, [27])\n\n def test_block_round_trip(self):\n BlockRoundTrip = ObjCClass(\"BlockRoundTrip\")\n instance = BlockRoundTrip.alloc().init()\n\n def block(a: int, b: int) -> int:\n return a + b\n\n returned_block = instance.roundTrip_(block)\n self.assertEqual(returned_block(8, 9), 17)\n","sub_path":"tests/test_rubicon.py","file_name":"test_rubicon.py","file_ext":"py","file_size_in_byte":45211,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"329429796","text":"import hashlib\nimport matplotlib.pyplot as plt\nimport pandas as pd\nimport numpy as np\nfrom scipy.stats.stats import pearsonr\nfrom scipy.stats.stats import spearmanr\n\nSEMEVAL_ANNOTADED_FILE = \"./semval_utils/it.test.data.annotated.tsv\"\nNASARI_PATH = \"./semval_utils/mini_NASARI.tsv\"\nSENSES2SYNSETS_PATH = \"semval_utils\\SemEval17_IT_senses2synsets.txt\"\nSEMEVAL_ANNOTADED_FILE_CONSEGNA2 = \"./semval_utils/it.test.data.annotated.consegna2.tsv\"\n#NASARI_PATH_TERMS = \"./summarizerutils/dd-small-nasari-15.txt\"\nNASARI_PATH_TERMS = \"./summarizerutils/dd-nasari.txt\"\n\ndef get_range(surname):\n nof_elements = 500\n base_idx = (abs(int(hashlib.sha512(surname.encode('utf-8')).hexdigest(), 16)) % 10)\n idx_intervallo = base_idx * 50+1\n return idx_intervallo\n\n#read the manually annotated file and return a dataframe\ndef readAnnotatedCouples():\n return pd.read_csv(SEMEVAL_ANNOTADED_FILE, sep='\t', names=['first','second','score'])\n\ndef read_nasari():\n nasari_df = pd.read_csv(NASARI_PATH, sep='$', names=['babel'])#fake separator\n nasari_df[['babel','terms']] = nasari_df[\"babel\"].str.split(\"\t\", 1, expand=True)\n nasari_df[['babel','lemma']] = nasari_df[\"babel\"].str.split(\"__\", 1, expand=True)\n nasari_df['lemma'] = nasari_df['lemma'].str.lower()\n return nasari_df\n\n#estrae i termini(embed) derivanti da ogni babelid dell'array topic\n#ritorna un array di array (uno per ogni babelid)\ndef getTermsFromBabelIds(topic, nasari_df):\n if len(topic)==0:\n return list()\n nasari_vect = []\n for single_topic in topic:\n nasari_terms = nasari_df.loc[nasari_df[nasari_df.columns[0]] == single_topic]['terms'].tolist()\n if len(nasari_terms) == 0:\n #nasari_vect.append([])\n continue\n else:\n nasari_vect.append(nasari_terms[0].split(\"\t\"))\n return nasari_vect\n\n#return a dict \ndef read_sense2synset():\n dict_to_ret = {}\n temp_synset_list = []\n last_term_seen = None\n first = True\n with open(SENSES2SYNSETS_PATH,encoding=\"utf-8\") as f:\n while True:\n line = f.readline().strip('\\n')\n if not line: \n break\n if line.startswith('#'):#è un termine\n if first:\n last_term_seen = line[1:]\n first = False\n else:\n dict_to_ret[last_term_seen] = temp_synset_list.copy()\n temp_synset_list.clear()\n last_term_seen = line[1:]\n else:#è un babelsynset\n temp_synset_list.append(line)\n\n return dict_to_ret\n\n#retrieve babel synsets terms related to columns first and second of annotated_couples dataframe\ndef getBabelTerms(annotated_couples):\n babel_term_synset_mapper = read_sense2synset()\n nasari_df = read_nasari()\n annotated_couples['first_syn_terms_embed'] = None\n annotated_couples['second_syn_terms_embed'] = None\n for i in annotated_couples.index:\n term1 = annotated_couples.iloc[i, :]['first']\n if not term1 in babel_term_synset_mapper:\n continue\n first_syns = babel_term_synset_mapper[term1]\n first_syn_terms = getTermsFromBabelIds(first_syns,nasari_df)\n\n if len(first_syn_terms) == 0:\n continue\n\n term2 = annotated_couples.iloc[i, :]['second']\n if not term2 in babel_term_synset_mapper:\n continue\n second_syns = babel_term_synset_mapper[term2]\n second_syn_terms = getTermsFromBabelIds(second_syns,nasari_df)\n\n if len(second_syn_terms) == 0:\n continue\n\n annotated_couples.at[i, 'first_syn_terms_embed'] = first_syn_terms\n annotated_couples.at[i, 'second_syn_terms_embed'] = second_syn_terms\n\n return annotated_couples,babel_term_synset_mapper\n\ndef cosine_similarity(x, y):\n x = [float(i) for i in x]\n y = [float(i) for i in y]\n return np.dot(x, y) / (np.sqrt(np.dot(x, x)) * np.sqrt(np.dot(y, y)))\n\n#calculate max cosine similarity between first_syn_terms_embed and second_syn_terms_embed\n#sens2syn_dict = a dict \ndef calculateNasariSimilarity(babelTerms,sens2syn_dict):\n babelTerms['nasari_cosin_similarity'] = None\n babelTerms['most_similar_syn1'] = None\n babelTerms['most_similar_syn2'] = None\n for i in babelTerms.index:\n first_syn_terms_embed = babelTerms.iloc[i, :]['first_syn_terms_embed']\n second_syn_terms_embed = babelTerms.iloc[i, :]['second_syn_terms_embed']\n maxSim = -100\n idx_max_synset1 = 0 #indice del babelsynset di first che massimizza la similarità\n idx_max_synset2 = 0 #indice del babelsynset di second che massimizza la similarità\n if (first_syn_terms_embed is None) or (second_syn_terms_embed is None):\n continue\n\n tmp_idx1 = 0\n tmp_idx2 = 0\n for term in first_syn_terms_embed:\n tmp_idx2 = 0\n for term2 in second_syn_terms_embed:\n sim = cosine_similarity(term,term2)\n if sim > maxSim:\n maxSim = sim\n idx_max_synset1 = tmp_idx1\n idx_max_synset2 = tmp_idx2\n tmp_idx2+=1\n tmp_idx1+=1\n\n babelTerms.at[i, 'nasari_cosin_similarity'] = maxSim\n babelTerms.at[i, 'most_similar_syn1'] = sens2syn_dict[babelTerms.iloc[i, :]['first']][idx_max_synset1]\n babelTerms.at[i, 'most_similar_syn2'] = sens2syn_dict[babelTerms.iloc[i, :]['second']][idx_max_synset2]\n\n return babelTerms\n\ndef printSpearmanPearson(list1, list2):\n\n #removing None similarity from list2\n res = [i for i in range(len(list2)) if list2[i] == None]\n print(res)\n for indexNone in sorted(res, reverse=True):\n del list1[indexNone]\n del list2[indexNone]\n\n print(\"pearson: \",pearsonr(list1,list2))\n print(\"spearman: \",spearmanr(list1,list2))\n\ndef consegna1():\n input_name = \"Coluccia\"\n\n values = []\n sx = get_range(input_name)\n values.append(sx)\n dx = sx+50-1\n intervallo = \"\" + str(sx) + \"-\" + str(dx)\n print('{:15}:\\tcoppie nell\\'intervallo {}'.format(input_name, intervallo))\n\n annotated_couples = readAnnotatedCouples()\n pd.to_numeric(annotated_couples['score'], errors='ignore')\n #normalizzo lo score annotato manualmente\n annotated_couples['score']=(annotated_couples['score']-annotated_couples['score'].min())/(annotated_couples['score'].max()-annotated_couples['score'].min())\n #print(annotated_couples)\n babelTerms,sens2syn_dict = getBabelTerms(annotated_couples)\n #print(babelTerms)\n nasari_sim = calculateNasariSimilarity(babelTerms,sens2syn_dict)\n print(nasari_sim)\n printSpearmanPearson(annotated_couples['score'].tolist(),nasari_sim['nasari_cosin_similarity'].tolist())\n #i coefficenti non evidenziano una forte correlazione --> secondo me perchè ci sono alcuni score molto distanti (soprattutto quelli che io ho messo a 0 o a 4)\n return nasari_sim #mi serve per la consegna2\n\n#-----------------------------------------------------------------------------------------------------------------------------------------------\n#-----------------------------------------------------------------------------------------------------------------------------------------------\n#-----------------------------------------------------------------------------------------------------------------------------------------------\n#-----------------------------------------------------------------------------------------------------------------------------------------------\n\ndef readNasariDfTerms():\n NASARI_DF = pd.read_csv(NASARI_PATH_TERMS, sep='$', names=['babel'])#fake separator\n NASARI_DF[['babel','terms']] = NASARI_DF[\"babel\"].str.split(\";\", 1, expand=True)\n NASARI_DF[['lemma','terms']] = NASARI_DF[\"terms\"].str.split(\";\", 1, expand=True)\n NASARI_DF['lemma'] = NASARI_DF['lemma'].str.lower()\n return NASARI_DF\n\ndef getTermsFromBabelIds_consegna2(single_topic,nasaridf):\n nasari_terms = nasaridf.loc[nasaridf[nasaridf.columns[0]] == single_topic]['terms'].tolist()\n \n nasari_terms_filtered = []\n for term in nasari_terms:\n if term is None:\n continue\n words = term.split(';')\n #words.pop(0)#remove first\n for word in words:\n if word == \"\":\n continue\n splitted = word.split(\"_\")\n if len(splitted) < 2:\n continue\n #print(splitted)\n nasari_terms_filtered.append(splitted[0])\n return nasari_terms_filtered\n\n#it returns a dataframe with this structure: 'term1','term2','babel1','babel2','terms_in_bs1', 'terms_in_bs2'\ndef readSynsetManuallyAnnotated():\n df = pd.read_csv(SEMEVAL_ANNOTADED_FILE_CONSEGNA2, sep='\t', names=['first','second','babel1','babel2'])\n df['terms_in_bs1'] = None\n df['terms_in_bs2'] = None\n\n nasari_df = readNasariDfTerms()\n\n for i in df.index:\n babel1 = df.iloc[i, :]['babel1']\n babel2 = df.iloc[i, :]['babel2']\n if babel1 is None:\n continue\n first_syn_terms = getTermsFromBabelIds_consegna2(babel1,nasari_df)\n\n if babel2 is None:\n continue\n second_syn_terms = getTermsFromBabelIds_consegna2(babel2,nasari_df)\n\n df.at[i, 'terms_in_bs1'] = first_syn_terms\n df.at[i, 'terms_in_bs2'] = second_syn_terms\n return df\n\ndef calculateBestSimilarityNasariSynset(annotated_df):\n #remove rows where at least one of the babel_terms list is empty\n filtered_df = annotated_df[(annotated_df.terms_in_bs1.map(len) > 0) & (annotated_df.terms_in_bs2.map(len) > 0)]\n filtered_df = filtered_df.reset_index()\n #print(filtered_df)\n #read nasari embed\n #nasari_df = read_nasari()\n\n #read sense2synset\n #babel_term_synset_mapper = read_sense2synset()\n\n filtered_df['first_syn_terms_embed'] = None\n filtered_df['second_syn_terms_embed'] = None\n\n #calculate cosine similarity for each row\n for i in filtered_df.index:\n #get embedded\n filtered_df,sens2syn_dict = getBabelTerms(filtered_df)\n nasari_sim_df = calculateNasariSimilarity(filtered_df,sens2syn_dict)\n #print(nasari_sim_df)\n #calculate accuracy over first babel\n most_similar_syn1 = nasari_sim_df['most_similar_syn1'].tolist()\n babel1 = nasari_sim_df['babel1'].tolist()\n correct = 0\n index = 0\n for nasari_babel in most_similar_syn1:\n my_babel = babel1[index]\n if nasari_babel == my_babel:\n correct+=1\n index+=1\n print(\"Accuracy over first babel: \",correct/len(most_similar_syn1))\n #calculate accuracy over second babel\n most_similar_syn2 = nasari_sim_df['most_similar_syn2'].tolist()\n babel2 = nasari_sim_df['babel2'].tolist()\n correct = 0\n index = 0\n for nasari_babel in most_similar_syn2:\n my_babel = babel2[index]\n if nasari_babel == my_babel:\n correct+=1\n index+=1\n print(\"Accuracy over second babel: \",correct/len(most_similar_syn2))\n #calculate accuracy over couple\n correct = 0\n index = 0\n for nasari_babel in most_similar_syn2:\n my_babel = babel2[index]\n my_babel1 = babel1[index]\n nasari_babel1 = most_similar_syn1[index]\n if nasari_babel == my_babel and nasari_babel1 == my_babel1:\n correct+=1\n index+=1\n print(\"Accuracy over couples: \",correct/len(most_similar_syn2))\n\ndef consegna2():\n annotated_df = readSynsetManuallyAnnotated()\n print(annotated_df)\n calculateBestSimilarityNasariSynset(annotated_df)\n\n\ndef main():\n consegna1_df = consegna1()\n print(\"################################################\\n##########################################\\n#########################\")\n consegna2()\n\nif __name__ == \"__main__\":\n main()","sub_path":"semval_lab.py","file_name":"semval_lab.py","file_ext":"py","file_size_in_byte":11777,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"349881933","text":"import re\r\nimport urllib.request\r\nimport urllib.error\r\n\r\nif __name__ == \"__main__\":\r\n page = 2\r\n url = \"http://www.qiushibaike.com/hot/page/\" + str(page)\r\n user_agent = 'Mozilla/4.0 (compatible; MSIE 5.5; Windows NT)'\r\n headers = {'User-Agent': user_agent}\r\n try:\r\n request = urllib.request.Request(url, headers=headers)\r\n response = urllib.request.urlopen(request)\r\n content = response.read().decode('utf-8')\r\n # re.S代表在匹配时为点任意匹配模式\r\n pattern = re.compile('(.*?).*?(.*?).*?', re.S)\r\n items = re.findall(pattern, content)\r\n for item in items:\r\n print(\"test\")\r\n print(item[0], item[1])\r\n except urllib.error.URLError as e:\r\n if hasattr(e, 'code'):\r\n print(e.code)","sub_path":"crawler_test/test_two.py","file_name":"test_two.py","file_ext":"py","file_size_in_byte":838,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"321227563","text":"from django.shortcuts import render\n\nfrom reader.overallPosition import overallPosition\n\nfrom dashboard.models import (Student, Payment, Record, Level, ScoreRemark,\n GeneralSetting, Subject, Bill\n )\n\nimport json\nimport re\n\n# mysql connection used in dashboard.views\n# from dashboard.views import mydb, c\n\n\ndef login(request):\n if request.method=='GET':\n students = Student.objects.all()\n context = {\n 'students': students,\n }\n return render(request, 'student_login.html', context)\n\n elif request.method =='POST':\n student_id = request.POST['student_id']\n student_name = request.POST['student_name']\n try:\n student = Student.objects.get(student_id=student_id, name=student_name)\n\n except:\n students = Student.objects.all()\n context = {\n 'students': students,\n 'error_message': 'Student Id and Student Name did match'\n }\n return render(request, 'student_login.html', context)\n\n if student:\n return get_profile(request, student)\n\ndef get_profile(request, student):\n student_name = student.name\n student_class_id = student.level.id\n student_name_id = student.id\n\n students = Student.objects.filter(has_left=False, level_id=student_class_id)\n scores_and_names =[]\n for student in students:\n records = Record.objects.filter(\n student_name_id=student.id, class_of_record_id=student_class_id)\n\n totals = 0\n for record in records:\n if record.total:\n totals += record.total\n else:\n pass\n scores_and_names.append((str(student), totals))\n student_position = overallPosition(scores_and_names)[student_name]\n\n student = Student.objects.get(level_id=student_class_id, name=student_name)\n records = Record.objects.all()\n # order by the scores in descending order\n score_remarks = ScoreRemark.objects.all().order_by('-score')\n # initializing empty array for scores/marks\n marks = []\n for score in score_remarks:\n marks.append(score.score)\n\n # for each of the records, find the total, remark and grades\n for record in records:\n # initializing i (used for iterating over the scores and their remarks)\n i = 0\n total = record.class_score + record.exam_score\n # calculating the total for each term and subjet\n if not record.total:\n # SQL update\n # sql = \"UPDATE dashboard_record SET total = %s WHERE class_score = %s AND exam_score = %s\"\n # val = (total, record.class_score, record.exam_score)\n # c.execute(sql, val)\n # mydb.commit()\n\n\n # Django update\n Record.objects.filter(class_score=record.class_score, exam_score=record.exam_score, student_name=record.student_name).update(total = total)\n # record_to_update.total = total\n # record_to_update.save()\n\n\n remarkNotSet = True\n while remarkNotSet:\n # if the total for a particular subjet is greater than the first score in marks array,\n # get that mark and its remarks (excellent, very good etc)\n if total >= marks[i]:\n remarks = ScoreRemark.objects.filter(score=marks[i])\n for r in remarks:\n grade = r.grade\n remark = r.remark\n # if the remark(excellent etc) or the grade (A, B. C etc) is not set, set it by updating the database\n if not record.remark or not record.grade:\n\n # SQL UPDATE\n # sql = \"UPDATE dashboard_record SET grade = %s, remark = %s WHERE total = %s\"\n # val = (grade, remark, record.total)\n # c.execute(sql, val)\n # mydb.commit()\n\n # Django UPDATE\n Record.objects.filter(total=record.total).update(grade = grade, remark = remark)\n # record_to_update.grade = grade\n # record_to_update.remark = remark\n # record_to_update.save()\n\n # stop the looping over the marks since the grade is found, if not it will iterate over smaller grade\n # and return wrong grade and remark\n remarkNotSet = False\n\n # else if the total mark is less than the first mark in the sorted mark array\n # go to the next mark by increasing the index(i)\n elif i < len(marks)-1:\n i = i + 1\n # if done with all the element in the mark array,\n # set the remark and grade for that total to be the least, (F, Fail)\n else:\n i = len(marks)-1\n remarks = ScoreRemark.objects.filter(score=marks[i])\n # getting the remark and grade for the least\n for r in remarks:\n grade = r.grade\n remark = r.remark\n if not record.remark or not record.grade:\n # SQL UPDATE\n # sql = \"UPDATE dashboard_record SET grade = %s, remark = %s WHERE total = %s\"\n # val = (grade, remark, record.total)\n # c.execute(sql, val)\n # mydb.commit()\n\n # DJANGO UPDATE\n Record.objects.filter(total=record.total, student_name_id=record.student_name_id).update(\n grade = grade, remark = remark\n )\n # record_to_update.grade = grade\n # record_to_update.remark = remark\n # record_to_update.save()\n\n remarkNotSet = False\n\n # get new record after the updates to be sent to the template for rendering\n records = Record.objects.filter(\n student_name_id=student_name_id, class_of_record_id=student_class_id, term=1)\n\n # student_class = Record.objects.filter(student_name=student_name)\n\n # get the classes for selection to view report\n all_classes = Level.objects.all()\n\n # initializing dictionary and arrays for drawing chart\n\n # subject and their various terminal totals\n subjectAndTotals = {}\n # list of subjects\n subjects = []\n # classes of record available in the records table\n classes = []\n\n # chart data for a student\n chartData = Record.objects.filter(student_name_id=student_name_id).order_by('id')\n\n # for that data, get the records for the classes, subjects and their total scores\n for data in chartData:\n # classes.append(data.class_of_record)\n subjects.append(data.subject.name) if data.subject.name not in subjects else None\n\n for subject in subjects:\n values = 0\n subject_id = Subject.objects.get(name=subject).id\n\n # Sql update\n # query = \"SELECT total FROM dashboard_record WHERE subject_id = %s AND student_name_id = %s ORDER BY id\"\n # val = (subject_id, student_name_id)\n # c.execute(query, val)\n # totals = c.fetchall()\n\n # Django update\n totals = Record.objects.filter(subject_id=subject_id, student_name_id=student_name_id).order_by('id')\n totalsFortheSubject = []\n\n for total in totals:\n classes.append(\"Record \"+(values+1).__str__()) if \"Record \" + \\\n (values+1).__str__() not in classes else None\n values += 1\n # [(93,), (78,), (69,), (47,), (8,)]\n totalsFortheSubject.append(total.total)\n\n # converting to strings to use regular expression and get values without brackets and commas\n # '[(90,), (78,), (69,), (47,), (8,)]'\n totalsFortheSubject = totalsFortheSubject.__str__()\n\n # finding the values using regular expression\n totalsFortheSubject = re.findall('[0-9]+', totalsFortheSubject)\n\n # new array for the storing the int totals\n intTotalsFortheSubject = []\n for stringTotal in totalsFortheSubject:\n # getting the totals in integer format\n intTotalsFortheSubject.append(int(stringTotal))\n\n subjectAndTotals.update({subject: intTotalsFortheSubject})\n payments = Payment.objects.filter(student_name_id=student_name_id,\n approved=True, deleted=False).order_by('-id')\n\n totalPayment = Payment.objects.filter(student_name_id=student.id,\n approved=True, deleted=False)\n payment_amount = 0\n bill_amount = 0\n for payment in totalPayment:\n payment_amount += payment.amount\n\n totalBill = Bill.objects.filter(student_id=student.id)\n for bill in totalBill:\n for item in bill.billitem_set.all():\n bill_amount += item.amount\n\n\n context = {\n 'settings': GeneralSetting.objects.all()[0],\n 'student_position': student_position,\n 'payments': payments,\n 'amount_owing': bill_amount-payment_amount,\n 'all_classes': all_classes,\n 'student': student,\n 'subjects': json.dumps(subjects),\n 'subjectAndTotals': json.dumps(subjectAndTotals),\n 'classes': json.dumps(classes),\n }\n return render(request, 'student_profile.html', context)\n","sub_path":"portal/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":9541,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"96968655","text":"# Minjoo Kim\n# ITP 115, Fall 2019\n# Assignment #3\n# 9/22/2019\n# minjook@usc.edu\n\n# Description:\n# This program calculates the largest, smallest, and average of given number.\n\n\ndef main():\n repeat = \"y\"\n\n while repeat == \"y\":\n number = 0\n large = -9999\n small = 9999\n total = 0\n count = 0\n\n print(\"Input an integer greater than or equal to 0 or -1 to quit:\")\n while number != -1:\n number = int(input())\n\n if number != -1:\n if number > large:\n large = number\n\n if number < small:\n small = number\n\n total += number\n count += 1\n\n average = (total + 1) / (count - 1)\n\n print(\"The largest number is \" + str(large))\n print(\"The smallest number is \" + str(small))\n print(\"The average number is \" + str(average))\n print(\"\\n\")\n repeat = input(\"Would you like to enter another set of numbers? (y/n): \")\n\n print(\"Goodbye!\")\n\n\nmain()\n","sub_path":"ITP 115 Assignments/ITP115_a3_Kim_Minjoo/ITP 115 Asn #3 - Largest, Smallest, and Average Number.py","file_name":"ITP 115 Asn #3 - Largest, Smallest, and Average Number.py","file_ext":"py","file_size_in_byte":1033,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"335660387","text":"# coding=utf-8\n\n'''\nAuthor: Amos.Li\nEmail: hpu120623@gmail.com\n\ndate: 2019/10/30 18:39\n'''\n\nfrom flask import Flask, jsonify\n\n# 创建flask的应用对象\n# __name__表示当前的模块名字\n# 模块名,flask以这个模块所在的目录为总目录,默认这个目录中的static为静态目录,templates为模板目录\napp = Flask(__name__,\n static_url_path='/python', # 访问静态资源的url前缀,默认值是static\n static_folder='static', # 静态文件的目录,默认就是static\n template_folder='templates', # 模板文件的目录,默认是templates\n )\n\n# 配置参数的使用方式\n# 1.使用配置文件\n# app.config.from_pyfile('config.cfg')\n\n# 2.使用对象配置参数(项目中使用)\nclass Config:\n DEBUG = True # 开启debug模式后,有修改会自动重启\n\n# app.config.from_object(Config)\n\n# 3.直接操作config的字典对象\n# app.config['DEBUG'] = True\n\n\n\n@app.route('/')\ndef index():\n \"\"\"定义视图函数\"\"\"\n # return jsonify({'result': 'hello flask'})\n a = 1 / 0\n return 'hello flask'\n\n\nif __name__ == '__main__':\n # 启动flask程序\n # app.run()\n app.run(host='192.168.1.197', port=8899)\n # app.run(host='0.0.0.0', port=8899)\n","sub_path":"hello.py","file_name":"hello.py","file_ext":"py","file_size_in_byte":1256,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"51885371","text":"import sys\n\nread = sys.stdin.readline\n\ndef find(x) :\n if par[x] == x : return par[x]\n else :\n par[x] = find(par[x])\n return par[x]\n \ndef merge(x,y) :\n p_x = find(x)\n p_y = find(y)\n\n if p_x == p_y : return\n par[p_x] = p_y\n\nT = int(read())\nfor _ in range(T) :\n N,M = map(int,read().split())\n par = [i for i in range(N+1)]\n cnt = 0\n for _ in range(M) :\n a,b = map(int,read().split())\n if find(a) == find(b) : continue\n merge(a,b)\n cnt += 1\n print(cnt)\n","sub_path":"BOJ/30_최소신장트리/9372_상근이의여행.py","file_name":"9372_상근이의여행.py","file_ext":"py","file_size_in_byte":527,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"504342243","text":"\nclass Utilities:\n \"\"\"This Class Serves as an helper class to Perform Grouping Checks\"\"\"\n\n @staticmethod\n def check_subscriptability(pod):\n \"\"\"Checks if an Application Grouping is Subscriptable\n\n :param str pod: Defines Pod objedct returned from getting deployment Pods.\n :return: str application group\n If the method is called asynchronously,\n returns the request thread.\n \"\"\"\n\n try:\n application_group = pod.metadata.labels['applicationGroup']\n except TypeError:\n application_group = None\n return application_group\n","sub_path":"utils/basic_utilities.py","file_name":"basic_utilities.py","file_ext":"py","file_size_in_byte":629,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"40384566","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Nov 6 12:16:58 2019\n\n@author: mackenziemitchell\n\"\"\"\n\nimport requests\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nfrom bs4 import BeautifulSoup\nimport pickle\nfrom RecFunctions import get_products_by_type,get_products_by_problem, categorical_columns, type_column\n\n#Functions Scrape Data and Get Info Df's For SkinStore\n\ncleansers=get_products_by_type('cleansers',14)\nexfoliators=get_products_by_type('exfoliators',6)\nremovers=get_products_by_type('makeup-removers',3)\ntoners=get_products_by_type('toners',6)\nmists=get_products_by_type('mists',3)\ntreatments=get_products_by_type('treatments',14)\nserums=get_products_by_type('serums',16)\nlotions=get_products_by_type('lotions',6)\nmoisturizers=get_products_by_type('moisturizers',23)\nbalms=get_products_by_type('balms',3)\noils=get_products_by_type('oils',5)\nmasks=get_products_by_type('masks',9)\npeels=get_products_by_type('peels',3)\nlips=get_products_by_type('lip-care',4)\neyes=get_products_by_type('eye-care',9)\nsupplements=get_products_by_type('supplements',1)\ntools=get_products_by_type('tools',5)\n\nacnedf,acne = get_products_by_problem('acne-blemishes',10)\nagedf,age = get_products_by_problem('anti-aging',30)\ndarkcdf,darkcircles = get_products_by_problem('dark-circles',5)\ndrydf,dryness = get_products_by_problem('dry-skin',19)\nncdf,norm= get_products_by_problem('normal-combination',17)\noilydf,oily=get_products_by_problem('oily-skin',17)\nsensitivedf,sensi=get_products_by_problem('sensitive-skin',17)\nreddf,redness=get_products_by_problem('redness-rosacea',17)\n\nbraa=[]\nratee=[]\npri=[]\nur=[]\nimgs=[]\nfor i,r,p,u in zip(acnedf.prodName,acnedf.rating,acnedf.price,acnedf.url):\n braa.append(i)\n ratee.append(r)\n pri.append(p)\n ur.append(u)\nfor i,r,p,u in zip(agedf.prodName,agedf.rating,agedf.price,agedf.url):\n braa.append(i)\n ratee.append(r)\n pri.append(p)\n ur.append(u)\nfor i,r,p,u in zip(darkcdf.prodName,darkcdf.rating,darkcdf.price,darkcdf.url):\n braa.append(i)\n ratee.append(r)\n pri.append(p)\n ur.append(u)\nfor i,r,p,u in zip(drydf.prodName,drydf.rating,drydf.price,drydf.url):\n braa.append(i)\n ratee.append(r)\n pri.append(p)\n ur.append(u)\nfor i,r,p,u in zip(ncdf.prodName,ncdf.rating,ncdf.price,ncdf.url):\n braa.append(i)\n ratee.append(r)\n pri.append(p)\n ur.append(u)\nfor i,r,p,u in zip(oilydf.prodName,oilydf.rating,oilydf.price,oilydf.url):\n braa.append(i)\n ratee.append(r)\n pri.append(p)\n ur.append(u)\nfor i,r,p,u in zip(sensitivedf.prodName,sensitivedf.rating,sensitivedf.price,sensitivedf.url):\n braa.append(i)\n ratee.append(r)\n pri.append(p)\n ur.append(u)\nfor i,r,p,u in zip(reddf.prodName,reddf.rating,reddf.price,reddf.url):\n braa.append(i)\n ratee.append(r)\n pri.append(p)\n ur.append(u)\ndatadict=[]\nfor i,r,p,u in zip(braa,ratee,pri,ur):\n datadict.append({'prodName':i,'rating':r,'price':p,'url':u})\nfinaldf=pd.DataFrame(datadict)\n\nfinaldf.rating=[float(i) for i in finaldf.rating]\nfinaldf.price=[float(i) for i in finaldf.price]\n\n#Finalize Full DF\ncategorical_columns('age',age,finaldf)\ncategorical_columns('darkcircles',darkcircles,finaldf)\ncategorical_columns('acne',acne,finaldf)\ncategorical_columns('dry',dryness,finaldf)\ncategorical_columns('redness',redness,finaldf)\ncategorical_columns('sensitive',sensi,finaldf)\ncategorical_columns('oily',oily,finaldf)\ncategorical_columns('normal',norm,finaldf)\ncategorical_columns('cleanser',cleansers,finaldf)\ncategorical_columns('exfoliator',exfoliators,finaldf)\ncategorical_columns('makeup-removers',removers,finaldf)\ncategorical_columns('toner',toners,finaldf)\ncategorical_columns('mist',mists,finaldf)\ncategorical_columns('treatment',treatments,finaldf)\ncategorical_columns('serum',serums,finaldf)\ncategorical_columns('lotion',lotions,finaldf)\ncategorical_columns('moisturizer',moisturizers,finaldf)\ncategorical_columns('balm',balms,finaldf)\ncategorical_columns('oil',oils,finaldf)\ncategorical_columns('mask',masks,finaldf)\ncategorical_columns('peel',peels,finaldf)\ncategorical_columns('lip',lips,finaldf)\ncategorical_columns('eye',eyes,finaldf)\ncategorical_columns('supplement',supplements,finaldf)\ncategorical_columns('tool',tools,finaldf)\n\n# with open('pickles/findf.pickle', 'wb') as f:\n# pickle.dump(finaldf, f, pickle.HIGHEST_PROTOCOL)\n\n#Get Review Info & Get Into DF\n\nratingdict=[]\nfor u in finaldf.url:\n response=requests.get('https://www.skinstore.com/the-ordinary-aha-30-bha-2-peeling-solution-30ml/{}.html'.format(u))\n soup=BeautifulSoup(response.content,'html.parser')\n titles=soup.findAll('h3',{'class':'productReviews_topReviewTitle'})\n ratings=soup.findAll('div',{'class':'productReviews_topReviewsRatingStarsContainer'})\n contents=soup.findAll('p',{'class':'productReviews_topReviewsExcerpt'})\n dates=soup.findAll('span',{'data-js-element':'createdDate'})\n users=soup.findAll('div',{'class':'productReviews_footerDateAndName'})\n brands=soup.find('div',{'data-information-component':'brand'})\n products=soup.find('h1',{'data-product-name':'title'})\n for t,r,c,d,i in zip(titles,ratings,contents,dates,users):\n ratingdict.append({'url':u,'brandName':brands.text.replace('\\n',''),'prodName':products.text,'title':t.text.replace('\\n',''),'rating':str(r).split('aria-label=')[1][1:2],'content':c.text.replace('\\n','').replace('\\r',''),'date':d.text,'user':i.text.replace('\\n','').split('by')[1].lower()})\nratingdf=pd.DataFrame(ratingdict)\nratingdf['rating']=[int(r) for r in ratingdf['rating']]\nratingdf.user=ratingdf.user.replace('','user')\nratingdf['brandName']=[r.replace('\\n','') for r in ratingdf['brandName']]\nratingdf.drop_duplicates(inplace=True)\nratingdf.content=[c.lower() for c in ratingdf.content]\nratingdf['date']=pd.to_datetime(ratingdf.date)\nfinaldf.drop(columns='rating',inplace=True)\nfinal=pd.merge(finaldf,ratingdf, on='url')\nfinal.drop(columns=['prodName_y'],inplace=True)\nfinal.rename(columns={'prodName_x':'prodName'},inplace=True)\n\n#Saving Final DF for Customized Rec Engine\n\nwith open('df1.pickle', 'wb') as f:\n pickle.dump(final, f, pickle.HIGHEST_PROTOCOL)\n","sub_path":"_GettingAllSkinStoreData.py","file_name":"_GettingAllSkinStoreData.py","file_ext":"py","file_size_in_byte":6167,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"471847037","text":"import acm\nfrom DealPackageDevKit import DealPackageDefinition, List, Action, Text, Settings\nfrom inspect import cleandoc\n\n@Settings(GraphApplicable=False,\n SheetApplicable=False)\nclass ListControlInteraction(DealPackageDefinition):\n \"\"\"\n Double click on one of the elements in the list, and it will move to \n the other list. You can also select one element and then click on the \n arrows in the middle to move an elemnt.\n \"\"\"\n \n left = List( defaultValue=['Cat', 'Dog', 'Mouse'],\n label='Left',\n elementDomain='FString',\n onSelectionChanged='@UpdateSelectedElement',\n onDoubleClick='@MouseMoveTo',\n addNewItem =['First', 'Sorted'],\n sortIndexCallback='@AnimalSortingCallback',\n _moveToDestination='right')\n \n right = List( label='Right',\n elementDomain='FString',\n onSelectionChanged='@UpdateSelectedElement',\n onDoubleClick='@MouseMoveTo',\n addNewItem =['First', 'Sorted'],\n sortIndexCallback='@AnimalSortingCallback',\n _moveToDestination='left')\n \n moveToRight = Action( label='>',\n action='@ButtonMoveTo',\n _moveToDestination='right',\n enabled='@IsLeftElementSelected',\n sizeToFit=True)\n \n moveToLeft = Action( label='<',\n action='@ButtonMoveTo',\n _moveToDestination='left',\n enabled='@IsRightElementSelected',\n sizeToFit=True)\n \n doc = Text( defaultValue=cleandoc(__doc__),\n editable=False,\n height=80) \n\n # ####################### #\n # Interface Overrides #\n # ####################### #\n \n def OnInit(self):\n self._selected = {}\n \n def CustomPanes(self):\n return self.GetCustomPanesFromExtValue('CustomPanes_ListControlInteraction_DPE')\n \n def IsValid(self, exceptionAccumulator, aspect):\n exceptionAccumulator('This example is used to demonstrate lists and can not be saved.')\n\n # ####################### #\n # Attribute Callbacks #\n # ####################### #\n \n def MouseMoveTo(self, attrName, selectedElement):\n self.UpdateSelectedElement(attrName, selectedElement)\n self.ButtonMoveTo(attrName)\n \n def ButtonMoveTo(self, attrName):\n destination = self.GetAttributeMetaData(attrName, '_moveToDestination')()\n self._MoveSelectedElementTo( destination )\n \n def UpdateSelectedElement(self, attrName, selectedElement):\n self._selected[attrName] = selectedElement\n \n def IsRightElementSelected(self, attrName):\n return self._GetSelectedElementInList('right') != None\n \n def IsLeftElementSelected(self, attrName):\n return self._GetSelectedElementInList('left') != None\n\n def AnimalSortingCallback(self, attrName, columnNbr, value1, formatter, obj):\n # Sort by reverse string\n return value1[::-1]\n \n # ####################### #\n # Convenience Methods #\n # ####################### # \n def _MoveSelectedElementTo(self, toList):\n fromList = self._GetOppositeListAttribute(toList)\n element = self._GetSelectedElementInList(fromList)\n if element != None:\n index = getattr(self, fromList).IndexOfFirstEqual(element)\n if index != -1:\n getattr(self, fromList).RemoveAt(index)\n getattr(self, toList).Add(element)\n self._selected['right'] = None\n self._selected['left'] = None\n \n def _GetOppositeListAttribute(self, attrName):\n return 'right' if attrName == 'left' else 'left'\n \n def _GetSelectedElementInList(self, listName):\n return self._selected.get(listName, None)\n","sub_path":"Extensions/Deal Package Examples/FPythonCode/ListControlInteraction_DPE.py","file_name":"ListControlInteraction_DPE.py","file_ext":"py","file_size_in_byte":4285,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"44329671","text":"from django.db import models\r\nfrom django.utils import encoding\r\n\r\nclass Product(models.Model):\r\n\tproductName = models.CharField(default='none', max_length=200)\r\n\ttcgId = models.IntegerField(default=-1, unique=True)\r\n\thiPrice = models.FloatField()\r\n\tlowPrice = models.FloatField()\r\n\tavgPrice = models.FloatField()\r\n\tlink = models.TextField(max_length=200)\r\n\t\r\n\tdef updateFieldsFromProduct(self, productInst):\r\n\t\tself.productName = productInst.productName\r\n\t\tself.tcgId = productInst.tcgId\r\n\t\tself.hiPrice = productInst.hiPrice\r\n\t\tself.lowPrice = productInst.lowPrice\r\n\t\tself.avgPrice = productInst.avgPrice\r\n\t\tself.link = productInst.link\r\n\t\treturn self\r\n\t\r\n\t@classmethod\r\n\tdef create(cls, jsonProductObject):\r\n\t\tproduct = cls()\r\n\t\tproduct.productName = jsonProductObject['name']\r\n\t\tproduct.tcgId = jsonProductObject.get('id', -1)\r\n\t\tproduct.hiPrice = jsonProductObject['hiprice']\r\n\t\tproduct.avgPrice = jsonProductObject['avgprice']\r\n\t\tproduct.lowPrice = jsonProductObject['lowprice']\r\n\t\tproduct.link = jsonProductObject['link']\r\n\t\treturn product\r\n\t\t\r\nclass CardSet(models.Model):\r\n\tsetCode = models.CharField(default='none', max_length=200, unique=True)\r\n\tgathererCode = models.CharField(default='none', max_length=200)\r\n\tname = models.CharField(default='none', max_length=200)\r\n\ttype = models.CharField(default='none', max_length=200)\r\n\tblock = models.CharField(default='none', max_length=200)\r\n\ttotal = models.IntegerField(default=0)\r\n\tcardIds = models.CharField(default='', max_length=500) #array of str \r\n\t\r\n\tdef updateFieldsFromSet(self, setInst):\r\n\t\tself.setCode = setInst.setCode\r\n\t\tself.gathererCode = setInst.gathererCode\r\n\t\tself.name = setInst.name\r\n\t\tself.type = setInst.type\r\n\t\tself.block = setInst.block\r\n\t\tself.cardIds = setInst.cardIds\r\n\t\treturn self\r\n\r\n\tdef getCardIds(self):\r\n\t\tcardIdArr = self.cardIds.replace(' ', '').split(',')\r\n\t\treturn cardIdArr\r\n\r\n\t@classmethod\r\n\tdef create(cls, jsonCardSetObject):\r\n\t\tcardSet = cls()\r\n\t\tcardSet.setCode = jsonCardSetObject['code']\r\n\t\tprint(cardSet.setCode)\r\n\t\tcardSet.gathererCode = jsonCardSetObject.get('gathererCode', 'none')\r\n\t\tcardSet.name = jsonCardSetObject['name']\r\n\t\tcardSet.type = jsonCardSetObject['type']\r\n\t\tcardSet.block = jsonCardSetObject.get('block', 'none')\r\n\t\tcards = []\r\n\t\tfor card in jsonCardSetObject['cards']:\r\n\t\t\t# try:\r\n\t\t\t\t# print(encoding.smart_text(card['name'], encoding='utf-8'))\r\n\t\t\t# except:\r\n\t\t\t\t# print('cannot encode card name')\r\n\t\t\tif(card.get('multiverseid', None) != None):\r\n\t\t\t\tcards.append(card['multiverseid'])\r\n\t\tcardSet.cardIds = str(cards).strip('[]')\r\n\t\treturn cardSet\r\n\t\t\r\nclass Card(models.Model):\r\n\tmultiverseId = models.CharField(default='none', max_length=200, unique=True)\r\n\tname = models.CharField(default='none', max_length=200)\r\n\tcolors = models.CharField(default='', max_length=200) #array of str\r\n\trarity = models.CharField(default='none', max_length=200)\r\n\tformats = models.CharField(default='', max_length=200) #array of str\r\n\tproduct = models.OneToOneField(Product, blank=True, null=True)\r\n\t\r\n\tdef updateFieldsFromCard(self, cardInst):\r\n\t\tmultiverseId = cardInst.multiverseId\r\n\t\tname = cardInst.name\r\n\t\tcolors = cardInst.colors\r\n\t\trarity = cardInst.rarity\r\n\t\tformats = cardInst.formats\r\n\t\tproduct = cardInst.product\r\n\t\treturn self\r\n\t\r\n\t@classmethod\r\n\tdef create(cls, jsonCardObject, product):\r\n\t\tcard = cls()\r\n\t\tcard.multiverseId = jsonCardObject.get('multiverseid', '-1')\r\n\t\tcard.name = jsonCardObject['name']\r\n\t\tcard.colors = str(jsonCardObject.get('colors', '')).strip('[]')\r\n\t\t#print(str(card.colors))\r\n\t\tcard.rarity = jsonCardObject['rarity']\r\n\t\tlegalities = jsonCardObject.get('legalities','')\r\n\t\tformats = []\r\n\t\tfor legalityKey in legalities:\r\n\t\t\tformats.append(str(legalityKey))\r\n\t\t#print('cardName: ' + card.name + ' , formats: ' + str(len(formats)))\r\n\t\tcard.formats = str(formats).strip('[]')\r\n\t\tif(product != None):\r\n\t\t\tcard.product = product\r\n\t\treturn card\r\n\t\r\n\t","sub_path":"backend/dealfinder/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":3894,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"445341777","text":"#!/usr/lical/bin/python3\n# ! -*- coding: utf-8 -*-\n\n# 自作の謎の(w)探索\n# 最大桁から1つずつ消していく。\n# O(10^n)かな?\n\nimport sys\nfrom decimal import Decimal\n\nY, M = (int(i) for i in sys.stdin.readline().split())\n\nYM = (Y-2013)*12+M\npre_i = 0\n\n\ndef cal_ym(q):\n q = Decimal(q)\n return (12*q+(q*(q+1))/Decimal(2))\n\n\ndef tansaku(start, end, step):\n pre_i = start\n for i in range(int(start), int(end), int(step)):\n if YM<=cal_ym(i):\n if step==1:\n return (i)\n else:\n return (tansaku(pre_i, end, step/10))\n else:\n pre_i = i\n\n\nj = tansaku(0, 10**17, 10**16)-1\nu_y = 2013+j\nu_m = YM-cal_ym(j)\nprint(u_y, int(u_m))\n","sub_path":"library_python/AtCoder_Event/utpc2013/utpc2013b_bekkai.py","file_name":"utpc2013b_bekkai.py","file_ext":"py","file_size_in_byte":729,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"207729039","text":"import collections\na=dict(collections.Counter('halloklempnerdasistfantastischfluggegecheimen'))\nprint(\"Введите стоп слово\",'\\n',sep='')\nc=str(input())\nf=[0]*len(c)\nb=str('halloklempnerdasistfantastischfluggegecheimen')\nt=float(1)\nq=0 \nfor i in range(len(c)):\n for j in range(len(b)):\n if (c[i]==b[j]):\n f[i]=f[i]+1 \nfor i in range(len(c)):\n if (f[i]==0):\n q=1 \nif (q==1):\n print('Сигизмунд не знает букву')\nelse:\n for i in range(len(c)):\n t=t*(a[c[i]]/len(b))\n print(\"Вероятность равна=\",t)\n \n","sub_path":"Practice/18/Python/PY 18.py","file_name":"PY 18.py","file_ext":"py","file_size_in_byte":596,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"109501162","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"Simple Python wrapper for runTagger.sh script for CMU's Tweet Tokeniser\nand Part of Speech tagger: http://www.ark.cs.cmu.edu/TweetNLP/\n\nPOS tags are represented with a single ASCII symbol. In brief:\n\n* __Nominal__\n `N` common noun\n `O` pronoun (personal/WH; not possessive)\n `^` proper noun\n `S` nominal + possessive\n `Z` proper noun + possessive\n* __Other open-class words__\n `V` verb incl. copula, auxiliaries\n `A` adjective\n `R` adverb\n `!` interjection\n* __Other closed-class words__\n `D` determiner\n `P` pre- or postposition, or subordinating conjunction\n `&` coordinating conjunction\n `T` verb particle\n `X` existential _there_, predeterminers\n* __Twitter/online-specific__\n `#` hashtag (indicates topic/category for tweet)\n `@` at-mention (indicates another user as a recipient of a tweet)\n `~` discourse marker, indications of continuation of a message across multiple tweets\n `U` URL or email address\n `E` emoticon\n* __Miscellaneous__\n `$` numeral\n `,` punctuation\n `G` other abbreviations, foreign words, possessive endings, symbols, garbage\n* __Other Compounds__\n `L` nominal + verbal (e.g. _i'm_), verbal + nominal (_let's_, _lemme_)\n `M` proper noun + verbal\n `Y` `X` + verbal\n\nModified August 2017 by John Meade\n\"\"\"\n\n\nimport shlex, pexpect\nfrom time import time, sleep\n\n\nprinthead = '[ {:^15} ] '.format( 'Tagger' )\ndef p( msg ): print( printhead + msg )\n\n\nclass TweetTagger:\n\n\n def __init__( self, java_opts='-XX:ParallelGCThreads=2 -Xmx500m', jarpath='ark-tweet-nlp-0.3.2/ark-tweet-nlp-0.3.2.jar' ):\n # NOTE default java options are directly lifted from original\n # java implementation. Example of the executed command:\n # java -XX:ParallelGCThreads=2 -Xmx500m -jar vendor/ark-tweet-nlp-0.3.2/ark-tweet-nlp-0.3.2.jar --output-format conll\n self.cmd = ' '.join([ 'java', java_opts, '-jar', jarpath, '--output-format', 'conll' ])\n self.proc = pexpect.spawn( self.cmd, echo=False )\n self.proc.expect('Listening on stdin for input\\. \\(\\-h for help\\)')\n\n\n def kill( self ):\n self.proc.kill( 1 )\n\n\n def __enter__( self ):\n return self\n\n\n def __exit__( self, typ, value, traceback ):\n self.kill()\n\n\n def _parse_raw_result( self, raw_result ):\n \"\"\"Parse the tab-delimited returned lines, modified from:\n https://github.com/brendano/ark-tweet-nlp/blob/master/scripts/show.py\n \"\"\"\n rows = raw_result.split('\\r\\n')\n for line in rows:\n line = line.strip() # remove '\\n'\n if len(line) > 0:\n if line.count( '\\t' ) == 2:\n parts = line.split( '\\t' )\n tokens = parts[0]\n tags = parts[1]\n confidence = float( parts[2] )\n yield tokens, tags, confidence\n\n\n def batch( self, tweets ):\n \"\"\"Call runTagger.sh on a list of tweets, parse the result, return lists of tuples of (term, type, confidence)\"\"\"\n\n # remove carriage returns and newlines, as they are interpretted as\n # tweet separators by the tagger\n tweets_cleaned = [ tw.replace('\\n', ' ').replace('\\r', ' ') for tw in tweets ]\n message = \"\\n\".join( tweets_cleaned )\n\n # force UTF-8 encoding (from internal unicode type) to avoid .communicate encoding error as per:\n # http://stackoverflow.com/questions/3040101/python-encoding-for-pipe-communicate\n # message = message.encode( 'utf-8' )\n\n # print(message, file=self.proc.stdin, flush=True)\n self.proc.write( (message + '\\n\\n').encode('utf-8') )\n # the output of the tagger will terminate with 4 newlines => use this\n # to detect batch completion\n try:\n self.proc.expect( '\\r\\n\\r\\n\\r\\n', timeout=30 )\n except:\n p('Exception while tagging tweets')\n return []\n\n # parse into a list of strings, ie a result for each input message\n raw = self.proc.before.strip().decode('utf-8')\n # occassionally there is a header to trim off...\n raw = ''.join( raw.split('Detected text input format') ).strip()\n # avoid missing result for empty lines?\n # pos_result = pos_result.replace( \"\\n\\n\", \"\\n\\n\\n\" )\n # split messages by double carriage returns\n raw_results = raw.split( '\\r\\n\\r\\n' )\n # parse each raw result into it's PoS tags\n return [ list( self._parse_raw_result( r ) ) for r in raw_results ]\n\n\nif __name__ == \"__main__\":\n with TweetTagger( jarpath='vendor/ark-tweet-nlp-0.3.2.jar' ) as tw_tag:\n print( \"\\nTweet PoS demo (first call will be slow while Java is booting up)\")\n\n def demo(tweets):\n print( '\\nProcessing: ' + str( tweets ) )\n ti = time()\n res = tw_tag.batch( tweets )\n tf = time()\n print( 'Results: ' + str( res ) )\n print( 'Took: {} seconds'.format( tf - ti ) )\n\n demo([ 'this is a message', 'and a second message' ])\n demo([ 'this is a third message', 'and a fourth message' ])\n demo([ 'this is a fifth message', '' ])\n","sub_path":"CMUTweetTagger.py","file_name":"CMUTweetTagger.py","file_ext":"py","file_size_in_byte":5156,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"183836609","text":"'''\nComputation of Sequence Charge Decoration\nAuthor: Yu Qi\n'''\n\ndef BranchedSeqCharge(b1, b2, tail, branchpoint):\n # `\n # ` <-- branch1 | b1 and b2 are branch sequences without the \"tail\"\n # ` V\n # ` (NC)_ _ _ _ <--- TAIL (starts after \"NC\") NC(from formula) is branchpoint\n # ,\n # ,\n # , ^\n # , <-- branch2 \\ tail begins at the 5th char (index 4) in this example\n\n total1 = 0\n total2 = 0\n total3 = 0\n\n pCharged = [\"K\", \"H\", \"R\"]\n nCharged = [\"D\", \"E\"]\n\n #Section 1\n #summation from m=2 (second letter of b1 sequence) to N1\n #summation from n=1 (first letter of b1) to m-1 (second to last letter of b1 sequence)\n\n #First we must add the first and second branches together, with NC in between:\n tempstring = b1 + branchpoint\n\n for x in range(0,len(b2)):\n tempstring += b2[len(b2) - x - 1]\n\n for m in range (1, len(tempstring)): #m is index number of the string - in formula it starts at 2 instead of 1\n if(tempstring[m] in pCharged): #assign Q values based on amino\n q1 = 1\n elif(tempstring[m] in nCharged):\n q1 = -1\n else:\n q1 = 0\n for n in range (0, m): #n can be up to m-1, which is essentially c (b/c m = c + 1), but still is the index number of the string\n if(tempstring[n] in pCharged):\n q2 = 1\n elif(tempstring[n] in nCharged):\n q2 = -1\n else:\n q2 = 0\n total1 += q1 * q2 * ((m-n)**0.5)\n print(\"Total1: \" + str(total1) + \"\\n\")\n\n #Section 2\n #to be determined based on sequence indexes, math part is complete.\n\n\n for m in range(0, len(tail)):\n if(tail[m] in pCharged):\n q1 = 1\n elif(tail[m] in nCharged):\n q1 = -1\n else:\n q1 = 0\n mIndex = len(b1) + m + 2\n for n in range(0, len(b1)): #if using for index, these are not proper index values\n if(b1[m] in pCharged):\n q2 = 1\n elif(b1[m] in nCharged):\n q2 = -1\n else:\n q2 = 0\n nIndex = n + 1\n nC = len(b1) + 1\n total2 += (q1*q2*((nC - nIndex)**2))/((mIndex-(nIndex))**(3/2))\n print(\"Total2: \" + str(total2) + \"\\n\")\n\n\n #Section 3\n #to be determined based on sequence indexes, math part is complete.\n\n #concatenate and reverse b2 with branching point\n # for instance \"ABC\"(b2) + \"D\" (branching point) becomes \"DCBA\"\n\n tempb2 = b2 + branchpoint\n tempb2reverse = \"\"\n\n for x in range(0,len(tempb2)):\n tempb2reverse += b2[len(b2) - x - 1]\n\n for m in range(0, len(tempb2reverse)): #if using for index, these are not proper index values\n if(tempb2reverse[m] in pCharged):\n q1 = 1\n elif(tempb2reverse[m] in nCharged):\n q1 = -1\n else:\n q1 = 0\n mIndex = len(b1) + 1 + m\n for n in range(0, len(tail)):\n if(tail[n] in pCharged):\n q2 = 1\n elif(tail[n] in nCharged):\n q2 = -1\n else:\n q2 = 0\n nIndex = len(b1) + n + 2\n nC = len(b1) + 1\n total3 += (q1*q2*((mIndex-nC)**2))/((mIndex+nIndex-(2*(nC)))**(3/2))\n print(\"Total3: \" + str(total3) + \"\\n\")\n\n return total1 + total2 + total3\n\n\n","sub_path":"BranchedSeqCharge.py","file_name":"BranchedSeqCharge.py","file_ext":"py","file_size_in_byte":3392,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"360226091","text":"#dot = Digraph(comment='The Test Table')\n#dot.node('Root', 'Root')\nimport re\nwith open(\"output_result1.txt\", \"r\") as f:\n a = f.readlines()\n \nfrom graphviz import Digraph\ndot = Digraph(comment='The Test Table')\n# 添加圆点A,A的标签是Dot A\n#dot.node('Root', 'Root')\n \ndata=[]\nFirst_node=[]\nTotal_node=[]\nedge_info=[]\nfor thing in a :\n thing=thing.replace(' ','').replace('\\n','').replace('_name:',' ')\n temp=thing.split(' ')\n id1=temp[0].split('id')[1]\n id1_name=temp[1].split('id')[0]\n id2=temp[1].split('id')[1]\n id2_name=temp[2].split('connection')[0]\n Total_node.append(id1+'+'+id1_name)\n Total_node.append(id2+'+'+id2_name)\n edge_info.append(id1+'+'+id2)\n#for index1,thing1 in enumerate(data):\n# if(index1<100):\n# temp=thing1.split(' ')\n# First_node.append(temp[0])\n# for index2,thing2 in enumerate(temp):\n# Total_node.append(thing2)\n# if(index2100\nnp.random.seed(101)\ndata = np.random.randint(1, 101, (100, 5))\nprint(data)\n\n# Create 2D vis w/Colorbar and Title\nplt.imshow(data, cmap=\"coolwarm\", aspect='auto')\nplt.title('title')\nplt.colorbar()\nplt.show()\n\n# Create pandas dateframe\ndf = pd.DataFrame(data)\nprint(df)\n\n# Show scatter plot of col 0 vs col 1\ndf.plot(x=0, y=1, kind='scatter')\nplt.show()\n\n# Scale data to have minimum of 0 and max of 1\nscaler = preprocessing.MinMaxScaler()\nscaled_data = scaler.fit_transform(data)\nprint(scaled_data)\n\n# Rename columns, split data into training and test\ndf.columns = ['f1', 'f2', 'f3', 'f4', 'label']\nX = df[['f1', 'f2', 'f3', 'f4']]\ny = df[['label']]\n\nX_train, X_test, y_train, y_test = model_selection.train_test_split(\n X, y, test_size=.33, random_state=42)\n\nprint(X_train)\n","sub_path":"review/_exercise.py","file_name":"_exercise.py","file_ext":"py","file_size_in_byte":948,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"313026566","text":"__author__ = 'PE20060014 Chen Wang'\n\nimport os\nimport math\nimport numpy as np\nimport random\nimport matplotlib.pyplot as plt\nfrom savegif import create_gif\n\n# 计算基尼系数\ndef gini(initial_money, people, wealth):\n cumul = 0\n ave = np.mean(wealth)\n #print(ave)\n max = ave * people * (people + 1) / 2\n for i in range(people):\n cumul += (people - i) * (ave - wealth[i])\n gini_index = cumul / max\n return gini_index\n\n# 玻尔兹曼分布曲线\ndef fit1(initial_money, people, step):\n x = np.arange(0.5, 0.5 + people)\n y = np.zeros(people)\n for i in range(people):\n y[i] = -initial_money * math.log(1 - x[i] * 1.0 / people)\n x = np.reshape(x, (len(x), 1))\n y = np.reshape(y, (len(y), 1))\n fitline = np.hstack((x, y))\n return fitline\n\n#S1:不允许负债的情况\ndef generate1(days):\n import random\n wealth = []\n order = []\n for i in range(100): #定义100个人的序号\n order.append(i+1)\n for i in range(100): #定义初始财富\n wealth.append(100)\n for j in range(days): #模拟22-65岁 共15695天\n for i in range(100): #模拟每天100个人的财富交换\n if wealth[i]==0: #不可负债\n #break #如果有人财富为0了 就停止游戏\n continue #财富为0时,不支出,游戏继续\n else:\n wealth[i] = wealth[i] - 1\n rand_guy=random.choice([x for x in range(0,100) ]) #100平均分布取出一个人\n wealth[rand_guy] = wealth[rand_guy] + 1\n wealth.sort()#对财富排序\n # 检验财富有无流失,return 10000则没有流失\n # print(sum(wealth))\n gini_index = gini(100, 100, wealth) # 计算基尼系数\n #print(gini_index)\n # 显示100个人的财富柱状图\n #fig1 = plt.figure()\n #玻尔兹曼分布曲线\n fitline = fit1(100, 100, j)\n if ( j % 200 == 0):\n plt.plot(fitline[:, 0], fitline[:, 1], 'r')\n plt.bar(order,wealth,0.5,color=\"blue\")\n plt.title('Scenario1:Debt Disabled\\nDays:' + str(j) + ' Gini Index:' + str(gini_index))\n plt.xlabel('Order of participants')\n plt.ylabel('Wealth')\n #plt.legend()\n plt.draw()\n plt.pause(0.00001)\n plt.savefig('../pythonProject/Scenario1/'+j//200*str(1)+'.png')\n if(j!=days): #保留最运行结束时的数据\n plt.close()\n gif_name = 'Scenario1.gif'\n pic_path = '../pythonProject/Scenario1/' # 指定文件路径\n duration = 0.1\n create_gif(gif_name, pic_path, duration)\n return wealth, order\n\n#S2:允许负债的情况\ndef generate2(days):\n wealth = []\n order = []\n for i in range(100):\n order.append(i+1)\n for i in range(100):\n wealth.append(100)\n for j in range(days):#可负债情况不需判断某人的金钱是否为0\n for i in range(100):\n wealth[i]=wealth[i]-1\n rand_guy=random.choice([x for x in range(0,100) if x!=i])\n wealth[rand_guy]=wealth[rand_guy]+1\n wealth.sort()\n gini_index = gini(100, 100, wealth) # 计算基尼系数\n # 显示100个人的财富柱状图\n # fig1 = plt.figure()\n if ( j % 200 == 0):\n plt.bar(order,wealth,0.5,color=\"blue\")\n plt.title('Scenario2:Debt Enabled\\nDays:' + str(j) + ' Gini Index:' + str(gini_index))\n plt.xlabel('Order of participants')\n plt.ylabel('Wealth')\n #plt.legend()\n plt.draw()\n plt.pause(0.00001)\n plt.savefig('../pythonProject/Scenario2/'+j//200*str(1)+'.png')\n if(j!=days): #保留最运行结束时的数据\n plt.close()\n gif_name = 'Scenario2.gif'\n pic_path = '../pythonProject/Scenario2/' # 指定文件路径\n duration = 0.1\n create_gif(gif_name, pic_path, duration)\n return wealth, order\n#wealth,order=generate()\n\n#S3:对富人收税的情况:\ndef generate3(days):\n import random\n wealth = []\n order = []\n for i in range(100):\n order.append(i+1)\n for i in range(100):\n wealth.append(100.0)\n for j in range(days):\n for i in range(100):\n wealth[i]=wealth[i]-1\n rand_guy=random.choice([x for x in range(0,100) if x!=i])\n if wealth[rand_guy]>=200.0:#对拥有大于200元的玩家征收20%的税\n wealth[rand_guy] = wealth[rand_guy] + 0.5\n wealth_transi = wealth[:]\n for m in range(5):#对最贫穷的2位玩家每人补贴0.25元\n ind=wealth_transi.index(min(wealth_transi))\n wealth[ind]=wealth[ind]+0.1\n wealth_transi[ind]=100000000\n else:\n wealth[rand_guy] = wealth[rand_guy] + 1\n wealth.sort()\n gini_index = gini(100, 100, wealth) # 计算基尼系数\n # 显示100个人的财富柱状图\n # fig1 = plt.figure()\n if (j % 200 == 0):\n plt.bar(order,wealth,0.5,color=\"blue\")\n plt.title('Scenario3:Debt and 50%TAX Enabled \\nDays:' + str(j) + ' Gini Index:' + str(gini_index))\n plt.xlabel('Order of participants')\n plt.ylabel('Wealth')\n #plt.legend()\n plt.draw()\n plt.pause(0.00001)\n plt.savefig('../pythonProject/Scenario3/' + j // 200 * str(1) + '.png')\n if(j!=days): #保留最运行结束时的数据\n plt.close()\n gif_name = 'Scenario3.gif'\n pic_path = '../pythonProject/Scenario3/' # 指定文件路径\n duration = 0.1\n create_gif(gif_name, pic_path, duration)\n return wealth, order\n#wealth,order=generate()\n\n#S4:富二代出现\ndef generate4(days):\n wealth = []\n order = []\n for i in range(100):\n order.append(i+1)\n for i in range(10):#前10位玩家是富二代,生来就比普通玩家多2倍财富\n wealth.append(400.0)\n for i in range(10,100):\n wealth.append(100.0)\n for j in range(days):#模拟15695轮财富分配\n for i in range(100):\n wealth[i]=wealth[i]-1\n rand_guy=random.choice([x for x in range(0,100) if x!=i])\n wealth[rand_guy]=wealth[rand_guy]+1\n # 排序\n wealth_transi = wealth[:]\n wealth_sort = []\n order_rich = []\n wealth_rich = []\n order_normal = []\n wealth_normal = []\n for i in range(100):\n ind_min = wealth_transi.index(min(wealth_transi))\n wealth_sort.append(min(wealth_transi))\n wealth_transi[ind_min] = float('inf')\n if ind_min in [0, 1, 2, 3, 4, 5, 6, 7, 8, 9]:\n order_rich.append(len(wealth_sort) - 1)\n wealth_rich.append(wealth_sort[-1])\n else:\n order_normal.append(len(wealth_sort) - 1)\n wealth_normal.append(wealth_sort[-1])\n gini_index = gini(100, 100, wealth_sort) # 计算基尼系数\n print(gini_index)\n # barchart\n # fig1 = plt.figure()\n if (j % 200 == 0):\n plt.bar(order_rich, wealth_rich, 0.5, color=\"red\")\n plt.bar(order_normal, wealth_normal, 0.5, color=\"blue\")\n plt.title('Scenario4:Born Rich with 400\\nDays:' + str(j) + ' Gini Index:' + str(gini_index))\n plt.xlabel('Order of participants')\n plt.ylabel('Wealth')\n plt.legend()\n plt.draw()\n plt.pause(0.00001)\n plt.savefig('../pythonProject/Scenario4/' + j // 200 * str(1) + '.png')\n if (j != days): # 保留最运行结束时的数据\n plt.close()\n # plt.show()\n gif_name = 'Scenario4.gif'\n pic_path = '../pythonProject/Scenario4/' # 指定文件路径\n duration = 0.1\n create_gif(gif_name, pic_path, duration)\n return wealth, order\n\n#S5更努力有用吗?\ndef generate5(days):\n wealth = []\n order = []\n for i in range(100):\n order.append(i+1)\n for i in range(10):\n wealth.append(50)\n for i in range(10,90):\n wealth.append(100)\n for i in range(90,100):\n wealth.append(200) #后10名为初始携带200元的富二代\n for j in range(days):\n for i in range(100):\n wealth[i]=wealth[i]-1\n rand_guylist1=[x for x in range(0,10) if x!=i]#定义0到9号玩家序号\n rand_guylist2 = [x for x in range(10, 100) if x != i]#定义10到99号玩家序号\n ind=random.randint(0,1000)\n # 前10位玩家得到钱的概率要微大于后90位玩家,几率大了将近万分之一,以此模拟10位更努力的玩家\n if ind<=100:#在0到9号玩家中取一个人的概率是11/1000\n rand_guy = random.choice(rand_guylist1)\n else:#在10到99号玩家中取一个人的概率是(1000-11)/1000\n rand_guy = random.choice(rand_guylist2)\n wealth[rand_guy]=wealth[rand_guy]+1\n #wealth.sort()#为展示前10位玩家财富变化,不进行排序\n # 排序\n wealth_transi = wealth[:]\n wealth_sort = []\n order_rich = []\n wealth_rich = []\n order_hard = []\n wealth_hard = []\n order_normal = []\n wealth_normal = []\n for i in range(100):\n ind_min = wealth_transi.index(min(wealth_transi))\n wealth_sort.append(min(wealth_transi))\n wealth_transi[ind_min] = float('inf')\n if ind_min in [0, 1, 2, 3, 4, 5, 6, 7, 8, 9]:\n order_hard.append(len(wealth_sort) - 1)\n wealth_hard.append(wealth_sort[-1])\n elif ind_min in [90, 91, 92, 93, 94, 95, 96, 97, 98, 99]:\n order_rich.append(len(wealth_sort) - 1)\n wealth_rich.append(wealth_sort[-1])\n else:\n order_normal.append(len(wealth_sort) - 1)\n wealth_normal.append(wealth_sort[-1])\n # print(len(order_hard)==len(order_hard),len(order_rich)==len(wealth_rich),len(order_normal)==len(wealth_normal))\n gini_index = gini(100, 100, wealth_sort) # 计算基尼系数\n # barchart\n # fig1 = plt.figure()\n if (j%200==0):\n plt.bar(order_rich, wealth_rich, 0.5, color=\"red\")\n plt.bar(order_hard, wealth_hard, 0.5, color=\"green\")\n plt.bar(order_normal, wealth_normal, 0.5, color=\"blue\")\n plt.title('Scenario5:Hard-working by 0.01% VS Born Rich 200%\\nDays:' + str(j) + ' Gini Index:' + str(gini_index))\n plt.xlabel('Order of participants')\n plt.ylabel('Wealth')\n plt.legend()\n plt.draw()\n plt.pause(0.00001)\n plt.savefig('../pythonProject/Scenario5/' + j // 200 * str(1) + '.png')\n if (j != days): # 保留最运行结束时的数据\n plt.close()\n # plt.show()\n gif_name = 'Scenario5.gif'\n pic_path = '../pythonProject/Scenario5/' # 指定文件路径\n duration = 0.1\n create_gif(gif_name, pic_path, duration)\n return wealth, order\n\n\n\ndef main():\n day=15695\n generate1(day)\n\nif __name__ == '__main__':\n main()","sub_path":"面向科学问题求解的编程实践/FinalProject/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":11269,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"483811044","text":"#encoding:utf-8\nimport re\n\n#Const_File_Format=[\"asp\",\"aspx\",\"html\",\"htm\",\"php\"]\np = re.compile(r'(i say,\\nhello)')\n\nfor line in open(\"list.txt\").readlines():\n\tline=line.strip('\\n')\n\tread_file = open(line, 'r').read()\n\tif \"百家乐\" in read_file:\n\t\twith open(\"log.txt\", 'a') as output:\n\t\t\toutput.write(line+\"\\n\")\n\topen(line, 'w').write(p.sub(r'', read_file))","sub_path":"regula/替换.py","file_name":"替换.py","file_ext":"py","file_size_in_byte":358,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"320337427","text":"from handler.base_plugin_command import CommandPlugin\nfrom vk.utils import Wait\nfrom vk.helpers import upload_photo\n\nimport asyncio\nimport aiohttp\nimport io\n\n\nclass DispatchPlugin(CommandPlugin):\n __slots__ = (\"admins\", )\n\n def __init__(self, *commands, prefixes=None, strict=False, admins=()):\n \"\"\"Allows admins to send out messages to users.\"\"\"\n\n super().__init__(*commands, prefixes=prefixes, strict=strict)\n\n self.admins = admins\n\n async def process_message(self, msg):\n if msg.user_id not in self.admins and not msg.meta.get(\"is_moder\"):\n return await msg.answer(\"Вы не администратор.\")\n\n cmd_len = len(msg.meta.get(\"__prefix\", \"\")) + len(msg.meta.get(\"__command\", \"\"))\n\n message = msg.full_text[cmd_len:].strip()\n attachment = \"\"\n\n for a in await msg.get_full_attaches():\n if a.type != \"photo\":\n attachment += str(a) + \",\"\n\n if a.type == \"photo\" and a.url:\n async with aiohttp.ClientSession() as sess:\n async with sess.get(a.url) as resp:\n new_a = await upload_photo(self.api, io.BytesIO(await resp.read()))\n\n if not new_a:\n continue\n\n attachment += str(new_a) + \",\"\n\n await msg.answer(\"Приступаю к рассылке!\")\n\n if await self.dispatch(message, attachment) is False:\n return await msg.answer(\"Ошибка при отправлении! Попробуйте позже!\")\n\n return await msg.answer(\"Рассылка закончена!\")\n\n async def dispatch(self, message, attachment):\n dialogs = await self.bot.api.messages.getDialogs(count=1, preview_length=1)\n\n if not dialogs or \"count\" not in dialogs:\n return False\n\n dialogs = dialogs[\"count\"]\n users = set()\n\n tasks = []\n\n with self.bot.api.mass_request():\n for i in range(int(dialogs / 200) + 1):\n tasks.append(await self.bot.api(wait=Wait.CUSTOM).messages.getDialogs(count=200, preview_length=1))\n\n future = asyncio.gather(*tasks, return_exceptions=True)\n\n await asyncio.wait_for(future, None)\n\n for r in future.result():\n if not r:\n continue\n\n for dialog in r.get(\"items\", []):\n if \"message\" not in dialog or \"user_id\" not in dialog[\"message\"]:\n continue\n\n users.add(int(dialog[\"message\"][\"user_id\"]))\n\n for i, u in enumerate(users):\n await self.bot.api(wait=Wait.NO).messages.send(user_id=u, message=message, attachment=attachment)\n\n if i % 25 == 0:\n await asyncio.sleep(0.2)\n","sub_path":"plugins/misc/dispatch.py","file_name":"dispatch.py","file_ext":"py","file_size_in_byte":2791,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"200869629","text":"from django.conf.urls import url\nfrom django.contrib.staticfiles.urls import staticfiles_urlpatterns\nfrom . import views\nurlpatterns = [\n url(r'^add-crawler', views.Add_Crawlers.as_view(),name='add-crawler'),\n url(r'^execute-crawler', views.Execute_Main_Crawler.as_view(),name='execute-crawler'),\n url(r'^test-add', views.Test_Sample_Data.as_view(),name='test-add'),\n url(r'^fetch-params-csv', views.Fetch_Params_From_Csv.as_view(), name='fetch-csv'),\n url(r'^pick-task-row', views.Pick_TaskRow_To_Execute.as_view(), name='pick-task-row'),\n url(r'^execute-linkedin-local', views.Execute_Linkedin_Local.as_view(), name='execute-linkedin-local'),\n url(r'^pick-task-row', views.Pick_TaskRow_To_Execute.as_view(), name='pick-task-row'),\n url(r'^clean-linkedin-data', views.Scrape_Linkedin_Data.as_view(), name='clean-linkedin-data'),\n url(r'^clean-wiki-data', views.Scrape_Wiki_Data.as_view(), name='clean-wiki-data'),\n #url(r'^clean-mygov-data', views.Scrape_Mygov_Data.as_view(), name='clean-mygov-data'),\n url(r'^clean-raw-data', views.Clean_Raw_Data.as_view(), name='clean-raw-data'),\n]","sub_path":"crawlers/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1119,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"46583236","text":"#!/usr/bin/env python\n\n\n\nimport rospy\nfrom nav_msgs.msg import OccupancyGrid\nfrom map_msgs.msg import OccupancyGridUpdate\nfrom group_msgs.msg import People, Person, Groups\nfrom geometry_msgs.msg import Pose, PoseArray\nfrom algorithm import SpaceModeling\nfrom obstacles import adapt_parameters\n\nfrom human_awareness_msgs.msg import PersonTracker, TrackedPersonsList\n\nimport tf\nimport math\nimport copy\nimport actionlib\nimport numpy as np\n\nimport matlab.engine\neng = matlab.engine.start_matlab()\neng.cd(r'/home/flash/catkin_ws/src/adaptive_social_layers/scripts', nargout=0)\n\nSTRIDE = 65 # in cm\nMDL = 8000\n\n# Relation between personal frontal space and back space\nBACK_FACTOR = 1.3\n\n# Robot radius\nROBOT_DIM = 100 # in cm\n\ndef calc_o_space(persons):\n \"\"\"Calculates the o-space center of the group given group members pose\"\"\"\n c_x = 0\n c_y = 0\n \n# Group size\n g_size = len(persons)\n \n for person in persons:\n c_x += person[0] + math.cos(person[2]) * STRIDE\n c_y += person[1] + math.sin(person[2]) * STRIDE\n\n center = [c_x / g_size, c_y / g_size]\n\n return center\n\ndef rotate(px, py, angle):\n \"\"\"\n Rotate a point counterclockwise by a given angle around a given origin.\n The angle should be given in radians.\n \"\"\"\n qx = math.cos(angle) * px - math.sin(angle) * py\n qy = math.sin(angle) * px + math.cos(angle) * py\n\n return qx, qy\n\ndef get_index(x, y, width):\n \"\"\" \"\"\"\n \n return (y * width) + x\n\nclass PeoplePublisher():\n \"\"\"\n \"\"\"\n def __init__(self):\n \"\"\"\n \"\"\"\n rospy.init_node('PeoplePublisher', anonymous=True)\n \n rospy.Subscriber(\"/human_awareness_tracker/trackers_list\",TrackedPersonsList,self.callback,queue_size=1)\n # https://answers.ros.org/question/207620/global-map-update/\n # We need to subscribe both costmap and costmap update topic\n rospy.Subscriber(\"/map\",OccupancyGrid , self.callbackCm, queue_size=1)\n self.loop_rate = rospy.Rate(rospy.get_param('~loop_rate', 10.0))\n\n self.map_received = False\n self.pose_received = False\n\n self.data = None\n self.pub = rospy.Publisher('/people', People, queue_size=10)\n self.pubg = rospy.Publisher('/groups', Groups, queue_size=10)\n\n\n def callback(self,data):\n \"\"\"\n \"\"\"\n \n self.data = data\n self.pose_received = True\n \n\n def callbackCm(self, data):\n \"\"\" Costmap Callback. \"\"\"\n\n self.map = data\n self.map_grid = list(data.data)\n self.map_received = True\n \n def publish(self):\n \"\"\"\n \"\"\"\n \n data = self.data\n groups = []\n group = []\n\n persons = []\n\n listener = tf.TransformListener()\n\n while not rospy.is_shutdown():\n try:\n (trans,rot) = listener.lookupTransform('/map', '/base_footprint', rospy.Time(0))\n break\n except (tf.LookupException, tf.ConnectivityException, tf.ExtrapolationException):\n continue\n\n tx = trans[0]\n ty = trans[1]\n (_, _, t_yaw) = tf.transformations.euler_from_quaternion(rot)\n\n if not data.personList:\n groups = []\n else:\n for poseinfo in data.personList:\n\n rospy.loginfo(\"Person Detected\")\n pose = poseinfo.body_pose\n quartenion = [pose.orientation.x, pose.orientation.y, pose.orientation.z, pose.orientation.w]\n (_, _, yaw) = tf.transformations.euler_from_quaternion(quartenion)\n\n \n # Pose transformation from base footprint frame to map frame\n (px, py) = rotate(pose.position.x, pose.position.y, t_yaw)\n pose_x = px + tx\n pose_y = py + ty\n pose_yaw = yaw + t_yaw\n\n\n pose_person = (pose_x * 100, pose_y * 100, pose_yaw)\n persons.append(pose_person)\n\n # Run GCFF gcff.m Matlab function \n if persons:\n groups = eng.gcff(MDL,STRIDE, matlab.double(persons))\n \n if groups:\n app = SpaceModeling(groups) # Space modeling works in cm\n pparams,gparams = app.solve()\n\n ####\n # Obstacles works in cm -> Convert to meters\n ox = self.map.info.origin.position.x * 100\n oy = self.map.info.origin.position.y * 100\n origin = [ox, oy]\n resolution = self.map.info.resolution * 100\n width = self.map.info.width \n height = self.map.info.height \n map = self.map.data\n\n pparams_adapt, gparams_adapt = adapt_parameters(groups, pparams, gparams, resolution, map, origin, width, ROBOT_DIM)\n \n\n p = People()\n p.header.frame_id = \"/map\"\n p.header.stamp = rospy.Time.now()\n\n g = Groups()\n g.header.frame_id = \"/map\"\n g.header.stamp = rospy.Time.now()\n \n for idx,group in enumerate(groups):\n aux_p = People()\n aux_p.header.frame_id = \"/map\"\n aux_p.header.stamp = rospy.Time.now()\n\n \n gvarx = float(gparams_adapt[idx][0]) / 100.0 # cm to m\n gvary = float(gparams_adapt[idx][1]) / 100.0 # cm to m\n \n \n \n ############## FIXED\n # sx = 0.9\n # sy = 0.9\n #########################\n for pidx, person in enumerate(group):\n\n p1 = Person()\n p1.position.x = person[0] / 100.0 # cm to m\n p1.position.y = person[1] / 100.0 # cm to m\n p1.orientation = person[2]\n\n sx = pparams_adapt[idx][pidx][\"sx\"]/ 100.0\n sy = pparams_adapt[idx][pidx][\"sy\"] / 100.0\n\n sx_back = pparams_adapt[idx][pidx][\"sx_back\"] / 100.0\n \n p1.sx = sx \n p1.sy = sy \n p1.sx_back = sx_back \n \n p1.ospace = False\n p.people.append(p1)\n\n \n aux_p.people.append(p1)\n \n # Only represent o space for +2 individuals\n if len(group) > 1:\n p1 = Person()\n center = calc_o_space(group)\n p1.position.x = center[0] / 100.0 # cm to m\n p1.position.y = center[1] / 100.0 # cm to m\n p1.orientation = math.pi\n\n \n p1.sx = gvarx\n p1.sy = gvary\n\n \n p1.ospace = True\n p.people.append(p1)\n\n aux_p.people.append(p1)\n\n g.groups.append(aux_p)\n\n self.pub.publish(p)\n \n self.pubg.publish(g)\n\n else:\n p = People()\n p.header.frame_id = \"/map\"\n p.header.stamp = rospy.Time.now()\n self.pub.publish(p)\n\n g = Groups()\n g.header.frame_id = \"/map\"\n g.header.stamp = rospy.Time.now()\n self.pubg.publish(g)\n\n def run_behavior(self):\n while not rospy.is_shutdown():\n if self.pose_received:\n self.pose_received = False\n\n if self.map_received:\n #self.map_received = False\n\n self.publish()\n \n\nif __name__ == '__main__':\n people_publisher = PeoplePublisher()\n people_publisher.run_behavior()\n eng.quit()","sub_path":"scripts/people_publisher_obstacles_vision.py","file_name":"people_publisher_obstacles_vision.py","file_ext":"py","file_size_in_byte":7701,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"424348518","text":"#!/usr/bin/env python3\nfirst_name = \"Casey\"\nlast_name = \"Jackson\"\nfull_name = first_name + last_name\nprint(full_name)\nfullName = \"Casey\" \" \" \"Jackson\"\nprint(fullName)\n\nstars = \"*\" * 12\npounds = 5 * \"#\"\nprint(stars, \":\", pounds)\nx = \"Hello there\"\nprint('t' in x, 'ell' in x, 'hell' in x)\n","sub_path":"script2.py","file_name":"script2.py","file_ext":"py","file_size_in_byte":287,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"319733162","text":"# CS39AB - Cloud Computing - Summer 2021\n# Instructor: Thyago Mota\n# Description: Activity 11 - Extract the dollar to real exchange rate, saving it into a database. All quotes are then displayed using a dynamically generated web page. \n\nimport requests\nfrom datetime import datetime\nfrom bs4 import BeautifulSoup\nimport mysql.connector\nimport os\nfrom http.server import BaseHTTPRequestHandler, HTTPServer\n\nEXCHANGE_RATE_URL = 'https://themoneyconverter.com/USD/BRL'\n\nclass MyHandler(BaseHTTPRequestHandler):\n\n def do_GET(self):\n\n # only accept self.path = \"/\"\n if self.path != '/':\n return \n\n # get quote and update db\n req = requests.get(EXCHANGE_RATE_URL)\n soup = BeautifulSoup(req.content, 'html.parser')\n el = soup.find('output')\n exch_rate = float(el.text.split(' ')[3])\n today = datetime.today().strftime(\"%Y-%m-%d %H:%M:%S\")\n sql = f\"INSERT INTO quotes VALUES ('{today}', {exch_rate})\"\n cursor = self.db.cursor()\n cursor.execute(sql)\n db.commit()\n\n # generate a response\n self.send_response(200)\n self.send_header(\"Content-type\", \"text/html\")\n self.end_headers()\n\n self.wfile.write(bytes('''\n \n \n Dollar to Real\n \n \n ''', \"utf-8\"))\n self.wfile.write(bytes('''\n \n
\n \n
\n
Date
\n
Time
\n
Exchange Rate
\n
\n \n \n ''', \"utf-8\"))\n sql = \"SELECT `datetime`, quote FROM quotes ORDER BY `datetime` DESC\"\n cursor = self.db.cursor(buffered = True)\n cursor.execute(sql)\n for date_time, quote in cursor:\n date = date_time.date()\n time = date_time.time()\n self.wfile.write(bytes(f\"
\n \n \n ''', \"utf-8\"))\n\nif __name__ == \"__main__\":\n\n # delete the following lines once you are satisfied with the db connection and before creating the docker image\n os.environ['DB_HOST'] = 'dollar2real.cvhpjdm21h9e.us-west-1.rds.amazonaws.com'\n os.environ['DB_NAME'] = 'dollar2real'\n os.environ['DB_USER'] = 'dollar2real'\n os.environ['DB_PASSWORD'] = '135791'\n\n # attempt to connect to MySQL\n db = mysql.connector.connect(\n host = os.getenv('DB_HOST'),\n database = os.getenv('DB_NAME'),\n user = os.getenv('DB_USER'),\n password = os.getenv('DB_PASSWORD')\n )\n\n # attempt to start a web server\n my_handler = MyHandler \n my_handler.db = db\n webServer = HTTPServer(('0.0.0.0', 8000), my_handler)\n print(\"Ready to serve!\")\n\n try:\n webServer.serve_forever()\n except KeyboardInterrupt:\n pass\n\n webServer.server_close()\n print(\"Server stopped.\")\n db.close()","sub_path":"hwk_04_docker/src/hw04.py","file_name":"hw04.py","file_ext":"py","file_size_in_byte":4489,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"161760878","text":"# -*- coding: utf-8 -*-\n\"\"\"\n Spatial AutoCorrelation - Run Script\n ~~~~~~~~~\n\n Run with:\n python run_morans.py shapefiles/my_shapes.shp \\\n 'Density_Col' 'Population_Col' 'etc_col' -f 'Filter_Col'\n\n :copyright: (c) 2015 by Joe Hand, Santa Fe Institute.\n :license: MIT\n\"\"\"\nimport argparse\nimport logging\nimport pickle\nimport time\n\nfrom datetime import timedelta\n\nimport pandas as pd\n\nfrom spatial_auto import run_moran_analysis\n\nparser = argparse.ArgumentParser(\n description=\"Run Moran's on Shapefile (optional filter)\")\nparser.add_argument('shapefile', type=str,\n help='source shapefile for analysis')\nparser.add_argument('analysis_vars', nargs='+',\n help='columns to run Morans I analysis on')\nparser.add_argument('-f', '--filter', type=str,\n help='Filter Shapefile by Column (Optional)')\nparser.add_argument('--show-logs', dest='log', action='store_true')\nparser.add_argument('--no-logs', dest='log', action='store_false')\nparser.set_defaults(log=True)\nparser.add_argument('--logs-file', dest='log_file', action='store_true')\nparser.set_defaults(log_file=False)\nargs = parser.parse_args()\n\nif __name__ == '__main__':\n t = time.process_time()\n if args.log_file:\n logging.basicConfig(format='%(asctime)s \\n \\t %(message)s',\n filename='morans.log', level=logging.DEBUG,\n datefmt='%m/%d/%Y %I:%M:%S %p')\n elif args.log:\n logging.basicConfig(format='%(asctime)s \\n \\t %(message)s',\n level=logging.DEBUG,\n datefmt='%m/%d/%Y %I:%M:%S %p')\n else:\n logging.basicConfig(format='%(asctime)s \\n \\t %(message)s',\n level=logging.WARNING,\n datefmt='%m/%d/%Y %I:%M:%S %p')\n if args.filter:\n filter_col = args.filter\n else:\n filter_col = None\n\n logging.info('Starting Analysis')\n try:\n results = run_moran_analysis(\n args.shapefile, args.analysis_vars, filter_column=filter_col)\n pickle.dump(results, open( \"results.p\", \"wb\" ))\n except (SystemExit, KeyboardInterrupt):\n raise\n except Exception as e:\n logging.exception('\\n\\nError: ')\n\n logging.info('Finished Calculations \\n\\n')\n\n try:\n results_df = []\n keys = []\n for shapefile, values in results:\n df = pd.DataFrame(values).transpose()\n if shapefile in keys:\n val = 1\n while True:\n shapefile = '{}_{}'.format(shapefile,val)\n if shapefile not in keys:\n break\n val += 1\n keys.append(shapefile)\n del(df['COLUMN']) # add this as a key later\n results_df.append(df)\n\n results_log = '{} RESULTS \\n'.format(shapefile.upper())\n results_log += df.to_string()\n results_log += '\\n'\n logging.debug(results_log)\n results_df = pd.concat(results_df, keys=keys, names=['CITY', 'COLUMN'], axis=0)\n results_df.to_csv('results.csv')\n except:\n logging.exception('Some error exporting results: ')\n logging.debug('Total elapsed time {}'.format(\n str(timedelta(seconds=time.process_time() - t))))\n","sub_path":"run_morans.py","file_name":"run_morans.py","file_ext":"py","file_size_in_byte":3339,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"199735316","text":"import copy\nimport datetime\n\nimport numpy as np\nimport pandas as pd\nfrom tqdm import tqdm\n\ndef smoothen_triplegs(triplegs, method='douglas-peucker', tolerance=1.0):\n \"\"\"reduces number of points while retaining structure of tripleg\n Parameters\n ----------\n triplegs: shapely file\n triplegs to be reduced\n method: method used to smoothen\n only the douglas-peucker method is available so far\n tolerance: float\n a higher tolerance removes more points; the units of tolerance are the same as the projection of the input geometry\n \"\"\"\n input_copy = copy.deepcopy(triplegs)\n origin_geom = input_copy.geom\n simplified_geom = origin_geom.simplify(tolerance, preserve_topology=False)\n input_copy.geom = simplified_geom\n\n return input_copy\n\ndef _temp_trip_stack_has_tripleg(temp_trip_stack):\n \"\"\"\n Check if a trip has at least 1 tripleg\n Parameters\n ----------\n temp_trip_stack : list\n list of dictionary like elements (either pandas series or\n python dictionary). Contains all elements\n that will be aggregated into a trip\n\n Returns\n -------\n Bool\n \"\"\"\n\n has_tripleg = False\n for row in temp_trip_stack:\n if row['type'] == 'tripleg':\n has_tripleg = True\n break\n\n return has_tripleg\n\n\ndef _create_trip_from_stack(temp_trip_stack, origin_activity, destination_activity):\n \"\"\"\n Aggregate information of trip elements in a structured dictionary\n\n Parameters\n ----------\n temp_trip_stack : list\n list of dictionary like elements (either pandas series or python dictionary). Contains all elements\n that will be aggregated into a trip\n origin_activity : dictionary like\n Either dictionary or pandas series\n destination_activity : dictionary like\n Either dictionary or pandas series\n\n Returns\n -------\n dictionary\n\n \"\"\"\n\n # this function return and empty dict if no tripleg is in the stack\n first_trip_element = temp_trip_stack[0]\n last_trip_element = temp_trip_stack[-1]\n\n # all data has to be from the same user\n assert origin_activity['user_id'] == last_trip_element['user_id']\n\n # double check if trip requirements are fulfilled\n assert origin_activity['activity'] == True\n assert destination_activity['activity'] == True\n assert first_trip_element['activity'] == False\n\n trip_dict_entry = {'user_id': origin_activity['user_id'],\n 'started_at': first_trip_element['started_at'],\n 'finished_at': last_trip_element['finished_at'],\n 'origin_staypoint_id': origin_activity['id'],\n 'destination_staypoint_id': destination_activity['id'],\n 'tpls': [tripleg['id'] for tripleg in temp_trip_stack if tripleg['type'] == 'tripleg'],\n 'spts': [tripleg['id'] for tripleg in temp_trip_stack if tripleg['type'] == 'staypoint']}\n \n return trip_dict_entry\n\n\ndef generate_trips(stps_input, tpls_input, gap_threshold=15, id_offset=0, print_progress=False):\n \"\"\" Generate trips based on staypoints and triplegs\n\n `generate_trips` aggregates the staypoints `stps_input` and `tpls_input` into `trips` which are returned\n in a new DataFrame. The function returns new versions of `stps_input` and `tpls_input` that are identically except\n for additional id's that allow the matching between staypoints, triplegs and trips.\n\n Parameters\n ----------\n stps_input : GeoDataFrame\n Staypoints that are used for the trip generation\n tpls_input : GeoDataFrame\n Triplegs that are used for the trip generation\n gap_threshold : float\n Maximum allowed temporal gap size in minutes. If tracking data is misisng for more than `gap_threshold`\n minutes, then a new trip begins after the gap.\n id_offset : int\n IDs for trips are incremented starting from this value.\n\n Returns\n -------\n (GeoDataFrame, GeoDataFrame, GeoDataFrame)\n the tuple contains (staypoints, triplegs, trips)\n\n Notes\n -----\n Trips are an aggregation level in transport planning that summarize all movement and all non-essential actions\n (e.g., waiting) between two relevant activities.\n The function returns altered versions of the input staypoints and triplegs. Staypoints receive the fields\n [`trip_id` `prev_trip_id` and `next_trip_id`], triplegs receive the field [`trip_id`].\n The following assumptions are implemented\n - All movement before the first and after the last activity is omitted\n - If we do not record a person for more than `gap_threshold` minutes, we assume that the person performed\n an activity in the recording gap and split the trip at the gap.\n - Trips that start/end in a recording gap can have an unknown origin/destination\n - There are no trips without a (recored) tripleg\n\n Examples\n ---------\n >>> staypoints, triplegs, trips = generate_trips(staypoints, triplegs)\n\n \"\"\"\n assert 'activity' in stps_input.columns, \"staypoints need the column 'activities' \\\n to be able to generate trips\"\n\n # we copy the input because we need to add a temporary column\n tpls = tpls_input.copy()\n spts = stps_input.copy()\n\n tpls['type'] = 'tripleg'\n spts['type'] = 'staypoint'\n\n # create table with relevant information from triplegs and staypoints.\n spts_tpls = spts[['started_at', 'finished_at', 'user_id', 'type', 'activity']].append(\n tpls[['started_at', 'finished_at', 'user_id', 'type']])\n\n # create ID field from index\n spts_tpls['id'] = spts_tpls.index\n\n # transform nan to bool\n spts_tpls['activity'] = spts_tpls['activity'] == True\n\n spts_tpls.sort_values(by=['user_id', 'started_at'], inplace=True)\n spts_tpls['started_at_next'] = spts_tpls['started_at'].shift(-1)\n spts_tpls['activity_next'] = spts_tpls['activity'].shift(-1)\n \n if print_progress:\n tqdm.pandas(desc='User trip generation')\n trips = spts_tpls.groupby(['user_id'], \n group_keys=False, \n as_index=False).progress_apply(_generate_trips_user, gap_threshold=gap_threshold).reset_index(drop=True)\n else:\n trips = spts_tpls.groupby(['user_id'], \n group_keys=False, \n as_index=False).apply(_generate_trips_user, gap_threshold=gap_threshold).reset_index(drop=True)\n trips['id'] = trips.index + id_offset\n \n # assign trip_id to tpls\n trip2tpl_map = trips[['id', 'tpls']].set_index('id').to_dict()['tpls']\n ls = []\n for key, values in trip2tpl_map.items():\n for value in values:\n ls.append([value, key])\n temp = pd.DataFrame(ls, columns=[tpls.index.name, 'trip_id']).set_index(tpls.index.name)\n tpls = tpls.join(temp, how='left')\n \n # assign trip_id to spts, for non-activity spts\n trip2spt_map = trips[['id', 'spts']].set_index('id').to_dict()['spts']\n ls = []\n for key, values in trip2spt_map.items():\n for value in values:\n ls.append([value, key])\n temp = pd.DataFrame(ls, columns=[spts.index.name, 'trip_id']).set_index(spts.index.name)\n spts = spts.join(temp, how='left')\n \n # assign prev_trip_id to spts\n temp = trips[['id', 'destination_staypoint_id']].copy()\n temp.rename(columns={\"id\":\"prev_trip_id\", \"destination_staypoint_id\":spts.index.name}, inplace=True)\n temp.set_index(spts.index.name, inplace=True)\n spts = spts.join(temp, how ='left')\n \n # assign next_trip_id to spts\n temp = trips[['id', 'origin_staypoint_id']].copy()\n temp.rename(columns={\"id\":\"next_trip_id\", \"origin_staypoint_id\":spts.index.name}, inplace=True)\n temp.set_index(spts.index.name, inplace=True)\n spts = spts.join(temp, how ='left')\n \n # final cleaning\n tpls.drop(columns=['type'], inplace=True)\n spts.drop(columns=['type'], inplace=True)\n trips.drop(columns = ['tpls', 'spts'], inplace=True)\n trips.set_index('id', inplace=True)\n \n return spts, tpls, trips\n\ndef _generate_trips_user(df, gap_threshold):\n # function called after groupby: should only contain records of one user\n user_id = df['user_id'].unique()\n assert len(user_id) == 1\n user_id = user_id[0]\n\n unknown_activity = {'user_id': user_id, 'activity': True, 'id': np.nan}\n origin_activity = unknown_activity\n temp_trip_stack = []\n in_trip = False\n trip_ls = []\n\n for _, row in df.iterrows():\n \n \n # check if we can start a new trip\n # (we make sure that we start the trip with the most recent activity)\n if in_trip is False:\n # If there are several activities in a row, we skip until the last one\n if row['activity'] and row['activity_next']:\n continue\n\n # if this is the last activity before the trip starts, reset the origin\n elif row['activity']:\n origin_activity = row\n in_trip = True\n continue\n\n # if for non-activities we simply start the trip\n else:\n in_trip = True\n \n if in_trip is True:\n # during trip generation/recording\n\n # check if trip ends regularly\n if row['activity'] is True:\n\n # if there are no triplegs in the trip, set the current activity as origin and start over\n if not _temp_trip_stack_has_tripleg(temp_trip_stack):\n origin_activity = row\n temp_trip_stack = list()\n in_trip = True\n\n else:\n # record trip\n destination_activity = row\n trip_ls.append(_create_trip_from_stack(temp_trip_stack, origin_activity,destination_activity))\n\n # set values for next trip\n if row['started_at_next'] - row['finished_at'] > datetime.timedelta(minutes=gap_threshold):\n # if there is a gap after this trip the origin of the next trip is unknown\n origin_activity = unknown_activity\n destination_activity = None\n temp_trip_stack = list()\n in_trip = False\n\n else:\n # if there is no gap after this trip the origin of the next trip is the destination of the\n # current trip\n origin_activity = destination_activity\n destination_activity = None\n temp_trip_stack = list()\n in_trip = False\n\n # check if gap during the trip\n elif row['started_at_next'] - row['finished_at'] > datetime.timedelta(minutes=gap_threshold):\n # in case of a gap, the destination of the current trip and the origin of the next trip\n # are unknown.\n\n # add current item to trip\n temp_trip_stack.append(row)\n\n # if the trip has no recored triplegs, we do not generate the current trip.\n if not _temp_trip_stack_has_tripleg(temp_trip_stack):\n origin_activity = unknown_activity\n in_trip = True\n temp_trip_stack = list()\n\n else:\n # add tripleg to trip, generate trip, start new trip with unknown origin\n destination_activity = unknown_activity\n\n trip_ls.append(_create_trip_from_stack(temp_trip_stack, origin_activity,destination_activity))\n origin_activity = unknown_activity\n destination_activity = None\n temp_trip_stack = list()\n in_trip = True\n\n else:\n temp_trip_stack.append(row)\n \n # if user ends generate last trip with unknown destination\n if (len(temp_trip_stack) > 0) and (_temp_trip_stack_has_tripleg(temp_trip_stack)):\n destination_activity = unknown_activity\n trip_ls.append(_create_trip_from_stack(temp_trip_stack, origin_activity,destination_activity,))\n \n # print(trip_ls)\n trips = pd.DataFrame(trip_ls)\n return trips","sub_path":"trackintel/preprocessing/triplegs.py","file_name":"triplegs.py","file_ext":"py","file_size_in_byte":12495,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"315578951","text":"\"\"\"Helper functions to deal with subprocess\"\"\"\n\nimport os\nimport logging\nimport subprocess\nimport select\nimport time\nimport errno\nimport commands\ntry:\n import fcntl\nexcept ImportError:\n pass\n\n\n# http://stackoverflow.com/questions/12270645/can-you-make-a-python-subprocess-output-stdout-and-stderr-as-usual-but-also-cap\n# !!! it won't work on Windows. select() accepts sockets only on Windows\ndef run(process, check=True, show_cmd=False):\n \"\"\"Run a Linux process and log stdout as debug, stderr as warning.\n Throw a exception if the process failed (returned non-zero)\n process - the process that should run\n check - by default the process exit status is checked to raise RuntimeError\n if the exit status is not zero.\n show_cmd - by default the command to run is logging as DEBUG. set it to True\n to logging as INFO, normally will also be printed to console\n The exit status will be returned as the function's return value.\n \"\"\"\n\n if show_cmd:\n logging.info(process)\n else:\n logging.debug(process)\n\n p = subprocess.Popen(process, shell=True,\n stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n\n # deal with subprocess output in another function\n # to avoid pylint warning Too many branches\n _nonblocking_read(p.stdout.fileno(), p.stderr.fileno(), _1, _2)\n\n returncode = p.wait()\n returnmsg = \"\"\"Call subprocess probably failed.\n The command line is {process}\n The return code is {retcode}\"\"\".format(process=process, retcode=returncode)\n\n if check and returncode:\n logging.debug(returnmsg)\n raise RuntimeError(returnmsg)\n elif returncode:\n logging.debug(returnmsg)\n\n return returncode\n\n\n# keep this function name short to avoid mess up the loggings in file\n# function name starts with '_' means it is an internal function\ndef _1(line):\n \"\"\"Log output line as debug\"\"\"\n logging.debug(line.rstrip())\n\n\ndef _2(line):\n \"\"\"Log output line as warning\"\"\"\n logging.warning(line.rstrip())\n\n\ndef _nonblocking_read(stdout_fd, stderr_fd, stdout_callback, stderr_callback, timeout=15):\n \"\"\"Read stdout and stderr simultaneously. This avoid the deadlock if the child\n process generates enough output to a stdout or stderr pipe such that it blocks\n waiting for the OS pipe buffer to accept more data.\n \"\"\"\n stdout_eof = False\n stderr_eof = False\n _set_nonblocking(stdout_fd)\n _set_nonblocking(stderr_fd)\n\n while True:\n rlist, _, _ = select.select([stdout_fd, stderr_fd], [], [])\n\n for fd in rlist:\n if fd == stdout_fd and not stdout_eof:\n line = _readline_with_timeout(stdout_fd, timeout)\n if line:\n stdout_callback(line)\n else:\n stdout_eof = True\n if fd == stderr_fd and not stderr_eof:\n line = _readline_with_timeout(stderr_fd, timeout)\n if line:\n stderr_callback(line)\n else:\n stderr_eof = True\n\n if stdout_eof and stderr_eof:\n break\n\n\ndef _set_nonblocking(fd):\n \"\"\"Turn fd into non blocking mode.\n Note afterward do not use python file object any more!\n \"\"\"\n flags = fcntl.fcntl(fd, fcntl.F_GETFL)\n fcntl.fcntl(fd, fcntl.F_SETFL, flags | os.O_NONBLOCK) # pylint: disable=no-member\n\n\ndef _readline_with_timeout(fd, timeout=15):\n \"\"\"file.readline with timeout. Read until '\\n' or until timeout, and return the\n read content. fd must be turned into non blocking mode. So do not use python\n file object any more! The advantage against file.readline is to avoid blocking\n if the subprocess does not output '\\n' for a long time.\n The default timeout value is well tuned that a user should see any output before\n he loses his patience and think the subprocess just \"dead\".\n \"\"\"\n content = _readline_non_blocking(fd) # fd must can read, so read it anyway\n if content:\n while content[-1] != '\\n' and timeout > 0:\n start = time.time()\n rlist, _, _ = select.select([fd], [], [], timeout)\n end = time.time()\n timeout = timeout - (end - start)\n if rlist:\n tmp = _readline_non_blocking(fd)\n if tmp:\n content += tmp\n else:\n break\n return content\n\n\ndef _readline_non_blocking(fd):\n \"\"\"call internally in _readline_with_timeout. best effort read until '\\n'.\n assume select was called before so that fd really has something can read!\n \"\"\"\n content = ''\n while True:\n try:\n tmp = os.read(fd, 1)\n if tmp:\n content += tmp\n if tmp == '\\n':\n break\n else:\n break\n except OSError as err:\n if err.errno == errno.EAGAIN:\n assert content # because by selecting fd, we know it must have content\n break\n else:\n break\n return content\n\n\ndef getoutput(cmd):\n \"\"\"A thin wrapper for commands.getoutput\"\"\"\n logging.debug('Exec shellcmd: ' + cmd)\n output = commands.getoutput(cmd)\n logging.debug('The output is: ' + output)\n return output\n\n\ndef getstatus(cmd):\n \"\"\"A thin wrapper for os.system\"\"\"\n logging.debug('Exec shellcmd: ' + cmd)\n status = os.system(cmd)\n logging.debug('The exit stat: ' + str(status))\n return status\n\n\ndef getstatusoutput(cmd):\n \"\"\"A thin wrapper for commands.getstatusoutput\"\"\"\n logging.debug('Exec shellcmd: ' + cmd)\n (status, output) = commands.getstatusoutput(cmd)\n logging.debug('The output is: ' + output)\n logging.debug('The exit stat: ' + str(status))\n return (status, output)\n","sub_path":"rda-base/src/main/python/apitools/subprocess_util.py","file_name":"subprocess_util.py","file_ext":"py","file_size_in_byte":5786,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"284182653","text":"'''\r\n 商城:\r\n 1.初始化钱包余额\r\n 2.推个空的购物车\r\n 3.正常购物:\r\n 输入商品的编号\r\n 看是否有这个商品\r\n 有:\r\n 看钱是否足够\r\n 够:\r\n 添加到购物车里\r\n 余额减去相对应的钱\r\n 不够:\r\n 温馨:穷鬼,别瞎弄!请买个其他商品\r\n 没有:\r\n 买个其他商品,别瞎弄!\r\n 4.打印购物小条\r\n 任务:\r\n 1.购物小条的商品重复打印问题\r\n 2. 10张联想电脑 0.5, 20老干妈优惠券 0.1 , 15 华为优惠券 0.6\r\n 随机抽取一张优惠券,在结算的时候进行打折,进行结算。\r\n'''\r\nshop = [\r\n [\"联想电脑\",5000],\r\n [\"苹果电脑\",12000],\r\n [\"华为手环\",2000],\r\n [\"机械革命\",15000],\r\n [\"老 干 妈\",7.5],\r\n [\"卫龙辣条\",3],\r\n [\"西 瓜\",2]\r\n]\r\n\r\n\r\n# 1.空的购物车\r\nmycart = []\r\n\r\n# 2.初始化您的余额\r\nmoney = input(\"请充值购物卡:\")\r\nmoney = int(money)\r\nmoney_1 = money\r\n# 随机抽取优惠券\r\nimport random\r\nbond = random.randint(1,45)\r\nif bond >=1 and bond <=10:\r\n print(\"恭喜获得联想电脑5折优惠券,购买时自动使用\")\r\n price = 5000\r\n rebat = 0.5\r\n bout = 1\r\n num = 0\r\nelif bond > 10 and bond <=30:\r\n print(\"恭喜获得:老干妈 1折优惠券,购买时自动使用\")\r\n price = 7.5\r\n rebat = 0.1\r\n bout = 1\r\n num = 4\r\nelse:\r\n print(\"恭喜获得:华为手环 6折优惠券,购买时自动使用\")\r\n price = 2000\r\n rebat = 0.6\r\n bout = 1\r\n num = 2\r\n\r\n# 3.正常购物\r\ni = 1\r\nwhile i <= 20:\r\n # 展示商品\r\n for key, value in enumerate(shop):\r\n print(key, value)\r\n chose = input(\"请输入您想要的商品:\")\r\n if chose.isdigit():\r\n chose = int(chose)\r\n if chose > len(shop): # len\r\n print(\"没有改号商品!请重新输入!\")\r\n else:\r\n # 钱够不够\r\n if money > shop[chose][1]:\r\n mycart.append(shop[chose])\r\n #使用优惠券\r\n if chose == num and bout > 0 :\r\n rebat_=(shop[chose][1]) * (1 - rebat)\r\n bout = bout - 1\r\n print(\"该商品已使用优惠券优惠了:\",rebat_,\"元\")\r\n else:\r\n rebat_ = 0\r\n money = money - (shop[chose][1]) + rebat_ # 减去价格\r\n print(\"恭喜,添加成功!您的余额还剩\",money)\r\n else:\r\n print(\"穷鬼,钱不够了,别瞎弄!买其他商品吧!\")\r\n elif chose == \"q\" or chose == \"Q\":\r\n print(\"结算中……\")\r\n break # 跳出循环\r\n else:\r\n print(\"对不起,您输入错误,别瞎弄!\")\r\n\r\n i = i + 1\r\n\r\n#消除重复项\r\n\r\n\r\n\r\nprint(\"以下是您的购物小条,请拿好!\")\r\nprint(\"-------------------------------------\")\r\nprint(\"编号 商品 单价 数量 合计\")\r\n#统计商品出现的次数\r\nfor h in range(len(shop)):\r\n sum = 0\r\n for key, value in enumerate(mycart):\r\n if value[1] == shop[h][1]:\r\n sum += 1\r\n if sum > 1:\r\n print(h,\"\\t\",shop[h][0],\"\\t\", shop[h][1],\"\\t\", sum,\"\\t\\t\",((shop[h][1])*sum))\r\n if sum == 1:\r\n print(h,\"\\t\", shop[h][0],\"\\t\", shop[h][1],\"\\t\",sum,\"\\t\\t\",((shop[h][1])*sum))\r\n\r\n\r\nprint(\"-------------------------------------\")\r\nif bout == 0:\r\n print(\"本次优惠券优惠:\",(price*(1-rebat)),)\r\nprint(\"您本次消费:\",(money_1 - money),\"元\")\r\nprint(\"您的余额还剩:\",money,\"元\")\r\nprint(\"欢迎下次光临!\")\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n","sub_path":"交互型商城系统.py","file_name":"交互型商城系统.py","file_ext":"py","file_size_in_byte":3876,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"167415141","text":"#!/usr/bin/env python\r\n\r\nimport os\r\nimport sys\r\nimport logging\r\nfrom argparse import ArgumentParser\r\n\r\n#============================= Function =================================\r\n##logging info\r\nDEBUG=\"\" #change it when debugging\r\nlogFormat = \"%(asctime)s [%(levelname)s] %(message)s\"\r\nlevel = \"DEBUG\" if DEBUG != \"\" else \"INFO\"\r\nlogging.basicConfig( stream=sys.stderr, level=level, format=logFormat )\r\n#========================================================================\r\n\r\ndef checkDir(Dirname):\r\n logging.info(\"Checking folder: '%s'\" % Dirname)\r\n dirname = os.path.abspath(Dirname)\r\n if not os.path.isdir(dirname):\r\n logging.error(\"Oops! Folder: '%s' does not exit. Please check!\" % Dirname)\r\n sys.exit(-1)\r\n if not os.access(dirname, os.W_OK):\r\n logging.error(\"Oops! Folder: '%s' is not writable. Please check!\" % Dirname)\r\n sys.exit(-1)\r\n\r\ndef checkFile(Filename):\r\n logging.info(\"Checking file: '%s'\" % Filename)\r\n filename = os.path.abspath(Filename)\r\n if not os.path.isfile(filename):\r\n logging.error(\"Oops! File: '%s' does not exit. Please check!\" % Filename)\r\n sys.exit(-1)\r\n if not os.access(filename, os.R_OK):\r\n logging.error(\"Oops! File: '%s' is not readable. Please check!\" % Filename)\r\n sys.exit(-1)\r\n\r\n\r\ndef renameFastaHeader(fasta, sampleName, delimiter, outDir):\r\n checkFile(fasta)\r\n checkDir(outDir)\r\n name = fasta.rsplit('.', 1)[0]\r\n fasta = os.path.abspath(fasta)\r\n outDir = os.path.abspath(outDir)\r\n outputfile = os.path.join(outDir, '%s.headerModified.fa' % name)\r\n logging.info(\"Start to rename the header ...\")\r\n with open (outputfile, 'w') as fd:\r\n with open (fasta, 'r') as fa:\r\n for line in fa: \r\n line = line.strip()\r\n if line.startswith(\">\"):\r\n line=\">%s%s%s\" % (sampleName, delimiter, line[1:])\r\n fd.write('%s\\n' % line)\r\n logging.info(\"Complete rename the header ...\")\r\n\r\n\r\nif __name__==\"__main__\":\r\n parser = ArgumentParser(description='rename the header of a given fasta file')\r\n parser.add_argument('--version', action='version', version='1.0')\r\n parser.add_argument('-f', dest='fasta', help='a fasta format file', type = str)\r\n parser.add_argument('-n', dest='name', help='name of the sample', type = str)\r\n parser.add_argument('-d', dest='delim', help=\"delimiter. Default: '||'\", type = str)\r\n parser.add_argument('-o', dest='output', help='the output directory', type=str)\r\n args = parser.parse_args()\r\n\r\n if None in [args.fasta, args.name, args.output]:\r\n print(parser.print_help())\r\n exit(-1)\r\n if args.delim == None:\r\n args.delim = \"||\"\r\n renameFastaHeader(args.fasta, args.name, args.delim, args.output)\r\n \r\n\r\n","sub_path":"panGraphViewerApp/scripts/renameFastaHeader.py","file_name":"renameFastaHeader.py","file_ext":"py","file_size_in_byte":2805,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"186118343","text":"import random\ndef swap(elementList, indexSwap1, indexSwap2):\n elementList[indexSwap1], elementList[indexSwap2] = elementList[indexSwap2], elementList[indexSwap1]\n\ndef quicksort(a,low=0, high=-1, worst = False):\n global globalCount\n if high == -1:\n high = len(a) -1\n if low < high:\n if worst:\n swap(a,low, a.index(min(a[low:high+1])))\n else:\n swap(a,low, random.randint(low,high))\n\n pivot = low\n for j in range(low+1,high+1):\n globalCount += 1\n if a[j] < a[low]:\n pivot += 1\n swap(a,pivot,j)\n swap(a,low,pivot)\n if pivot > 0:\n quicksort(a,low,pivot-1)\n quicksort(a,pivot+1,high)\n\nglobalCount = 0\nunsortedList = list(random.randint(0,10001) for _ in range(10001))\nquicksort(unsortedList)\nprint(globalCount)\n\nglobalCount = 0\nquicksort(unsortedList, worst=True)\nprint(globalCount)\n\n","sub_path":"2.5.py","file_name":"2.5.py","file_ext":"py","file_size_in_byte":931,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"397369557","text":"\nimport numpy as np\n\n\ndef ewma(x, halflife):\n \"\"\"\n Exponentially Weighted Moving Average\n It is expected that the numbers passed as x will be finite, halflife is\n expected to be a finite, non negative number.\n >>> ewma(np.arange(5), halflife=2)\n array([ 0. , 0.58578644, 1.22654092, 1.91911977, 2.65947261])\n \"\"\"\n assert np.isfinite(halflife) and 0 < halflife\n\n decay_coefficient = np.exp(np.log(0.5) / halflife)\n out = np.empty_like(x, dtype=np.float64)\n\n for i in range(out.shape[0]):\n if i == 0:\n out[i] = x[i]\n sum_prior = 1\n else:\n sum_i = sum_prior + np.power(decay_coefficient, i)\n out[i] = (decay_coefficient * out[i - 1] * sum_prior + x[i]) / sum_i\n sum_prior = sum_i\n\n return out\n\n\nif __name__ == '__main__':\n import pytest\n pytest.main([__file__])\n","sub_path":"fastats/maths/ewma.py","file_name":"ewma.py","file_ext":"py","file_size_in_byte":881,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"79735244","text":"#\n#\n# Obsolete classes and functions. Moved here for sanity. Not checked if it can run as is. Need to check imports.\n#\n\nimport os, sys, pprint\nfrom fastkml import kml,styles\nfrom pygeoif import geometry\n \nFAKE_DATA = {\n \"clusters\" : 5, \n \"outlets\" :10,\n \"activations\":50,\n \"sellouts\": 50,\n \"stocks\":70\n}\n\n\nclass KMLReader:\n def __init__(self, k):\n self._k = k\n self.root = k.features().next()\n \n self.__styles()\n \n def __styles(self):\n if isinstance(self.root, kml.Document):\n self.__styles_dict, self.__stylemap_dict = self.__get_styles()\n \n def __get_styles (self):\n _gs_style_dict = {}\n _gs_stylemap_dict = {}\n for s in self.root.styles():\n if isinstance(s, styles.StyleMap):\n _gs_stylemap_dict[s.id] = { 'normal_url' : s.normal.url, 'highlight_url' : s.highlight.url }\t\t\t\t\n elif isinstance(s, styles.Style):\n for i_s in s.styles():\n if isinstance(i_s, styles.IconStyle):\t\t\t\t\t\n _ = _gs_style_dict.setdefault(s.id, {})\n _['icon_href'] = i_s.icon_href\n \n # end if\n if isinstance(i_s, styles.LineStyle):\n _ = _gs_style_dict.setdefault(s.id, {})\n _['lineColor'] = i_s.color\n _['lineWidth'] = i_s.width\n\n # end if\n if isinstance(i_s, styles.PolyStyle):\n _ = _gs_style_dict.setdefault(s.id, {})\n _.update({ 'polyColor' : i_s.color, 'polyColorMode' : i_s.colorMode, 'polyFill' : i_s.fill, 'polyOutline' : i_s.outline })\n # _gs_style_dict[s.id] = { 'color' : i_s.color, 'colorMode' : i_s.colorMode, 'fill' : i_s.fill, 'outline' : i_s.outline }\n # app.logger.debug(\"Found poly style: %s\", _)\n #end if\n # end for\n # end if\n return _gs_style_dict, _gs_stylemap_dict\n # end for\n def __get_color(self, p_style_url):\n # app.logger.debug(\"stylemap_dict: %s\", pprint.pformat(p_stylemap_dict))\n _s_map = self.__stylemap_dict.get(p_style_url.strip('#'), '')\n # app.logger.debug(\"Style Url: %s, %s\",p_style_url.strip('#'), _s_map)\n if _s_map:\t\t\t\n _s = self.__styles_dict.get(_s_map['normal_url'].strip('#'), '')\n # app.logger.debug(\"Style: %s\", _s)\n if _s:\n return _s['polyColor'], _s['lineColor'], _s['lineWidth']\n # end if\n # end if\n return ''\n # end def _get_color\n\n def __get_icon(self, p_style_url):\n # app.logger.debug(\"stylemap_dict: %s\", pprint.pformat(p_stylemap_dict))\n _s_map = self.__stylemap_dict.get(p_style_url.strip('#'), '')\n # app.logger.debug(\"Style Url: %s, %s\",p_style_url.strip('#'), _s_map)\n if _s_map:\t\t\t\n _s = self.__styles_dict.get(_s_map['normal_url'].strip('#'), '')\n # app.logger.debug(\"Style: %s\", _s)\n if _s:\n return {'href' : _s['icon_href'] }\n # end if\n # end if\n return ''\n def __get_polygons(self, features, selected_elements = [], recurse = False):\n polygons = []\n for c in features: #placemark\n if isinstance(c, kml.Placemark) and isinstance(c.geometry, geometry.Polygon) :\n style_url = c.styleUrl\n _poly_color, _line_color, _line_width = self.__get_color(style_url)\n\n if not selected_elements or c.name in selected_elements: \n polygons.append({\n 'name' : c.name, \n 'coords' :c.geometry.exterior.coords, \n 'description': simplejson.loads(c.description.replace('\\n','')) if c.description else FAKE_DATA, \n 'color': {\n 'polyColor' : _poly_color, \n 'lineColor' : _line_color, \n 'lineWidth' : _line_width,\n }\n })\n \n if isinstance(c, kml.Folder) and recurse:\n _p = self.__get_polygons(c, recurse)\n polygons.extend(_p)\n return polygons\n\n def __get_points(self, features, recurse = False):\n points = []\n for c in features: #placemark\n\n if isinstance(c, kml.Placemark) and isinstance(c.geometry, geometry.Point) :\n\n style_url = c.styleUrl\n # _poly_color, _line_color, _line_width = self.__get_color(style_url)\n i = self.__get_icon(style_url)\n points.append({ 'name':c.name, 'coords' : c.geometry.coords, 'icon' : i, 'description' : simplejson.loads(c.description.replace('\\n','')) if c.description else FAKE_DATA })\n # polygons.append({'name' : c.name, 'coords' :c.geometry.exterior.coords, 'color': { 'polyColor' : _poly_color, 'lineColor' : _line_color, 'lineWidth' : _line_width} })\n \n if isinstance(c, kml.Folder) and recurse:\n _p = self.__get_points(c, recurse)\n points.extend(_p)\n return points\n\n def getPoints(self, folder_name = None):\n points = []\n if isinstance(self.root, kml.Document):\n for b in self.root.features(): # folder\n\n if isinstance(b, kml.Folder) and b.name == folder_name:\n \n points = self.__get_points(b.features()) \n # end for\n # end if\n # end for\n # end if\t\n\n return points\n\n def getPolygons(self, folder_name = None, selected_elements = []):\n polygons = []\n if isinstance(self.root, kml.Document):\n for b in self.root.features(): # folder\n if isinstance(b, kml.Folder) and b.name == folder_name:\n polygons = self.__get_polygons(b.features()) if not selected_elements else self.__get_polygons(b.features(), selected_elements)\n # end for\n # end if\n # end for\n # end if\t\n\n return polygons\n \n\n \n# Utility class for getting Polygon from fastkml.KML objects.\n# DEPRECATED use KMLReader instead\nclass Polygon:\t\n #\n # get the KML object from fastkml, and get the Polygons only\n # returns list of dict.\n # dict contents: name, coords, color.\n\n @staticmethod\n def digest(k):\n # get the color of the style url of style_map from styles\n def _get_color(p_style_url, p_stylemap_dict, p_styles_dict):\n # app.logger.debug(\"stylemap_dict: %s\", pprint.pformat(p_stylemap_dict))\n _s_map = p_stylemap_dict.get(p_style_url.strip('#'), '')\n \n # app.logger.debug(\"Style Url: %s, %s\",p_style_url.strip('#'), _s_map)\n if _s_map:\t\t\t\n _s = p_styles_dict.get(_s_map['normal_url'].strip('#'), '')\n # app.logger.debug(\"Style: %s\", _s)\n if _s:\n \n return _s['polyColor'], _s['lineColor'], _s['lineWidth']\n # end if\n # end if\n return ''\n # end def _get_color\n \n \n # return dictionary of styles, and style map\n # style: key = style url\n #\t value = dict of normal_url and highlight_url\n #\n # style_map: key = style url\n # value: dict of color, colorMode, fill, outline\t\t\n def _get_styles (p_document):\n _gs_style_dict = {}\n _gs_stylemap_dict = {}\n for s in p_document.styles():\n if isinstance(s, styles.StyleMap):\n _gs_stylemap_dict[s.id] = { 'normal_url' : s.normal.url, 'highlight_url' : s.highlight.url }\t\t\t\t\n elif isinstance(s, styles.Style):\n for i_s in s.styles():\n if isinstance(i_s, styles.IconStyle):\t\t\t\t\t\n pass\n # end if\n if isinstance(i_s, styles.LineStyle):\n _ = _gs_style_dict.setdefault(s.id, {})\n _['lineColor'] = i_s.color\n _['lineWidth'] = i_s.width\n\n # end if\n if isinstance(i_s, styles.PolyStyle):\n _ = _gs_style_dict.setdefault(s.id, {})\n _.update({ 'polyColor' : i_s.color, 'polyColorMode' : i_s.colorMode, 'polyFill' : i_s.fill, 'polyOutline' : i_s.outline })\n # _gs_style_dict[s.id] = { 'color' : i_s.color, 'colorMode' : i_s.colorMode, 'fill' : i_s.fill, 'outline' : i_s.outline }\n # app.logger.debug(\"Found poly style: %s\", _)\n #end if\n # end for\n # end if\n # end for\n # app.logger.debug(\"styles_dict: %s\", pprint.pformat( _gs_style_dict ))\n # app.logger.debug(\"stylemap_dict: %s\", pprint.pformat(_gs_stylemap_dict))\n \n return _gs_style_dict, _gs_stylemap_dict\n # end def _get_styles\n \n polygons = []\n styles_dict = {}\n stylemap_dict = {}\n \n top_level = k.features()\n \n for a in k.features(): # document\n if isinstance(a, kml.Document):\n styles_dict, stylemap_dict = _get_styles(a)\n for b in a.features(): # folder\n if isinstance(b, kml.Folder):\n for c in b.features(): #placemark\n if isinstance(c, kml.Placemark):\n style_url = c.styleUrl\n _poly_color, _line_color, _line_width = _get_color(style_url, stylemap_dict, styles_dict )\n polygons.append({'name' : c.name, 'coords' :c.geometry.exterior.coords, 'color': { 'polyColor' : _poly_color, 'lineColor' : _line_color, 'lineWidth' : _line_width} })\n # end if\n # end for\n # end if\n # end for\n # end if\t\n # end for\n return polygons\n","sub_path":"apps/app/deprecated/kml.py","file_name":"kml.py","file_ext":"py","file_size_in_byte":10371,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"266597270","text":"import pandas as pd\nimport numpy as np\n\n\nfrom hinpy.classes.object_group_class import *\nfrom time import time as TCounter\n\n\ndef RandomRecommender(start_object_group,end_object_group,parameters,verbose=False):\n\n\n\n start_objects = start_object_group.GetNames()\n end_objects = end_object_group.GetNames()\n\n start_group = start_object_group.name\n end_group = end_object_group.name\n\n relation_name=''\n timestamp=pd.Timestamp('')\n\n if verbose:\n t=TCounter()\n VerboseMessage(verbose,'Computing Random Recommendations of %s for %s...'%(end_group,start_group))\n\n recommended_table=pd.DataFrame(columns=['relation','start_group', 'start_object', 'end_group', 'end_object',\n 'value','timestamp'])\n\n # For each start object...\n counter=0\n for start_obj in start_objects:\n # We select random topK_predictions objects to recommend\n user_list = np.random.choice(end_objects,size=parameters['topK_predictions'])\n for end_obj in user_list:\n recommended_table.loc[counter] = [relation_name,start_group,start_obj,end_group,end_obj,'',timestamp]\n counter+=1\n\n if verbose:\n VerboseMessage(verbose,'Random Recommendations computed in %s.'%(ETSec2ETTime(TCounter()-t)))\n\n return recommended_table,{};\n","sub_path":"hinpy/rs/random_rs.py","file_name":"random_rs.py","file_ext":"py","file_size_in_byte":1292,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"71121799","text":"#=========================================================================\n# This is OPEN SOURCE SOFTWARE governed by the Gnu General Public\n# License (GPL) version 3, as described at www.opensource.org.\n# Copyright (C)2016 William H. Majoros (martiandna@gmail.com).\n#=========================================================================\nfrom __future__ import (absolute_import, division, print_function,\n unicode_literals, generators, nested_scopes, with_statement)\nfrom builtins import (bytes, dict, int, list, object, range, str, ascii,\n chr, hex, input, next, oct, open, pow, round, super, filter, map, zip)\nfrom Bed3Record import Bed3Record\nfrom Bed6Record import Bed6Record\nimport re\n\n#=========================================================================\n# Attributes:\n# fh : file handle\n# Instance Methods:\n# reader=BedReader(filename)\n# record=reader.nextRecord() # Bed3Record or Bed6Record\n# reader.close()\n# list=BedReader.readAll(filename)\n# hash=BedReader.hashBySubstrate(filename) # chr -> list of records\n# Class Methods:\n# \n#=========================================================================\nclass BedReader:\n \"\"\"BedReader reads bed3 and/or bed6 files\"\"\"\n def __init__(self,filename):\n self.fh=open(filename,\"r\")\n\n @classmethod\n def readAll(cls,filename):\n reader=BedReader(filename)\n array=[]\n while(True):\n record=reader.nextRecord()\n if(not record): break\n array.append(record)\n reader.close()\n return array\n\n @classmethod\n def hashBySubstrate(cls,filename):\n list=cls.readAll(filename)\n hash={}\n for rec in list:\n if(hash.get(rec.chr,None) is None):\n hash[rec.chr]=[]\n hash[rec.chr].append(rec)\n return hash\n\n def close(self):\n self.fh.close()\n\n def nextRecord(self):\n while(True):\n line=self.fh.readline()\n if(not line): return None\n if(not re.search(\"\\S\",line)): continue\n line=line.rstrip()\n line=line.lstrip()\n fields=line.split()\n n=len(fields)\n if(n==3):\n return Bed3Record(fields[0],int(fields[1]),int(fields[2]))\n if(n==4):\n return Bed6Record(fields[0],int(fields[1]),int(fields[2]),\n fields[3],0.0,\".\")\n if(n==5):\n return Bed6Record(fields[0],int(fields[1]),int(fields[2]),\n fields[3],float(fields[4]),\".\")\n if(n==6):\n return Bed6Record(fields[0],int(fields[1]),int(fields[2]),\n fields[3],float(fields[4]),fields[5])\n raise Exception(\"wrong number of fields in bed file: \"+line)\n\n","sub_path":"BedReader.py","file_name":"BedReader.py","file_ext":"py","file_size_in_byte":2811,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"305156703","text":"from PaddleController import *\nfrom src.data_models.KeyListener import *\n\n\nclass PlayerPaddleController(PaddleController, KeyListener):\n\n STD_KEY_SET_LETTERS = [pygame.K_w, pygame.K_s] # when on_key_up/down method is called, it is checked if one of these keys is pressed\n STD_KEY_SET_ARROWS = [pygame.K_UP, pygame.K_DOWN]\n\n def __init__(self, paddle, key_set): # [UP_KEY, DOWN_KEY]\n PaddleController.__init__(self, paddle)\n self.key_set = key_set\n\n def on_key_up(self, event):\n if event.key == self.key_set[0]:\n self.get_paddle().on_up_activate()\n elif event.key == self.key_set[1]:\n self.get_paddle().on_down_activate()\n\n def on_key_down(self, event):\n if event.key == self.key_set[0]:\n self.get_paddle().on_up_deactivate()\n elif event.key == self.key_set[1]:\n self.get_paddle().on_down_deactivate()\n\n def update(self, dt):\n pass\n","sub_path":"Pong/src/game/controllers/PlayerPaddleController.py","file_name":"PlayerPaddleController.py","file_ext":"py","file_size_in_byte":944,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"598557274","text":"import numpy as np\nimport pandas as pd\n\nfrom matplotlib import pyplot as plt\nfrom nltk.corpus import stopwords\nfrom nltk.stem import PorterStemmer\nfrom nltk.tokenize import TreebankWordTokenizer\nfrom numpy.linalg import norm\nfrom sklearn.cluster import KMeans\nfrom sklearn.feature_extraction.text import CountVectorizer\n\ntokenizer = TreebankWordTokenizer()\neng_stopwords = tuple(stopwords.words(\"english\"))\nstemmer = PorterStemmer()\n\n\ndef tokenization(sent):\n \"\"\"\n @author: Mihir Gadgil\n Tokenizer for vectorizing sentences.\n \"\"\"\n # Tokenize sentence\n tokens = tokenizer.tokenize(sent)\n # Remove stopwords and stemming\n processed_tokens = [stemmer.stem(token) for token in tokens if token not in eng_stopwords]\n return processed_tokens\n\n\n# Text vectorizer for CosineSimilarity\nvectorizer = CountVectorizer(tokenizer=tokenization, binary=True)\n\n\ndef JaccardSimilarity(sent1, sent2):\n \"\"\"\n @author: Mihir Gadgil\n Sentence Jaccard similarity.\n \"\"\"\n tokens1 = set(tokenizer.tokenize(sent1))\n tokens2 = set(tokenizer.tokenize(sent2))\n\n return len(tokens1.intersection(tokens2)) / len(tokens1.union(tokens2))\n\n\ndef CosineSimilarity(sent1, sent2):\n \"\"\"\n @author: Mihir Gadgil\n Sentence cosine similarity.\n \"\"\"\n sents = vectorizer.fit_transform([sent1, sent2]).toarray()\n return np.matmul(sents[0], sents[1].T) / (norm(sents[0]) * norm(sents[1]))\n\n\nclass StemmerTokenizer(object):\n def __init__(self):\n self.porter_stemmer = PorterStemmer()\n self.treebank_tokenizer = TreebankWordTokenizer()\n\n def __call__(self, sentence):\n return [self.porter_stemmer.stem(token) for token in self.treebank_tokenizer.tokenize(sentence)\n if token not in eng_stopwords]\n\n\ndef KMeansClusteringElbowCurve(quote_dict):\n \"\"\"Shows an elbow curve plot to determine the appropriate number of k-means clusters.\"\"\"\n count_vectorizer = CountVectorizer(tokenizer=StemmerTokenizer(), lowercase=True,\n stop_words=stopwords.words('english'), binary=True)\n X = count_vectorizer.fit_transform(quote_dict.values())\n distorsions = []\n for k in range(1, 4):\n kmeans_model = KMeans(n_clusters=k)\n kmeans_model.fit(X)\n distorsions.append(kmeans_model.inertia_)\n fig = plt.figure(figsize=(15, 5))\n plt.plot(range(1, 4), distorsions)\n plt.title('Elbow Curve')\n plt.show()\n\n\ndef KMeansClustering(quote_dict, clusters=2):\n \"\"\"Returns a pandas data frame containing the quote_dict and cluster label.\"\"\"\n count_vectorizer = CountVectorizer(tokenizer=StemmerTokenizer(), lowercase=True)\n X = count_vectorizer.fit_transform(quote_dict.values()).toarray()\n kmeans_model = KMeans(n_clusters=clusters).fit(X)\n y = kmeans_model.predict(X)\n kmeans_df = pd.DataFrame.from_dict(quote_dict, orient='index', columns=['sentence'])\n kmeans_df[\"cluster\"] = kmeans_model.labels_\n return X, y, kmeans_model, kmeans_df\n\n\ndef KMeansClusteringPlot(X, y, kmeans_model, quote_dict):\n \"\"\"Show clusters with centroids from k-means.\"\"\"\n plt.scatter(X[:, 0][0], X[:, 1][0], s=200, color='blue', label=[k for k in quote_dict.keys()][0])\n plt.scatter(X[:, 0][1], X[:, 1][1], s=200, color='red', label=[k for k in quote_dict.keys()][1])\n plt.scatter(X[:, 0][2], X[:, 1][2], s=200, color='green', label=[k for k in quote_dict.keys()][2])\n centers = kmeans_model.cluster_centers_\n plt.scatter(centers[:, 0], centers[:, 1], c='black', s=100, alpha=0.6)\n plt.legend()\n plt.show()\n\n\nif __name__ == \"__main__\":\n quote_dict = {'cnn': 'witch hunt', 'fox': 'donald trump says this is a witch hunt',\n 'bbc': 'donald trump is a crookity crook who should be impeached'}\n\n kmeans_elbow = KMeansClusteringElbowCurve(quote_dict)\n\n X, y, kmeans_model, kmeans_df = KMeansClustering(quote_dict)\n print(kmeans_df)\n\n kmeans_plot = KMeansClusteringPlot(X, y, kmeans_model, quote_dict)\n","sub_path":"library/metrics.py","file_name":"metrics.py","file_ext":"py","file_size_in_byte":3956,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"167603059","text":"from . import ZObject\n\n\nclass FilterRule(ZObject):\n \"\"\" A mailbox filter rule object\n \"\"\"\n TAG_NAME = 'filter'\n ATTRNAME_PROPERTY = 'name'\n\n\nclass Identity(ZObject):\n \"\"\"An identity object\n \"\"\"\n SELECTORS = ('name', 'id')\n TAG_NAME = 'identity'\n ATTRNAME_PROPERTY = 'name'\n\n def to_creator(self):\n \"\"\" Returns the dict suitable for CreateIdentity or ModifyIdentity\n \"\"\"\n o = {}\n\n for prop in ('name', 'id'):\n if hasattr(self, prop):\n o[prop] = getattr(self, prop)\n\n try:\n if len(self.a) > 0:\n o['a'] = []\n for node in self._unparse_a_tags(self._a_tags):\n o['a'].append(node)\n except AttributeError:\n pass\n return o\n\n def is_default(self):\n \"\"\" Is it the default identity ? \"\"\"\n # it's not just a convention : default identity name cannot be\n # changed...\n return self.name == 'DEFAULT'\n\n def to_selector(self):\n \"\"\" For some reason, the selector for is\n\n \n\n rather than\n\n \n \"\"\"\n\n for i in self.SELECTORS:\n if hasattr(self, i):\n val = getattr(self, i)\n selector = i\n break\n\n return {selector: val}\n\n\nclass Signature(ZObject):\n TAG_NAME = 'signature'\n SELECTORS = ('id', 'name')\n\n @classmethod\n def from_dict(cls, d):\n \"\"\" Override default, adding the capture of content and contenttype.\n \"\"\"\n o = super(Signature, cls).from_dict(d)\n if 'content' in d:\n # Sometimes, several contents, (one txt, other html), take last\n try:\n o._content = d['content']['_content']\n o._contenttype = d['content']['type']\n except TypeError:\n o._content = d['content'][-1]['_content']\n o._contenttype = d['content'][-1]['type']\n\n return o\n\n def to_selector(self):\n \"\"\" For some reason, the selector for is\n\n \n\n rather than\n\n \n \"\"\"\n\n for i in self.SELECTORS:\n if hasattr(self, i):\n val = getattr(self, i)\n selector = i\n break\n\n return {selector: val}\n\n def get_content(self):\n return self._content\n\n def set_content(self, content, contenttype='text/html'):\n self._content = content\n self._contenttype = contenttype\n\n def to_creator(self, for_modify=False):\n \"\"\" Returns a dict object suitable for a 'CreateSignature'.\n\n A signature object for creation is like :\n\n \n My signature content\n \n\n which is :\n\n {\n 'name' : 'unittest',\n 'content': {\n 'type': 'text/plain',\n '_content': 'My signature content'\n }\n }\n\n Note that if the contenttype is text/plain, the content with text/html\n will be cleared by the request (for consistency).\n \"\"\"\n signature = {}\n\n if for_modify:\n try:\n # we should have an ID\n signature['id'] = self.id\n except AttributeError:\n raise AttributeError('a modify request should specify an ID')\n # Case where we change or set a name\n if hasattr(self, 'name'):\n signature['name'] = self.name\n\n else:\n # a new signature should have a name\n signature['name'] = self.name\n\n if self.has_content():\n # Set one, flush the other (otherwise, we let relief behind...)\n if self._contenttype == 'text/plain':\n plain_text = self._content\n html_text = ''\n else:\n html_text = self._content\n plain_text = ''\n\n content_plain = {'type': 'text/plain', '_content': plain_text}\n content_html = {'type': 'text/html', '_content': html_text}\n\n signature['content'] = [content_plain, content_html]\n\n else:\n # A creation request should have a content\n if not for_modify:\n raise AttributeError(\n 'too little information on signature, '\n 'run setContent before')\n\n return signature\n\n def has_content(self):\n return (hasattr(self, '_content') and hasattr(self, '_contenttype'))\n\n def get_content_type(self):\n return self._contenttype\n","sub_path":"zimsoap/zobjects/account.py","file_name":"account.py","file_ext":"py","file_size_in_byte":4765,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"260047924","text":"import tensorflow as tf\nimport numpy as np\n\nbandits = [0.5, -0.3, 0, -0.2, 0.5]\nnum_bandits = len(bandits)\n\ndef pullBandit(bandit):\n\tresult = np.random.randn(1)\n\tif result > bandit:\n\t\t# Return positive reward\n\t\treturn 1\n\telse:\n\t\t# Return negative reward\n\t\treturn -1\n\ntf.reset_default_graph()\n\n# Feed-forward part of the network. Does the choosing\nweights = tf.Variable(tf.ones([num_bandits])) # tf variable with 1 weights\nchosen_action = tf.argmax(weights,0)\n\n# Establish training procedure. Feed reward and action into network\nreward_holder = tf.placeholder(shape=[1],dtype=tf.float32)\naction_holder = tf.placeholder(shape=[1],dtype=tf.int32)\nresponsible_weight = tf.slice(weights,action_holder,[1])\nloss = -(tf.log(responsible_weight)) * reward_holder\noptimizer = tf.train.GradientDescentOptimizer(learning_rate=0.001)\nupdate = optimizer.minimize(loss)\n\ntotal_episodes = 1000 # total number of iterations for training\ntotal_reward = np.zeros(num_bandits) # scoreboard for bandits\ne = 0.1 # Chance of random action\n\ninit = tf.global_variables_initializer()\n\n# Launch tensorflow graph\nwith tf.Session() as sess:\n\tsess.run(init)\n\ti = 0\n\twhile i < total_episodes:\n\t\tif np.random.rand(1) < e:\n\t\t\taction = np.random.randint(num_bandits)\n\t\telse:\n\t\t\taction = sess.run(chosen_action)\n\n\t\treward = pullBandit(bandits[action]) # Get reward for picking bandit\n\n\t\t# Update network\n\t\t_, resp, ww = sess.run([update, responsible_weight, weights], feed_dict={reward_holder:[reward],action_holder:[action]})\n\t\t# update running tally of scores\n\t\ttotal_reward[action] += reward\n\t\tif i % 50 == 0:\n\t\t\tprint(\"Running reward for the \" + str(num_bandits) + \" bandits \" + str(total_reward))\n\t\ti += 1\n\n\tprint(\"The agent thinks bandit \" + str(np.argmax(ww)+1) + \" is the most promising...\")\n\tif np.argmax(ww) == np.argmax(-np.array(bandits)):\n\t\tprint(\"... and it was right!\")\n\telse:\n\t\tprint(\"... and it was wrong!\")","sub_path":"My Sandbox/armed_bandits.py","file_name":"armed_bandits.py","file_ext":"py","file_size_in_byte":1890,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"353950485","text":"#!/usr/bin/env python3\n\"\"\"\n Run commands from .cmd files, storing output in .out files\n\"\"\"\nfrom __future__ import print_function\n\nimport argparse\nimport os\nimport sys\nfrom subprocess import PIPE, Popen\n\n\ndef expand_path(path):\n \"\"\"Expand variables in and provide absolute version of the given 'path'\"\"\"\n\n return os.path.abspath(os.path.expanduser(os.path.expandvars(path)))\n\n\ndef update_file(fpath, content):\n \"\"\"Writes 'content' to 'fpath'\"\"\"\n\n with open(fpath, \"w+\") as output:\n output.write(content)\n\n\ndef cmd_run(cmd, args):\n \"\"\"Execute the given command and return stdout, stderr, and returncode\"\"\"\n\n with Popen(\n cmd, shell=True, stdout=PIPE, stderr=PIPE, executable=args.shell\n ) as process:\n out, err = process.communicate()\n\n return out, err, process.returncode\n\n\ndef cmd_from_file(fpath):\n \"\"\"Produces a 'cmd' as a list of strings from the given 'fpath'\"\"\"\n\n # Grab commands\n with open(fpath) as cmdfd:\n cmds = [line.strip() for line in cmdfd.readlines()]\n\n # Merge those line-continuations\n cmds = \"\\n\".join(cmds).replace(\"\\\\\\n\", \"\").splitlines()\n\n if not cmds:\n fname = os.path.basename(fpath)\n cmds = [fname.replace(\".uone\", \"\").replace(\".cmd\", \"\")]\n\n return cmds\n\n\ndef produce_cmd_output(args):\n \"\"\"Do the actual work\"\"\"\n\n for root, _, fnames in os.walk(args.path):\n if args.recursive and root != args.path:\n continue\n\n for fname in sorted(fname for fname in fnames if fname.endswith(\".cmd\")):\n if args.exclude and args.exclude in fname:\n continue\n\n cmd_fpath = os.sep.join([root, fname])\n\n out_fpath = cmd_fpath.replace(\".cmd\", \".out\")\n err_fpath = cmd_fpath.replace(\".cmd\", \".err\")\n uone = cmd_fpath.endswith(\".uone.cmd\")\n output = []\n errored = False\n\n for cmd in cmd_from_file(cmd_fpath):\n stdout, stderr, rcode = cmd_run(cmd, args)\n\n output.append(stdout)\n output.append(stderr)\n\n err = bool(rcode) and not uone\n errored |= err\n\n yield out_fpath, cmd_fpath, cmd, rcode, uone, err\n\n if errored:\n update_file(err_fpath, \"\\n\".join([o.decode(\"utf-8\") for o in output]))\n\n if not errored or uone:\n update_file(out_fpath, \"\\n\".join([o.decode(\"utf-8\") for o in output]))\n\n\ndef parse_args():\n parser = argparse.ArgumentParser(\n description=\"Run commands from .cmd files, storing output in .out files\"\n )\n parser.add_argument(\"path\", type=str, help=\"Path to DIR containing .cmd files\")\n parser.add_argument(\"-r\", \"--recursive\", action=\"store_true\", help=\"go deepah!\")\n parser.add_argument(\"-s\", \"--shell\", help=\"Absolute path to the Shell to use\")\n parser.add_argument(\"-x\", \"--exclude\", help=\"Exclude command-files matching this\")\n\n args = parser.parse_args()\n args.path = expand_path(args.path)\n\n return args\n\n\ndef main():\n \"\"\"Entry point\"\"\"\n\n args = parse_args()\n\n nerrs = 0\n\n try:\n print(\"args:\")\n print(\" path: %r\" % args.path)\n print(\" recursive: %r\" % args.recursive)\n print(\"results:\")\n for out_fp, cmd_fp, cmd, rcode, uone, err in produce_cmd_output(args):\n nerrs += int(err)\n\n print(\"- out_fp: %r\" % out_fp)\n print(\" cmd_fp: %r\" % cmd_fp)\n print(\" cmd: %r\" % cmd)\n print(\" rcode: %r\" % rcode)\n print(\" uone: %r\" % uone)\n print(\" err: %r\" % err)\n\n except OSError as exc:\n print(\"# err(%s)\" % exc)\n return 1\n\n print(\"nerrs: %r\" % nerrs)\n\n return nerrs\n","sub_path":"src/kmdo/cli.py","file_name":"cli.py","file_ext":"py","file_size_in_byte":3724,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"207691526","text":"import re\nimport requests\nfrom urllib import error\nfrom bs4 import BeautifulSoup\nimport os\n\ndownload_pic_index = 0\n\nIMG_EXTENSIONS = [\n '.jpg', '.JPG', '.jpeg', '.JPEG',\n '.png', '.PNG', '.ppm', '.PPM', '.bmp', '.BMP',\n]\n\n\ndef Find(url, max_page=39):\n List = []\n print('正在检测图片总数,请稍等.....')\n t = 2\n s = 0\n while t < max_page:\n Url = url + str(t)\n try:\n Result = requests.get(Url, timeout=7)\n except BaseException:\n t = t + 1\n continue\n else:\n result = Result.text\n html_urls = re.findall('\"_blank\" href=\"(.*?)\"', result, re.S)\n\n for html_url in html_urls:\n if \"login\" not in html_url:\n try:\n Result = requests.get(Url, timeout=7)\n result = Result.text\n pic_urls = re.findall('src=\"(.*?)\"', result, re.S)\n\n s += len(pic_urls)\n if len(pic_urls) == 0:\n break\n else:\n List.append(pic_urls)\n\n except BaseException:\n continue\n t = t + 1\n\n\n\n\n\n return List, s\n\n\ndef dowmloadPicture(url_list, savepath, limit):\n global download_pic_index\n\n if not os.path.exists(savepath):\n os.mkdir(savepath)\n\n for eachhtml in url_list:\n for picurl in eachhtml:\n if \"900.png\" not in picurl:\n continue\n print('正在下载第' + str(download_pic_index + 1) + '张图片,图片地址:' + str(picurl))\n try:\n if picurl is not None:\n pic = requests.get(picurl, timeout=7)\n else:\n continue\n except BaseException:\n print('错误,当前图片无法下载')\n continue\n else:\n filetail = \".\" + picurl.split(\".\")[-1]\n\n if any(filetail == extension for extension in IMG_EXTENSIONS):\n download_file_path = 'pic_' + str(download_pic_index) + filetail\n else:\n download_file_path = 'pic_' + str(download_pic_index) + '.' + \"jpg\"\n\n download_file_path = os.path.join(savepath, download_file_path)\n fp = open(download_file_path, 'wb')\n fp.write(pic.content)\n fp.close()\n download_pic_index += 1\n if download_pic_index >= limit:\n return\n\n\ndef goToFind(savepath, limit):\n #11690 39,933 17,6570 19,414 , 88\n label = [\"11690\", \"933\", \"6570\", \"414\"]\n max_page = [39, 17, 19, 88]\n for index in range(3):\n\n url = 'https://ku.pzhan.com/'+label[index]+'/p'\n savepath = savepath + label[index]\n url_list, pic_count = Find(url, max_page[index])\n\n if not os.path.exists(savepath):\n os.mkdir(savepath)\n dowmloadPicture(url_list, os.path.join(savepath), limit)\n\n\n\nif __name__ == '__main__': # 主函数入口\n # https://safebooru.donmai.us/posts?page=2\n # 文件保存位置\n savepath = \"/media/letmesleep/LENOVO/datasets/cartoon_dataset/\"\n\n #下载最多不超过多少张图片\n limit = 30000\n\n goToFind(savepath, limit)\n\n print(\"total \" + str(download_pic_index) + \" pictures\")\n\n","sub_path":"datasets/data_utils/Crawlers/Pzhan.py","file_name":"Pzhan.py","file_ext":"py","file_size_in_byte":3412,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"508389305","text":"import sys\nimport numpy as np\nfrom itertools import zip_longest, product, chain, repeat\nimport viz_sequence\nimport train_TFmodel\nimport helper\nfrom scipy.signal import correlate2d\nfrom colour import Color\nimport matplotlib.pyplot as plt\n\none_hot_encoder = np.fromstring('acgt', np.uint8)\n\n# some sequence encoding methods\ndef encode_to_string(seq):\n \"return a string from string, uint8, or onehot\"\n if isinstance(seq, str):\n return seq\n elif isinstance(seq, np.ndarray):\n if seq.dtype == np.uint8:\n #uint8 array\n return seq.tostring().decode('UTF-8')\n else:\n #onehot array\n indicies = np.argmax(seq, axis=1)\n return np.asarray([one_hot_encoder[i] for i in indicies]).tostring().decode('UTF-8')\n else:\n raise TypeError('Sequence is not an accepted type')\n\ndef encode_to_uint8(seq):\n \"return a uint8 from string, uint8, or onehot\"\n if isinstance(seq, str):\n return np.fromstring(seq.lower(), dtype=np.uint8)\n elif isinstance(seq, np.ndarray):\n if seq.dtype == np.uint8:\n #uint8 array\n return seq\n else:\n #onehot array\n indicies = np.argmax(seq, axis=1)\n return np.asarray([one_hot_encoder[i] for i in indicies])\n else:\n raise TypeError('Sequence is not an accepted type')\n\ndef encode_to_onehot(seq):\n \"return a onehot from string, uint8, or onehot\"\n if isinstance(seq, str):\n return np.asarray([np.equal(char, one_hot_encoder) for char in np.fromstring(seq.lower(), dtype=np.uint8)])\n elif isinstance(seq, np.ndarray):\n if seq.dtype == np.uint8:\n #uint8 array\n return np.asarray([np.equal(char, one_hot_encoder) for char in seq])\n else:\n #onehot array\n return seq\n else:\n raise TypeError('Sequence is not an accepted type')\n\ndef rc(seq):\n \"\"\"Takes a seq to its reverse complement of same type.\"\"\"\n onehot = encode_to_onehot(seq)\n rc = onehot[:, ::-1, ::-1]\n if isinstance(seq, str):\n return encode_to_string(rc)\n elif isinstance(seq, np.ndarray):\n if seq.dtype == np.uint8:\n #uint8 array\n return encode_to_uint8(rc)\n else:\n #onehot array\n return rc\n else:\n raise TypeError('Sequence is not an accepted type')\n\nclass Sequence(object):\n \"\"\" Encoding and variations on a sequence.\n\n Attributes:\n seq -- onehot encoding of the sequence.\n \"\"\"\n \n def __init__(self, nucleotides):\n \"\"\" Create a sequence object.\n \n Arguments:\n nucleotides -- Sequence in string, np.uint8, or one-hot form.\n \"\"\"\n self.seq= encode_to_onehot(nucleotides) \n\n def __string__(self):\n \"\"\"ACTG representation of the sequence.\"\"\"\n return encode_to_string(self.seq)\n\n def __repr__(self):\n \"\"\"Information about the sequence.\"\"\"\n return 'Sequence() length ' + str(self.seq.shape[0])\n \n def logo(self, start=None, end=None):\n \"\"\"Plot a sequence logo from start to end.\"\"\"\n viz_sequence.plot_weights(self.seq[start:end])\n\n def model_input(self):\n return self.seq\n\n def sequential_mutant_gen(self):\n \"\"\"Generate sequences with a blank mutation.\"\"\"\n for idx in range(self.seq.shape[0]):\n new_seq = np.copy(self.seq)\n new_seq[idx] = np.fromstring('x', np.uint8)\n yield new_seq\n\n def ngram_mutant_gen(self, n=1, padding='valid'):\n \"\"\" Generate ngram mutants trying every possible amino acid combination of a length n in a sequence.\n\n Keywords:\n n -- width of the motif to mutate.\n padding -- valid or same, similar to keras funcitonality.\n \"\"\"\n done = False\n if padding != 'valid':\n print('Alternative padding not yet supported')\n while not done:\n for idx in range(len(self.seq)):\n if n//2 <= idx <= len(self.seq) - n//2 - 1:\n first = idx-n//2\n last = idx+(n+1)//2 \n #standard case\n ngrams = product(one_hot_encoder, repeat=n)\n for gram in ngrams:\n new_seq = np.copy(self.seq)\n new_seq[first:last] = encode_to_onehot(np.asarray(gram))\n yield new_seq\n done = True\n\n def double_mutant_gen(self, n=1):\n \"\"\"Generate every possible double mutant.\"\"\"\n for mut1_seq in self.ngram_mutant_gen(n=n):\n for mut2_seq in Sequence(mut1_seq).ngram_mutant_gen(n=n):\n yield mut2_seq\n\n def insertion_mutant_gen(self, n=1):\n \"\"\"Generate every n length insertion.\"\"\"\n done = False\n while not done:\n for idx in range(len(self.seq)):\n ngrams = product(one_hot_encoder, repeat=n)\n for gram in ngrams:\n new_seq = np.insert(self.seq, idx, encode_to_onehot(np.asarray(gram)), axis=0)\n yield new_seq[:256]\n done = True\n\n def deletion_mutant_gen(self, n=1):\n \"\"\"Generate every deletion mutant.\"\"\"\n done = False\n while not done:\n ngrams = product(one_hot_encoder, repeat=n)\n gram = next(ngrams)\n for start_idx in range(len(self.seq)-n):\n del_idx = range(start_idx, start_idx+n)\n new_seq = np.delete(self.seq, del_idx, axis=0)\n new_seq = np.append(new_seq, encode_to_onehot(np.asarray(gram)), axis=0)\n yield new_seq\n done = True\n\n def motif_insert_gen(self, motif, mode='same'):\n \"\"\"Insert a given motif at every position.\"\"\"\n #have i track the middle of the insertion\n for i in range(self.seq.shape[0]):\n new_seq = self.seq.copy()\n if i-motif.shape[0]//2 < 0: # too early\n if mode == 'same':\n new_seq[0:i-motif.shape[0]//2 + motif.shape[0]] = motif[motif.shape[0]//2 - i:]\n yield new_seq\n elif i-motif.shape[0]//2 + motif.shape[0] > new_seq.shape[0]: # too late\n if mode == 'same':\n new_seq[i-motif.shape[0]//2:new_seq.shape[0]] = motif[:new_seq.shape[0]-i+motif.shape[0]//2]\n yield new_seq\n else: # just right\n new_seq[i-motif.shape[0]//2:i-motif.shape[0]//2 + motif.shape[0]] = motif\n yield new_seq\n\n def find_pwm(self, meme_library=None, viz=False):\n \"\"\" Convolute a meme with the sequence.\n \n Keywords:\n meme_library -- list of memes to use.\n viz -- sequence logo of importance?\n Output:\n meme -- SeqDist() of the best matching meme.\n position -- start position of the hit.\n score -- correlation score.\n \"\"\"\n if meme_library==None:\n meme_library = CTCF_memes\n # find the meme and location of the best match.\n score = -np.inf\n position = 0\n meme = meme_library[0]\n for test_meme in meme_library:\n corr = correlate2d(self.seq, test_meme.pwm, mode='valid')\n if np.nanmax(corr) > score:\n score = np.nanmax(corr)\n position = np.nanargmax(corr)\n meme = test_meme\n if viz:\n print('Weighted log-odds of the Sequence Distribution')\n insert = np.zeros(self.seq.shape)\n insert[position:position+meme.pwm.shape[0]] = meme.pwm\n overlap = insert * self.seq\n viz_sequence.plot_weights(overlap)\n return meme, position, score\n \n def run_pwm(self, meme=None, position=None, viz=False):\n \"\"\"Get the pwm correlation score with a sequence.\n\n Keywords:\n meme -- SeqDist() of the best matching meme, or library of memes to test.\n position -- start position of the hit.\n viz -- sequence logo of importance?\n Outputs:\n overlap -- overlap which can be summed for the score.\n \"\"\"\n if meme==None:\n # we need to find everything\n meme, position, score = self.find_pwm()\n elif position==None or isinstance(meme, list):\n # we have the meme/memelist\n meme, position, score = self.find_pwm(meme_library=meme)\n # just get the score\n insert = np.zeros(self.seq.shape)\n insert[position:position+meme.pwm.shape[0]] = meme.pwm\n overlap = insert * self.seq\n if viz:\n print('Weighted log-odds of the Sequence Distribution')\n viz_sequence.plot_weights(overlap)\n return overlap\n\nclass SeqDist(Sequence):\n \"\"\"A sequence, but as a probability distribution.\n\n Attributes:\n seq -- probability distribution of bases. \n \"\"\"\n\n def __init__(self, distribution):\n \"\"\"Create a new sequence distribution object.\"\"\"\n if isinstance(distribution, np.ndarray) and not (distribution.dtype == np.uint8):\n # right type!\n self.seq = helper.softmax(np.log(distribution)) \n else:\n raise TypeError('Sequence is not an accepted type')\n \n def __repr__(self):\n \"\"\"Information about the sequence.\"\"\"\n return 'SeqDist() length ' + str(self.seq.shape[0])\n\n def logo(self, start=None, end=None):\n \"\"\"Plot a sequence logo from start to end.\"\"\"\n viz_sequence.plot_icweights(self.seq[start:end])\n\n def discrete_gen(self):\n \"\"\"Create a generator of discrete sequences.\"\"\"\n while True: \n yield self.discrete_seq()\n\n def discrete_seq(self):\n \"\"\"Return a discrete sequence samples from the continuous distribuiton.\"\"\"\n discrete = [np.random.choice(one_hot_encoder, p=base) for base in self.seq]\n return encode_to_onehot(np.asarray(discrete))\n\nclass Meme(SeqDist):\n \"\"\"A position weight matrix.\n \n Attirbutes:\n seq -- frequency representation of the seqeunce.\n pwm -- log-odds representaiton of the motif. \n \"\"\"\n\n def __init__(self, dist, pwm):\n \"\"\"Create a new Meme object.\"\"\"\n self.seq = helper.softmax(np.log(dist))\n self.pwm = pwm\n\n def __repr__(self):\n \"\"\"Information about the sequence.\"\"\"\n return 'Meme() length ' + str(self.seq.shape[0])\n\nclass ATACSeq(Sequence):\n \"\"\" A Sequence and matching atac counts.\"\"\"\n\n def __init__(self, nucs, atac_counts=None):\n if atac_counts == None:\n super().__init__(nucs[:, 1:])\n self.atac_counts = nucs[:, 0]\n else:\n super().__init__(nucs)\n self.atac_counts = atac_counts\n\n def model_input(self):\n return np.insert(self.seq.astype(np.float32), 0, self.atac_counts, axis=1)\n\n def sequential_mutant_gen(self):\n s = super().sequential_mutant_gen\n for nucs in s:\n yield np.insert(nucs.astype(np.float32), 0, self.atac_counts, axis=1)\n\n def ngram_mutant_gen(self, n=1, padding='valid'):\n n = super().ngram_mutant_gen(n=n, padding=padding)\n for nucs in n:\n yield np.insert(nucs.astype(np.float32), 0, self.atac_counts, axis=1) \n \n def double_mutant_gen(self, n=1):\n s = super().double_mutant_gen(n)\n for nucs in s:\n yield np.insert(nucs.astype(np.float32), 0, self.atac_counts, axis=1)\n\n def logo(self, top=None, bottom=None):\n colors = list(Color(\"blue\").range_to(Color(\"white\"), 50))\n [colors.append(c) for c in (Color(\"white\").range_to(Color(\"red\"), 51))]\n if top == None:\n top = np.amax(self.atac_counts)\n if bottom == None:\n bottom = np.amin(self.atac_counts)\n \n #get hightlights!\n color_weights = [int((x-bottom)/(top-bottom)*100) for x in self.atac_counts]\n highlight=dict()\n for i in range(len(self.atac_counts)):\n w = color_weights[i]\n highlight[(colors[w].rgb[0],colors[w].rgb[1], colors[w].rgb[2], .3)] = [(i, i+1)]\n #plot things out\n viz_sequence.plot_weights(self.seq, highlight=highlight)\n\n def graph(self):\n plt.figure(figsize=(20, 2))\n plt.title('ATAC counts per base')\n plt.plot(self.atac_counts)\n plt.show()\n\n\n\nclass ATACDist(SeqDist):\n \"\"\" A sequence distribution and matching atac counts.\"\"\"\n\n def __init__(self, nucs, atac_counts=None):\n if atac_counts == None:\n super().__init__(nucs[:, 1:])\n self.atac_counts = nucs[:, 0]\n else:\n super().__init__(nucs)\n self.atac_counts = atac_counts\n\n def discrete_seq(self):\n return np.insert(super().discrete_seq().astype(np.float32), 0, self.atac_counts, axis=1)\n\n \ndef process_meme(meme_path, transform=False, verb=False):\n \"\"\"Extract a meme distribution and process.\n \n Arguments:\n meme_path -- file path to a .meme file.\n Keywords:\n transform -- apply normalization and a log transform or use the pre-generated log-odds matrix.\n Outputs:\n meme_list -- List of SeqDist() meme and reverse complements.\n \"\"\"\n with open(meme_path, 'r') as infile:\n meme_length = -1\n meme_dists = list()\n meme_lods = list()\n # read for the frequencies\n for line in infile.readlines():\n if 'letter-probability matrix' in line:\n meme_length = int(line.split()[5])\n if verb:\n print('found meme')\n this_meme_lines = list()\n elif meme_length > 0:\n this_meme_lines.append([float(item.strip()) for item in line.split()])\n meme_length = meme_length - 1\n elif meme_length == 0:\n this_meme = np.asarray(this_meme_lines)\n meme_dists.append(this_meme)\n meme_length = -1\n if meme_length == 0:\n this_meme = np.asarray(this_meme_lines)\n meme_dists.append(this_meme)\n meme_length = -1\n # add rcs of memes\n rcs = list()\n for meme in meme_dists:\n rcs.append(meme[::-1, ::-1])\n meme_dists = meme_dists + rcs\n with open(meme_path, 'r') as infile:\n # read for the pwms\n for line in infile.readlines():\n if 'log-odds matrix' in line:\n meme_length = int(line.split()[5])\n this_meme_lines = list()\n elif meme_length > 0:\n this_meme_lines.append([float(item.strip()) for item in line.split()])\n meme_length = meme_length - 1\n elif meme_length == 0:\n this_meme = np.asarray(this_meme_lines)\n meme_lods.append(this_meme)\n meme_length = -1\n if meme_length == 0:\n this_meme = np.asarray(this_meme_lines)\n meme_lods.append(this_meme)\n meme_length = -1\n # add rcs of memes\n rcs = list()\n for meme in meme_lods:\n rcs.append(meme[::-1, ::-1])\n meme_lods = meme_lods + rcs\n if len(meme_lods) == 0:\n #transofrm the memes\n if verb:\n print('using manual log-odds calculation')\n psuedocount=0.005\n for meme in meme_dists:\n # add the pseudocount so probabilities don't zero out\n meme = meme*(.98) + psuedocount\n #norms = np.repeat(np.linalg.norm(meme, axis=1), 4).reshape((-1, 4))\n meme = np.log(meme) - np.log(.25)\n meme_lods.append(meme)\n #make distribution objects\n meme_list = [Meme(distribution, log_odds) for distribution, log_odds in zip(meme_dists, meme_lods)]\n return meme_list\n\nCTCF_memes = process_meme('/home/kal/TF_models/data/memes/CTCF.meme')\nmystery_memes = process_meme('/home/kal/TF_models/data/memes/mystery_motif.meme') \n","sub_path":"bin/atacseq.py","file_name":"atacseq.py","file_ext":"py","file_size_in_byte":15928,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"278945719","text":"#!/usr/bin/python3\n\nfrom subprocess import PIPE, call, Popen\nimport os, sys\nfrom time import sleep\nimport re\nimport fileinput\n\nDN = open(os.devnull, 'w')\n\nopenssh = b'^openssh-server'\n\ndef search(pkg):\n\tproc = Popen(['apt-cache', 'search', pkg],stdout=PIPE, stderr=DN)\n\n\tlistPkg = proc.communicate()[0].split(b'\\n')\n\n\tfor package in listPkg:\n\t\tif len(package) == 0:\n\t\t\tcontinue\n\t\tif (package[0]) != b' '[0]:\n\t\t\tglobal getPackage\n\t\t\tgetPackage = package[:package.find(b' ')]\n\t\t\tif re.match(pkg, getPackage, re.IGNORECASE):\n\t\t\t\tprint('Package %s available to install...' % (getPackage.decode(\"utf-8\")))\n\t\t\telse:\n\t\t\t\tprint('Package %s not available, please add another repository package...' % (getPackage))\n\n\treturn getPackage\n\ndef configSSH():\n\t'''\n\t\tFunction of configuration OpenSSH-Server\n\t'''\n\n\tos.system('clear')\n\tprint('Installation Process Done...')\n\tprint('Configuration OpenSSH-Server\\n')\n\tconfigFile = '/etc/ssh/sshd_config'\n\tif os.path.isfile(configFile) and os.access(configFile, os.R_OK):\n\t\t# code here\n\t\ttry:\n\t # config = open(configFile, 'r')\n\t\t\twith open(configFile, 'r') as searchconfig:\n\t\t\t\tfor port in searchconfig:\n\t\t\t\t\tif 'Port' in port:\n\t\t\t\t\t\told_port = port\n\t\t\t\t# for rootlogin in searchconfig:\n\t\t\t\t# \tif 'PermitRootLogin' in root_login:\n\t\t\t\t# \t\troot_access = root_login\n\n\t # print(root_access)\n\n\t\t\twith fileinput.FileInput(configFile, inplace=True, backup='.bak') as conf:\n\t\t\t\tnew_port = input('Set Port SSH(default Port 22): ')\n\t\t\t\tdefault = '22'\n\n\t\t\t\tfor configPort in conf:\n\t\t\t\t\t\tif not new_port:\n\t\t\t\t\t\t\tprint(configPort.replace(old_port, ('Port %s\\n' % default)), end=\"\")\n\t\t\t\t\t\telse:\n\t\t\t\t\t\t\tprint(configPort.replace(old_port, ('Port %s\\n' % new_port)), end=\"\")\n\t\t\tconf.close()\n\t\texcept IOError:\n\t\t\tprint('Something wrong')\n\t# else:\n\t# \treturn search(openssh)\n\nif __name__ == '__main__':\n\ttry:\n\t\tif not os.geteuid() == 0:\n\t\t\texit('Please run as r00t...\\n')\n\n\t\t# checkPkg(openssh)\n\n\t\tsearch(openssh)\n\n\t\tchar = re.sub(b'[\\^]', b'', openssh)\n\t\tif getPackage == char:\n\t\t\tdoInstall = call(['apt-get', 'install', getPackage], stderr=DN)\n\t\t\tconfigSSH()\n\t\t\tprint('Configuration Success...')\n\t\t\t# print(getPackage.decode(\"utf-8\"))\n\texcept SyntaxError as error:\n\t\tprint('Something wrong to execution')\n","sub_path":"openssh.py","file_name":"openssh.py","file_ext":"py","file_size_in_byte":2232,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"236914203","text":"def insere_atividade():\r\n n = input('insira o numero de atividades a serem cadastradas: ')\r\n\r\n # atividade = {'titulo': 't', 'materia': 'm', 'pontuação': 'p', 'data': 'd' }\r\n i = int(n)\r\n lista = list()\r\n while i > 0:\r\n lista.append({'título': input('Insira o título da atividade: '), 'materia': input('Insira a materia: '),\r\n 'pontuacao': int(input('Insira a pontuacao: ')), 'data': input('Insira a data de entrega: ')})\r\n i -= 1\r\n lista_ordenada = sorted(lista, key=lambda k: k['pontuacao'], reverse=True)\r\n with open('Lista_atividades', 'a+') as file:\r\n for n in range(len(lista_ordenada)):\r\n file.writelines(str(lista_ordenada[n]).split(sep=',{'))\r\n file.write('\\n')\r\n#file.writelines(str(lista_ordenada).split(sep=',{'))\r\n return lista_ordenada\r\n\r\n\r\ndef deleta_atividade(lista):\r\n n = input('Entre com a posição, entre espaços, das atividades a eliminar da lista: ')\r\n leng = n.split()\r\n t = len(leng)\r\n i = t\r\n while i > 0:\r\n i -= 1\r\n del (lista[int(leng[i])])\r\n\r\n with open('Lista_atividades', 'w+') as file:\r\n for n in range(len(lista)):\r\n if str(lista[n]) != '\\n':\r\n file.writelines(str(lista[n]).split(sep=',{'))\r\n# file.write('\\n')\r\n\r\n\r\ndef imprime_atividade():\r\n file = open('Lista_atividades')\r\n print(file.read())\r\n\r\n\r\n\r\np = True\r\nwhile p is True:\r\n arquivo = open('Lista_atividades')\r\n lista_univ = arquivo.readlines()\r\n c = input(\r\n 'Entre com uma opção de execução:\\n[1] - Inserir atividade\\n[2] - Deletar atividade\\n[3] - Imprimir lista de '\r\n 'tarefas\\n[4] - Sair\\n')\r\n if int(c) == 1:\r\n lista_univ = insere_atividade()\r\n if int(c) == 2:\r\n deleta_atividade(lista_univ)\r\n if int(c) == 3:\r\n imprime_atividade()\r\n if int(c) == 4:\r\n p = False\r\n","sub_path":"Lista_1_Q2.py","file_name":"Lista_1_Q2.py","file_ext":"py","file_size_in_byte":1905,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"494540368","text":"# 696. Count Binary Substrings\n# DescriptionHintsSubmissionsDiscussSolution\n# Give a string s, count the number of non-empty (contiguous) substrings that have the same number of 0's and 1's, and all the 0's and all the 1's in these substrings are grouped consecutively.\n# Substrings that occur multiple times are counted the number of times they occur.\n# Example 1:\n# Input: \"00110011\"\n# Output: 6\n# Explanation: There are 6 substrings that have equal number of consecutive 1's and 0's: \"0011\", \"01\", \"1100\", \"10\", \"0011\", and \"01\".\n# Notice that some of these substrings repeat and are counted the number of times they occur.\n# Also, \"00110011\" is not a valid substring because all the 0's (and 1's) are not grouped together.\n# Example 2:\n# Input: \"10101\"\n# Output: 4\n# Explanation: There are 4 substrings: \"10\", \"01\", \"10\", \"01\" that have equal number of consecutive 1's and 0's.\n# Note:\n# s.length will be between 1 and 50,000.\n# s will only consist of \"0\" or \"1\" characters.\n\n\nclass Solution:\n # 我想出来的方法一,首先将二进制字符串分割,把连续的0或者连续的1放在一起,并存储在一个list中(只需要存储串的长度)。\n # 遍历该list,在每一对相邻两个元素中,取其较小者,累加即可得到结果。\n # 该方法需要遍历两次(相当于)数组。可以改进\n def countBinarySubstrings(self, s):\n \"\"\"\n :type s: str\n :rtype: int\n \"\"\"\n i, l, count = 0, len(s), 0\n ls = []\n while i < l:\n j = i + 1\n # 其实不需要用‘0’来判断,后面只要判断s[j] == s[i]就行了\n if s[i] == '0':\n while j < l and s[j] == '0':\n j += 1\n else:\n while j < l and s[j] == '1':\n j += 1\n ls.append(j - i)\n i = j\n i = 1\n while i < len(ls):\n count += min(ls[i], ls[i - 1])\n i += 1\n return count\n\n # 尝试遍历一次字符串得到结果.\n def method2(self, s):\n i = count = 0\n l = len(s)\n a, b = -1, 0\n while i < l:\n j = i + 1\n while j < l and s[j] == s[i]:\n j += 1\n if a == -1:\n a = j - i\n else:\n b = j - i\n count += min(a, b)\n a = b\n i = j\n return count\n\n\nif __name__ == '__main__':\n s = Solution()\n s_ = '01010100011'\n print(s.countBinarySubstrings(s_))\n print(s.method2(s_))\n\n\nconsecutively.So","sub_path":"python/Leetcode/count_binary_string.py","file_name":"count_binary_string.py","file_ext":"py","file_size_in_byte":2584,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"49670243","text":"from PyQt5 import QtWidgets,QtGui,QtCore\nimport sys\nclass M_tray(QtWidgets.QSystemTrayIcon):\n def __init__(self,aa):\n super(M_tray, self).__init__(aa)\n self.setIcon(QtGui.QIcon(\"Res/Tray.ico\"))\n #托盘被点击icocliced\n self.activated.connect(self.icoclicked)\n self.icon = self.MessageIcon()\n self.showmenu()\n def showmenu(self):\n pw=self.parent()\n self.mainmenu=QtWidgets.QMenu()\n self.mainAction = QtWidgets.QAction(\"显示主界面\", self, triggered=pw.show)\n self.settinAction = QtWidgets.QAction(\"设置\", self, triggered=pw.aa.show)\n self.quitAction = QtWidgets.QAction(\"退出\", self, triggered=self.quit)\n self.mainmenu.addAction(self.mainAction)\n self.mainmenu.addAction(self.settinAction)\n self.mainmenu.addAction(self.quitAction)\n self.setContextMenu(self.mainmenu)\n def icoclicked(self,reason):\n pw=self.parent()\n if reason==2:\n if pw.isVisible():\n pw.hide()\n else:\n pw.show()\n def quit(self):\n self.setVisible(False)\n #注意close()、quit()、与exit()的区别\n self.parent().close()\n sys.exit()\n","sub_path":"Models/M_tray.py","file_name":"M_tray.py","file_ext":"py","file_size_in_byte":1236,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"306284843","text":"#\n# Licensed to the Apache Software Foundation (ASF) under one or more\n# contributor license agreements. See the NOTICE file distributed with\n# this work for additional information regarding copyright ownership.\n# The ASF licenses this file to You under the Apache License, Version 2.0\n# (the \"License\"); you may not use this file except in compliance with\n# the License. You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\n\"\"\"\n\nMade to run with directrunner and a local PostgreSQL database in docker.\n\nReads data from BQ, writes to two tables in PostgreSQL.\n\nexample based on https://github.com/apache/beam/blob/master/sdks/python/apache_beam/examples/cookbook/bigquery_tornadoes.py\n\npython bq-postgres.py --temp_location gs://PROJECT_ID-test --project PROJECT_ID\n\n\"\"\"\n\nfrom __future__ import absolute_import\nfrom beam_nuggets.io import relational_db\nfrom sqlalchemy import Table, Integer, String, Column\n\nimport datetime;\n\nimport argparse\nimport logging\n\nimport apache_beam as beam\n\n\ndef count_categories(input_data):\n \"\"\"\n\n \"\"\"\n ts = datetime.datetime.now().strftime(\"%Y%m%d%-%H%M%S\")\n\n return (\n input_data\n | 'count ' >> beam.FlatMap(\n lambda row: [(int(row['category']),1)])\n | 'count inputs' >> beam.CombinePerKey(sum)\n | 'formatoutput' >>\n beam.Map(lambda k_v: {\n 'category_ts': str(k_v[0])+ts, 'count': k_v[1]\n }))\n\n\ndef run(argv=None):\n parser = argparse.ArgumentParser()\n parser.add_argument(\n '--input',\n default='PROJECT_ID:demos.small_teams',\n help=(\n 'Input BigQuery table to process specified as: '\n 'PROJECT:DATASET.TABLE or DATASET.TABLE.'))\n parser.add_argument(\n '--output',\n# required=True,\n required=False,\n help=(\n 'Output BigQuery table for results specified as: '\n 'PROJECT:DATASET.TABLE or DATASET.TABLE.'))\n\n parser.add_argument(\n '--gcs_location',\n required=False,\n help=('GCS Location to store files to load '\n 'data into Bigquery'))\n\n known_args, pipeline_args = parser.parse_known_args(argv)\n\n source_config = relational_db.SourceConfiguration(\n drivername='postgresql+pg8000',\n host='localhost',\n port=5432,\n username='postgres',\n password='pwd',\n database='postgres'\n\n )\n\n\n table_config_teams = relational_db.TableConfiguration(\n name='teams',\n create_if_missing=True, # automatically create the table if not there\n primary_key_columns=['id'] # and use 'id' column as primary key\n )\n\n table_config_category = relational_db.TableConfiguration(\n name='category',\n create_if_missing=True, # automatically create the table if not there\n primary_key_columns=['category_ts'] # and use 'category_ts' column as primary key\n )\n\n with beam.Pipeline(argv=pipeline_args) as p:\n # Read the table rows into a PCollection.\n rows = p | 'read' >> beam.io.ReadFromBigQuery(\n query=\"\"\"\n SELECT id, category FROM `PROJECT_ID.demos.small_teams` limit 1500\"\"\",\n use_standard_sql=True)\n counted= count_categories(rows)\n\n\n # Write the output using a \"Write\" transform that has side effects.\n\n rows | 'Write Teams' >> relational_db.Write(\n source_config=source_config,\n table_config=table_config_teams )\n counted | 'Write Counts' >> relational_db.Write(\n source_config=source_config,\n table_config=table_config_category )\n\n # Run the pipeline (all operations are deferred until run() is called).\n\n\nif __name__ == '__main__':\n logging.getLogger().setLevel(logging.INFO)\n run()\n","sub_path":"bq-postgres.py","file_name":"bq-postgres.py","file_ext":"py","file_size_in_byte":3948,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"341184521","text":"import os\nimport sys\nimport requests\n\nSERVER_URL = \"https://gitlab.kwant-project.org\"\nPROJECT_ID = 334\nJOB_NAME = \"build singularity image\"\nIMAGE_PATH = \"build/Singularity.simg\"\nTARGET_FILENAME = \"/etc/singularity_url\"\n\ntry:\n token = os.environ[\"GITLAB_API_TOKEN\"]\n pipeline_id = os.environ[\"CI_PIPELINE_ID\"]\nexcept KeyError as ex:\n print(f\"{ex.args[0]} is undefined, not resolving the Singularity container\",\n file=sys.stderr)\n sys.exit(0)\n\nreq = requests.get(f\"{SERVER_URL}/api/v4/projects/{PROJECT_ID}/pipelines/{pipeline_id}/jobs\",\n headers={\"PRIVATE-TOKEN\": token})\n\nfor job in reversed(req.json()):\n if job[\"name\"] == JOB_NAME:\n with open(TARGET_FILENAME, \"w\") as f:\n print(f\"{SERVER_URL}/api/v4/projects/{PROJECT_ID}/jobs/{job['id']}/artifacts/{IMAGE_PATH}\", file=f)\n sys.exit(0)\n\nprint(f\"Job \\\"{JOB_NAME}\\\" is not found in the CI job\", file=sys.stderr)\nsys.exit(250)\n","sub_path":"main_image/resolve_singularity.py","file_name":"resolve_singularity.py","file_ext":"py","file_size_in_byte":941,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"468077459","text":"# -*- coding: utf-8 -*-\n\"\"\"HTTP exception object.\"\"\"\n\nfrom .http import HTTP_STATUS_CODES\n\n\nclass HTTPException(Exception):\n\n \"\"\"This can be raised from middleware to render an error page.\"\"\"\n\n def __init__(self, code=None, message=None, exception=None):\n if not isinstance(code, int) or code not in HTTP_STATUS_CODES:\n code = 500\n if message is None:\n message = HTTP_STATUS_CODES[code]\n super(HTTPException, self).__init__(message)\n self.status_code = code\n self.message = message\n self.exception = exception\n","sub_path":"malt/exceptions.py","file_name":"exceptions.py","file_ext":"py","file_size_in_byte":580,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"385227581","text":"def plusOne1(digits):\n return [int(i) for i in str(int(\"\".join([str(i) for i in digits]))+1)]\n\ndef plusOne2(digits):\n s = \"\".join([str(i) for i in digits])\n leftzeros = len(s) - len(str(int(s)))\n return [0]*leftzeros + [int(i) for i in str(int(s)+1)]\n\ndef plusOne(digits):\n if digits[-1] != 9:\n digits[-1] += 1\n else:\n i=-1\n while -i <= len(digits) and (digits[i] == 9):\n digits[i] = 0\n i -= 1\n if i+1 == -len(digits):\n digits = [1] + digits\n else:\n digits[i] += 1\n return digits\n\nlst = [9,9]\nprint(plusOne1(lst))","sub_path":"20201114.py","file_name":"20201114.py","file_ext":"py","file_size_in_byte":613,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"8683873","text":"from functions import functionObj, rosenbrock\nfrom models.optimizers import GoldenSectionSearch\n\nf_x = lambda x: x**2 - 4*x + 4\nf_x_obj = functionObj(f_x)\n\nopt = GoldenSectionSearch(f_x_obj, xtol = 1e-6, maxIter=2e10)\n\nx_min, _ = opt._line_search()\nprint('X: %.9f \\nF_x: %.9f'%(x_min, f_x_obj(x_min)))\nprint('Function evals: %d'%(f_x_obj.fevals - 1))\n","sub_path":"tests/testGR.py","file_name":"testGR.py","file_ext":"py","file_size_in_byte":351,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"642044675","text":"import fileinput\n\nfrom setuptools import setup\nfrom subprocess import run, CalledProcessError, PIPE, DEVNULL\n\n\nPACKAGE_NAME = 'pytib'\n\n\ndef _untagged_dev_version():\n try:\n # parameters to get correct info if light weight tags are used\n r = run(('git', 'describe', '--tags', '--abbrev=0'),\n stdout=PIPE, stderr=DEVNULL, check=True)\n _dev_version = r.stdout.decode().strip()\n except CalledProcessError:\n _dev_version = '0.1.0'\n\n # Used for automatic development versions only!\n r = run(('git', 'rev-parse', 'HEAD'), stdout=PIPE, check=True)\n\n return f'{_dev_version}+git{r.stdout.decode().strip()}'\n\n\ntry:\n # Releases must be done with git tags\n r = run(('git', 'tag', '-l', '--points-at', 'HEAD'),\n check=True, stdout=PIPE, stderr=DEVNULL)\n _version = r.stdout.decode().strip()\n\n # If not, create a development version based on git commit\n if _version == '':\n _version = _untagged_dev_version()\nexcept CalledProcessError:\n _version = _untagged_dev_version()\n\nwith open(f'{PACKAGE_NAME}/__init__.py', 'a') as f:\n f.write(\"__version__ = '%s'\" % _version)\n\ntry:\n setup(\n name='pytib',\n version=_version,\n description='Produce Tibetan unicode from latin script',\n url='https://github.com/ironhouzi/pytib',\n author='Robin Skahjem-Eriksen',\n author_email='robin@skahjem-eriksen.no',\n license='MIT',\n packages=[\n 'pytib',\n ],\n scripts=['ptib'],\n install_requires=[\n 'click',\n ],\n include_package_data=True,\n zip_safe=False\n )\nfinally:\n # remove injected __version__ line so version control is unaffected by build\n for line in fileinput.input(f'{PACKAGE_NAME}/__init__.py', inplace=True):\n if line.startswith('__version__'):\n print('', end='')\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1882,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"342181486","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\n\"\"\"\nPython code provided as is.\nMade by Vincent Wieczny, from Chemistry Department, ENS de Lyon, France\nThis code is under licence CC-BY-NC-SA. It enables you to reuse the code by mentioning the orginal author and without making profit from it.\n\"\"\"\n\n#Librairies\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport widgets\nimport scipy.constants as constants\nfrom matplotlib import rc\nimport matplotlib.patches as patches\n\n################################\n### Paramater initialization ###\n################################\n\n#Physical constants\nh=6.62607004e-34 #Planck constant (m2.kg.s-1)\nR=8.314 #Gas constant (J/K/mol)\nT=298.0 #Temperature (K)\neV=1.60e-19 #1 eV in J\n\n#EPR physical constants\ng=2.0023 #Landé g-factor\nmuB=9.274009994e-24 #Bohr magneton (J.T-1)\nnu=9388.2e6 #X-band frequency (Hz)\n\n#EPR magnetic field\nBmin=0 #T\nBmax=0.5 #T\nDeltaBmax=1e-1 #T\n\n# Modulated parameters\nparameters = {'DeltaB' : widgets.FloatSlider(value=0.05, description='$B_1$ $\\mathrm{(T)}$', min=0.03, max=DeltaBmax),\n 'B0' : widgets.FloatSlider(value=0.1, description='$B_0$ $\\mathrm{(T)}$', min=Bmin, max=Bmax)}\n\n#################\n### Functions ###\n#################\n\n#Down-state energy \ndef E_down(B0):\n return -0.5*g*muB*B0\n\n#Up-state energy \ndef E_up(B0):\n return 0.5*g*muB*B0\n\n#Transition energy \ndef E_trans():\n return h*nu\n\n#Resonant magnetic field\ndef B_trans():\n return E_trans()/(g*muB)\n\n#Sigma\ndef sigma(DeltaB):\n return DeltaB/6\n\n\ndef signal_abs(B0,DeltaB):\n return 1/(sigma(DeltaB)*np.sqrt(2*3.1416))*np.exp(-(B0-B_trans())**2/(2*sigma(DeltaB)**2))\n\n#Derivative signal\ndef signal_der(B0,DeltaB):\n return 1/(sigma(DeltaB)*np.sqrt(2*3.1416))*-(B0-B_trans())/sigma(DeltaB)**2*signal_abs(B0,DeltaB)\n\n#===========================================================\n# --- Plot of the updated curves ---------------------------\n#===========================================================\n\n\n## This function is called when the sliders are changed \ndef plot_data(B0,DeltaB):\n \n lines['Absorption spot'].set_data(B0,signal_abs(B0,DeltaB))\n lines['First derivative spot'].set_data(B0,signal_der(B0,DeltaB))\n truc['$Abs_courbe$'].set_data(B,signal_abs(B,DeltaB))\n truc['$Der_courbe$'].set_data(B,signal_der(B,DeltaB))\n truc['$E_\\mathrm{trans}$'].set_data([B0,B0],[-1,1])\n r1.set_transform(mpl.transforms.Affine2D().translate(B0-DeltaB/2,-E_trans()/2)+ax1.transData)\n r1.set_width(DeltaB)\n \n fig.canvas.draw_idle()\n\n\n##===========================================================\n## --- Initialization of the plot ---------------------------\n##===========================================================\n\n#Plot definition\nfig=plt.figure(figsize=(18,8))\n\nax1 = fig.add_axes([0.2, 0.2, 0.35, 0.7])\nax2 = fig.add_axes([0.60, 0.6, 0.35, 0.3])\nax3 = fig.add_axes([0.60, 0.2, 0.35, 0.3])\n\n\n#Plot comments\nfig.suptitle(r'Simulation of an EPR spectrum at X waveband for a free electron',weight='bold')\n\nfig.text(0.01,0.9,r'EPR magnetic field', multialignment='left', verticalalignment='top',weight='bold')\nfig.text(0.01,0.85,r'$B=B_0+B_1 \\ \\cos{(2 \\, \\pi \\, \\nu \\, t)}$', multialignment='left', verticalalignment='top')\nfig.text(0.01,0.82,r'with $\\nu=100 \\ \\mathrm{kHz}$', multialignment='left', verticalalignment='top')\nfig.text(0.01,0.77,r'EPR X band frequency', multialignment='left', verticalalignment='top',weight='bold')\nfig.text(0.01,0.72,r'$\\nu_\\mathrm{X}=9388.2 \\ \\mathrm{MHz}$', multialignment='left', verticalalignment='top')\nfig.text(0.01,0.67,r'EPR spin level energies', multialignment='left', verticalalignment='top',weight='bold')\nfig.text(0.01,0.62,r'Up-state', multialignment='left', verticalalignment='top')\nfig.text(0.01,0.59,r'$E_\\mathrm{up}=\\frac{1}{2} \\, g \\, \\mu_\\mathrm{B} \\, B$', multialignment='left', verticalalignment='top')\nfig.text(0.01,0.54,r'Down-state', multialignment='left', verticalalignment='top')\nfig.text(0.01,0.51,r'$E_\\mathrm{down}=-\\frac{1}{2} \\, g \\, \\mu_\\mathrm{B} \\, B$', multialignment='left', verticalalignment='top')\n\n\nB=np.arange(Bmin,Bmax,0.0005)\n\n\n\n\nif __name__=='__main__':\n \n ax1.plot(B,E_up(B),lw=2,color='red',label='Up-state energy')\n ax1.plot(B,E_down(B),lw=2,color='blue',label='Down-state energy')\n ax1.plot([Bmin,Bmax],[E_trans()/2,E_trans()/2],':',lw=2,color='grey',label='X-band energy') \n ax1.plot([Bmin,Bmax],[-E_trans()/2,-E_trans()/2],':',lw=2,color='grey') \n \n \n ax1.set_xlim(Bmin,Bmax)\n ax1.set_xlabel('$B_0$ $\\mathrm{(T)}$')\n ax1.set_ylabel('$E$ $\\mathrm{(J)}$')\n \n \n ax2.set_xlim(Bmin,Bmax)\n ax2.set_ylim(-10,150)\n ax2.set_yticklabels([])\n ax2.set_xlabel('$B_0$ $\\mathrm{(T)}$')\n ax2.set_ylabel('$Absorption \\ intensity$')\n \n \n ax3.set_xlim(Bmin,Bmax)\n ax3.set_ylim(-1000000,1000000)\n ax3.set_yticklabels([])\n ax3.set_xlabel('$B_0$ $\\mathrm{(T)}$')\n ax3.set_ylabel('$First \\ derivative \\ intensity$')\n \n truc={}\n \n truc['$E_\\mathrm{trans}$'], = ax1.plot([], [],'--',lw=2, color='gray',label='$B_0$')\n truc['$Abs_courbe$'], = ax2.plot([],[],lw=2,color='red',label='Absorption signal')\n truc['$Der_courbe$'], = ax3.plot([],[],lw=2,color='red',label='First derivative signal')\n r1 = ax1.add_patch(patches.Rectangle((0, 0),DeltaBmax,E_trans(), edgecolor = '#000000', facecolor = '#dddddd', fill=True,label='Excitation band'))\n\n lines = {}\n\n lines['Absorption spot'], = ax2.plot([],[],'o',color='black',lw=2)\n lines['First derivative spot'], = ax3.plot([],[],'o',color='black',lw=2)\n \n ax1.legend()\n ax2.legend()\n ax3.legend()\n \n param_widgets = widgets.make_param_widgets(parameters, plot_data, slider_box=[0.20, 0.05, 0.35, 0.05])\n choose_widget = widgets.make_choose_plot(lines,box=[0.01,0.2,0.12, 0.1])\n reset_button = widgets.make_reset_button(param_widgets,box=[0.85, 0.05, 0.10, 0.05])\n \n plt.show()\n","sub_path":"EPR/free_electron_EPR_spectrum/free_electron_EPR_spectrum.py","file_name":"free_electron_EPR_spectrum.py","file_ext":"py","file_size_in_byte":5901,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"163997946","text":"import time\nfrom django.db import models\nfrom django.forms import ModelForm, forms\nfrom grants.settings import UPLOAD_USER_FOLDER\nfrom extuser.models import OrgUser\n\n\ndef user_directory_path(instance, filename):\n\t# завантаження відбувається в деректорію MEDIA_ROOT/user_/\n\treturn UPLOAD_USER_FOLDER + '{0}/{1}'.format(instance.user.id, filename)\n\n\nclass FileUser(models.Model):\n\tclass Meta:\n\t\tverbose_name = u'Файл користувача'\n\t\tverbose_name_plural = u'Файли користувачів'\n\t\tdb_table = 'file_name'\n\n\tuser = models.ForeignKey(OrgUser)\n\n\tname_file = models.FileField(\n\t\tverbose_name=u'Повний Шлях Файлу',\n\t\tupload_to=user_directory_path\n\t)\n\tname = models.FileField(\n\t\tverbose_name=u'Назва файлу',\n\t\tdefault=''\n\t)\n\tdate_joined = models.DateTimeField(\n\t\tverbose_name=u'Дата додавання',\n\t\tdefault=time.strftime(\"%Y-%m-%d %H:%M:%S\")\n\t)\n\n\tdef __str__(self):\n\t\treturn u'Файл: {}'.format(str(self.name_file).split('/')[-1])\n\n\tdef __repr__(self):\n\t\treturn u'<{}>'.format(str(self.name_file).split('/')[-1])\n\n\nclass FormFileUser(ModelForm):\n\tclass Meta:\n\t\tmodel = FileUser\n\t\tfields = ['name_file']\n\n\tdef clean_name_file(self):\n\t\tname = self.cleaned_data['name_file']\n\t\tallowed_file = ['pdf', 'PDF', 'jpg', 'jpeg', 'png', 'gif', 'bmp']\n\t\texc = str(name.name.split('.')[-1])\n\t\tname.name = str(name.name).lower().replace(' ', '-').replace('_', '-')\n\t\terror = ''\n\t\tif len(name.name) < 2:\n\t\t\terror += u'Довжина файла не може бути меншою 2х символів!'\n\t\t\traise forms.ValidationError(error)\n\t\texc = [ae for ae in allowed_file if exc == ae]\n\t\tif not exc:\n\t\t\terror += u'Файли з таким розширенням не дозволені!'\n\t\t\traise forms.ValidationError(error)\n\t\treturn name\n\n","sub_path":"fileuser/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":1838,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"577650186","text":"from collections import deque\n\nimport pytest\n\nfrom hashtable import Hashtable\n\n\ndef test_hash(hashtable):\n \"\"\"\n Hash function gives expected values.\n \"\"\"\n\n key = 'cat'\n\n expected = 8\n actual = hashtable.hash(key)\n\n assert actual == expected\n\n key = 'tac'\n\n actual = hashtable.hash(key)\n\n assert actual == expected\n\n\ndef test_add(hashtable):\n \"\"\"\n Can add a key-value pair to a hashtable.\n \"\"\"\n\n key = 'cat'\n value = 9\n hashtable.add(key, value)\n\n d = deque()\n d.append([key, value])\n\n expected = d[0]\n actual = hashtable.buckets[8][0]\n\n assert actual == expected\n\n\ndef test_add_collision(hashtable):\n \"\"\"\n Can add a key-value pair to a hashtable with collision.\n \"\"\"\n\n keys = ['cat', 'act']\n values = [9, 11]\n d = deque()\n for i in range(len(keys)):\n hashtable.add(keys[i], values[i])\n d.append([keys[i], values[i]])\n\n expected = d[i]\n actual = hashtable.buckets[8][i]\n\n assert actual == expected\n\n\ndef test_contains(hashtable):\n \"\"\"\n Can check a hashtable for a key-value pair.\n \"\"\"\n\n key = 'cat'\n value = 9\n hashtable.add(key, value)\n\n expected = True\n actual = hashtable.contains(key)\n\n assert actual == expected\n\n\ndef test_contains_collision(hashtable):\n \"\"\"\n Can check a hashtable for a key-value pair with hashtable collision.\n \"\"\"\n\n keys = ['cat', 'act']\n values = [9, 11]\n for i in range(len(keys)):\n hashtable.add(keys[i], values[i])\n\n expected = True\n actual = hashtable.contains(keys[1])\n\n assert actual == expected\n\n\ndef test_no_contains(hashtable):\n \"\"\"\n Not all key-value pairs need to be in a hashtable.\n \"\"\"\n\n key = 'cat'\n\n expected = False\n actual = hashtable.contains(key)\n\n assert actual == expected\n\n\ndef test_get(hashtable):\n \"\"\"\n Can get a value from a hashtable.\n \"\"\"\n\n key = 'cat'\n value = 9\n hashtable.add(key, value)\n\n expected = value\n actual = hashtable.get(key)\n\n assert actual == expected\n\n\ndef test_get_collision(hashtable):\n \"\"\"\n Can get a value from a hashtable with a collision\n \"\"\"\n\n keys = ['cat', 'act']\n values = [9, 11]\n for i in range(len(keys)):\n hashtable.add(keys[i], values[i])\n\n expected = values[1]\n actual = hashtable.get(keys[1])\n\n assert actual == expected\n\n\ndef test_no_get(hashtable):\n \"\"\"\n Can't get a value from a hashtbale which isn't in the hashtable.\n \"\"\"\n\n key = 'cat'\n\n expected = None\n actual = hashtable.get(key)\n\n assert actual == expected\n\n\n# Fixtures\n\n\n@pytest.fixture\ndef hashtable():\n \"\"\"\n Hashtable instance.\n \"\"\"\n\n return Hashtable()\n","sub_path":"python/challenges/hashtable/test_hashtable.py","file_name":"test_hashtable.py","file_ext":"py","file_size_in_byte":2696,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"533878057","text":"from main import db\nfrom flask import Blueprint\n\ndb_commands = Blueprint(\"db-custom\", __name__)\n\n@db_commands.cli.command(\"create\")\ndef create_db():\n db.create_all()\n print(\"Tables created!\")\n\n@db_commands.cli.command(\"drop\")\ndef drop_db():\n db.drop_all()\n print(\"Tables deleted\")\n \n@db_commands.cli.command(\"seed\")\ndef seed_db():\n from models.Artists import Artists\n from models.User import User\n from main import bcrypt\n from faker import Faker\n import random\n\n faker = Faker()\n # users = []\n\n # for i in range(5):\n # user = User()\n # user.email = f\"test{i}@test.com\"\n # user.password = bcrypt.generate_password_hash(\"123456\").decode(\"utf-8\")\n # db.session.add(user)\n # users.append(user)\n\n # db.session.commit()\n\n for i in range(20):\n artist = Artists()\n artist.name = faker.catch_phrase()\n db.session.add(artist)\n \n db.session.commit()\n print(\"Tables seeded\")","sub_path":"src/commands.py","file_name":"commands.py","file_ext":"py","file_size_in_byte":973,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"9893741","text":"# -*- coding:utf-8 _*- \n\"\"\" \n@author:Administrator\n@file: merge_train_data_process.py\n@time: 2018/11/20\n\"\"\"\nimport pandas as pd\n\n\ndata = pd.read_csv('./merge_train_data.csv')\ndata_orgin = data\n\n# print(data.head())\nprint(data.shape)\ndata = data[abs(data.trainPrediction-data.daysOnMarket)<20]\n# print(data.head())\nprint(data.shape)\ndata.to_csv('../input/treb_toronto_3to8_1.csv')\ndata_orgin.to_csv('./orgin_data.csv')\n","sub_path":"first_reporter_task_one_week_finish/test_treb2/test_treb/merge_data_bak/merge_train_data_process.py","file_name":"merge_train_data_process.py","file_ext":"py","file_size_in_byte":419,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"348577486","text":"## CREATED BY:\n## Gert Sterenborg; gertsterenborg@gmail.com\n## 21-01-2015\n\n##imports:\nimport os\nimport json\nimport urllib2\nimport osgeo.ogr, osgeo.osr\n\ndef getLatLng(place):\n ## Fetches coordinates from the google api\n url = \"http://maps.googleapis.com/maps/api/geocode/json?address=\"+place\n response = urllib2.urlopen(url)\n jsonF = json.loads(response.read())\n if jsonF['status'] == \"OK\":\n lat = jsonF['results'][0]['geometry']['location']['lat']\n lng = jsonF['results'][0]['geometry']['location']['lng']\n return lat,lng\n\ndef storeShp(placeDic):\n ## current file location\n path = os.path.dirname(os.path.realpath(__file__))+\"/\"\n ## remove shapefile\n removeShp(path,'places')\n spatialReference = osgeo.osr.SpatialReference()\n spatialReference.ImportFromEPSG(4326) ##WGS84 degrees coordinates\n driver = osgeo.ogr.GetDriverByName('ESRI Shapefile') # will select the driver foir our shp-file creation.\n shapeData = driver.CreateDataSource(path) #so there we will store our data\n layer = shapeData.CreateLayer('places', spatialReference, osgeo.ogr.wkbPoint) #this will create a corresponding layer for our data with given spatial information.\n layer_defn = layer.GetLayerDefn() # gets parameters of the current shapefile\n new_field = osgeo.ogr.FieldDefn('PLACE', osgeo.ogr.OFTString)\n layer.CreateField(new_field)\n point = osgeo.ogr.Geometry(osgeo.ogr.wkbPoint)\n i = 0\n for place in placeDic:\n point.AddPoint(placeDic[place]['lng'],placeDic[place]['lat']) #create a new point at given ccordinates\n featureIndex = i\n feature = osgeo.ogr.Feature(layer_defn)\n feature.SetGeometry(point)\n feature.SetFID(featureIndex)\n j = feature.GetFieldIndex(\"PLACE\")\n feature.SetField(j, place)\n layer.CreateFeature(feature)\n i+= 1\n shapeData.Destroy() #lets close the shapefile\n\ndef removeShp(path,fileName):\n ## removes the exsisting shapefile\n extensions = [\"shp\",\"shx\",\"prj\",\"dbf\"]\n for extension in extensions:\n command = \"rm \"+path+fileName+\".\"+extension\n os.system(command)\n\nif __name__ == \"__main__\":\n placeDic = {} ## dictionary where all the coordinates and places will be stored in\n with open(\"places.txt\") as f:\n for line in f:\n lineSplit = line.split(',')\n for place in lineSplit:\n lat,lng = getLatLng(place.strip())\n placeDic[place.strip()] = {\n 'lat':lat,\n 'lng':lng}\n storeShp(placeDic)\n try: ## show the result in qgis\n os.system(\"qgis places.shp\")\n except:\n pass\n\n \n","sub_path":"PlacesToShape.py","file_name":"PlacesToShape.py","file_ext":"py","file_size_in_byte":2674,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"433653077","text":"#!/usr/bin/env python3\n\nimport sys, os\nsys.path.insert(0, os.path.abspath('../lib'))\n\nimport time\nimport random\nimport numpy as np\n\nfrom common.core import BaseWidget, run, lookup\nfrom common.gfxutil import topleft_label, CEllipse, KFAnim, AnimGroup\n\nfrom kivy.uix.image import Image\nfrom kivy.core.image import Image as Img\nfrom kivy.uix.widget import Widget\nfrom kivy.core.window import Window\nfrom kivy.graphics import Color, Ellipse, Rectangle, Line\nfrom kivy.graphics.instructions import InstructionGroup\n\n\nclass InteractiveImage(Image):\n def __init__(self, **kwargs):\n super(InteractiveImage, self).__init__(**kwargs, keep_data=True, allow_stretch=True, keep_ratio=False)\n self.callback = None\n Window.bind(mouse_pos=self.on_mouse_pos)\n\n def set_callback(self, callback):\n self.callback = callback\n\n def collide_point(self, x, y):\n try:\n # Adjust x and y to reflect coordinates within the image\n x = (x - self.x) * self._coreimage.width / self.width\n y = (self.height - (y - self.y)) * self._coreimage.height / self.height\n color = self._coreimage.read_pixel(x, y)\n except:\n color = 0, 0, 0, 0\n if color[-1] > 0:\n return True\n return False\n\n def on_mouse_pos(self, window, pos):\n if self.collide_point(*pos):\n self.color = [1, 1, 1, 0.5]\n else:\n self.color = [1, 1, 1, 1]\n\n def on_touch_down(self, touch):\n if self.collide_point(*touch.pos):\n if not self.callback is None:\n self.callback()\n\n\nclass FadingMusicNote(InstructionGroup):\n def __init__(self, pos=(0, 0)):\n super(FadingMusicNote, self).__init__()\n self.body = Rectangle(pos=pos, size=(50, 50), texture=Img('./data/scene/eightnote.png').texture)\n self.pop_anim = KFAnim((0, self.body.size[0]), (.5, self.body.size[0]), (1.0, 0))\n mag = random.uniform(20, 30)\n theta = random.uniform(0, 2*np.pi)\n dx, dy = mag * np.cos(theta), mag * np.sin(theta)\n self.pos_anim = KFAnim((0, pos[0], pos[1]), (.5, pos[0] + dx, pos[1] + dy))\n self.add(self.body)\n self.time = 0\n self.active = True\n self.on_update(0)\n\n def on_update(self, dt):\n # the disappearing animation just reduces the size\n new_size = self.pop_anim.eval(self.time)\n new_pos = self.pos_anim.eval(self.time)\n self.body.size = (new_size, new_size)\n self.body.pos = new_pos\n self.time += dt\n return self.pop_anim.is_active(self.time)\n\n def start_anim(self):\n self.active = True\n\nclass FlyingCarWidget(Image):\n def __init__(self, init_pos, size, velocity, **kwargs):\n super(FlyingCarWidget, self).__init__(**kwargs)\n self.init_pos = init_pos\n self.size = size\n\n self.velocity = velocity\n self.t = time.time()\n\n self.anim_delay = 0.5\n\n def is_visible(self):\n x, y = self.pos\n w, h = self.size\n\n if self.velocity > 0:\n return x < Window.width\n else:\n return x + w > 0\n\n def on_update(self):\n t = time.time() - self.t\n x_0, y_0 = self.init_pos\n self.pos = (x_0 + self.velocity * t, y_0)\n\nclass FlyingCarGeneratorWidget(BaseWidget):\n\n car_assets = (\n (\"./data/scene/food_truck.gif\", \"./data/scene/food_truck_reverse.gif\"),\n (\"./data/scene/nyan_cat.gif\", \"./data/scene/nyan_cat_reverse.gif\"),\n (\"./data/scene/warp_ship.gif\", \"./data/scene/warp_ship_reverse.gif\"),\n (\"./data/scene/superman.png\", \"./data/scene/superman_reverse.png\"),\n (\"./data/scene/flying_delorean.png\", \"./data/scene/flying_delorean_reverse.png\"),\n )\n\n def __init__(self, y_range):\n super(FlyingCarGeneratorWidget, self).__init__()\n\n self.y_range = y_range\n self.speed = 90 # pixels/sec\n self.cars = []\n\n self.t_next_car = 0\n self.max_cars = 5\n\n def generate_car(self):\n\n # Choose random car asset\n forward, backward = random.choice(self.car_assets)\n\n # Choose randomly between forward and backward\n direction = random.choice([\"forward\", \"backward\"])\n if direction == \"forward\":\n source = forward\n velocity = self.speed * random.uniform(0.8, 1.2)\n x_0 = 0\n else:\n source = backward\n velocity = -self.speed * random.uniform(0.8, 1.2)\n x_0 = Window.width\n\n # Randomly select starting y coordinate\n y_0 = random.uniform(*self.y_range)\n\n # Construct car widget\n car = FlyingCarWidget((x_0, y_0), (100, 100), velocity)\n car.source = source\n self.cars.append(car)\n self.add_widget(car)\n\n\n def on_update(self):\n\n # Process updates for each car\n for car in self.cars:\n car.on_update()\n\n # Check for cars that have gotten\n cars_to_delete = []\n for i, car in enumerate(self.cars):\n if not car.is_visible():\n cars_to_delete.append(i)\n\n # Delete complete cars\n for i in cars_to_delete:\n self.remove_widget(self.cars[i])\n del self.cars[i]\n\n # Generate new cars if there's room and if we haven't\n # recently created a new car\n t_now = time.time()\n if len(self.cars) < self.max_cars and \\\n t_now > self.t_next_car:\n self.generate_car()\n self.t_next_car = t_now + random.uniform(1, 5)\n\nclass BackgroundWidget(BaseWidget):\n def __init__(self):\n super(BackgroundWidget, self).__init__()\n\n # Background\n self.background = Image(allow_stretch=True, keep_ratio=False)\n self.background.source = \"./data/scene/background.png\"\n self.add_widget(self.background)\n\n self.car_generator = FlyingCarGeneratorWidget((100, 500))\n self.add_widget(self.car_generator)\n\n def on_layout(self, win_size):\n self.background.size = win_size\n\n def on_update(self):\n pass\n\nclass ForegroundWidget(BaseWidget):\n def __init__(self):\n super(ForegroundWidget, self).__init__()\n\n # Foreground\n self.foreground = Image(allow_stretch=True, keep_ratio=False)\n self.foreground.source = \"./data/scene/foreground.png\"\n self.add_widget(self.foreground)\n\n # Amp\n self.amp = InteractiveImage()\n self.amp.source = \"./data/scene/amp.png\"\n self.amp.set_callback(lambda: print(\"amp\"))\n self.add_widget(self.amp)\n\n # Guitar\n self.guitar = InteractiveImage()\n self.guitar.source = \"./data/scene/guitar.png\"\n self.guitar.set_callback(lambda: print(\"guitar\"))\n self.add_widget(self.guitar)\n\n # Mic\n self.mic = InteractiveImage()\n self.mic.source = \"./data/scene/mic.png\"\n self.mic.set_callback(lambda: print(\"mic\"))\n self.add_widget(self.mic)\n \n # Radio\n self.radio = InteractiveImage()\n self.radio.source = \"./data/scene/radio.png\"\n self.radio.set_callback(lambda: print(\"radio\"))\n self.add_widget(self.radio)\n \n # File cabinet\n self.storage = InteractiveImage()\n self.storage.source = \"./data/scene/storage.png\"\n self.storage.set_callback(lambda: print(\"storage\"))\n self.add_widget(self.storage)\n\n def on_layout(self, win_size):\n self.foreground.size = win_size\n self.amp.size = win_size\n self.guitar.size = win_size\n self.mic.size = win_size\n self.radio.size = win_size\n self.storage.size = win_size\n \n\nclass Scene(BaseWidget):\n def __init__(self):\n super(Scene, self).__init__()\n\n self.background = BackgroundWidget()\n self.add_widget(self.background)\n\n self.foreground = ForegroundWidget()\n self.add_widget(self.foreground)\n\n # Flying music notes\n self.anim_group = AnimGroup()\n self.canvas.add(self.anim_group)\n self.anim_group.add(FadingMusicNote())\n\n def on_layout(self, win_size):\n self.background.on_layout(win_size)\n self.foreground.on_layout(win_size)\n\n def on_update(self):\n self.anim_group.on_update()\n\n def add_note_sprite(self):\n self.anim_group.add(FadingMusicNote((320, 80)))\n\n\nif __name__ == \"__main__\":\n run(Scene())\n","sub_path":"graphics.py","file_name":"graphics.py","file_ext":"py","file_size_in_byte":8388,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"415652401","text":"\n#ejercicio 01 \"validacion de un entero\"\ndef validar_un_entero(entero): #se aplica la funcion validar un entero\n if entero.isalnum()==True: #condicion doble\n validacion=\"El numero \"+str(entero)+\" si es entero\" #se guarda un valor en la variable validacion\n return validacion #retorna la variable validacion\n else:\n return False #retorna falso si la condicion es falsa\n #fin_def\n\n#ejercicio 2 \"validacion de una cadena\"\ndef validar_cadena(msg): #se aplica la funcion validar cadena:\n if msg.isalpha()==True:# #se aplica una condicional doble\n validacion_cadena=\"El valor \"+msg+\" si es una cadena \" #se guarda un valor en la variable validacion de cadena\n return validacion_cadena #retorna el valor de la variable validacion cadena\n else:\n return False #retorna falso si la condicion es falsa\n #fin_def\n\n\n#ejercicio 03 \"validacion de ataque y recompensa de un videojuego\"\ndef validar_dano(ataque): #funcion def\n print(\"INDICAR EL TIPO A QUIEN ATACA MELE o RANGO\") #se imprime un comentario inicial\n if ataque.isdigit()==False: #se valida si no es un numero condicion doble\n if ataque==\"MELE\": #condiciones multiples para ver que recommpensa se ganara\n rm=35\n recompensa_mele=\"Su recompensa es de: \"+str(rm)\n return recompensa_mele\n elif ataque==\"RANGO\":\n rr=43\n recompensa_rango=\"Su recompensa es de: \"+str(rr)\n return recompensa_rango\n else:\n comando=\"el comando ingresado es FALSO\"\n return comando\n else:\n return False\n#fin_def\n\n\n#ejercicio 04 \"validacion de impresion de numeros de 3 digitos\"\ndef validar_numero_tres_cifras(numero):#funcion def\n #validando si es de tres cifras\n if len(numero)==3:\n #validando si es entero\n if numero.isdigit()==True:\n return True\n\n else:\n return False\n else:\n return False\n#fin_def\n\n\n#ejercicio 05 \"validacion de una vocal\"\ndef validar_vocal(vocal):\n #validando la longitud\n if len(vocal)==1:\n #validando si es una vaocal ingresada\n if vocal==\"a\" or vocal==\"e\" or vocal==\"i\" or vocal==\"o\" or vocal==\"u\" :\n return True\n else:\n return False\n else:\n return False\n\n #fin_def\n\n#ejercicio 06 \"validacion de DNI\"\ndef validar_dni(DNI):\n #primero validamos la longitud de la cadena\n if len(DNI)==8:\n #valdidamos que los datos ingresados sean puros numero enteros\n if DNI.isdigit()==True:\n return True\n else:\n return False\n else:\n return False\n\n#ejercicio 07 \"validar Ruc SUNAT\"\ndef validar_ruc(RUC):\n fragmento=validar_dni(\"\")\n #primero validamos la longitud de la cadena que consta de 11 digitos\n if len(RUC)==11:\n #segundo validamos las condiciones de la cadena\n if RUC[0:2]==10:\n print(\"la persona es un trabajador fisico o natural\")\n #una ves que validamos todas las condiciones de la cadena sigue validar el numero de dni\n #para eso creamos una variable a al que llamemos a la funcion validar dni\n if RUC[2:11]==fragmento:\n #por ultimo hacemos la validacion delultimo codigo que siempre debe ser de 8\n if RUC[11]==8:\n print(\"el codigo es correcto\") #una ves que hicimos la validacion de la cadena lo que haremos finalmente es verificar si es int\n if RUC.isdigit()==True:\n return True\n else:\n return False\n else:\n return False\n else:\n return False\n elif RUC[0:2]==20:\n print(\"la persona es una trabajador fisico\")\n if RUC[2:11]==fragmento:\n if RUC[11]==8:\n print(\"el codigo es correcto\")\n if RUC.isdigit()==True:\n return True\n else:\n return False\n else:\n return False\n elif RUC[0:2]==15:\n print(\"La persona es una socidad \")\n if RUC[2:11]==fragmento:\n if RUC[11]==8:\n print(\"el codigo es correcto\")\n if RUC.isdigit()==True:\n return True\n else:\n return False\n else:\n return False\n elif RUC[0:2]==16:\n print(\"se menciona como valido\")\n if RUC[2:11]==fragmento:\n if RUC[11]==8:\n print(\"el codigo es correcto\")\n if RUC.isdigit()==True:\n return True\n else:\n return False\n else:\n return False\n elif RUC[0:2]==17:\n print(\"inscripcion durante 2019 a 2025\")\n if RUC[2:11]==fragmento:\n if RUC[11]==8:\n print(\"el codigo es correcto\")\n if RUC.isdigit()==True:\n return True\n else:\n return False\n else:\n return False\n else:\n\n return False\n\n else:\n return False\n\n\n\n#ejercicio 08 \"validar un numero capicua\"\n\ndef validar_capicua(capicua):\n #primero validamos que sea int\n if capicua.isdigit()==True:\n #segundo para que un numero sea capicua tiene que ser igual que su inversa entonces\n if capicua==capicua[::-1]:\n return True\n else:\n return False\n else:\n return False\n\n #fin_si\n#fin_def\n\n\n#ejercicio 09 \"validar edad\"\ndef validar_edad(edad):\n #primero validar si es entero\n if edad.isdigit()==True:\n if edad>0 and edad<120:\n return True\n else:\n return False\n\n else:\n return False\n #fin_si\n\n#fin_def\n\n#ejercicio 10 \"validar codigo de alumno de UNPRG\"\ndef validar_codigo_unprg(codigo):\n #primero validamos la longitud de la cadena\n if len(codigo)==7:\n #validacion de los dos primero digitos\n if codigo[0:2]==19:\n #validamos la ultima parte de la cadena\n if codigo[7].isdigit()==True:\n #validamos que una parte de la cadena sea numeros\n if codigo[0:7].isdigit()==True:\n return True\n else:\n return False\n #fin_si\n else:\n return False\n #fin_si\n else:\n return False\n #fin_si\n else:\n return False\n #fin_si\n\n#fin_def\n\n#ejercicio 11 \"validacion de una fuerza realizada\"\ndef validar_fuerza(masa,aceleracion):\n #primero validamos el numero ingresado que sea un real\n fuerza=int(masa)*int(aceleracion)\n fuerza_cero=\"La fuerza realizada es cero ya que la aceleracion es igual a 0\"\n if masa.isdigit()==True:\n if aceleracion.isdigit()==True:\n if fuerza==0:\n return fuerza_cero\n elif fuerza>0:\n return fuerza\n else:\n return -1*fuerza\n\n else:\n return False\n else:\n return False\n #fin_si\n#fin_def\n\n#ejercicio 12 \"validar un interruptor\"\ndef validar_interruptor(comando):\n #validar que sea un str:\n on=\"prendido\"\n oof=\"apagado\"\n if comando.isalpha()==True:\n #una ves que validamos que sea un str colocamos las condiciones\n if comando.upper()==\"ON\":\n return on\n elif comando.upper()==\"OOF\":\n return oof\n else:\n return False\n else:\n return False\n\n\n #fin_si\n\n#ejercicio 13 \"lanzamiento de un balon de basket\"\ndef validar_lanzamiento(distancia):\n #validamos primeramente la longitud de la cadena\n if len(distancia[0])>0 and len(distancia[0])<=3:\n #validamos que sea alfanumerico\n\n if distancia.isalnum()==True:\n\n if len(distancia)==2:\n #validamos que tenga el signo al final\n if distancia[1]==\"m\":\n if int(distancia[0])>=6:\n print(\"fue anotacion de 3\")\n else:\n print(\"FUE anotacion de 2\")\n else:\n return False\n elif len(distancia)==3:\n if distancia[2]==\"m\":\n if int(distancia[0])>=6:\n print(\"fue anotacion 3\")\n else:\n print(\"fue anotacion de 2\")\n else:\n return False\n else:\n return False\n else:\n return False\n else:\n return False\n\n#ejercicio 14 \"numero telefonico de peru ejemplo \"+51973396201\" \"\ndef validar_numero_peru(telefono):\n #validamos la longitud de la cadena\n if len(telefono)==12:\n #validamos el valor inicial\n if telefono[0]==\"+\":\n #validamos los dos siguientes digitos\n if telefono[1:3]==51:\n if telefono[1:].isdigit()==True:\n return True\n else:\n return False\n else:\n return False\n else:\n return False\n else:\n return False\n\n\n#ejercicio 15 \"validar un factorial\"\n\ndef validar_factorial(cifras,factorial):\n cif=1\n #validamos la longitud de cualquier factorial y a la vez estamos validando de que no sea un numero negativo\n if cifras+1==len(factorial):\n #validamos el ingreso de factoriales\n if factorial[cifras+1]==\"!\":\n #validamos que sean numeros\n if factorial[0:cifras+1].isdigit()==True:\n #una vez que validamos esttablecemos condiciones\n #si es facotrial d euno o cero retorna automaticamente un 1 si no se hara el calculo respectivoy retornara la variable factor\n if factorial==\"1!\" or factorial==\"0!\":\n return 1\n else:\n for i in range(2,factorial[0:cifras+1]):\n cif*=i\n return cif\n else:\n return False\n else:\n return False\n else:\n return False\n #FIN_SI\n\n#fin_def\n\n#ejercicio 16 \"validar el comando de un videojuego\"\ndef valida_salto(hop):\n pos=1\n letra=\"\"\n numero=\"\"\n direccion=\"\"\n if len(hop)==13:\n for iteam in hop.split(\" \"):\n if pos==1:\n iteam=letra\n if letra.isdigit()==False:\n return True\n else:\n return False\n if pos==2:\n iteam=numero\n if numero.isalpha()==False:\n return True\n else:\n return False\n if pos==3:\n iteam=direccion\n if direccion.isdigit()==False:\n return True\n else:\n return False\n else:\n return False\n\n\n#ejercicio 17 \"validar la hora ingresada ejemplo 03:34 am\"\ndef validar_hora(horario):\n pico_del_dia=\"12:00 m\"\n #validamos la longitud\n if len(horario)==8:\n #validamos la estructura de la cadena\n if horario[0:2].isdigit()==True:\n if horario[2]==\":\":\n if horario[3:5].isdigit()==True:\n if horario[5]==\" \":\n if horario[6:]==\"am\":\n return True\n elif horario[6:]==\"pm\":\n return True\n else:\n return False\n else:\n return False\n else:\n return False\n else:\n return False\n else:\n return False\n elif len(horario)==7:\n return pico_del_dia\n else:\n return False\n\n\n\n#ejercicio 18 \"validar un numero mayo de 2 cifras en donde si es mayor retorna el numero mayor\"\ndef validar_mayor(num1,num2):\n #validamos la longitud\n if len(num1)==2 and len(num2):\n #validamos si la cadena esta compuesta de digitos\n if num1.isdigit()==True and num2.isdigit()==True:\n #validamos que sea mayor\n if num1>num2:\n return num1\n else:\n return False\n else:\n return False\n else:\n return False\n\n\n#ejercicio 19 \"validar un numero menor de 2 cifras en donde si es menor retorna el menor\"\ndef validar_menor(numero1,numero2):\n #validamos la longitud\n if len(numero1)==2 and len(numero2):\n #VALIDAMOS QUE LA CADENA ESTE COMPUESTA DE NUMEROS\n if numero1.isdigit()==True and numero2.isdigit()==True:\n #validamos que sea menor\n if numero1 6:\n raise UserError(\n _(\"You can't have more than six assigned positive tags.\"))\n\n if self.search_count(\n [('type', '=', 'negative'), ('assigned', '=', True)]) > 6:\n raise UserError(\n _(\"You can't have more than six assigned negative tags.\"))\n","sub_path":"pivotino_general_feedback/models/general_feedback_tags.py","file_name":"general_feedback_tags.py","file_ext":"py","file_size_in_byte":1073,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"86484817","text":"from rest_framework import serializers\n\nfrom .models import Category, Comment, Genre, Review, Title\n\n\nclass CategorySerializer(serializers.ModelSerializer):\n '''Serializer for Category model'''\n\n class Meta:\n fields = ('name', 'slug')\n model = Category\n lookup_field = 'slug'\n\n\nclass GenreSerializer(serializers.ModelSerializer):\n '''Serializer for Genre model'''\n\n class Meta:\n fields = ('name', 'slug')\n model = Genre\n lookup_field = 'slug'\n\n\nclass CategoryReprField(serializers.SlugRelatedField):\n '''Serializer for Category model'''\n\n def to_representation(self, value):\n return {'name': value.name, 'slug': value.slug}\n\n\nclass GenreReprField(serializers.SlugRelatedField):\n '''GenreReprField Serializer'''\n\n def to_representation(self, value):\n return {'name': value.name, 'slug': value.slug}\n\n\nclass TitleSerializer(serializers.ModelSerializer):\n '''Serializer for Title model.'''\n category = CategoryReprField(slug_field='slug',\n queryset=Category.objects.all())\n genre = GenreReprField(slug_field='slug',\n queryset=Genre.objects.all(),\n many=True)\n\n class Meta:\n fields = (\n 'id',\n 'name',\n 'year',\n 'rating',\n 'description',\n 'genre',\n 'category',\n )\n model = Title\n\n\nclass ReviewSerializer(serializers.ModelSerializer):\n '''Serializer for Review model. Slug related field author.'''\n author = serializers.SlugRelatedField(\n slug_field='username',\n read_only=True\n )\n\n def get_serializer_context(self):\n return {'title_id': self.kwargs['title_id'], 'request': self.request}\n\n def validate(self, data):\n '''Call the instance's validate() method and\n raise error if user has already added a review for this tittle.\n '''\n title_id = self.context.get('request').parser_context['kwargs']['title_id']\n if (Review.objects.filter(title_id=title_id, author=self.context['request'].user).exists()\n and self.context['request'].method == 'POST'):\n raise serializers.ValidationError('This user has already added review for this title')\n return data\n\n class Meta:\n fields = ('id', 'text', 'author', 'score', 'pub_date',)\n model = Review\n\n\nclass CommentSerializer(serializers.ModelSerializer):\n '''Serializer for Comment model. Slug related field author.'''\n author = serializers.SlugRelatedField(\n slug_field='username',\n read_only=True\n )\n\n class Meta:\n model = Comment\n fields = ('id', 'text', 'author', 'pub_date')\n\n\nclass EmailSerializer(serializers.Serializer):\n '''Email Serializer'''\n email = serializers.EmailField(required=True)\n","sub_path":"api/serializers.py","file_name":"serializers.py","file_ext":"py","file_size_in_byte":2874,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"453666278","text":"import collections\n\n\ndef flatten(x):\n result = []\n for el in x:\n if isinstance(x, collections.Iterable) and not isinstance(el, dict):\n result.extend(flatten(el))\n else:\n result.append(el)\n return result\n\npseudo_parameters = [\n 'AWS::AccountId',\n 'AWS::NotificationARNs',\n 'AWS::NoValue',\n 'AWS::Partition',\n 'AWS::Region',\n 'AWS::StackId',\n 'AWS::StackName',\n 'AWS::URLSuffix'\n]","sub_path":"cloud_formation_viz/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":447,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"141366041","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport socket\nimport struct\nimport json\nimport os\n\n\nclass FTPClient:\n address_family = socket.AF_INET\n socket_type = socket.SOCK_STREAM\n max_package_size = 8192\n coding = 'utf-8'\n client_dir = 'C:\\\\Users\\Administrator\\\\PycharmProjects\\\\fullstack'\n\n def __init__(self, server_address, connect=True):\n \"\"\"Constructor. May be extended, do not override.\"\"\"\n self.server_address = server_address\n self.connect = connect\n self.socket = socket.socket(self.address_family, self.socket_type)\n\n if connect:\n try:\n self.client_connect()\n except:\n self.client_close()\n raise\n\n def client_connect(self):\n \"\"\"Called by constructor to bind the socket.\n\n May be overridden.\n \"\"\"\n self.socket.connect_ex(self.server_address)\n\n def client_close(self):\n \"\"\"Called to clean-up the server.\n May be overridden.\n \"\"\"\n self.socket.close()\n\n def run(self):\n \"\"\"merely processing client command\"\"\"\n if not self.connect:\n self.client_connect()\n while True: # 通信循环\n cmd = input('>>>:').strip()\n if not cmd:\n continue\n cmd_list = cmd.split(' ')\n command = cmd_list[0]\n if hasattr(self, command):\n func = getattr(self, command)\n func(cmd_list)\n else:\n print('命令格式输入有误')\n\n def put(self, args):\n # 规范化文件路径,os.path.normpath在linux平台无效\n file_path = os.path.normpath(args[-1])\n if not os.path.exists(file_path):\n print('要上传的文件不存在')\n # 此处调用return是为了防止else语句块过大\n return\n else:\n file_size = os.stat(file_path).st_size\n # 发送头信息\n headers_dict = {'command': args[0], 'filename': os.path.basename(file_path), 'filesize': file_size}\n headers_json = json.dumps(headers_dict)\n headers_bytes = headers_json.encode(self.coding)\n # 注意,struct发的是长度,先把头的长度再发过去,然后在send头信息,server取固定的头长度得到头信息长度,根据\n # 头长度取得头信息,因此这里的两次send不会粘包,因为有头信息来控制\n self.socket.send(struct.pack('i', len(headers_bytes)))\n self.socket.send(headers_bytes)\n\n send_size = 0\n with open(file_path, 'rb') as f:\n for line in f:\n already_send_size = len(line)\n self.socket.send(line)\n send_size += already_send_size\n # print(send_size)\n else:\n print('upload success')\n\n def get(self, args):\n headers_dict = {'command': args[0], 'filename': args[-1]}\n headers_json = json.dumps(headers_dict)\n headers_bytes = headers_json.encode(self.coding)\n headers_length = len(headers_bytes)\n self.socket.send(struct.pack('i', headers_length))\n self.socket.send(headers_bytes)\n\n headers_struct = self.socket.recv(4)\n headers_length = struct.unpack('i', headers_struct)[0]\n headers_str = self.socket.recv(headers_length)\n headers_dict = json.loads(headers_str)\n err_msg = headers_dict.get('err_msg', None)\n if err_msg:\n print(err_msg)\n return\n file_path = os.path.join(self.client_dir, args[-1])\n print(file_path)\n filesize = headers_dict.get('file_size')\n already_recv_size = 0\n with open(file_path, 'wb') as f:\n while already_recv_size < filesize:\n recv_data = self.socket.recv(self.max_package_size)\n f.write(recv_data)\n already_recv_size += len(recv_data)\n\nif __name__ == '__main__':\n obj = FTPClient(('192.168.0.98', 13140))\n obj.run()\n\n\n\n","sub_path":"day39/ftpclient.py","file_name":"ftpclient.py","file_ext":"py","file_size_in_byte":4036,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"115545724","text":"import os\r\n\r\ninput_file = open(\"input_large.in\", \"r\")\r\noutput_file = open(\"output_large.txt\", \"w\")\r\n\r\ncases = int(input_file.readline())\r\n\r\nfor i in range(cases):\r\n\tstack = list(input_file.readline()[0:-1])\r\n\tlast = \"\"\r\n\tclean_stack = []\r\n\tfor p in stack:\r\n\t\tif p != last:\r\n\t\t\tclean_stack.append(p)\r\n\t\tlast = p\r\n\r\n\tfrowns = clean_stack.count(\"-\") * 2\r\n\tif clean_stack[0] == \"-\":\r\n\t\tfrowns -= 1\r\n\r\n\toutput_file.write(\"Case #\" + str(i+1) + \": \" + str(frowns) + \"\\n\")","sub_path":"codes/CodeJamCrawler/16_0_2_neat/16_0_2_PartlyGloudy_Revenge of the Pancakes.py","file_name":"16_0_2_PartlyGloudy_Revenge of the Pancakes.py","file_ext":"py","file_size_in_byte":464,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"355570550","text":"import matplotlib.pyplot as plt\r\nimport numpy as np\r\n\r\nfrom matplotlib import style\r\n\r\nstyle.use(\"ggplot\")\r\n\r\nclass support_vector_machine:\r\n def __init__ (self, visualisation = True):\r\n self.visualisation = visualisation\r\n self.colors = {1: \"r\", -1: \"b\"}\r\n if self.visualisation:\r\n self.fig = plt.figure()\r\n self.ax = self.fig.add_subplot(1, 1, 1)\r\n \r\n def fit (self, data):\r\n self.data = data\r\n # {||w||: [w, b]} \r\n opt_dict = {}\r\n transforms = [[1, 1], [1, -1], [-1, 1], [-1, -1]]\r\n \r\n all_data = []\r\n for y_i in self.data:\r\n for feature_set in self.data[y_i]:\r\n for feature in feature_set:\r\n all_data.append(feature)\r\n \r\n self.max_feature_value = max(all_data)\r\n self.min_feature_value = min(all_data)\r\n all_data = None\r\n \r\n step_sizes = [self.max_feature_value * 0.1, self.max_feature_value * 0.01, self.max_feature_value * 0.001]\r\n b_range_multiple = 5\r\n b_multiple = 5\r\n \r\n latest_optimum = self.max_feature_value * 10\r\n \r\n for step in step_sizes: \r\n w = np.array([latest_optimum, latest_optimum])\r\n # SVM is always a convex optimisation problem\r\n optimised = False\r\n \r\n while not optimised:\r\n for b in np.arange(-1 * (self.max_feature_value * b_range_multiple), self.max_feature_value * b_range_multiple, step * b_multiple):\r\n for transformation in transforms:\r\n w_t = w * transformation\r\n found_option = True\r\n for i in self.data:\r\n for x_i in self.data[i]:\r\n y_i = i\r\n if not y_i * (np.dot(w_t, x_i) + b) >= 1:\r\n found_option = False\r\n \r\n if found_option:\r\n opt_dict[np.linalg.norm(w_t)] = [w_t, b]\r\n \r\n if w[0] < 0:\r\n optimised = True\r\n print(\"Optimised a step\")\r\n else:\r\n w = w - step\r\n \r\n norms = sorted([n for n in opt_dict])\r\n opt_choice = opt_dict[norms[0]]\r\n \r\n self.w = opt_choice[0]\r\n self.b = opt_choice[1]\r\n latest_optimum = opt_choice[0][0] + step * 2\r\n \r\n \r\n def predict (self, features):\r\n classification = np.sign(np.dot(np.array(self.w), self.w) + self.b)\r\n \r\n if classification != 0 and self.visualisation:\r\n self.ax.scatter(features[0], features[1], s = 200, marker = \"*\", color = self.colors[classification])\r\n \r\n return classification\r\n \r\n def visualise (self):\r\n [[self.ax.scatter(x[0], x[1], s = 100, color = self.colors[i]) for x in data_dict[i]] for i in data_dict]\r\n \r\n def hyper_plane (x, w, b, v):\r\n return (-w[0] * x - b + v) / w[1]\r\n \r\n data_range = (self.min_feature_value * 0.9, self.max_feature_value * 1.1)\r\n hyp_x_min = data_range[0]\r\n hyp_x_max = data_range[1]\r\n \r\n psv_1 = hyper_plane(hyp_x_min, self.w, self.b, 1)\r\n psv_2 = hyper_plane(hyp_x_max, self.w, self.b, 1)\r\n self.ax.plot([hyp_x_min, hyp_x_max], [psv_1, psv_2])\r\n \r\n nsv_1 = hyper_plane(hyp_x_min, self.w, self.b, -1)\r\n nsv_2 = hyper_plane(hyp_x_max, self.w, self.b, -1)\r\n self.ax.plot([hyp_x_min, hyp_x_max], [nsv_1, nsv_2])\r\n \r\n db_1 = hyper_plane(hyp_x_min, self.w, self.b, 0)\r\n db_2 = hyper_plane(hyp_x_max, self.w, self.b, 0)\r\n self.ax.plot([hyp_x_min, hyp_x_max], [db_1, db_2])\r\n \r\n plt.show()\r\n \r\n \r\ndata_dict = {-1: np.array([[1, 7], [2, 8], [3, 8]]), 1: np.array([[5, 1], [6, -1], [7, 3]])}\r\nsvm = support_vector_machine()\r\nsvm.fit(data = data_dict)\r\nsvm.visualise()","sub_path":"svm_from_scratch.py","file_name":"svm_from_scratch.py","file_ext":"py","file_size_in_byte":4081,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"244263475","text":"import numpy as np\n\nclass Base:\n\n gen = None\n\n def __init__(self, generator):\n print(\"Importing base...\")\n self.gen = generator\n self.gen.functionList.append(\"addall [r] [g] [b] [a]\")\n self.gen.functionList.append(\"setall [r] [g] [b] [a]\")\n self.gen.functionList.append(\"set [x] [y] [r] [g] [b] [a]\")\n self.gen.commands[\"addall\"] = self.addall\n self.gen.commands[\"setall\"] = self.setall\n self.gen.commands[\"set\"] = self.set\n \n def addall(self, r, g, b, a):\n colorVector = np.fromstring(r + \" \" + g + \" \" + b + \" \" + a, dtype=int, sep=' ')\n \n print(\"adding \" + str(colorVector) + \" to all pixels...\") # DEBUG\n\n self.gen.imgArray = self.gen.imgArray + colorVector\n self.gen.imgArray = np.clip(self.gen.imgArray, 0, 255)\n\n def setall(self, r, g, b, a):\n colorVector = np.fromstring(r + \" \" + g + \" \" + b + \" \" + a, dtype=int, sep=' ')\n \n print(\"setting all pixels to \" + str(colorVector) + \"...\") # DEBUG\n \n for y in range(0, self.gen.imgHeight):\n for x in range(0, self.gen.imgWidth):\n self.gen.imgArray[y][x] = colorVector\n \n def set(self, x, y, r, g, b, a, verbose=True):\n colorVector = np.fromstring(r + \" \" + g + \" \" + b + \" \" + a, dtype=int, sep=' ')\n print(\"setting pixel at (\" + str(x) + \",\" + str(y) + \") to \" + str(colorVector) + \"...\") # DEBUG\n\n self.gen.imgArray[int(y)][int(x)] = colorVector\n","sub_path":"py/base.py","file_name":"base.py","file_ext":"py","file_size_in_byte":1496,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"440543777","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Oct 26 13:14:30 2014\n\n@author: Scott Will, SUNY Buffalo, Department of Electrical Engineering\n\"\"\"\n\nimport cv2\nimport numpy as np\n\ncv2.destroyAllWindows()\n\n# Read in the target image\nfilename = '../../data/m7/IMG_0290.JPG'\nimg = cv2.imread(filename)\n\n# Convert BGR to HSV\nhsv = cv2.cvtColor(img, cv2.COLOR_BGR2HSV)\n\n# define range of red color in HSV\nlower_red = np.array([0, 0, 0])\nupper_red = np.array([5, 255, 255])\n\n# Threshold the HSV image to get only red colors\nmask = cv2.inRange(hsv, lower_red, upper_red)\n\n# Bitwise-AND mask and original image\nres = cv2.bitwise_and(img, img, mask=mask)\n\n# Convert to grayscale\ngray = cv2.cvtColor(res, cv2.COLOR_BGR2GRAY)\ngray = np.float32(gray)\n\n# Run Harris corner detection\n# Arguments: (image, blocksize, sobel aperture size, free parameter)\ndst = cv2.cornerHarris(gray, 2, 3, 0.04)\n\n# Result is dilated for marking the corners, not important\ndst = cv2.dilate(dst, None)\n\n# Threshold for an optimal value, it may vary depending on the image.\nimg[dst > 0.01*dst.max()] = [0, 0, 255]\n\ncv2.imshow('dst', img)\ncv2.waitKey()\n","sub_path":"src/legacy/harris.py","file_name":"harris.py","file_ext":"py","file_size_in_byte":1110,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"172706750","text":"import xlrd\nimport datetime\n\nprovinces = {'黑龙江','青海','陕西','重庆','辽宁','贵州','西藏','福建','甘肃','湖南','湖北','海南','浙江','河南','河北','江西','江苏','新疆','广西','广东','山西','山东','安徽','宁夏','天津','四川','吉林','北京','内蒙古','云南','上海','31省'}\nprovince_ids = {'451','971','290','230','240','851','891','591','931','731','270','898','571','371','311','791','250','991','771','200','351','531','551','951','220','280','431','100','471','871','210'}\n\ndef hasProvince(channels):\n for province in provinces:\n if channels.find(province) > -1:\n return True\n\nchannel_delete = \"DELETE FROM PRODUCT.PM_PRODUCT_CHANNEL_CFG WHERE PRODUCT_ID = {0} AND CHANNEL_ID = {1};\"\nchannel_insert = \"INSERT INTO PRODUCT.PM_PRODUCT_CHANNEL_CFG (PRODUCT_ID, CHANNEL_ID, INURE_TIME, EXPIRE_TIME, OPR_CODE, EFFT_TYPE) VALUES ({0}, {1}, SYSDATE, TO_DATE('3000-01-01 00:00:00', 'YYYY-MM-DD HH24:MI:SS'), {2}, '{3}');\"\ndef openChannel(product_id,opr_list,out):\n for province_id in province_ids:\n print(channel_delete.format(product_id,province_id),file=out)\n for opr in opr_list:\n for i in range(1,5):\n print(channel_insert.format(product_id,province_id,opr,i),file=out)\n print(\"\",file=out)\n\n\nsync_delete = \"DELETE FROM CUSTOMER.CC_SYNC_ORDER_CONFIG WHERE PRODUCT_ID = '{0}' AND ROUTE_VALUE = '{1}';\"\nsync_insert = \"INSERT INTO CUSTOMER.CC_SYNC_ORDER_CONFIG (CONFIG_ID, PRODUCT_ID, ROUTE_VALUE, STATUS) VALUES (CUSTOMER.SEQ_SYNC_ORDER_CONFIG.NEXTVAL, '{0}', '{1}', '1');\"\ndef openProvince(product_id,out):\n for province_id in province_ids:\n print(sync_delete.format(product_id,province_id),file=out)\n print(sync_insert.format(product_id,province_id),file=out)\n print(\"\",file=out)\n\nopr_map = {\n 5: '7', # 资源勘查\n 6: '8', # 资源预占\n 7: '9', # 预占延期\n 8: '10', # 预占取消\n 9: '1', # 产品开通\n 10: '5', # 资费变更\n 11: '6', # 资源变更\n 12: '23', # 产品续订\n 13: '3', # 业务暂停\n 14: '4', # 业务恢复\n 15: '2', # 业务注销\n 16: '25', # 成员管理\n 17: '44', # 密码重置\n 18: '11', # 系统暂停\n 19: '80', # 产品审批\n 20: '12' # 系统恢复\n}\n\ndef getAllProducts(sheet):\n product_list = set()\n for rowx in range(1,sheet.nrows):\n product_id = sheet.cell_value(rowx,2)\n channels = sheet.cell_value(rowx,3)\n if product_id != '':\n if hasProvince(channels):\n product_list.add(product_id)\n return product_list\n\ndef getAllOprations(product_list,sheet):\n channels = { x : [] for x in product_list }\n for rowx in range(1,sheet.nrows):\n pid = sheet.cell_value(rowx,1)\n if pid in channels:\n oprs = channels[pid]\n opr = sheet.cell_value(rowx,2)\n effType = sheet.cell_value(rowx,4)\n oprs.append([opr,effType])\n return channels\n\n\ndef openChannel1(opr_map,out):\n for p_id in opr_map:\n for province_id in province_ids:\n print(channel_delete.format(p_id,province_id),file=out)\n for oprs in opr_map[p_id]:\n print(channel_insert.format(p_id,province_id,oprs[0],oprs[1]),file=out)\n print(\"\",file=out)\n\n\ntimestamp = datetime.datetime.now().strftime('%Y-%m-%d_%H-%M-%S')\nprint_file1 = r'D:/Workspaces/channel_{}.sql'.format(timestamp)\nprint_file2 = r'D:/Workspaces/province_{}.sql'.format(timestamp)\n\ndef printSQL(file1,file2):\n ''' print sql to file''' \n excel = xlrd.open_workbook(r'D:\\Workspaces\\SVN\\EBOSS\\trunk\\01_需求分析\\02-产品需求\\政企EBOSS产品树.xlsx')\n sheet = excel.sheet_by_name(u'产品受理操作')\n with open(file1, encoding='utf8', mode='a') as a_file,open(file2,encoding='utf8',mode='a') as b_file:\n for rowx in range(1,sheet.nrows):\n product_id = sheet.cell_value(rowx,2)\n channels = sheet.cell_value(rowx,3)\n oprs = [ opr_map[x] for x in range(5,21) if sheet.cell_value(rowx,x) == 'Y' ]\n if product_id != '':\n if hasProvince(channels):\n openChannel(product_id,oprs,a_file)\n openProvince(product_id,b_file)\n print(\"\",file=a_file)\n print(\"\",file=b_file)\n\ndef printChannelSQL(file):\n excel = xlrd.open_workbook(r'D:\\Workspaces\\SVN\\EBOSS\\trunk\\01_需求分析\\02-产品需��\\政企EBOSS产品树.xlsx')\n sheet1 = excel.sheet_by_name(u'产品受理操作')\n product_list = getAllProducts(sheet1)\n sheet2 = excel.sheet_by_name(u'产品订购及账期生效规则')\n opr_map = getAllOprations(product_list,sheet2)\n with open(file, encoding='utf8', mode='a') as a_file:\n openChannel1(opr_map,a_file)\n\nprintChannelSQL(print_file1)","sub_path":"python/product/AllProvince.py","file_name":"AllProvince.py","file_ext":"py","file_size_in_byte":4881,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"118585212","text":"import webbrowser\nimport json\nimport requests\nimport requests_oauthlib\nfrom functools import reduce\nimport csv\nfrom eb_data import CLIENT_ID, CLIENT_SECRET, personal_token\nimport sys\nfrom datetime import datetime\n\nDATETIME_FORMAT = \"%Y-%m-%d %H:%M:%S.%f\"\nAUTHORIZATION_URL = 'https://www.eventbrite.com/oauth/authorize'\nTOKEN_URL = 'https://www.eventbrite.com/oauth/token'\nREDIRECT_URI = 'https://www.programsinformationpeople.org/runestone/oauth'\n\nHARVEY_CACHE_FNAME = \"harvey_cache_contents.json\"\nCONCERT_CACHE_FNAME = \"concert_cache_contents.json\"\n\n#--------------------------------------------------\n# Load cache files: data and credentials\n#--------------------------------------------------\n# Load data cache\ndef check_if_cached(fname):\n try:\n with open(fname, 'r') as cache_file:\n cache_json = cache_file.read()\n CACHE_DICTION = json.loads(cache_json)\n except:\n CACHE_DICTION = {}\n return CACHE_DICTION\n\ndef has_cache_expired(timestamp_str, expire_in_days):\n \"\"\"Check if cache timestamp is over expire_in_days old\"\"\"\n # gives current datetime\n now = datetime.now()\n\n # datetime.strptime converts a formatted string into datetime object\n cache_timestamp = datetime.strptime(timestamp_str, DATETIME_FORMAT)\n\n # subtracting two datetime objects gives you a timedelta object\n delta = now - cache_timestamp\n delta_in_days = delta.days\n\n # now that we have days as integers, we can just use comparison\n # and decide if cache has expired or not\n if delta_in_days > expire_in_days:\n return True # It's been longer than expiry time\n else:\n return False\n\n# This is just for testing\nHARVEY_CACHE_DICTION = check_if_cached(HARVEY_CACHE_FNAME)\nCONCERT_CACHE_DICTION = check_if_cached(CONCERT_CACHE_FNAME)\n\ndef get_saved_token():\n with open('token.json', 'r') as f:\n token_json = f.read()\n token_dict = json.loads(token_json)\n\n return token_dict\n\n\ndef save_token(token_dict, expire_in_days):\n token_dict['timestamp'] = datetime.now().strftime(DATETIME_FORMAT)\n token_dict['expire_in_days'] = expire_in_days\n with open('token.json', 'w') as f:\n token_json = json.dumps(token_dict)\n f.write(token_json)\n\n\ndef get_eventbrite_cache(search_params, CACHE_FNAME, expire_in_days=7, force_download=False):\n CACHE_DICTION = check_if_cached(CACHE_FNAME)\n token_expired = False\n # if we need to get an oauth2 session started\n if CACHE_DICTION == {} or force_download:\n # see if we have the token\n try:\n token = get_saved_token()\n except FileNotFoundError:\n token = None\n\n if token:\n if not has_cache_expired(token['timestamp'], token['expire_in_days']):\n print('Token already saved and not expired')\n oauth2inst = requests_oauthlib.OAuth2Session(CLIENT_ID, token=token)\n else:\n print('token has expired, will need to get a new one')\n token_expired=True\n\n if token is None or token_expired:\n print('Getting token the long way')\n oauth2inst = requests_oauthlib.OAuth2Session(CLIENT_ID, redirect_uri=REDIRECT_URI) # Create an instance of an OAuth2Session\n\n # get the authorization url to send the user to\n authorization_url, state = oauth2inst.authorization_url(AUTHORIZATION_URL)\n\n # Opening auth URL for you to sign in to the EventBrite service\n webbrowser.open(authorization_url) \n authorization_response = input('Authenticate and then enter the full callback URL: ').strip() # Need to get the full URL in order to parse the response\n\n # The OAuth2Session instance has a method that extracts what we need from the url, and helps do some other back and forth with EB\n token = oauth2inst.fetch_token(TOKEN_URL, authorization_response=authorization_response, client_secret=CLIENT_SECRET)\n save_token(token, expire_in_days=expire_in_days)\n \n\n\n print('Token saved. Getting search results')\n r = oauth2inst.get('https://www.eventbriteapi.com/v3/events/search/', params=search_params)\n\n # the result is now a dictionary\n response_diction = json.loads(r.text)\n with open(CACHE_FNAME, 'w') as cache_file:\n print('caching result as:', CACHE_FNAME)\n for event in response_diction['events']:\n CACHE_DICTION[event['id']] = event\n cache_json = json.dumps(CACHE_DICTION, indent=2)\n cache_file.write(cache_json)\n else:\n print(\"{} already saved as cache, will return it\".format(CACHE_FNAME))\n \n return CACHE_DICTION\n\n\n\n\nclass Event(object):\n def __init__(self, event_dict):\n self.event = event_dict\n self.id = self.event.get('id')\n self.name = self.event.get('name', {}).get('text')\n\n self.capacity = self.event.get('capacity')\n self.url = self.event.get('url')\n self.is_free = self.event.get('is_free')\n self.description = self.event.get('description', {}).get('text')\n\n self.get_data()\n\n def get_data(self, key_list=[('id',), ('name', 'text'), ('capacity',),\n ('url',), ('is_free',),\n ('description', 'text')]):\n self.data = {}\n for key in key_list:\n try:\n self.data[','.join(key)] = reduce(dict.get, key, self.event)\n except:\n self.data[','.join(key)] = None\n\n def __str__(self):\n return \"{0}: {1}\".format(self.id, self.name)\n\n\ndef write_to_csv(event_list, filename):\n with open(filename, 'w') as outfile:\n outwriter = csv.writer(outfile, delimiter=',')\n keys = event_list[0].data.keys()\n header = list(keys)\n outwriter.writerow(header)\n\n for event in event_list:\n row = list(event.data.values())\n outwriter.writerow(row)\n\n\nif __name__ == '__main__':\n try:\n force_download = sys.argv[1].lower() == 'true'\n except:\n force_download = False\n\n harvey_search_params = {'q':'Hurricane Harvey',\n \"location.address\":'6100 Main St, Houston, TX 77005',\n 'location.within':'30mi'}\n concert_search_params = {'q':'concert',\n 'location.address': \"500 S State St, Ann Arbor, MI 48109\",\n 'location.within':'20mi'}\n\n harvey_response = get_eventbrite_cache(harvey_search_params, \n HARVEY_CACHE_FNAME,\n force_download=force_download)\n\n concert_response = get_eventbrite_cache(concert_search_params,\n CONCERT_CACHE_FNAME,\n force_download=force_download)\n\n harvey_event_list = [Event(event_dict) for event_dict in harvey_response.values()]\n concert_event_list = [Event(event_dict) for event_dict in concert_response.values()]\n print('writing to csv')\n write_to_csv(harvey_event_list, 'harvey.csv')\n write_to_csv(concert_event_list, 'um_concert.csv')\n","sub_path":"SI507project5_code.py","file_name":"SI507project5_code.py","file_ext":"py","file_size_in_byte":7205,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"209277738","text":"# -*- coding: utf-8 -*-\n# !/usr/bin/env python\nimport gi\ngi.require_version('Gst', '1.0')\nfrom gi.repository import GLib, Gst\nfrom json import loads\n\n\nclass BaseStreamer(object):\n def __init__(self, desc=None):\n if not Gst.is_initialized():\n if not Gst.init_check(None):\n raise BaseException\n self.config_name = ''\n self.playing = False\n self.pool = dict()\n self.pipeline = Gst.Pipeline()\n self.bus = self.pipeline.get_bus()\n self.bus.add_signal_watch()\n self.bus.connect('message::eos', self.on_eos)\n self.bus.connect('message::error', self.on_error)\n if desc:\n self._create_pipeline(desc)\n\n def _create_pipeline(self, desc):\n if isinstance(desc, str):\n desc = loads(desc)\n\n for (name, props) in desc.iteritems():\n etype = props.pop('type', 'audioresample')\n link_to = props.pop('link', None)\n caps = props.pop('caps', None)\n self.pool[name] = Gst.ElementFactory.make(etype, None)\n self.pipeline.add(self.pool[name])\n if caps:\n self.pool[name].set_property('caps', Gst.caps_from_string(caps))\n self.pool[name].link_to = link_to\n for (prop, val) in props.iteritems(): # set all the properties\n self.pool[name].set_property(prop, val)\n\n for elem in self.pool.itervalues():\n if elem.link_to:\n elem.link(self.pool[elem.link_to])\n\n def _destroy_pipeline(self):\n self.stop()\n for name in self.pool:\n with self.pool.pop(name) as elem:\n elem.unlink()\n self.pipeline.remove(elem)\n\n def set_config(self, desc):\n self._destroy_pipeline()\n self._create_pipeline(desc)\n\n def play(self):\n self.pipeline.set_state(Gst.State.NULL)\n self.pipeline.set_state(Gst.State.PLAYING)\n self.playing = True\n\n def stop(self):\n self.playing = False\n return self.pipeline.set_state(Gst.State.NULL)\n\n def state(self):\n return str(self.pipeline.get_state(0)[1]).split()[1]\n\n def on_eos(self, bus, msg):\n print('on_eos()')\n\n def on_error(self, bus, msg):\n print('on_error():', msg.parse_error())\n\n\n\n\nfrom gi.repository import GObject, Gst, Gtk\n# Needed for window.get_xid(), xvimagesink.set_window_handle(), respectively:\nfrom gi.repository import GdkX11, GstVideo\n\n# GObject.threads_init()\n# Gst.init(None)\n\nclass VideoPlayer(BaseStreamer):\n def __init__(self):\n super\n self.window = Gtk.Window()\n self.window.connect('destroy', self.stop)\n self.window.set_default_size(640, 480)\n\n self.drawingarea = Gtk.DrawingArea()\n self.window.add(self.drawingarea)\n\n # This is needed to make the video output in our DrawingArea:\n self.bus.enable_sync_message_emission()\n self.bus.connect('sync-message::element', self.on_sync_message)\n\n self.playbin = Gst.ElementFactory.make('playbin', None)\n self.pipeline.add(self.playbin)\n\n def show(self):\n self.window.show_all()\n # You need XID after window.show_all(). Don't get it in on_sync_message() because threading causes segfaults.\n self.xid = self.drawingarea.get_property('window').get_xid()\n self.window.fullscreen()\n\n def play(self, ):\n self.pipeline.set_state(Gst.State.NULL)\n self.pipeline.set_state(Gst.State.PLAYING)\n self.playing = True\n\n def stop(self, window=None):\n self.pipeline.set_state(Gst.State.NULL)\n self.playing = False\n #Gtk.main_quit()\n\n def set_config(self, cfg=''):\n self.playbin.set_property('uri', 'file://' + cfg)\n\n def state(self):\n return str(self.pipeline.get_state(0)[1]).split()[1]\n\n def on_sync_message(self, bus, msg):\n if msg.get_structure().get_name() == 'prepare-window-handle':\n print('prepare-window-handle')\n msg.src.set_window_handle(self.xid)\n\n def on_eos(self, bus, msg):\n print('on_eos(): seeking to start of video')\n self.pipeline.seek_simple(Gst.Format.TIME, Gst.SeekFlags.FLUSH | Gst.SeekFlags.KEY_UNIT, 0)\n\n def on_error(self, bus, msg):\n print('on_error():', msg.parse_error())\n\n\n\n\nif __name__ == \"__main__\":\n description = '{\"src\": {\"multicast-iface\": \"eth0\", \"auto-multicast\": true, \"caps\": \"application/x-rtp,media=(string)audio,clock-rate=(int)48000,encoding-name=(string)X-GST-OPUS-DRAFT-SPITTKA-00\", \"type\": \"udpsrc\", \"port\": 3333, \"multicast-group\": \"224.1.1.1\"}, \"sink\": {\"device\": \"hw:0,0\", \"link\": \"dec\", \"type\": \"alsasink\"}, \"dec\": {\"link\": \"rtp\", \"type\": \"opusdec\", \"use-inband-fec\": false}, \"rtp\": {\"link\": \"jtr\", \"type\": \"rtpopusdepay\"}, \"jtr\": {\"link\": \"src\", \"latency\": 200, \"type\": \"rtpjitterbuffer\", \"do-retransmission\": true}}'\n sink = BaseStreamer(description)\n\n","sub_path":"helpers/streamer.py","file_name":"streamer.py","file_ext":"py","file_size_in_byte":4916,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"534157185","text":"#!/usr/bin/env python\n#-*- coding: utf-8 -*-\nimport base64\nimport urllib2\nimport inkex\n\n\nclass UnsplashPlaceholder(inkex.Effect):\n def __init__(self):\n inkex.Effect.__init__(self)\n self.OptionParser.add_option('-W', '--width', action='store', type='string', dest='width',\n default='800', help='Set image width')\n self.OptionParser.add_option('-H', '--height', action='store', type='string', dest='height',\n default='680', help='Set image height')\n self.OptionParser.add_option('-C', '--category', action='store', type='string', dest='category',\n default='', help='Set image category')\n\n def effect(self):\n image = self._get_image()\n node = self._create_image_node(image)\n\n self.document.getroot().append(node)\n\n def _get_image(self):\n url = 'https://lorempixel.com/{width}/{height}/{category}'.format(\n width=self.options.width,\n height=self.options.height,\n category=self.options.category\n )\n response = urllib2.urlopen(url)\n data = response.read()\n return data\n\n def _create_image_node(self, data):\n attribs = {\n 'height': self.options.height,\n 'width': self.options.width,\n 'category': self.options.category,\n 'x': '0',\n 'y': '0',\n 'preserveAspectRatio': 'None',\n inkex.addNS('href', 'xlink'): u'data:image/jpeg;base64,' + base64.encodestring(data)\n }\n node = inkex.etree.Element(inkex.addNS('image','svg'), attribs)\n return node\n\n\nif __name__ == '__main__':\n placeholder = UnsplashPlaceholder()\n placeholder.affect()\n","sub_path":"lorempixel.py","file_name":"lorempixel.py","file_ext":"py","file_size_in_byte":1769,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"130929434","text":"# -*- coding: utf-8 -*-\r\nimport struct\r\nfrom PIL import Image\r\nimport tensorflow as tf\r\nimport numpy as np\r\nimport datetime\r\nimport os\r\nfrom inference_images import inference_images\r\nfrom inference_labels import inference_labels\r\n\r\n#train_epoch:2 accuracy:0.4951\r\nif __name__ == \"__main__\":\r\n\tprint(\"Begin inference!\")\r\n\t#base_path = \"/home/mnist_dataset\"\r\n\tbase_path = os.getcwd()\r\n\tbase_inference_path = os.path.join(base_path,\"test_data\")\r\n\tinference_image_path = os.path.join(base_inference_path,\"t10k-images-idx3-ubyte\")\r\n\tinference_label_path = os.path.join(base_inference_path,\"t10k-labels-idx1-ubyte\")\r\n\tinference_labels = inference_labels(inference_label_path)\r\n\tinference_images = inference_images(inference_image_path)\r\n\tinput_image_size = int(inference_images.get_row_number())*int(inference_images.get_column_number())\r\n\tright_count = 0\r\n\tbatchsize = 1\r\n\twith tf.Session() as sess:\r\n\t\tsaver = tf.train.import_meta_graph(os.path.join(base_path,\"train_data/checkPoint/trainModel.meta\"))\r\n\t\tsaver.restore(sess, tf.train.latest_checkpoint(os.path.join(base_path,\"train_data/checkPoint\")))\r\n\r\n\t\titerations = inference_images.get_images_number()/batchsize\r\n\t\tfor step in range(iterations):\r\n\t\t\tlabel_vals = inference_labels.read_labels(batchsize)\r\n\t\t\tinference_image_pixs = inference_images.read_images(batchsize)\r\n\t\t\tinference_y_label = []\r\n\t\t\tfor item in label_vals:\r\n\t\t\t\tinference_sub_y_label = []\r\n\t\t\t\tfor i in range(10):\r\n\t\t\t\t\tif item != i:\r\n\t\t\t\t\t\tinference_sub_y_label.append(0)\r\n\t\t\t\t\telse:\r\n\t\t\t\t\t\tinference_sub_y_label.append(1)\r\n\t\t\t\t\tinference_y_label.append(inference_sub_y_label)\r\n\t\t\tinference_x = np.array(inference_image_pixs,dtype=np.float32)\r\n\t\t\tinference_y = np.array(inference_y_label,dtype=np.float32)\r\n\t\t\t# 获取需要进行计算的operator\r\n\t\t\tYs = sess.graph.get_tensor_by_name('Ys:0')\r\n\t\t\tX = sess.graph.get_tensor_by_name('X:0')\r\n\t\t\tY_ = sess.graph.get_tensor_by_name('Y_:0')\r\n\t\t\tresults = sess.run(Ys,feed_dict={X:inference_x, Y_:inference_y})\r\n\t\t\tfor image_number in range(batchsize):\r\n\t\t\t\tmaxindex = np.argmax(results[image_number])\r\n\t\t\t\ttrue_label = np.argmax(inference_y[image_number])\r\n\t\t\t\tif maxindex == true_label:\r\n\t\t\t\t\tright_count = right_count + 1\r\n\r\n\t\tprint(\"right_count is:{}\".format(right_count))\r\n\t\tprint(\"total dataset is:{}\".format(inference_images.get_images_number()))\r\n\t\tprint(\"accuracy is:{}\".format(float(right_count)/inference_images.get_images_number()))\r\n\r\n\t\t# maxindex = np.argmax(sess.run(op,feed_dict={X:inference_x, Y_:inference_y}))\r\n\t\t# print maxindex \r\n\t\t# print np.argmax(inference_y[0])","sub_path":"mnist/mnist_FC/mnist_inference.py","file_name":"mnist_inference.py","file_ext":"py","file_size_in_byte":2554,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"408030434","text":"#! /usr/bin/python\n\nimport httplib,urllib\n\nparams = urllib.urlencode({\"question_2755\" : \"10493\"})\n\nheaders = {\"Origin\":\"http://cgi.mmog.163.com:8088\",\n \"Accept-Encoding\":\"gzip,deflate,sdch\",\n \"Accept-Language\":\"en-US,en;q=0.8,zh-CN;q=0.6,zh;q=0.4\",\n \"User-Agent\":\"Mozilla/5.0 (X11; Linux i686) AppleWebKit/537.36 (KHTML, like Gecko) Ubuntu Chromium/34.0.1847.116 Chrome/34.0.1847.116 Safari/537.36\",\n \"Content-Type\":\"application/x-www-form-urlencoded\",\n \"Accept\":\"application/json, text/javascript, */*\",\n \"Referer\":\"http://cgi.mmog.163.com:8088/v4a/show_vote/1106/?6\",\n \"X-Requested-With\":\"XMLHttpRequest\",\n \"Connection\":\"keep-alive\"}\n\n\nconn = httplib.HTTPConnection(\"cgi.mmog.163.com:8088\")\n\nconn.request(\"POST\", \"/v4a/show_vote/1106/\", params, headers)\n\nresponse = conn.getresponse()\n\nprint(response.status, response.reason)\n\ndata = response.read()\nfh = open('rece.html', 'w')\nfh.write(data)\nfh.close()\n\nconn.close()\n","sub_path":"http/post.py","file_name":"post.py","file_ext":"py","file_size_in_byte":944,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"202565482","text":"def count(arr,x):\r\n\tcount = 0\r\n\tfor y in arr:\r\n\t\tif x==y:\r\n\t\t\tcount+=1\r\n\treturn count\r\n\r\ndef greatest(arr,strings):\r\n\tmaxVal = -1\r\n\tmaxIndex = -1\r\n\tfor i in range(len(arr)):\r\n\t\tif arr[i]>maxVal:\r\n\t\t\tmaxVal = arr[i]\r\n\t\t\tmaxIndex = i\r\n\t\telif(arr[i]==maxVal):\r\n\t\t\tif(len(strings[i]) 0:\n # 有命中,进一步判断 confidence 是否达到要求\n confidence = respond[0][\"confidence\"]\n if confidence >= self.threshold:\n # 命中该问题,返回回答\n answer = respond[0][\"answer\"]\n if utils.validjson(answer):\n answer = random.choice(json.loads(answer))\n logger.info(\"{} 回答:{}\".format(self.SLUG, answer))\n return answer\n # 没有命中,走兜底\n if self.secondary != \"null\" and self.secondary is not None:\n try:\n ai = get_robot_by_slug(self.secondary)\n return ai.chat(texts, parsed)\n except Exception:\n logger.critical(\n \"Secondary robot {} failed to response for {}\".format(\n self.secondary, msg\n )\n )\n return get_unknown_response()\n else:\n return get_unknown_response()\n except Exception:\n logger.critical(\"AnyQ robot failed to response for %r\", msg, exc_info=True)\n return \"抱歉, 我的大脑短路了,请稍后再试试.\"\n\n\nclass OPENAIRobot(AbstractRobot):\n\n SLUG = \"openai\"\n\n def __init__(self, openai_api_key,model, temperature, max_tokens,top_p,frequency_penalty,presence_penalty,stop_ai):\n \"\"\"\n OpenAI机器人\n openai.api_key = os.getenv(\"OPENAI_API_KEY\")\n \"\"\"\n super(self.__class__, self).__init__()\n self.openai_api_key = openai_api_key\n openai.api_key=self.openai_api_key\n logger.info(self.openai_api_key)\n self.model = model\n self.temperature = temperature\n self.max_tokens = max_tokens\n self.top_p = top_p\n self.frequency_penalty = frequency_penalty\n self.presence_penalty = presence_penalty\n self.stop_ai = stop_ai\n\n @classmethod\n def get_config(cls):\n # Try to get anyq config from config\n return config.get(\"openai\", {})\n\n def chat(self, texts, parsed):\n \"\"\"\n 使用OpenAI机器人聊天\n\n Arguments:\n texts -- user input, typically speech, to be parsed by a module\n \"\"\"\n msg = \"\".join(texts)\n msg = utils.stripPunctuation(msg)\n try:\n response = openai.Completion.create(\n model=self.model,\n prompt=msg,\n temperature=self.temperature,\n max_tokens=self.max_tokens,\n top_p=self.top_p,\n frequency_penalty=self.frequency_penalty,\n presence_penalty=self.presence_penalty,\n stop=self.stop_ai\n )\n logger.debug(response)\n logger.debug(response.choices[0].text)\n respond=response.choices[0].text\n logger.info(\"openai response: {}\".format(respond))\n return respond\n \n except Exception:\n logger.critical(\"openai robot failed to response for %r\", msg, exc_info=True)\n return \"抱歉, 我的大脑短路了,请稍后再试试.\"\n\n\ndef get_unknown_response():\n \"\"\"\n 不知道怎么回答的情况下的答复\n\n :returns: 表示不知道的答复\n \"\"\"\n results = [\"抱歉,我不会这个呢\", \"我不会这个呢\", \"我还不会这个呢\", \"我还没学会这个呢\", \"对不起,你说的这���,我还不会\"]\n return random.choice(results)\n\n\ndef get_robot_by_slug(slug):\n \"\"\"\n Returns:\n A robot implementation available on the current platform\n \"\"\"\n if not slug or type(slug) is not str:\n raise TypeError(\"Invalid slug '%s'\", slug)\n\n selected_robots = list(\n filter(\n lambda robot: hasattr(robot, \"SLUG\") and robot.SLUG == slug, get_robots()\n )\n )\n if len(selected_robots) == 0:\n raise ValueError(\"No robot found for slug '%s'\" % slug)\n else:\n if len(selected_robots) > 1:\n logger.warning(\n \"WARNING: Multiple robots found for slug '%s'. \"\n + \"This is most certainly a bug.\" % slug\n )\n robot = selected_robots[0]\n logger.info(\"使用 {} 对话机器人\".format(robot.SLUG))\n return robot.get_instance()\n\n\ndef get_robots():\n def get_subclasses(cls):\n subclasses = set()\n for subclass in cls.__subclasses__():\n subclasses.add(subclass)\n subclasses.update(get_subclasses(subclass))\n return subclasses\n\n return [\n robot\n for robot in list(get_subclasses(AbstractRobot))\n if hasattr(robot, \"SLUG\") and robot.SLUG\n ]\n","sub_path":"robot/AI.py","file_name":"AI.py","file_ext":"py","file_size_in_byte":9199,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"445914381","text":"from os import listdir, makedirs\nfrom os.path import isfile, join\nfrom pathlib import Path\nfrom pprint import pprint\nfrom time import sleep\n\nfrom requests import post\n\nfrom utils import to_color\n\n\nclass Team():\n def __init__(self, token, team_id):\n self.token = token\n self.id = None\n self.cookies = None\n self.owner = None\n self.compos = None\n self.path = 'https://leekwars.com/api/'\n self.team_getPrivate(team_id)\n\n def team_getPrivate(self, team_id):\n '''\n compos = r.json()[\"team\"][\"compositions\"] \n team_fights = compo[\"leeks\"][0][\"team_fights\"]\n '''\n res = post(self.path + 'team/get-private/',\n headers={'Authorization': \"Bearer \"+self.token},\n data={\"team_id\": team_id}).json()\n for leek in res[\"members\"]:\n if leek[\"grade\"] == 'owner':\n self.owner = leek[\"name\"] # or leek[\"id\"]\n break\n self.compos = res[\"compositions\"]\n\n def team_registerTournaments(self):\n for compo in self.compos:\n res = self.__team_registerTournament(compo[\"id\"])\n if \"error\" in res:\n print(\"Can't register ({}): {}\".format(compo[\"name\"], res[\"error\"]))\n else:\n print(\"Successfully registered ({}).\".format(compo[\"name\"]))\n\n def __team_registerTournament(self, composition_id):\n return post(self.path + 'team/register-tournament/',\n headers={'Authorization': \"Bearer \"+self.token},\n data={\"composition_id\": composition_id}).json()\n\n def __garden_getCompositionOpponents(self, composition):\n r = post(self.path + 'garden/get-composition-opponents/',\n headers={'Authorization': \"Bearer \"+self.token},\n data={\"composition\": composition})\n self.cookies = r.cookies\n return r.json()\n \n def __garden_startTeamFight(self, composition_id, target_id):\n return post(self.path + 'garden/start-team-fight/',\n headers={'Authorization': \"Bearer \"+self.token},\n data={\"composition_id\": composition_id, \"target_id\": target_id},\n cookies=self.cookies).json()\n\n def __fight_get(self, fight_id):\n return post(self.path + 'fight/get/',\n headers={'Authorization': \"Bearer \"+self.token},\n data={\"fight_id\": fight_id},\n cookies=self.cookies).json()\n \n def wait_fight_result(self, fight_id):\n \"\"\"Wait the result of a fight.\n\n Args:\n fight_id (int): The id of the fight to wait\n \"\"\"\n nbr_wait = 0\n while True:\n res = self.__fight_get(fight_id)\n fight = res['fight']\n winner = fight['winner']\n if winner == -1: # Fight isn't resolved yet\n print(\"Waiting (\"+ str(1 + nbr_wait * 2) + \"s) .\" + \".\" * nbr_wait, end=\"\\r\", flush=True)\n nbr_wait += 1\n sleep(2)\n continue\n elif winner >= 0:\n WARNING = '\\033[93m'\n OKGREEN = '\\033[92m'\n FAIL = '\\033[91m'\n ENDC = '\\033[0m'\n win = \"WTF?\"\n if winner == 0:\n win = to_color(\"DRAW\", None, True)\n #win = WARNING + \"DRAW\" + ENDC\n elif winner == 1:\n win = to_color(\"WIN \", True, False)\n #win = OKGREEN + \"WIN \" + ENDC\n elif winner == 2:\n win = to_color(\"LOSE\", False, False)\n #win = FAIL + \"LOSE\" + ENDC\n\n team1_name = fight['report']['team1']['name']\n team1_talent = fight['report']['team1']['talent'] + fight['report']['team1']['talent_gain']\n team2_name = fight['report']['team2']['name']\n team2_talent = fight['report']['team2']['talent'] + fight['report']['team2']['talent_gain']\n\n print(\" \" * (nbr_wait + 15), end=\"\\r\", flush=True)\n print(\"Team {} {} ({}) vs {} ({})\".format(win, team1_name, team1_talent, team2_name, team2_talent), flush=True)\n return\n\n def startTeamFights(self):\n fightIds = []\n for compo in self.compos:\n team_fights = compo[\"leeks\"][0][\"team_fights\"]\n print(\"\")\n print(\"Team [{}], {} leeks, level {}, {} talent\".format(\n compo[\"name\"], len(compo[\"leeks\"]), compo[\"total_level\"], compo[\"talent\"]))\n print(\"team_fights\", team_fights)\n for _ in range(0, team_fights):\n res = self.__garden_getCompositionOpponents(compo[\"id\"])\n if res[\"opponents\"]:\n # TODO: Select best opponent\n #pprint(res[\"opponents\"])\n opponent = res[\"opponents\"][0]\n #print(\"Start team Fight vs '{}', level {}, {} talent\".format(opponent[\"name\"], opponent[\"total_level\"], opponent[\"talent\"]))\n res = self.__garden_startTeamFight(compo[\"id\"], opponent[\"id\"])\n fightIds.append(res['fight'])\n self.wait_fight_result(res['fight'])\n #break\n","sub_path":"team.py","file_name":"team.py","file_ext":"py","file_size_in_byte":5309,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"73692326","text":"import unittest\nfrom Wpp.WppCore import WppCore\nfrom out.OutContextMemoryStream import OutContextMemoryStream\n\nclass TestWppInterface(unittest.TestCase):\n\tdef testExport(self):\n\t\tsource = \"\"\"\ninterface public A\ninterface public B\n\textends A\n\t\t\"\"\"\n\t\tmodule = WppCore.createMemModule(source, 'root.fake')\n\t\toutContext = OutContextMemoryStream()\n\t\tmodule.export(outContext)\n\t\tself.assertEqual(str(outContext), module.strPack(source))\n\n\tdef testInvalidParent(self):\n\t\tsource = \"\"\"\nclass public A\ninterface public B\n\textends A\n\t\t\"\"\"\n\t\twith self.assertRaises(RuntimeError) as cm:\n\t\t\tmodule = WppCore.createMemModule(source, 'root.fake')\n\t\tself.assertEqual(cm.exception.args[0], 'Invalid parent root.A:Class')\n","sub_path":"src1/Wpp/tests/testWppInterface.py","file_name":"testWppInterface.py","file_ext":"py","file_size_in_byte":703,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"170202640","text":"###########\n# Imports #\n###########\n\nfrom flask import Flask, render_template, request, redirect, url_for\nfrom sqlalchemy.orm import sessionmaker\nfrom aa_run_first_database_setup import Base, Restaurant, User\nfrom flask import session as login_session\nimport json\nfrom sqlalchemy import create_engine\n\n###############\n# Setup Flask #\n###############\n\napp = Flask(__name__)\n\nCLIENT_ID = json.loads(\n open('z_client_test_services.json', 'r').read())['web']['client_id']\nAPPLICATION_NAME = \"Restaurant Menu Application\"\n\n# Connect to Database and create database session # small change\nengine = create_engine('sqlite:///restaurantmenuwithusers.db')\nBase.metadata.bind = engine\n\nDBSession = sessionmaker(bind=engine)\nsession = DBSession()\n\n\n##############################\n# Initial Required Functions #\n##############################\n\n# User Helper Functions\ndef create_user(login_session_):\n new_user = User(name=login_session_['username'], email=login_session_['email'], picture=login_session_['picture'])\n session.add(new_user)\n session.commit()\n user = session.query(User).filter_by(email=login_session_['email']).one()\n return user.id\n\n\ndef get_user_info(user_id):\n user = session.query(User).filter_by(id=user_id).one()\n return user\n\n\ndef get_user_id(email):\n try:\n user = session.query(User).filter_by(email=email).one()\n return user.id\n except:\n return None\n\n\n###########\n# Library #\n###########\n\ndef edit_one_restaurant(restaurant_id):\n edit_restaurant = session.query(Restaurant).filter_by(id=restaurant_id).one()\n logged_in_user = login_session['username']\n\n if 'username' not in login_session:\n return redirect('/ohnommy/login')\n\n if edit_restaurant.user_id != login_session['user_id']:\n return (\"\")\n\n if request.method == 'POST':\n if request.form['name']:\n edit_restaurant.name = request.form['name']\n if request.form['cuisines']:\n edit_restaurant.rtype = request.form['cuisines']\n if request.form['style']:\n edit_restaurant.diningStyle = request.form['style']\n if request.form['contact']:\n edit_restaurant.contactNo = request.form['contact']\n if request.form['location']:\n edit_restaurant.location = request.form['location']\n if request.form['website']:\n edit_restaurant.website = request.form['website']\n if request.form['openHour']:\n edit_restaurant.openTime = request.form['openHour']\n if request.form['closedHour']:\n edit_restaurant.closeTime = request.form['closedHour']\n if request.form['bsHour']:\n edit_restaurant.bsHours = request.form['bsHour']\n if request.form['beHour']:\n edit_restaurant.beHours = request.form['beHour']\n if request.form['lsHour']:\n edit_restaurant.lsHours = request.form['lsHour']\n if request.form['leHour']:\n edit_restaurant.leHours = request.form['leHour']\n if request.form['dsHour']:\n edit_restaurant.dsHours = request.form['dsHour']\n if request.form['deHour']:\n edit_restaurant.deHours = request.form['deHour']\n if request.form['aboutrestaurant']:\n edit_restaurant.about = request.form['aboutrestaurant']\n session.commit()\n return redirect(url_for('show_restaurant_page', restaurant_id=restaurant_id))\n else:\n return render_template('edit-restaurant.html', loggedInUser=logged_in_user, restaurant_id=restaurant_id,\n editRestaurant=edit_restaurant)\n\n\ndef delete_one_restaurant(restaurant_id):\n delete_restaurant = session.query(Restaurant).filter_by(id=restaurant_id).one()\n if 'username' in login_session:\n logged_in_user = login_session['username']\n else:\n return redirect('/ohnommy/login')\n\n if delete_restaurant.user_id != login_session['user_id']:\n return (\"\")\n\n if request.method == 'POST':\n session.delete(delete_restaurant)\n session.commit()\n return redirect(url_for('opening_page'))\n else:\n return render_template('delete-restaurant.html', loggedInUser=logged_in_user, restaurant_id=restaurant_id,\n deleteRestaurant=delete_restaurant)\n","sub_path":"Oh_Nommy/directory/g_edit_and_delete_restaurant.py","file_name":"g_edit_and_delete_restaurant.py","file_ext":"py","file_size_in_byte":4687,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"122168164","text":"s=input(\"정수를 입력하세요 : \")\nl=s.split()#자동 리스트 작성\nl=[eval(i) for i in l]#list comprehension\nmax=0\nl.sort()\nfor i in range(0,l[-1]):\n if l.count(i)>=max:\n max=l.count(i)\n\nfor i in range(0,l[-1]):\n if l.count(i)==max:\n print(\"가장 많이 나온 숫자는 : {0}\".format(i))","sub_path":"untitled/14.2.py","file_name":"14.2.py","file_ext":"py","file_size_in_byte":318,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"214746256","text":"import numpy as np\nimport random\nimport math\n\nMAX = 2147483647\n\n# mean-normalization\n# turn to [-1, 1]\n# returns\n# X, y matrix\n# rx, ry ranges\ndef meanNormalize(X, y):\n\trxmx = [-MAX for _ in range(len(X[0]))]\n\trxmn = [MAX for _ in range(len(X[0]))]\n\trymx = [-MAX]\n\trymn = [MAX]\n\tfor _ in X:\n\t\trxmx = [max(a, b) for a, b in zip(rxmx, _)]\n\t\trxmn = [min(a, b) for a, b in zip(rxmn, _)]\n\tfor _ in y:\n\t\trymx = [max(a, b) for a, b in zip(rymx, _)]\n\t\trymn = [min(a, b) for a, b in zip(rymn, _)]\n\n\trx = [[a, b] if b - a > 0.1 else [a - 0.1, b + 0.1] for a, b in zip(rxmn, rxmx)]\n\try = [rymn[0], rymx[0]] if rymx[0] - rymn[0] > 0.1 else [rymn[0] - 0.1, rymx[0] + 0.1]\n\n\tfor i in range(len(X)):\n\t\tX[i] = [(v - m[0]) / (m[1] - m[0]) * 2.0 - 1 for v, m in zip(X[i], rx)]\n\n\tfor i in range(len(y)):\n\t\ty[i] = (y[i][0] - ry[0]) / (ry[1] - ry[0]) * 2.0 - 1\n\treturn np.mat(X), np.mat(y).T, rx, ry\n\ndef gradient(X, y, theta):\n\tdiff = np.dot(X, theta) - y\n\treturn 1. / (X[0].size) * (np.transpose(X) * diff)\n\n# gradient descent\n# return theta\ndef gradientDescent(X, y):\n\ttimes = 5000\n\talpha = 0.2\n\ttheta = np.zeros([X[0].size, 1])\n\tprint(theta)\n\tfor _ in range(times):\n\t\tgrad = gradient(X, y, theta)\n\t\ttheta = theta - alpha * grad\n\treturn theta\n\n\ndef lr(X, y):\n\tm = len(X[0])\n\tX, y, rx, ry = meanNormalize(X, y)\n\ttheta = gradientDescent(X, y)\n\ttmp = np.copy(theta)\n\ttheta[0] = tmp[0] * (ry[1] - ry[0]) / (rx[0][1] - rx[0][0])\n\tfor i in range(1, m):\n\t\ttheta[i] = tmp[i] * (ry[1] - ry[0]) / (rx[i][1] - rx[i][0])\n\t\ttheta[0] = theta[0] - tmp[i] * rx[i][0] * (ry[1] - ry[0]) / (rx[i][1] - rx[i][0]) + ry[0]\n\treturn theta\n\nif __name__ == '__main__':\n print(\"hello world!\")","sub_path":"linerregression/liner.py","file_name":"liner.py","file_ext":"py","file_size_in_byte":1650,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"435296264","text":"import json\nimport boto3\nimport decimal\nfrom botocore.vendored import requests as requests\nfrom uuid import uuid1\n\n\n# Helper class to convert a DynamoDB item to JSON.\nclass DecimalEncoder(json.JSONEncoder):\n def default(self, o):\n if isinstance(o, decimal.Decimal):\n return int(o) if o % 1 == 0 else float(o)\n \n return super(DecimalEncoder, self).default(o)\n\n\ndef lambda_handler(event, context):\n \n task_id = event['pathParameters'].get('task_id')\n sofier = event['queryStringParameters'].get('sofier')\n \n status, data = set_start_execution(task_id, sofier)\n \n return {\n 'statusCode': status,\n 'body': json.dumps(data, cls=DecimalEncoder),\n 'headers': {\n 'Access-Control-Allow-Origin': '*' \n } \n }\n\nURL = 'https://mysofie.com/api/v2/micro_task/execution/{task_id}/start'\n\nHEADERS = {'Authorization': 'Bearer RVVfU09VX0FfTEVOREE='}\n\ndef set_start_execution(task_id: str, sofier: str) -> tuple:\n \"\"\"\n \n [X] - Recupera o nome da tarefa atrelada à tarefa\n [X] - Recupera o fluxo de execução da tarefa\n [X] - Sinaliza à Bússola o status da tarefa\n \"\"\"\n status, buffer = 500, None\n\n #: MARCA NA TABELA DE TAREFAS QUE A TAREFA ESTÁ EM EXECUÇÃO, RECUPERANDO AS VARIABLES\n response_task = boto3.resource('dynamodb').Table('table_micro_task_in_person').get_item(\n Key={'task_id': task_id},\n ProjectionExpression='task.#name, variables',\n ExpressionAttributeNames={'#name': 'name'}\n )\n\n #: RECUPERA O FLUXO DE EXECUÇÃO DA TAREFA\n response_flow = boto3.resource('dynamodb').Table('table_micro_task_flows').get_item(\n Key={'name': response_task['Item']['task']['name'], 'version': 1}\n )\n\n #: SINALIZA AO BACKEND LEGADO DE QUE A TAREFA FOI INICIADA\n response = requests.post(URL.format(task_id=task_id), headers=HEADERS, params={'sofier': sofier})\n if response.status_code != 200:\n return response.status_code, response.json()\n\n #: Formatando a resposta final\n status = 200\n response = {\n 'task_id': task_id,\n 'execution_id': str(uuid1()),\n 'task_flow': response_flow['Item']['task_flow'],\n 'variables': response_task['Item'].get('variables', dict()),\n 'task_info': {\n 'name': response_flow['Item']['name'],\n 'version': response_flow['Item']['version']\n }\n }\n \n return status, response","sub_path":"serverless_aws/LAMBDA FUNCTIONS/micro_task-start/lambda_function.py","file_name":"lambda_function.py","file_ext":"py","file_size_in_byte":2462,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"70291784","text":"from DAO import *\nimport config as WebConfig \nimport xlsxwriter\nimport json\n\nclass Report:\n def DownloadReport(BiceId,FromDate,ToDate,path):\n response=DataAccess.DownloadReport(BiceId,FromDate,ToDate)\n print(path)\n FileName=Report.CeateExcel(response,path)\n return FileName\n def CeateExcel(data,path):\n filename=''\n col=['Candidate_Id', 'Salutation', 'First_Name', 'Middle_Name', 'Last_Name', 'Bice_Name', 'Door_No_Street', 'City_Name', 'State_Name', 'Country_Name', 'Pincode', 'Dob', 'Mobile_Number', 'Max_Edu', 'Edu_Others', 'Id_Proof', 'Id_Number', 'Id_Others', 'Has_Bank', 'Bank_Name', 'Bank_Acc_Number', 'Created_By', 'Created_On', 'Bank_Image', 'Id_Proof_Image', 'Candidate_Image']\n ImagePath=WebConfig.FilePath\n try:\n workbook = xlsxwriter.Workbook(path)\n\n header_format = workbook.add_format({\n 'bold': True,\n #'text_wrap': True,\n 'align': 'top',\n 'valign': 'center',\n 'fg_color': '#D7E4BC',\n 'border': 1})\n\n write_format = workbook.add_format({\n 'border': 1,\n 'align': 'top',\n 'valign': 'top'})\n\n url_format = workbook.add_format({\n 'border': 1,\n 'align': 'top',\n 'valign': 'top',\n 'font_color': 'blue',\n 'underline': 1})\n worksheet = workbook.add_worksheet('Candidate Report')\n for i in range(len(col)):\n worksheet.write(0,i ,col[i], header_format) \n for j in range(len(data)) : \n for k in range(len(col)-3):\n worksheet.write(j+1,k ,data.iloc[j,k],write_format) \n for l in range(k+1,len(col)):\n if data.iloc[j,l]==None or data.iloc[j,l]=='':\n worksheet.write(j+1,l ,'NA',write_format)\n else:\n worksheet.write_url(j+1,l,ImagePath+data.iloc[j,l], url_format, string='Image', tip='Click to open image')\n \n \n workbook.close()\n except Exception as e:\n #filename='Error creating excel '+ str(e)\n print(str(e))\n \n return filename\n\n","sub_path":"Models/Report.py","file_name":"Report.py","file_ext":"py","file_size_in_byte":2330,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"55"}
+{"seq_id":"66952742","text":"import gevent, gevent.server\nfrom telnetsrv.green import TelnetHandler, command\n\nclass MyTelnetHandler(TelnetHandler):\n WELCOME = \"Welcome to my server.\"\n\n @command(['echo', 'copy', 'repeat'])\n def command_echo(self, params):\n '''\n Echo text back to the console.\n\n '''\n self.writeresponse( ' '.join(params) )\n\n @command('timer')\n def command_timer(self, params):\n '''