diff --git "a/5338.jsonl" "b/5338.jsonl"
new file mode 100644--- /dev/null
+++ "b/5338.jsonl"
@@ -0,0 +1,638 @@
+{"seq_id":"35341006","text":"from app.objects.c_obfuscator import Obfuscator\nfrom app.utility.base_world import BaseWorld\nfrom plugins.stockpile.app.stockpile_svc import StockpileService\n\nname = 'Stockpile'\ndescription = 'A stockpile of abilities, adversaries, payloads and planners'\naddress = '/plugin/stockpile/gui'\naccess = BaseWorld.Access.APP\n\n\nasync def enable(services):\n stockpile_svc = StockpileService(services)\n services.get('app_svc').application.router.add_route('GET', '/plugin/stockpile/gui', stockpile_svc.splash)\n await services.get('file_svc').add_special_payload('.donut', 'plugins.stockpile.app.donut.donut_handler')\n await stockpile_svc.data_svc.store(\n Obfuscator(name='plain-text',\n description='Does no obfuscation to any command, instead running it in plain text',\n module='plugins.stockpile.app.obfuscators.plain_text')\n )\n await stockpile_svc.data_svc.store(\n Obfuscator(name='base64',\n description='Obfuscates commands in base64',\n module='plugins.stockpile.app.obfuscators.base64_basic')\n )\n await stockpile_svc.data_svc.store(\n Obfuscator(name='base64jumble',\n description='Obfuscates commands in base64, then adds characters to evade base64 detection. '\n 'Disclaimer: this may cause duplicate links to run.',\n module='plugins.stockpile.app.obfuscators.base64_jumble')\n )\n await stockpile_svc.data_svc.store(\n Obfuscator(name='caesar cipher',\n description='Obfuscates commands through a caesar cipher algorithm, which uses a randomly selected '\n 'shift value.',\n module='plugins.stockpile.app.obfuscators.caesar_cipher')\n )\n await stockpile_svc.data_svc.store(\n Obfuscator(name='base64noPadding',\n description='Obfuscates commands in base64, then removes padding',\n module='plugins.stockpile.app.obfuscators.base64_no_padding')\n )\n await stockpile_svc.data_svc.store(\n Obfuscator(name='steganography',\n description='Obfuscates commands through image-based steganography',\n module='plugins.stockpile.app.obfuscators.steganography')\n )\n","sub_path":"hook.py","file_name":"hook.py","file_ext":"py","file_size_in_byte":2298,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"578033595","text":"from telegram.ext import Updater, MessageHandler, Filters\nfrom telegram.ext import CallbackContext, CommandHandler\nimport time as times\n\n\ndef time(update, context):\n update.message.reply_text(times.asctime().split()[3])\n\n\ndef date(update, context):\n ans = ' '.join(times.asctime().split()[:3]) + \" \" + times.asctime().split()[-1]\n update.message.reply_text(ans)\n\n\ndef task(context):\n job = context.job\n context.bot.send_message(job.context, text='Вернулся')\n\n\ndef set_timer(update, context):\n chat_id = update.message.chat_id\n try:\n due = (context.args[0])\n if due < 0:\n update.message.reply_text('Прошлое не круто')\n return\n if 'job' in context.chat_data:\n old_job = context.chat_data['job']\n old_job.shedule_removal()\n new_job = context.job_queue.run_once(task, due, context=chat_id)\n context.chat_data['job'] = new_job\n update.message.reply_text(f'Вернусь через {due} секунд')\n except (IndexError, ValueError):\n update.message.reply_text('Глупышка')\n\ndef echo(update, context):\n text = update.message.text\n update.message.reply_text(f'Я получил сообщение \"{text}\"')\n\n\ndef main():\n updater = Updater('1121846344:AAH6Xmry_2HfNc9i6YjxvSoBnMJbJ0jKeLo',\n use_context=True)\n dp = updater.dispatcher\n dp.add_handler(CommandHandler(\"set_timer\", set_timer))\n dp.add_handler(CommandHandler(\"date\", date))\n dp.add_handler(CommandHandler(\"time\", time))\n text_handler = MessageHandler(Filters.text, echo)\n dp.add_handler(text_handler)\n updater.start_polling()\n updater.idle()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1742,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"339285265","text":"#!/usr/bin/env python\n#\n# 2012-2014 (BSD) Ioannis Filippidis\n\"\"\"\ncreate/delete junctions based on \"dirlink\" files\n(win, mac, linux)\n\"\"\"\nimport logging\nlogging.basicConfig(level=logging.WARNING)\nlogger = logging.getLogger()\n\nimport sys\nimport os\nimport subprocess\nimport argparse\nimport shlex\nimport fnmatch\nimport platform\n\ndef locate(pattern, root=os.curdir):\n \"\"\"Recursively find files under given directory matching pattern.\n \"\"\"\n for path, dirs, files in os.walk(os.path.abspath(root)):\n for filename in fnmatch.filter(files, pattern):\n yield os.path.join(path, filename)\n\ndef create_delete_dirlinks_darwin(dirlink, args):\n fid = open(dirlink, 'r')\n igot = fid.readlines()\n fid.close()\n \n path = dirlink.replace('dirlink', '')\n path = path.replace('\\\\', '/')\n logger.debug('current dirlink file in: ' + str(path))\n \n n = 0;\n for line in igot:\n n = n +1\n logger.debug('Line No.' + str(n))\n \n line = line.replace('\\\\', '/')\n if line.find('target') > -1:\n target = line.replace('target = ', '')\n logger.debug('The target is: ' + target)\n elif line.find('link') > -1:\n junction = line.replace('link = ', '')\n logger.debug('The Link is: ' + junction)\n \n junction = path + junction\n \n if '$' in target:\n target = os.path.expandvars(target)\n else:\n target = path + target\n \n if args.create:\n args = shlex.split('ln -s ' + target + ' ' + junction);\n elif args.delete:\n args = shlex.split('rm -f -v ' + junction);\n else:\n raise sys.exit(\n 'Unknown operation. '\n 'Available operations: '\n '-c | --create | -d | --delete'\n )\n \n logger.debug(args)\n subprocess.call(args)\n\ndef create_delete_dirlinks_windows(dirlink, args):\n fid = open(dirlink, 'r')\n igot = fid.readlines()\n fid.close()\n \n path = dirlink.replace('dirlink', '')\n path = path.replace('\\\\', '\\\\\\\\')\n logger.info('current dirlink file in: ' + path)\n \n n = 0;\n for line in igot:\n n = n +1\n logger.debug('Line No.' + str(n))\n \n line = line.replace('\\\\', '\\\\\\\\')\n if line.find('target') > -1:\n target = line.replace('target = ', '')\n logger.info('The target is: ' + target)\n elif line.find('link') > -1:\n junction = line.replace('link = ', '')\n logger.info('The Link is :' + junction)\n \n junction = path +junction\n target = path +target\n \n if args.create:\n args = shlex.split('ln1 --junction ' + junction +\n ' ' + target)\n elif args.delete:\n args = shlex.split('junction -d ' + junction)\n else:\n raise sys.exit(\n 'Unknown operation. '\n 'Available operations: '\n '-c | --create | -d | --delete'\n )\n \n subprocess.call(args)\n\nif __name__ == '__main__':\n desc = (\n 'locate() is used in case of exporting only to a PDF '\n '(w/o latex export). '\n 'Only in that case \\includegraphics{} is still '\n 'able to find the PDF without a path. '\n 'If latex export is used, '\n 'then the produced .pdf_tex should be used, '\n 'with an \\input{} command in latex and '\n 'a relative path is mandatory for '\n 'the \\input{} command to work.'\n )\n \n parser = argparse.ArgumentParser(description=desc)\n \n g = parser.add_mutually_exclusive_group()\n \n g.add_argument('-c', '--create', help='add symbolic links',\n action='store_true')\n g.add_argument('-d', '--delete', help='erase symbolic links',\n action='store_true')\n \n parser.add_argument('-f', '--filename', help='filename',\n default='dirlink')\n parser.add_argument('-v', '--verbose', help='increase output verbosity',\n action='store_true')\n \n args = parser.parse_args()\n \n # no args ?\n if len(sys.argv)==1:\n parser.print_help()\n sys.exit(1)\n \n if args.verbose:\n logger.setLevel(logging.DEBUG)\n \n filename = args.filename\n logger.debug('instruction in filename = ' + str(filename))\n \n # create or delete junctions\n file_generator = locate(filename, './')\n osname = platform.system()\n for f in file_generator:\n flag = 1\n logger.debug('Found file named: ' + str(f))\n \n if osname == 'Darwin':\n create_delete_dirlinks_darwin(f, args)\n else:\n create_delete_dirlinks_windows(f, args)\n","sub_path":"junctions.py","file_name":"junctions.py","file_ext":"py","file_size_in_byte":4945,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"66165241","text":"import cv2\nimport numpy as np\nfrom display_utils import *\n\nA = 0.75\nB = 1.25\n\ndef render_circle(mask, cx, cy, radius):\n y,x = np.ogrid[-radius:radius, -radius:radius]\n index = x**2 + y**2 <= radius**2\n mask[cy-radius:cy+radius, cx-radius:cx+radius][index] = 1\n\ndef render_butterworth_curve(mask, radius):\n sy,sx = mask.shape\n radius2 = radius ** 2\n sx /= 2\n sy /= 2\n y,x = np.ogrid[-sy:sy, -sx:sx]\n x2 = x ** 2\n y2 = y ** 2\n mask[:,:] = 1 / (1 + ((x2 + y2) / radius2) ** 3)\n\ndef render_gaussian_curve(mask, radius):\n sy,sx = mask.shape\n radius2 = 2 * (radius ** 2)\n sx /= 2\n sy /= 2\n y,x = np.ogrid[-sy:sy, -sx:sx]\n x2 = x ** 2\n y2 = y ** 2\n mask[:,:] = np.exp(-(x2 + y2) / radius2)\n\n\nimg = cv2.imread('trees.jpg')\norig_img = np.copy(img)\nimg = np.sum(img, 2) / 3\nimg = np.log1p(img)\n\nfft = np.fft.fft2(img)\nfshift = np.fft.fftshift(fft)\n\ncutoff_radius = 2\n\ndef scale(thing, clip=True):\n if clip:\n thing = np.maximum(0, np.minimum(thing, 255))\n else:\n thing -= np.min(thing)\n thing /= np.max(thing)\n thing *= 255\n thing = thing.astype(np.uint8)\n return thing\n\ndef apply_mask(mask):\n masked_fshift = (A + B * mask) * fshift\n return scale(\n np.exp(\n np.real(\n np.fft.ifft2(\n np.fft.ifftshift(masked_fshift)))) - 1)\n\ndef ideal(radius = cutoff_radius):\n mask = np.zeros(img.shape, np.uint8)\n center_x = img.shape[1] // 2\n center_y = img.shape[0] // 2\n render_circle(mask, center_x, center_y, radius)\n rmask = np.uint8(mask != 1)\n return (\n rmask * 255, apply_mask(rmask),\n )\n\ndef butterworth(radius = cutoff_radius):\n mask = np.zeros(img.shape, np.float)\n render_butterworth_curve(mask, radius)\n rmask = 1 - mask\n return (\n np.uint8(rmask * 255), apply_mask(rmask),\n )\n\ndef gaussian(radius = cutoff_radius):\n mask = np.zeros(img.shape, np.float)\n render_gaussian_curve(mask, radius)\n rmask = 1 - mask\n return (\n np.uint8(rmask * 255), apply_mask(rmask),\n )\n\n\ncv2.namedWindow('Output', cv2.WINDOW_NORMAL)\ncv2.resizeWindow('Output', 1000, 1000)\n\nview = nxm_matrix_view(range(6), ['ideal', 'butterworth', 'gaussian'], 1, 2)\ncv2.imshow('Output', side_by_side(\n layout_with_names({\n \"Homomorphic Fourier Filters - AliMPFard\": Horizontal(6,6),\n '': view,\n }, Vertical), *ideal(), *butterworth(), *gaussian(), orig_img\n))\n\n\nwait_for_key('q', lambda: cv2.waitKey(0), cv2.destroyAllWindows)\n","sub_path":"homomorphic-filter.py","file_name":"homomorphic-filter.py","file_ext":"py","file_size_in_byte":2541,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"650633241","text":"\nfrom HTMLParser import HTMLParser\n\nclass LinkExtractor(HTMLParser):\n\n def reset(self):\n HTMLParser.reset(self)\n self.links = []\n\n def setPattern(self, url, mask, type):\n self.mask = mask\n self.type = type\n if \"/\" in url:\n self.base = url[0:url.rfind(\"/\")] + \"/\"\n else:\n self.base = url + \"/\"\n\n def handle_starttag(self, tag, attrs):\n if tag == \"a\":\n attrs = dict(attrs) # store attributes in object\n if tag == \"a\" and \"href\" in attrs:\n href = attrs[\"href\"]\n if \"/\" in href:\n file = href[href.rfind(\"/\")+1:].lower()\n else:\n file = href.lower()\n# if href.lower().endswith(\".\"+self.type) and (href.lower().maskswith(self.mask) or (\"/\"+self.mask) in href.lower()):\n if file.endswith(\".\" + self.type) and fnmatch.fnmatch(file, self.mask):\n if not href.lower().startswith(\"http://\") and not href.lower().startswith(\"https://\"):\n href = self.base + href\n self.links.append(href)\n\n","sub_path":"openaddresses/lib/extractor.py","file_name":"extractor.py","file_ext":"py","file_size_in_byte":1113,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"208393199","text":"from __future__ import annotations\n\nfrom typing import Any\n\nfrom checkov.common.models.enums import CheckCategories, CheckResult\nfrom checkov.kubernetes.checks.resource.base_spec_check import BaseK8Check\n\n\nclass ShareHostPID(BaseK8Check):\n def __init__(self) -> None:\n # CIS-1.3 1.7.2\n # CIS-1.5 5.2.2\n name = \"Containers should not share the host process ID namespace\"\n id = \"CKV_K8S_17\"\n # Location: Pod.spec.hostPID\n # Location: CronJob.spec.jobTemplate.spec.template.spec.hostPID\n # Location: *.spec.template.spec.hostPID\n supported_kind = (\n \"Pod\",\n \"Deployment\",\n \"DaemonSet\",\n \"StatefulSet\",\n \"ReplicaSet\",\n \"ReplicationController\",\n \"Job\",\n \"CronJob\",\n )\n categories = (CheckCategories.KUBERNETES,)\n super().__init__(name=name, id=id, categories=categories, supported_entities=supported_kind)\n\n def scan_spec_conf(self, conf: dict[str, Any]) -> CheckResult:\n spec = {}\n\n if conf[\"kind\"] == \"Pod\":\n if \"spec\" in conf:\n spec = conf[\"spec\"]\n elif conf[\"kind\"] == \"CronJob\":\n spec = conf.get(\"spec\")\n if spec:\n job_template = spec.get(\"jobTemplate\")\n if job_template:\n job_template_spec = job_template.get(\"spec\")\n if job_template_spec:\n template = job_template_spec.get(\"template\")\n if template:\n if \"spec\" in template:\n spec = template[\"spec\"]\n else:\n inner_spec = self.get_inner_entry(conf, \"spec\")\n spec = inner_spec if inner_spec else spec\n if spec:\n if \"hostPID\" in spec:\n if spec[\"hostPID\"]:\n return CheckResult.FAILED\n\n # This value is by default set to false\n return CheckResult.PASSED\n\n\ncheck = ShareHostPID()\n","sub_path":"checkov/kubernetes/checks/resource/k8s/ShareHostPID.py","file_name":"ShareHostPID.py","file_ext":"py","file_size_in_byte":2031,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"186651478","text":"import cv2\nimport numpy as np\nfrom matplotlib import pyplot as plt\nfrom tensorflow import keras\n\nclass pictionary():\n\n\n\n# function to predict the drawing using the model\n def predict(self,im1):\n pred_probab=self.model.predict(im1)[0]\n pred_class = list(pred_probab).index(max(pred_probab))\n if(max(pred_probab)>0.5):\n print(max(pred_probab), self.pred[pred_class-1])\n\n# mouse callback function\n def pic_draw(self,event,former_x,former_y,flags,param):\n\n if event==cv2.EVENT_LBUTTONDOWN:\n self.drawing=True\n self.current_former_x,self.current_former_y=former_x,former_y\n\n elif event==cv2.EVENT_MOUSEMOVE:\n if self.drawing==True:\n if self.mode==True:\n cv2.line(self.im,(self.current_former_x,self.current_former_y),(former_x,former_y),(0,0,255),1)\n self.current_former_x = former_x\n self.current_former_y = former_y\n\n\n elif event==cv2.EVENT_LBUTTONUP:\n self.drawing=False\n if self.mode==True:\n cv2.line(self.im,(self.current_former_x,self.current_former_y),(former_x,former_y),(0,0,255),1)\n self.current_former_x = former_x\n self.current_former_y = former_y\n # return former_x,former_y\n\n\n# im=np.zeros((255,255), dtype='float32')\n\n def __init__(self):\n self.drawing=False # true if mouse is pressed\n self.mode=True # if True, draw rectangle.\n self.pred=[\"alarm clock\",\"bicycle\",\"bed\",\"airplane\",\"apple\",\"belt\",\"banana\",\"cake\"]\n self.model=keras.models.load_model('/home/diksha/pictionary/my_model2.h5')\n\n self.im=cv2.imread(r'./white.jpg')\n self.img1=cv2.imread(r'./black.jpg')\n cv2.namedWindow(\"Pictionary\")\n cv2.setMouseCallback('Pictionary',self.pic_draw)\n\n def run(self):\n while(1):\n cv2.imshow('Pictionary',self.im)\n img=cv2.cvtColor(self.im, cv2.COLOR_BGR2GRAY)\n th, img = cv2.threshold(img, 127, 255, cv2.THRESH_BINARY_INV)\n blur1 = cv2.GaussianBlur(img, (5, 5), 0)\n cnts, heir= cv2.findContours(blur1, cv2.RETR_TREE, cv2.CHAIN_APPROX_NONE)\n if(len(cnts)>1):\n cv2.drawContours(self.img1, cnts,-1, (255,255,255),5)\n\n cnt = max(cnts, key=cv2.contourArea)\n\n print(cv2.contourArea(cnt))\n if cv2.contourArea(cnt) > 2000:\n x, y, w, h = cv2.boundingRect(cnt)\n self.im1 = self.img1[y:y + h, x:x + w]\n\n self.im1 = cv2.resize(self.im1, (28, 28))\n self.im1 = np.array(self.im1, dtype=np.float32)\n self.im1 = np.reshape(self.im1, (-1, 28, 28, 1))\n self.predict(self.im1)\n \n\n k=cv2.waitKey(1)&0xFF\n if k==27:\n break\n\n elif k==97:\n self.im=cv2.imread(r'./white.jpg')\n self.img1=cv2.imread(r'./black.jpg')\n\n cv2.destroyAllWindows()\n\nif __name__ == '__main__':\n pictionary().run()\n\n# /kaggle/input/quick-draw/full_numpy_bitmap_alarm clock.npy\n# /kaggle/input/quick-draw/full_numpy_bitmap_bicycle.npy\n# /kaggle/input/quick-draw/full_numpy_bitmap_bed.npy\n# /kaggle/input/quick-draw/full_numpy_bitmap_airplane.npy\n# /kaggle/input/quick-draw/full_numpy_bitmap_apple.npy\n# /kaggle/input/quick-draw/full_numpy_bitmap_belt.npy\n# /kaggle/input/quick-draw/full_numpy_bitmap_banana.npy\n# /kaggle/input/quick-draw/full_numpy_bitmap_birthday cake.npy\n","sub_path":"a.py","file_name":"a.py","file_ext":"py","file_size_in_byte":3561,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"380654231","text":"# TO-DO: complete the helper function below to merge 2 sorted arrays\ndef merge(arrA, arrB):\n elements = len(arrA) + len(arrB)\n merged_arr = [0] * elements #creates a list of 0's \n #[0 for _ in range(elements)] is another way to do this\n\n # Your code here\n #start pointers at the start of both lists\n #compare value of pointers and smallest gets added to merge list\n #increment pointer to move onto next list\n \n #initialize each list for both pointers\n a = 0\n b = 0\n \n #loop through merged_arr\n for i in range(len(merged_arr)):\n #make sure pointer is in range of array\n #if it is out of range, copy other array\n if a >= len(arrA):\n merged_arr[i] = arrB[b]\n b += 1\n elif b >= len(arrB):\n merged_arr[i] = arrA[a]\n a += 1\n elif arrA[a] < arrB[b]: #both indices are in range of their own array\n merged_arr[i] = arrA[a]\n a += 1\n else:\n merged_arr[i] = arrB[b]\n b += 1\n\n \n return merged_arr\n\n# while a < len(arrA) and b < len(arrB):\n# if arrA[a] < arrB[b]:\n# combined.append(arrA[a])\n# a += 1\n# else:\n# combined.append(arrB[b])\n# b +=1\n \n# #this will create the combined list\n \n# while a < len(arrA):\n# combined.append(arrA[a])\n# a += 1\n# while b < len(arrB):\n# combined.append(arrB[b])\n# b += 1\n \n# return combined\n\n\n# TO-DO: implement the Merge Sort function below recursively\ndef merge_sort(arr):\n # Your code here\n if len(arr) > 1:\n left = merge_sort(arr[:len(arr) // 2])\n right = merge_sort(arr[len(arr) // 2:])\n arr = merge(left, right)\n\n return arr\n\n\n# STRETCH: implement the recursive logic for merge sort in a way that doesn't \n# utilize any extra memory\n# In other words, your implementation should not allocate any additional lists \n# or data structures; it can only re-use the memory it was given as input\n# def merge_in_place(arr, start, mid, end):\n# # Your code here\n# pass\n\n# def merge_sort_in_place(arr, l, r):\n# # Your code here\n# pass\n","sub_path":"src/sorting/sorting.py","file_name":"sorting.py","file_ext":"py","file_size_in_byte":2255,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"172952026","text":"from _Helper import *\n\n# \n# Use three times of multiplication to calculate complex number, inspired by Strassen algorithm.\n# \n\n# \n# TIME: O(1)\n# SPACE: O(1)\n# \n\n\n# \nclass Solution(object):\n def CalculateComplexNumStrassen(self, a, b, c, d):\n A = a * (c + d)\n B = b * (c - d)\n C = c * (b - a)\n return (B - C, A + C)\n\n def CalculateComplexNumNaive(self, a, b, c, d):\n return (a * c - b * d, a * d + b * c)\n\n\n# \n\n# \ns = Solution()\na = 0.5\nb = 1\nc = 1.5\nd = 2\nexpected = (-1.25, 2.5)\nactual = s.CalculateComplexNumStrassen(a, b, c, d)\nnaive = s.CalculateComplexNumNaive(a, b, c, d)\nAreEqual(expected, actual, a, b, c, d)\nAreEqual(expected, naive, a, b, c, d)\n# another pair\ns = Solution()\na = -2\nb = 1\nc = 4\nd = 2\nexpected = (-10.0, 0.0)\nactual = s.CalculateComplexNumStrassen(a, b, c, d)\nnaive = s.CalculateComplexNumNaive(a, b, c, d)\nAreEqual(expected, actual, a, b, c, d)\nAreEqual(expected, naive, a, b, c, d)\n# ","sub_path":"assets/Introduction-to-Algorithms/Python/4_2_CalculateComplexNum.py","file_name":"4_2_CalculateComplexNum.py","file_ext":"py","file_size_in_byte":1028,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"327459382","text":"import sys\nimport os.path\nsys.path.append(\n os.path.abspath(os.path.join(os.path.dirname(__file__), os.path.pardir)))\nfrom cork import *\n\nEngine.init(\"Hello World!\", 800, 600)\nclass HelloWorld(Entity):\n def __init__(self):\n self.w = 180\n self.h = 50\n self.x = Engine.window[\"width\"]/2 - (self.w / 2)\n self.y = Engine.window[\"height\"]/2 - (self.h / 2)\n self.drawable = Text(\"Hello World\", self)\n super(HelloWorld, self).__init__()\n\nhw = HelloWorld()\n\nEngine.run()\n","sub_path":"examples/helloworld.py","file_name":"helloworld.py","file_ext":"py","file_size_in_byte":511,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"266609711","text":"from tiger.stork import Stork, StorkConfigurationError\nfrom tiger.stork.component import BaseComponent\nfrom tiger.stork.font import FontComponent\nfrom tiger.stork.swatch import SwatchComponent\nfrom nose.tools import raises\n\n\ndef test_component_groups_contain_components():\n panels = Stork(config_path='stork/tests/fixtures/valid.yml')\n for panel in panels:\n for group in panel:\n for component in group:\n assert isinstance(component, BaseComponent)\n\n\ndef test_components_have_unique_keys():\n panels = Stork(config_path='stork/tests/fixtures/valid.yml')\n for panel in panels:\n for group in panel:\n for component in group:\n assert component.key == '%s-%s' % (panel.name.lower(), component.name.lower())\n\n\n@raises(StorkConfigurationError)\ndef test_no_component_type_fails():\n panels = Stork(config_path='stork/tests/fixtures/no_component_type.yml')\n\n\n@raises(StorkConfigurationError)\ndef test_invalid_component_type_fails():\n panels = Stork(config_path='stork/tests/fixtures/invalid_component_type.yml')\n\n\ndef test_all_components_are_in_cache():\n panels = Stork(config_path='stork/tests/fixtures/valid.yml')\n for panel in panels:\n for group in panel:\n for component in group:\n assert component in [v for k, v in panels.component_cache]\n assert panels[component.key] == component\n\n\ndef test_correct_subclass_is_assigned():\n panels = Stork(config_path='stork/tests/fixtures/valid.yml')\n assert isinstance(panels['foo-baz'], SwatchComponent)\n assert isinstance(panels['bar-quux'], FontComponent)\n","sub_path":"tiger/stork/tests/test_components.py","file_name":"test_components.py","file_ext":"py","file_size_in_byte":1636,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"169636419","text":"from tkinter import *\n\nclass Text_Settings(object):\n def __init__(self,master):\n tx = self.tx = Toplevel(master,bd=15)\n tx.title(\"Text Settings\")\n pad_x = 5\n pad_y = 5\n\n self.size_m = Message(tx,text = 'Input Text Size:',aspect = 500, justify = LEFT).pack(anchor = CENTER)\n self.size_e = Entry(tx)\n self.size_e.pack(anchor=CENTER)\n\n self.okay_button = Button(tx,text = \"OKAY\", command = self.cleanup)\n self.okay_button.pack(anchor=S,side = BOTTOM,pady=pad_y)\n\n def cleanup(self):\n text_size = self.size_e.get()\n self.settings = [text_size]\n self.tx.destroy()\n","sub_path":"Separation/Text_Settings.py","file_name":"Text_Settings.py","file_ext":"py","file_size_in_byte":649,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"376466695","text":"# Copyright 2015 OpenStack Foundation\n# Copyright 2015 Hewlett-Packard Development Company, L.P.\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nimport collections\nimport json\nimport logging\nimport random\nimport socket\nimport threading\nimport time\n\nlogging.basicConfig(level=logging.DEBUG)\n\n\ndef default_echo_handler(message, ovsconn):\n logging.debug(\"responding to echo\")\n ovsconn.send({\"result\": message.get(\"params\", None),\n \"error\": None, \"id\": message['id']})\n\n\ndef default_message_handler(message, ovsconn):\n ovsconn.responses.append(message)\n\n\nclass OVSDBConnection(threading.Thread):\n \"\"\"Connects to an ovsdb server that has manager set using\n\n ovs-vsctl set-manager ptcp:5000\n\n clients can make calls and register a callback for results, callbacks\n are linked based on the message ids.\n\n clients can also register methods which they are interested in by\n providing a callback.\n \"\"\"\n\n def __init__(self, IP, PORT, **handlers):\n super(OVSDBConnection, self).__init__()\n self.socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n self.socket.connect((IP, PORT))\n self.responses = []\n self.callbacks = {}\n self.read_on = True\n self.handlers = handlers or {\"echo\": default_echo_handler}\n self.start()\n\n def send(self, message, callback=None):\n if callback:\n self.callbacks[message['id']] = callback\n self.socket.send(json.dumps(message))\n\n def response(self, id):\n return [x for x in self.responses if x['id'] == id]\n\n def set_handler(self, method_name, handler):\n self.handlers[method_name] = handler\n\n def _on_remote_message(self, message):\n try:\n json_m = json.loads(message,\n object_pairs_hook=collections.OrderedDict)\n handler_method = json_m.get('method', None)\n if handler_method:\n self.handlers.get(handler_method, default_message_handler)(\n json_m, self)\n elif json_m.get(\"result\", None) and json_m['id'] in self.callbacks:\n id = json_m['id']\n if not self.callbacks[id](json_m, self):\n self.callbacks.pop(id)\n\n else:\n default_message_handler(message, self)\n except Exception as e:\n logging.exception(\n \"exception [%s] in handling message [%s]\", e.message, message)\n\n def __echo_response(message, self):\n self.send({\"result\": message.get(\"params\", None),\n \"error\": None, \"id\": message['id']})\n\n def run(self):\n\n chunks = []\n lc = rc = 0\n while self.read_on:\n try:\n response = self.socket.recv(4096)\n if response:\n response = response.decode('utf8')\n message_mark = 0\n for i, c in enumerate(response):\n if c == '{':\n lc += 1\n elif c == '}':\n rc += 1\n\n if rc > lc:\n raise Exception(\"json string not valid\")\n\n elif lc == rc and lc is not 0:\n chunks.append(response[message_mark:i + 1])\n message = \"\".join(chunks)\n self._on_remote_message(message)\n lc = rc = 0\n message_mark = i + 1\n chunks = []\n\n chunks.append(response[message_mark:])\n except Exception:\n # Pass to avoid EOF error\n pass\n\n def stop(self, force=False):\n self.read_on = False\n if force:\n self.socket.close()\n\n def select_table(self, table):\n select_dict = {\"op\": \"select\", \"table\": table, \"where\": []}\n op_id = str(random.getrandbits(128))\n params = ['hardware_vtep']\n params.append(select_dict)\n query_select = {\"method\": \"transact\",\n \"params\": params,\n \"id\": op_id}\n return query_select\n\n def find_row(self, net_id, count, resp_dec):\n for i in range(count):\n row = str(resp_dec['result'][0]['rows'][i])\n if net_id in row:\n return row\n\n def get_response(self, OVSDB_IP, OVSDB_PORT, table):\n query = self.select_table(table)\n self.send(query)\n time.sleep(2)\n resp = self.responses\n resp = str(resp[0])\n return resp\n","sub_path":"networking_l2gw/tests/scenario/ovsdb_connections.py","file_name":"ovsdb_connections.py","file_ext":"py","file_size_in_byte":5203,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"484309155","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.test import TestCase\nfrom decimal import Decimal\nimport datetime\nfrom .models import (\n RequestHeader,\n SelfReportedCashFlow,\n Address,\n Business,\n CFApplicationData,\n Owner,\n LoanApp,\n)\n\n# Request Header Model Test (Convert request_date to datetime)\nclass RequestHeaderModelTest(TestCase):\n def setUp(self):\n RequestHeader.objects.create(\n cf_request_id=\"500653901\",\n request_date=\"2019-06-26T23:05:41.2898238Z\",\n cf_api_user_id=None,\n cf_api_password=None,\n is_test_lead=True,\n )\n\n def test_model_created(self):\n request_header = RequestHeader.objects.first()\n self.assertEquals(request_header.cf_request_id, \"500653901\")\n self.assertEquals(request_header.request_date, \"2019-06-26T23:05:41.2898238Z\")\n self.assertEquals(request_header.cf_api_user_id, None)\n self.assertEquals(request_header.cf_api_password, None)\n self.assertTrue(request_header.is_test_lead)\n\n\nclass SelfReportedCashFlowModelTest(TestCase):\n def setUp(self):\n SelfReportedCashFlow.objects.create(\n annual_revenue=49999999.01,\n monthly_average_bank_balance=94941.0,\n monthly_average_credit_card_volume=18191.0,\n )\n\n def test_model_created(self):\n self_reported_cash_flow = SelfReportedCashFlow.objects.first()\n self.assertEquals(\n self_reported_cash_flow.annual_revenue, Decimal(\"49999999.01\")\n )\n self.assertEquals(\n self_reported_cash_flow.monthly_average_bank_balance, Decimal(\"94941.00\")\n )\n self.assertEquals(\n self_reported_cash_flow.monthly_average_credit_card_volume,\n Decimal(\"18191.00\"),\n )\n\n\nclass AddressModelTest(TestCase):\n def setUp(self):\n Address.objects.create(\n address1=\"1234 Red Ln\",\n address2=None,\n city=\"Santa Monica\",\n state=\"CA\",\n zipcode=\"45321\",\n )\n\n def test_model_created(self):\n address = Address.objects.first()\n self.assertEquals(address.address1, \"1234 Red Ln\")\n self.assertEquals(address.address2, None)\n self.assertEquals(address.city, \"Santa Monica\")\n self.assertEquals(address.state, \"CA\")\n self.assertEquals(address.zipcode, \"45321\")\n\n\nclass BusinessModelTest(TestCase):\n\n self_reported_cash_flow = SelfReportedCashFlow.objects.create(\n annual_revenue=49999999.01,\n monthly_average_bank_balance=94941.0,\n monthly_average_credit_card_volume=18191.0,\n )\n\n address = Address.objects.create(\n address1=\"1234 Red Ln\",\n address2=None,\n city=\"Santa Monica\",\n state=\"CA\",\n zipcode=\"45321\",\n )\n\n def setUp(self):\n self.self_reported_cash_flow.save()\n self.address.save()\n Business.objects.create(\n name=\"Wow Inc\",\n self_reported_cash_flow=self.self_reported_cash_flow,\n address=self.address,\n tax_id=\"839674398\",\n phone=\"6573248876\",\n naics=\"79232\",\n has_been_profitable=True,\n has_been_bankrupted_in_last_7_years=False,\n inception_date=datetime.datetime.now(),\n )\n\n def test_model_created(self):\n business = Business.objects.first()\n self.assertEquals(business.name, \"Wow Inc\")\n self.assertEquals(\n business.self_reported_cash_flow, self.self_reported_cash_flow\n )\n self.assertEquals(business.address, self.address)\n self.assertEquals(business.tax_id, \"839674398\")\n self.assertEquals(business.phone, \"6573248876\")\n self.assertEquals(business.naics, \"79232\")\n self.assertEquals(business.has_been_profitable, True)\n self.assertEquals(business.has_been_bankrupted_in_last_7_years, False)\n # self.assertTrue(business.inception_date < datetime.datetime.now())\n\n\nclass CFApplicationDataModelTest(TestCase):\n def setUp(self):\n CFApplicationData.objects.create(\n requested_loan_amount=\"49999999\",\n stated_credit_history=1,\n legal_entity_type=\"LLC\",\n filter_id=\"897079\",\n )\n\n def test_model_created(self):\n cf_application_data = CFApplicationData.objects.first()\n self.assertEquals(cf_application_data.requested_loan_amount, \"49999999\")\n self.assertEquals(cf_application_data.stated_credit_history, 1)\n self.assertEquals(cf_application_data.legal_entity_type, \"LLC\")\n self.assertEquals(cf_application_data.filter_id, \"897079\")\n\n","sub_path":"loanapplicationapi/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":4686,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"115729582","text":"# index的admin文件\r\n# 修改title和header\r\nfrom django.contrib import admin\r\nfrom .models import *\r\n\r\n# Register your models here.\r\n# 一:直接将模型注册到admin后台\r\n# admin.site.register(Product)\r\n# admin.site.register(Type)\r\n\r\n# 修改title和header\r\nadmin.site.site_header = 'MyDjango'\r\nadmin.site.site_title = 'MyDjango 后台管理'\r\n\r\n\r\n# 二:自定义ProductAdmin类并集成ModelAdmin\r\n# 注册方法一,使用Python装饰器讲ProductAdmin和模型Product绑定并注册到后台\r\n@admin.register(Product)\r\nclass ProductAdmin(admin.ModelAdmin):\r\n # 设置显示的字段\r\n list_display = ['id','name','weight','size','type',]\r\n list_display.append('colored_type')\r\n # 设置可搜索的字段并在Admin后台数据生成搜索框,如有外键,因使用双下划线连接两个模型地方字段\r\n search_fields = ['id', 'name', 'type__type_name']\r\n # 设置过滤器,在后台数据的右侧生成导航栏,如有外键,因使用双下划线连接两个模型的字段\r\n list_filter = ['name', 'type__type_name']\r\n # 设置排序方式,['id']为升序,降序为['-id']\r\n ordering = ['id']\r\n # 设置时间选择器,如字段中有时间格式才可以使用\r\n # date_hierarchy = Field\r\n # 添加新数据时,设置可添加数据的字段\r\n fields = ['name', 'weight', 'size', 'type']\r\n # 设置可读字段,在修改或新增数据时使其无法设置\r\n readonly_fields = ['name']\r\n # 重写get_readonly_fields()函数,设设置超级用户和普通用户 的权限\r\n def get_readonly_fields(self, request, obj=None):\r\n if request.user.is_superuser:\r\n self.readonly_fields = []\r\n return self.readonly_fields\r\n def get_queryset(self, request):\r\n qs = super(ProductAdmin,self).get_queryset(request)\r\n if request.user.is_superuser:\r\n return qs\r\n else:\r\n return qs.filter(id__lt=6)\r\n\r\n@admin.register(Type)\r\nclass TypeAdmin(admin.ModelAdmin):\r\n list_display = ['id', 'type_name']\r\n\r\n# 注册方法二\r\n# admin.site.register(Product, ProductAdmin)","sub_path":"django/MyDjango/index/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":2118,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"353567573","text":"import random, os, sys\n\ncardName = {1: 'As', 2: 'Dos', 3: 'Tres', 4: 'Quatre', 5: 'Cinc', 6: 'Sis', 7: 'Set', 8: 'Vuit', 9: 'Nou',\n 10: 'Deu', 11: 'Jota', 12: 'Reina', 13: 'Rei'}\ncardSuit = {'c': 'Piques', 'h': 'Cors', 's': 'Espases', 'd': 'Diamants'}\n\nclass color:\n PURPLE = '\\033[95m'\n CYAN = '\\033[96m'\n DARKCYAN = '\\033[36m'\n BLUE = '\\033[94m'\n GREEN = '\\033[92m'\n YELLOW = '\\033[93m'\n RED = '\\033[91m'\n BOLD = '\\033[1m'\n UNDERLINE = '\\033[4m'\n END = '\\033[0m'\n\nclass Card:\n\n def __init__(self, rank, suit):\n self.rank = rank\n self.suit = suit\n\n def __str__(self):\n return (cardName[self.rank] + \" de \" + cardSuit[self.suit])\n\n def getRank(self):\n return (self.rank)\n\n def getSuit(self):\n return (self.suit)\n\n def BJValue(self):\n if self.rank > 9:\n return (10)\n else:\n return (self.rank)\n\n\ndef showHand(hand):\n for card in hand:\n print(\" \" + str(card))\n\n\ndef showCount(hand):\n print(\"Puntuació: \" + str(handCount(hand)))\n\n\ndef handCount(hand):\n handCount = 0\n for card in hand:\n handCount += card.BJValue()\n return (handCount)\n\ndef gameEnd(score):\n print(color.BOLD +\"Blackjack!! \"+ color.END + \"-*Puntuació*- maquina: \" + str(score['maquina']) + \" Puntuació jugador: \" + str(score['jugador']))\n sys.exit(0)\n\n\ndeck = []\nsuits = ['c', 'h', 'd', 's']\nscore = {'maquina': 0, 'jugador': 0}\nhand = {'maquina': [], 'jugador': []}\n\nfor suit in suits:\n for rank in range(1, 14):\n deck.append(Card(rank, suit))\n\nkeepPlaying = True\n\nwhile keepPlaying:\n\n random.shuffle(deck)\n random.shuffle(deck)\n random.shuffle(deck)\n\n # Deal Cards\n\n hand['jugador'].append(deck.pop(0))\n hand['maquina'].append(deck.pop(0))\n\n hand['jugador'].append(deck.pop(0))\n hand['maquina'].append(deck.pop(0))\n\n playHuman = True\n bustedHuman = False\n\n while playHuman:\n print(color.RED + \"---------------------------------------------------------\" + color.END)\n print(color.BOLD + \"_-*Blackjack*-_ \"+ color.BLUE + \"Maquina: \" + str(score['maquina']) + color.PURPLE + \" - Jugador: \" + str(score['jugador']) + color.END)\n\n print()\n\n print(color.BOLD + \"La maquina ha robat: \" + color.END + str(hand['maquina'][-1]) )\n print()\n\n print(color.BOLD + \"La teva ma:\" + color.END)\n\n showHand(hand['jugador'])\n print()\n showCount(hand['jugador'])\n\n print()\n\n inputCycle = True\n userInput = ''\n\n while inputCycle:\n userInput = input(\"(U)na carta mes, (R)es mes, or (S)ortir: \").upper()\n if userInput == 'U' or 'S' or 'Q':\n inputCycle = False\n\n if userInput == 'U':\n hand['jugador'].append(deck.pop(0))\n if handCount(hand['jugador']) > 21:\n playHuman = False\n bustedHuman = True\n elif userInput == 'R':\n playHuman = False\n else:\n gameEnd(score)\n\n playComputer = True\n bustedComputer = False\n\n while not bustedHuman and playComputer:\n print(handCount(hand['maquina']))\n if handCount(hand['maquina']) < 17:\n hand['maquina'].append(deck.pop(0))\n else:\n playComputer = False\n\n if handCount(hand['maquina']) > 21:\n playComputer = False\n bustedComputer = True\n\n if bustedHuman:\n print(color.PURPLE + 'El jugador ha perdut' + color.END)\n score['maquina'] += 1\n elif bustedComputer:\n print(color.BLUE + 'La maquina ha perdut' + color.END)\n score['jugador'] += 1\n elif handCount(hand['jugador']) > handCount(hand['maquina']):\n print(color.PURPLE +'El jugador guanya' + color.END)\n score['jugador'] += 1\n else:\n print(color.BLUE + 'La maquina guanya' + color.END)\n score['maquina'] += 1\n\n print()\n print('Ma de la maquina:')\n showHand(hand['maquina'])\n showCount(hand['maquina'])\n\n print()\n print('Ma de l\\'usuari')\n showHand(hand['jugador'])\n showCount(hand['jugador'])\n print()\n if input(\"(S)ortir o intro per jugar una nova ronda\").upper() == 'S':\n gameEnd(score)\n\n deck.extend(hand['maquina'])\n deck.extend(hand['jugador'])\n\n del hand['maquina'][:]\n del hand['jugador'][:]","sub_path":"blackjack.py","file_name":"blackjack.py","file_ext":"py","file_size_in_byte":4344,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"48636531","text":"import sys\nimport os\nos.chdir('C:\\\\adb')\nif len(sys.argv) != 2:\n print('Enter the Filename: ')\n raise SystemExit(1)\nf = open(sys.argv[1])\nlines = f.readlines()\nf.close()\nflines = [float(line) for line in lines]\n\nprint('the max of two numbers is ' + max(flines))\nprint('the min of two numbers is' + min(flines))","sub_path":"Scripts-Pyc/argv.py","file_name":"argv.py","file_ext":"py","file_size_in_byte":316,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"190803242","text":"\"\"\"\nKurtosis-specific routines\n\"\"\"\nimport warnings\n\nimport numpy as np\nfrom scipy.signal import lfilter\nfrom obspy.core.stream import Stream\nfrom obspy.core.trace import Trace\n\nfrom .utils import smooth_filter\nfrom .pick_candidate import PickCandidate\n# from .logger import log\n\n\nclass Kurtosis_pick():\n \"\"\"\n Holds one Kurtosis-based pick and the values that chose it\n \"\"\"\n def __init__(self, time, k_jump, kurtosis):\n \"\"\"\n :param time: kurtosis pick time\n :param k_jump: jump in kurtosis associated with the pick\n :param kurtosis: kurtosis value at pick\n \"\"\"\n self.time = time\n self.k_jump = k_jump\n self.kurtosis = kurtosis\n\n\nclass Kurtosis():\n \"\"\"\n Class for picking seismograms using the Kurtosis\n\n Kurtosis picking is a n-step process whose goal is to identify jumps\n in the Kurtosis:\n 1) calculate the Kurtosis\n 2) calculate the cumulative sum of the kurtosis (kurto_cum)\n 3) remove a linear fit from the beginning to the end of the kurto_cum\n (corr_kurto_cum)\n 4) Identify peaks (highs) and subtract this from the preceding data. Set\n all positive values to zero\n 5) identify where the slope of the resulting function changes from positive\n to negative\n \"\"\"\n def __init__(self, freq_bands, wind_lengths, n_smooth, extrem_smooths=None,\n plot=False):\n \"\"\"\n :param FBs: list of [min, max] frequency bands\n :param window_lengths: list of window lengths (seconds)\n :param n_smooth: smoothings to apply in calculating Kurtosis (samples)\n :param extrem_smooth: smoothing sequence to use when calculating times\n of extrema\n :param plot: make a plot of Kurtosis parameters\n \"\"\"\n assert isinstance(freq_bands, list), \"freq_bands is not a list\"\n assert isinstance(wind_lengths, list), \"wind_lengths is not a list\"\n assert isinstance(n_smooth, int), \"n_smooth is not an int\"\n\n self.freq_bands = freq_bands\n self.wind_lengths = wind_lengths\n self.n_smooth = n_smooth\n self.plot = plot\n self.extrem_smooths = None # smoothings for extrema following\n\n # Mean trace over freq_bands, wind_lengths & smoothing\n self.mean_kurtosis = None\n # Mean trace, cumulated and detrended\n self.mean_cumulative_kurtosis = None\n # Gradients of mean_cumulative_kurtosis, for each extrem_smoothing\n self.kurto_gradients = None\n\n def __str__(self):\n s = \"Kurtosis:\\n\"\n s += f\" freq bands = {self.freq_bands}\\n\"\n s += f\" window_lengths = {self.wind_lengths}\\n\"\n s += f\" n_smooth = {self.n_smooth}\\n\"\n s += f\" extrem_smooths = {self.extrem_smooths}\\n\"\n s += f\" mean_kurtosis = {self.mean_kurtosis}\\n\"\n s += f\" mean_cumulative kurtosis = {self.mean_cumulative_kurtosis}\\n\"\n s += f\" len{self.kurto_gradients} kurto_gradients\"\n return s\n\n def pick_trace(self, trace, n_candidates, starttime=None, endtime=None,\n extrem_type='mini', extrem_smooths=[1, 5, 10],\n extrem_normalize=False, extrem_which='max'):\n \"\"\"\n Pick a trace using the Kurtosis\n\n :param trace: waveform trace\n :param n_candidates: maximum number of candidates to return\n :param starttime: first time of interest\n :param endtime: last time of interest\n :param extrem_type: 'mini' or 'maxi', depending on wheter you want to\n follow minima or maxima\n :param extrem_smooths: smoothing to apply when following extrema\n :param extrem_normalize: normalize the cumulative Kurtosis gradient\n :param extrem_which:\n 'first': select first 'n_extrema' extrema > 0.1, order right to\n left\n 'max': select the 'n_extrema' biggest extrema, order biggest\n to smallest\n :returns: list of PickCandidates\n \"\"\"\n self.extrem_smooths = extrem_smooths\n self.calc_kurtocum(trace, starttime, endtime)\n candidates = self.follow_extrem(type_=extrem_type,\n n_extrema=n_candidates,\n normalize=extrem_normalize,\n sense=extrem_which)\n return candidates\n\n def calc_kurtocum(self, trace, starttime=None, endtime=None, debug=False):\n \"\"\"\n Calculate cumulative kurtosis over bandwidths and windows\n\n Puts mean kurtosis over all windows and freq_bands in\n self.mean_cumulative_kurtosis\n :param trace: the raw trace (works one trace, not a stream)\n :param starttime: is the first time of interest\n :param endtime: is the last time of interest\n \"\"\"\n # trace2FWkurto.m:26\n assert isinstance(trace, Trace), \"trace is not an obspy Trace\"\n\n # sr = trace.stats.sampling_rate\n # data_length = trace.stats.endtime - trace.stats.starttime\n if starttime is None:\n starttime = trace.stats.starttime\n elif starttime < trace.stats.starttime:\n starttime < trace.stats.starttime\n if endtime is None:\n endtime = trace.stats.endtime\n elif endtime > trace.stats.endtime:\n endtime = trace.stats.endtime\n\n # Filter traces in different frequency bands\n B = []\n for FB in self.freq_bands:\n f = trace.copy()\n if debug:\n print(f'trace2FWkurto: filtering from {FB[0]} to {FB[1]} Hz')\n f.detrend('demean')\n f.filter(type='bandpass', freqmin=FB[0], freqmax=FB[1], corners=3)\n f = f.slice(starttime, endtime)\n f.data[np.arange(0, 50)] = 0\n B.append(f)\n\n # 2-level lists: : 1st dim: window_lengths, 2nd dim: freq bands\n K, C = [], [] # all kurtoses and all cumulative, detrended kurtoses\n for win_len in self.wind_lengths:\n corr_cums, kurtos = self.calc_cum_kurtoses(B, win_len)\n C.append(corr_cums)\n K.append(kurtos)\n self.mean_kurtosis = _mean_trace(K)\n self.mean_cumulative_kurtosis = _mean_trace(C)\n\n def calc_cum_kurtoses(self, B, win_len):\n \"\"\"\n Calculate kurtoses and cumulative kurtoses for given window length\n\n :param B: list of data, filtered in different bands\n :win_len: window length in seconds\n \"\"\"\n corr_cums, kurtos = [], []\n sr = B[0].stats.sampling_rate\n data_length = B[0].stats.endtime - B[0].stats.starttime\n if win_len > data_length:\n warnings.warn('Kurtosis window > data window ('\n f'{win_len:.3g}s > {data_length:.3g}s), skipping!')\n return [], []\n else:\n win_samps = int(np.floor(win_len * sr)) + 1\n win_samps = min(win_samps, len(B[0].data))\n for tr, fb in zip(B, self.freq_bands):\n k = _fast_kurtosis(tr, win_samps)\n filt = smooth_filter(k, self.n_smooth)\n corr_cum, _ = _f_cumul(filt)\n corr_cum.detrend('simple')\n # corr_cum = kurto_cum.copy()\n # line = _f_segment(corr_cum)\n # corr_cum.data -= line.data\n if self.plot:\n self._plot_kurtosis(win_len, fb, tr, k, corr_cum)\n kurtos.append(k)\n corr_cums.append(corr_cum)\n return corr_cums, kurtos\n\n @staticmethod\n def _plot_kurtosis(wl, fb, trace, kurtosis, corr_cum):\n print(f'Plot kurtosis for win_len={wl}, f_band={fb}', 'debug')\n cor = corr_cum.copy()\n cor.stats.channel = 'COR'\n kurtosis.stats.channel = 'KUR'\n Stream([trace, kurtosis, cor]).plot(\n size=(600, 600), equal_scale=False)\n\n def follow_extrem(self, type_='mini', n_extrema=2,\n normalize=False, sense=None, debug=False):\n \"\"\"\n Return extrema of the cumulative kurtosis\n\n Steps:\n 1) Smooth the cumulative kurtosis (self.mean_cumulative_kurtosis)\n using the windows specified in self.extrem_smooths\n 2) locate the first 'n' first extremas of the smoothest function\n 3) refine these extrema step by step through less and less smooth\n functions\n\n Uses self.mean_cumulative kurtosis\n :param type: 'mini' or 'maxi': follow minima or maxima\n :param n_extrema: number of extrema to follow\n :param normalize: normalize the gradient?\n :param sense:\n 'first': select first 'n_extrema' extrema > 0.1, ordered right\n to left\n 'max': select from biggest to smallest\n :returns: list of PickCandidates\n \"\"\"\n # follow_extrem.m:29\n # Parameters\n assert not len(self.mean_cumulative_kurtosis) == 0,\\\n 'no mean cumulative kurtosis!'\n st = self.mean_cumulative_kurtosis.stats.starttime\n sr = self.mean_cumulative_kurtosis.stats.sampling_rate\n\n all_extrema, self.kurto_gradients, _ =\\\n _get_extrema(self.extrem_smooths, self.mean_cumulative_kurtosis,\n type_, normalize)\n\n selected_extrema = _select_extrema(all_extrema[0],\n n_extrema, sense)\n\n # Sharpen the indices/values using the smaller smoothing values\n sharp_extrema = []\n for e in selected_extrema:\n extrema = e\n for finer_extrema in all_extrema[1:]:\n c_extrem = _find_close_extrema(\n extrema, finer_extrema, 40)\n if c_extrem is not None:\n extrema = c_extrem\n sharp_extrema.append(extrema)\n\n return [PickCandidate(st + x['index']/sr,\n 'kurtosis',\n x['value'],\n sampling_rate=sr)\n for x in sharp_extrema]\n\n\ndef _find_close_extrema(best_extrem, finer_extrema, max_diff=40):\n if len(finer_extrema) == 0:\n warnings.warn('No extrema found for {:d}-sample smoothing'.format(\n finer_extrema['smoothing']))\n return None\n i_diff = abs(best_extrem['index'] - [x['index'] for x in finer_extrema])\n if np.any(i_diff <= max_diff):\n return finer_extrema[np.argmin(i_diff)]\n return None\n\n\ndef _get_extrema(smoothing_list, cumul_k, sense, normalize, debug=False):\n \"\"\"\n Returns a list of extrema, from smoothest to roughest\n\n :param smoothing list: list of smoothings in number of samples\n :param cumul_k: cumulative kurtosis Trace\n :param sens: 'maxi' or 'mini', passed on to ext_indices\n :param normalize: passed on to _cum2grad()\n :returns: list of {indices: value, trace: value} from smoothest to\n roughest. value is the approximate height of the kurtosis jump\n \"\"\"\n extrema, gradients = [], []\n # Put the strongest smoothing first\n sorted_smooth = sorted(smoothing_list, reverse=True)\n for smoothing in sorted_smooth:\n v_smooth = smooth_filter(cumul_k, smoothing)\n v_smooth = _cum2grad(v_smooth, normalize)\n ext_indices, _ = _loca_ext(v_smooth, None, None, sense)\n extrema.append([{'index': i, 'value': -v_smooth.data[i]}\n for i in ext_indices])\n gradients.append(v_smooth)\n # create a list of {index: value:} dictionaries with all of the\n # detected extrema in the smoothest trace\n # extrema = [{'index': i, 'value': extrema[0]['trace'].data[i]}\n # for i in extrema[0]['indices']]\n return extrema, gradients, sorted_smooth\n\n\ndef _select_extrema(extrema, N, sense='max', threshold=0.1):\n \"\"\"\n Return N extrema, ordered by size or time\n :param extrema: list of {'value': val, 'index': i}\n :param sense: 'max' return from largest to smallest\n 'first': return from first to last\n :param threshold: minimum value to accept for sense=='first'\n \"\"\"\n if sense == 'first':\n big_extrema = [x for x in extrema if x['value'] >= threshold]\n if len(big_extrema) > 1:\n ext = sorted(big_extrema, key=lambda k: k['index'])\n else:\n ext = sorted(extrema, key=lambda k: k['index'])\n try:\n selected = [x for x in ext[:N:-1]]\n except Exception:\n selected = [x for x in ext[::-1]]\n elif sense == 'max':\n ext = sorted(extrema, key=lambda k: np.abs(k['value']), reverse=True)\n try:\n selected = [x for x in ext[:N]]\n except Exception:\n selected = [x for x in ext]\n else:\n raise NameError(\"sense not 'max' or 'first'\")\n return selected\n\n\ndef _fast_kurtosis(trace, win_samps):\n \"\"\"\n Compute kurtosis really quickly using \"filter\" function\n\n could I just use obspy kurtosis?\n\n :param trace: one trace\n :param win_samps: number of samples in the sliding window\n :returns: Kurtosis trace\n \"\"\"\n assert isinstance(trace, Trace), \"trace is not an obspy Trace\"\n win_samps = int(round(win_samps))\n # log(win_samps, 'debug')\n # fast_kurtosis.m:11\n if win_samps == 1:\n win_samps = 2\n\n # Set NaNs to 0 for computational stability\n f = trace.copy()\n f.detrend(type='demean')\n # f.data[np.isnan(f.data)] = 0\n\n # Compute kurtosis\n a = np.divide(np.ones(win_samps), float(win_samps))\n b = 1.\n m_2 = lfilter(a, b, f.data**2)\n m_4 = lfilter(a, b, f.data**4)\n out = trace.copy()\n out.data = np.divide(m_4, (m_2 ** 2))\n # Protect against edge effect\n out.data[:win_samps] = out.data[win_samps]\n\n # Set any kurtosis value to nan for any indices within win_samples of\n # an NaN in the original data.\n # I think this is outdated, should just trim\n for i in np.nonzero(trace.data == np.nan)[0]:\n if i:\n out.data[i: i + win_samps] = np.nan\n return out\n\n\ndef _f_cumul(f):\n \"\"\"\n Calculate the positive gradient cumulative of f\n\n If the gradient is positive we take the cumulative, if the gradient is\n negative then, as long as the gradient is negative, the value is that\n of the last positive gradient. The output has then only positive\n gradients\n ___/\n /\n ___/\n :param f: trace, stream or list of traces containing the data\n :returns: trace or list of cumulative output traces, first value = 0\n list of cumulative output traces\n \"\"\"\n # f_cumul.m:14\n bare_trace = False\n if isinstance(f, Trace):\n f = [f]\n bare_trace = True\n g = f.copy()\n for t in g:\n tdata = t.data.copy() # Backup copy for info\n t.data[np.isnan(t.data)] = 0\n t.differentiate(method='gradient')\n t.data[t.data < 0] = 0\n t.data = np.cumsum(t.data)\n t.data[np.isnan(tdata)] = np.nan\n\n p = g.copy()\n\n # Subtract first non-nan value from g\n for t in g:\n ind = np.nonzero(np.isfinite(t.data))[0]\n if len(ind) > 0:\n t.data -= t.data[ind[0]]\n\n if bare_trace:\n return g[0], p[0]\n else:\n return g, p\n\n\ndef _f_segment(f):\n \"\"\"\n Return a line segment between the first and last values of function.\n SHOULD BE REPLACEABLE BY DETREND('linear')\n\n Goes from first non-nan value to last non-nan value\n Input and output have the same size.\n :param f: trace, stream or list of data traces\n :returns: line segment trace or list of traces\n \"\"\"\n bare_trace = False\n if isinstance(f, Trace):\n f = [f]\n bare_trace = True\n assert isinstance(f[0], Trace), f'f is list of {type(f[0])}s!'\n # f_segment.m:12\n segments = []\n # for i in arange(1,n).reshape(-1):\n for trace in f:\n # print(type(trace), trace)\n # clear('a','b','ya','yb','lin')\n a = np.nonzero(np.isfinite(trace.data))[0][0]\n b = np.nonzero(np.isfinite(trace.data))[0][-1]\n ya = trace.data[a]\n yb = trace.data[b]\n lin = np.linspace(start=ya, stop=yb, num=b-a+1)\n segment = trace.copy()\n # Ugly handling of different cases\n before = np.empty(a)\n before[:] = np.nan\n after = np.empty(len(trace.data)-b-1)\n after[:] = np.nan\n segment.data = np.concatenate((before, lin, after))\n segments.append(segment)\n if bare_trace:\n return segments[0]\n else:\n return segments\n\n\ndef _cum2grad(f_in, normalize=False, debug=False):\n \"\"\"\n Transform cumulative into function that is shifted below 0 for all maxima\n\n Finds peaks, subtracts the next peak value from all data, gets rid of\n all values > 0. Give negative values that \"jump\" up to zero at the next\n peak\n\n :param f_in: cumulative trace\n :param normalize: 'True' will divide the output by it's min value\n :returns: f_out\n \"\"\"\n # cum2grad.m:7\n assert not len(f_in) == 0, 'f_in is empty!'\n\n tycalpha = np.zeros(len(f_in.data))\n tikxs, _ = _loca_ext(f_in, None, None, 'maxi')\n a = int(np.nonzero(np.isfinite(f_in.data))[0][0])\n b = int(np.nonzero(np.isfinite(f_in.data))[0][-1])\n # tikxs = tikxs.to_list()\n if len(tikxs) == 0:\n tikxs = np.array([a, b])\n else:\n tikxs = np.array([a] + tikxs.tolist() + [b])\n\n tikys = f_in.data[tikxs]\n # function equal to the next peak\n for ilo, ihi, y in zip(tikxs[:-2], tikxs[1:], tikys[1:]):\n tycalpha[ilo:ihi] = y\n # tycalpha[np.arange(tikx(j), tikx(j + 1))] = tiky(j + 1)\n\n f_out = f_in.copy()\n f_out.data -= tycalpha # input minus the next peak value\n f_out.data[f_out.data > 0] = 0 # Get rid of everything above zero\n if normalize:\n f_out.data /= abs(min(f_out.data))\n if debug:\n Stream([f_in, f_out]).plot()\n return f_out\n\n\ndef _loca_ext(trace, starttime, endtime, type_, debug=False):\n \"\"\"\n Returns local extrema\n\n Actually just returns where the trace slope changes from positive to\n negative (type_=='maxi'), or vice versa (type='mini')\n :param trace: waveform trace\n :param start_time: start of window to look at\n :param end_time: end of window to look at\n :param type_: 'maxi' or 'mini'\n :returns: indice, trace_value, diff_value\n \"\"\"\n diff = trace.copy()\n diff.data = np.diff(np.sign(np.diff(trace.data)))\n diff = diff.slice(starttime, endtime)\n if debug:\n diff.plot()\n if type_ == 'maxi':\n loc = (diff.data < 0)\n else:\n loc = (diff.data > 0)\n\n i_extremes = np.nonzero(loc)[0] + 1\n return i_extremes, trace.data[i_extremes]\n\n\ndef _mean_trace(traces):\n \"\"\"\n Calculate the mean trace\n\n :param traces: stream or list of traces\n :returns: trace object\n \"\"\"\n if isinstance(traces, Trace):\n return traces.copy()\n if isinstance(traces[0], Trace):\n if len(traces) == 1:\n return traces[0].copy()\n else:\n assert isinstance(traces[0][0], Trace),\\\n \"traces not a Trace, list of Traces, or list of lists of Traces\"\n traces = [t for x in traces for t in x]\n\n data_len = len(traces[0].data)\n for tr in traces[1:]:\n assert len(tr.data) == data_len, 'traces are not the same length'\n\n mean_tr = traces[0].copy()\n for tr in traces[1:]:\n mean_tr.data += tr.data\n mean_tr.data /= len(traces)\n return mean_tr\n\n\nif __name__ == '__main__':\n pass\n","sub_path":"build/lib/ps_picker/kurtosis.py","file_name":"kurtosis.py","file_ext":"py","file_size_in_byte":19446,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"446850005","text":"# Definition for a binary tree node.\n# class TreeNode:\n# def __init__(self, x):\n# self.val = x\n# self.left = None\n# self.right = None\n\nclass Solution:\n def minDiffInBST(self, root):\n \"\"\"\n :type root: TreeNode\n :rtype: int\n \"\"\"\n v = []\n def travel(tree):\n if not tree:\n return\n v.append(tree.val)\n travel(tree.left)\n travel(tree.right)\n travel(root)\n res = 1000000\n v.sort()\n for i in range(1, len(v)):\n if v[i] - v[i-1] < res:\n res = v[i] - v[i-1]\n if res == 1:\n break\n return res","sub_path":"181231/Minimum_Distance_Between_BST_Nodes.py","file_name":"Minimum_Distance_Between_BST_Nodes.py","file_ext":"py","file_size_in_byte":699,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"338666598","text":"from django.test import LiveServerTestCase\ntry:\n from django.urls import reverse\nexcept:\n from django.core.urlresolvers import reverse\n\n\nclass TestMainUsage(LiveServerTestCase):\n def setUp(self):\n from selenium import webdriver\n self.selenium = webdriver.Chrome()\n self.selenium.implicitly_wait(10)\n\n def test_pay(self):\n from selenium.common.exceptions import NoSuchElementException\n self.selenium.get(self.live_server_url + reverse('home') + '?order_id=1')\n self.assertRaises(NoSuchElementException, self.selenium.find_element_by_css_selector, '.errorlist') # noqa\n\n self.selenium.find_element_by_css_selector('[type=submit]').click()\n self.assertEquals(self.selenium.current_url, 'https://secure.futubank.com/testing-pay/') # noqa\n self.assertEquals(self.selenium.title, '[ТЕСТ] Оплата покупки')\n\n def tearDown(self):\n self.selenium.quit()\n\n\n","sub_path":"example/app/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":948,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"310719356","text":"#!/usr/bin/env python\n# -*- coding:utf-8 -*-\n\n#author: tp7309\n#usage:\n#python showjar.py *.apk/*.aar/*.dex/*.jar\n\n\nfrom __future__ import print_function\n\nimport os\nimport shutil\nimport subprocess\nimport sys\nimport zipfile\n\ndex2jar=\"dex2jar-2.1-SNAPSHOT\\d2j-dex2jar.bat\"\nif not os.name == 'nt': dex2jar = \"./dex2jar-2.1-SNAPSHOT/d2j-dex2jar.sh\"\njdgui=\"jd-gui-1.4.0.jar\"\napktool=\"apktool_2.3.0.jar\"\nneedUnzipFiles=['patch.jar']\nneedDecompileResources=0\n\n#unpack\n#java -jar apktool_2.2.4.jar d test.apk\n#repack\n#java -jar apktool_2.2.4.jar b test\n\n\ndef sh(command):\n p = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)\n print(p.stdout.read())\n\n\nif __name__==\"__main__\":\n f = sys.argv[1]\n # fix windows path\n if \":\\\\\" in f and not \":\\\\\\\\\" in f:\n f = f.replace(\"\\\\\", \"\\\\\\\\\")\n dexes = []\n jars = []\n if f.endswith(\".apk\") or f in needUnzipFiles:\n print(\"unzip %s...\"%(f))\n tempDir = os.path.splitext(f)[0]\n with zipfile.ZipFile(f, 'r') as zip:\n zip.extractall(tempDir)\n dexes = [f for f in os.listdir(tempDir) if f.endswith('.dex')]\n print(\"founded dexes: \" + ', '.join(dexes))\n for file in os.listdir(tempDir):\n if file.endswith('-dex2jar.jar') and os.path.exists(file):\n os.remove(file)\n for dex in dexes:\n sh(\"%s -f %s\"%(dex2jar, os.path.join(tempDir, dex)))\n jars.append(os.path.splitext(dex)[0] + \"-dex2jar.jar\")\n shutil.rmtree(tempDir)\n elif f.endswith(\".aar\"):\n print(\"unzip %s...\"%(f))\n tempDir = os.path.splitext(f)[0]\n with zipfile.ZipFile(f, 'r') as zip:\n \tzip.extractall(tempDir)\n dstPath = os.path.join(os.path.dirname(os.path.abspath(__file__)), \"classes.jar\")\n shutil.move(os.path.join(tempDir, \"classes.jar\"), dstPath)\n shutil.rmtree(tempDir)\n jars.append(\"classes.jar\")\t\n elif f.endswith(\".dex\"):\n sh(\"%s -f %s\"%(dex2jar, f))\n jars.append(os.path.splitext(f)[0] + \"-dex2jar.jar\")\n elif f.endswith(\".jar\"):\n jars.append(f)\n else:\n print(\"error file extension!\")\n exit\n\n if needDecompileResources and f.endswith(\".apk\") or f in needUnzipFiles:\n print(\"decompile resources...\")\n sh(\"java -jar %s d %s\"%(apktool, f))\n print(\"decompile resources done\")\n \n sh(\"java -jar %s %s\"%(jdgui, ' '.join(jars) if len(jars)> 0 else jars[0]))\n print(\"Done\")\n","sub_path":"showjar.py","file_name":"showjar.py","file_ext":"py","file_size_in_byte":2486,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"197161228","text":"from django.conf.urls import patterns, include, url\n\n# Uncomment the next two lines to enable the admin:\n# from django.contrib import admin\n# admin.autodiscover()\n\nurlpatterns = patterns('',\n url(r'^$', 'teacher_tool.views.index', name='index'),\n url(r'^my_cources/$', 'teacher_tool.views.my_cources_view', name='my_cources_list'),\n url(r'^groups/$', 'teacher_tool.views.groups_view', name='groups_list'),\n\n\n # url(r'^teacher_tool/', include('teacher_tool.foo.urls')),\n\n # Uncomment the admin/doc line below to enable admin documentation:\n # url(r'^admin/doc/', include('django.contrib.admindocs.urls')),\n\n # Uncomment the next line to enable the admin:\n # url(r'^admin/', include(admin.site.urls)),\n)\n","sub_path":"teacher_tool/teacher_tool/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":726,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"572862234","text":"# Default Imports\n\nimport pandas as pd\nimport numpy as np\n\nfrom greyatomlib.logistic_regression_project.q02_data_cleaning_all.build import data_cleaning\nfrom greyatomlib.logistic_regression_project.q01_outlier_removal.build import outlier_removal\n\nloan_data = pd.read_csv('data/loan_prediction_uncleaned.csv')\nloan_data = loan_data.drop('Loan_ID', 1)\nloan_data = outlier_removal(loan_data)\nX, y, X_train, X_test, y_train, y_test = data_cleaning(loan_data)\n\n\ndef data_cleaning_2(X_train,X_test,y_train,y_test) :\n\n for i in X_train.select_dtypes(include=[np.number]).columns.tolist() :\n np.sqrt(X_train[i])\n\n for i in X_test.select_dtypes(include=[np.number]).columns.tolist() :\n np.sqrt(X_test[i])\n\n X_test = pd.get_dummies(X_test, drop_first = True )\n #print(X_test.shape)\n\n X_train = pd.get_dummies(X_train, drop_first = True )\n\n #X_train.drop(mylist, axis=1, inplace = True)\n #print(X_train.shape)\n #print(list(X_test['Dependents_1'].value_counts()))\n\n return X_train, X_test, y_train, y_test\n\n# Write your solution here :\n","sub_path":"q02_data_cleaning_all_2/build.py","file_name":"build.py","file_ext":"py","file_size_in_byte":1066,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"261670648","text":"import os\nfrom flask import Flask, render_template, request, redirect, url_for\nfrom flask_sqlalchemy import SQLAlchemy\nfrom flask_wtf import FlaskForm\nfrom wtforms import StringField, SubmitField, RadioField\nfrom flask_migrate import Migrate\nfrom datetime import datetime\n\napp = Flask(__name__)\napp.config['SECRET_KEY'] = 'mysecretkey'\nbasedir = os.path.abspath(os.path.dirname(__file__))\napp.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:///' + os.path.join(basedir,'data.sqlite')\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\ndb = SQLAlchemy(app)\nMigrate(app,db)\n\ndef decode_resp(s):\n pairs_with_equalto = s.split('&')\n resp = {}\n for item in pairs_with_equalto:\n key, value = item.split('=')\n resp[key] = value \n return resp\ndef day_in_month(month,year):\n month = int(month)\n year = int(year)\n if ((year%100!=0 and year%4==0) or (year%400==0)) and month == 2:\n return 29\n elif month==2:\n return 28\n elif month in [1,3,5,7,8,10,12]:\n return 31\n else:\n return 30\n\ndef blank_divs(year,month):\n return ['' for i in range(int(datetime(int(year),int(month),1).strftime(\"%w\")))]\n\ndef month_name(month):\n return datetime(2020,int(month),1).strftime(\"%B\")\n\nclass GoTo(FlaskForm):\n year = StringField('Year')\n month = StringField('Month')\n submit = SubmitField('GoTo')\n\nclass Task(db.Model):\n id = db.Column(db.Integer, primary_key=True)\n year = db.Column(db.String)\n month = db.Column(db.String)\n day = db.Column(db.String)\n task = db.Column(db.Integer,default=0)\n def __init__(self,year,month,day,task):\n self.year = year\n self.month = month\n self.day = day\n self.task = task\n\n\n\n\n\n@app.route('/', methods=['GET','POST'])\ndef index():\n datalist = []\n year = str(datetime.now().strftime(\"%Y\"))\n month = str(datetime.now().strftime(\"%m\"))\n x_home = str(year) + str(int(month))\n days = day_in_month(month,year) \n form = GoTo()\n if request.method=='POST':\n try:\n # Insert new data\n data = decode_resp(request.data.decode('ascii'))\n db.session.add(Task(data['year'],data['month'],data['day'], 1))\n db.session.commit()\n except:\n # goto another month\n if form.validate_on_submit():\n year = form.year.data \n month = form.month.data \n \n \n for day in range(1,days+1):\n x = Task.query.filter_by(day=day,month=str(int(month)),year=(year)).first()\n \n if x:\n datalist.append([day,x.task])\n \n else:\n \n datalist.append([day,0])\n \n y_home = str(year) + str(int(month))\n return render_template('calender.html', year=year, month=int(month), homebutton=(x_home==y_home) ,fullNameOfMonth=month_name(month),datalist=datalist, form=form, blank_divs=blank_divs(year,month))\n\n\nif __name__ == \"__main__\":\n app.run(debug=True)","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":2965,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"61636452","text":"import re\nimport os\nimport sys\nimport logging\nimport functools\nfrom tests.common.devices.base import AnsibleHostBase\nfrom ansible.utils.unsafe_proxy import AnsibleUnsafeText\n\n# If the version of the Python interpreter is greater or equal to 3, set the unicode variable to the str class.\nif sys.version_info[0] >= 3:\n unicode = str\n\nlogger = logging.getLogger(__name__)\n\nSAMPLE_COMMAND_DATA = '''\nRP/0/RP0/CPU0:vlab-01#show operational LLDP NodeTable Node/NodeName/Rack=0;\nSlot=0;Instance=CPU0 Neighbors DeviceTable Device/DeviceID=vlab-02/Interf$\nWed Aug 10 08:45:43.126 UTC\n......\n \"Operational\": {\n \"LLDP\": {\n \"@MajorVersion\": \"1\",\n \"@MinorVersion\": \"2\",\n \"NodeTable\": {\n \"Node\": {\n \"Naming\": {\n \"NodeName\": {\n \"Rack\": \"0\",\n \"Slot\": \"0\",\n \"Instance\": \"CPU0\"\n }\n },\n \"Neighbors\": {\n \"DeviceTable\": {\n \"Device\": {\n \"Naming\": {\n \"DeviceID\": \"vlab-02\",\n \"InterfaceName\": \"GigabitEthernet0/0/0/1\"\n },\n \"Entry\": {\n \"ReceivingInterfaceName\": \"GigabitEthernet0/0/0/1\",\n \"ReceivingParentInterfaceName\": \"Bundle-Ether1\",\n \"DeviceID\": \"vlab-02\",\n \"ChassisID\": \"5254.0085.5c1c\",\n \"PortIDDetail\": \"fortyGigE0/4\",\n \"HeaderVersion\": \"0\",\n \"HoldTime\": \"120\",\n \"EnabledCapabilities\": \"B,R\",\n \"Detail\": {\n......\n'''\n\n\ndef adapt_interface_name(func):\n \"\"\"Decorator to adapt interface name used in topology to cisco interface name.\"\"\"\n @functools.wraps(func)\n def _decorated(self, *args):\n args_list = list(args)\n new_list = []\n for item in args_list:\n new_item = item\n if isinstance(new_item, str) or isinstance(new_item, unicode) or isinstance(new_item, AnsibleUnsafeText):\n if 'Ethernet' in new_item and 'GigabitEthernet' not in new_item:\n new_item = re.sub(r'(^|\\s)Ethernet', 'GigabitEthernet0/0/0/', new_item)\n elif 'Port-Channel' in new_item:\n new_item = re.sub(r'(^|\\s)Port-Channel', 'Bundle-Ether', new_item)\n new_list.append(new_item)\n new_args = tuple(new_list)\n return func(self, *new_args)\n return _decorated\n\n\nclass CiscoHost(AnsibleHostBase):\n \"\"\"\n @summary: Class for Cisco host\n \"\"\"\n def __init__(self, ansible_adhoc, hostname, ansible_user, ansible_passwd):\n '''Initialize an object for interacting with cisco device using ansible modules\n Args:\n ansible_adhoc (): The pytest-ansible fixture\n hostname (string): hostname of the cisco device\n ansible_user (string): Username for accessing the cisco CLI interface\n ansible_passwd (string): Password for the ansible_user\n '''\n self.ansible_user = ansible_user\n self.ansible_passwd = ansible_passwd\n AnsibleHostBase.__init__(self, ansible_adhoc, hostname)\n # Reserved for execute ansible commands in local device\n self.localhost = ansible_adhoc(inventory='localhost', connection='local', host_pattern=\"localhost\")[\"localhost\"]\n\n def __getattr__(self, module_name):\n if module_name.startswith('iosxr_'):\n evars = {\n 'ansible_connection': 'network_cli',\n 'ansible_network_os': module_name.split('_', 1)[0],\n 'ansible_user': self.ansible_user,\n 'ansible_password': self.ansible_passwd,\n 'ansible_ssh_user': self.ansible_user,\n 'ansible_ssh_pass': self.ansible_passwd,\n }\n else:\n raise Exception(\"Does not have module: {}\".format(module_name))\n self.host.options['variable_manager'].extra_vars.update(evars)\n return super(CiscoHost, self).__getattr__(module_name)\n\n def __str__(self):\n return ''.format(self.hostname)\n\n def __repr__(self):\n return self.__str__()\n\n def commands(self, *args, **kwargs):\n return self.iosxr_command(*args, **kwargs)\n\n def config(self, *args, **kwargs):\n return self.iosxr_config(*args, **kwargs)\n\n @adapt_interface_name\n def shutdown(self, interface_name=None):\n out = self.config(\n lines=['shutdown'],\n parents=['interface {}'.format(interface_name)])\n logging.info('Shut interface [%s]' % interface_name)\n return out\n\n def shutdown_multiple(self, interfaces):\n intf_str = ','.join(interfaces)\n return self.shutdown(interface_name=intf_str)\n\n @adapt_interface_name\n def no_shutdown(self, interface_name):\n out = self.config(\n lines=['no shutdown'],\n parents=['interface {}'.format(interface_name)])\n logging.info('No shut interface [%s]' % interface_name)\n return out\n\n def no_shutdown_multiple(self, interfaces):\n intf_str = ','.join(interfaces)\n return self.no_shutdown(intf_str)\n\n @adapt_interface_name\n def rm_member_from_channel_grp(self, interface_name, channel_group):\n out = self.config(\n lines=['no bundle id {} mode active'.format(channel_group)],\n parents=['interface {}'.format(interface_name)])\n logging.info('Rm interface {} from bundle-ethernet {}'.format(interface_name, channel_group))\n return out\n\n @adapt_interface_name\n def add_member_to_channel_grp(self, interface_name, channel_group):\n out = self.config(\n lines=['bundle id {} mode active'.format(channel_group)],\n parents=['interface {}'.format(interface_name)])\n logging.info('Add interface {} to bundle-ethernet {}'.format(interface_name, channel_group))\n return out\n\n @adapt_interface_name\n def check_intf_link_state(self, interface_name):\n show_int_result = self.commands(\n commands=['show interfaces %s' % interface_name])\n return 'line protocol is up' in show_int_result['stdout_lines'][0]\n\n @adapt_interface_name\n def set_interface_lacp_rate_mode(self, interface_name, mode):\n if mode == 'fast':\n command = 'lacp period short'\n else:\n command = 'no lacp period'\n\n out = self.config(\n lines=[command],\n parents='interface %s' % interface_name)\n return out\n\n def get_lldp_neighbor(self, local_iface=None, remote_device=None):\n try:\n if (local_iface is not None and remote_device is not None):\n command = 'show operational LLDP NodeTable ' \\\n 'Node/NodeName/Rack=0;Slot=0;Instance=CPU0 Neighbors DeviceTable ' \\\n 'Device/DeviceID={}/InterfaceName={} json'.format(local_iface, remote_device)\n else:\n command = 'show operational LLDP json'\n output = self.commands(\n commands=[command],\n module_ignore_errors=True)\n logger.debug('cisco lldp output: %s' % (output))\n return output['stdout_lines'][0]['Response']['Get']['Operational'] if output['failed'] is False else False\n except Exception as e:\n logger.error('command {} failed. exception: {}'.format(command, repr(e)))\n return False\n\n def config_key_chain(self, name, key):\n # create key chain\n output = self.config(\n lines=['key chain {} key 1'.format(name)])\n logger.debug('config key chain: %s' % (output))\n\n # configure key chain parameters\n output = self.config(\n lines=['accept-lifetime 00:00:00 december 01 2014 infinite',\n 'send-lifetime 00:00:00 december 01 2014 infinite',\n 'cryptographic-algorithm HMAC-MD5',\n 'key-string clear {}'.format(key)],\n parents=['key chain {} key 1'.format(name)])\n logger.debug('config key chain parameters: %s' % (output))\n\n def remove_key_chain(self, name):\n # remove key chain\n output = self.config(lines=['no key chain {}'.format(name)])\n logger.debug('remove key chain: %s' % (output))\n\n def isis_config_auth(self, key):\n key_chain_name = 'ISIS'\n self.config_key_chain(key_chain_name, key)\n\n # configure key chain to isis\n output = self.config(\n lines=['lsp-password keychain {} level 2'.format(key_chain_name),\n 'interface Bundle-Ether1 hello-password keychain {}'.format(key_chain_name)],\n parents=['router isis test'])\n logger.debug('config key chain to isis: %s' % (output))\n\n def isis_remove_auth(self, key):\n key_chain_name = 'ISIS'\n # remove key chain from isis\n output = self.config(\n lines=['no lsp-password keychain {} level 2'.format(key_chain_name),\n 'no interface Bundle-Ether1 hello-password keychain {}'.format(key_chain_name)],\n parents=['router isis test'])\n logger.debug('remove key chain from isis: %s' % (output))\n\n self.remove_key_chain(key_chain_name)\n\n def ping_dest(self, dest):\n try:\n command = 'ping {} count 5'.format(dest)\n output = self.commands(commands=[command])\n logger.debug('ping result: %s' % (output))\n return re.search('!!!!!', output['stdout'][0]) is not None if output['failed'] is False else False\n except Exception as e:\n logger.error('command {} failed. exception: {}'.format(command, repr(e)))\n return False\n\n def show_command_to_json(self, command, lookup_key=None, lookup_val=None):\n \"\"\"\n This function will pull the show operational command output as json string and convert it json object and return\n \"\"\"\n try:\n json_command = command + \" json\"\n output = self.commands(commands=[json_command])\n if all([lookup_key, lookup_val]):\n return self.extract_key_val_pair_from_json(output['stdout_lines'], lookup_key)\n elif lookup_key is not None and lookup_val is None:\n return self.extract_val_from_json(output['stdout_lines'], lookup_key)\n else:\n return output['stdout_lines']\n except Exception as e:\n return {\"error\": e}\n\n def extract_key_val_pair_from_json(self, data, lookup_key):\n \"\"\"\n Function to recursivly match provided key in all levels and return list of same level data\n \"\"\"\n result = []\n\n def help(data, lookup_key, result):\n if isinstance(data, dict):\n for k, v in data.items():\n if k == lookup_key:\n result.append(data)\n elif isinstance(v, (list, dict)):\n sub_result = help(v, lookup_key, result)\n if sub_result:\n result.append(sub_result)\n elif isinstance(data, list):\n for ele in data:\n if isinstance(ele, (list, dict)):\n sub_result = help(ele, lookup_key, result)\n if sub_result:\n result.append(sub_result)\n help(data, lookup_key, result)\n return result\n\n def extract_val_from_json(self, json_data, lookup_key):\n \"\"\"\n Function to recursivly match provided key in all levels and return matched key's value into a list\n \"\"\"\n result = []\n\n def help(data, lookup_key, result):\n if isinstance(data, dict):\n for k, v in data.items():\n if k == lookup_key:\n result.append(v)\n elif isinstance(v, (list, dict)):\n sub_result = help(v, lookup_key, result)\n if sub_result:\n result.append(sub_result)\n elif isinstance(data, list):\n for ele in data:\n if isinstance(ele, (list, dict)):\n sub_result = help(ele, lookup_key, result)\n if sub_result:\n result.append(sub_result)\n help(json_data, lookup_key, result)\n return result\n\n def _has_cli_cmd_failed(self, cmd_output_obj):\n err_out = False\n if 'stdout' in cmd_output_obj:\n stdout = cmd_output_obj['stdout']\n msg = stdout[-1] if type(stdout) == list else stdout\n err_out = 'Cannot advertise' in msg\n\n return ('failed' in cmd_output_obj and cmd_output_obj['failed']) or err_out\n\n def load_configuration(self, config_file, backup_file=None):\n if backup_file is None:\n out = self.config(\n src=config_file,\n replace='config',\n )\n else:\n out = self.config(\n src=config_file,\n replace='line',\n backup='yes',\n backup_options={\n 'filename': os.path.basename(backup_file),\n 'dir_path': os.path.dirname(backup_file),\n }\n )\n return not self._has_cli_cmd_failed(out)\n\n @adapt_interface_name\n def get_portchannel_by_member(self, member_intf):\n try:\n command = 'show lacp {}'.format(member_intf)\n output = self.commands(commands=[command])['stdout'][0]\n regex_pc = re.compile(r'Bundle-Ether([0-9]+)', re.U)\n for line in [item.strip().rstrip() for item in output.splitlines()]:\n if regex_pc.match(line):\n return re.sub('Bundle-Ether', 'Port-Channel', line)\n except Exception as e:\n logger.error('Failed to get PortChannel for member interface \"{}\", exception: {}'.format(\n member_intf, repr(e)\n ))\n return None\n\n @adapt_interface_name\n def no_isis_interface(self, isis_instance, interface):\n out = self.config(\n lines=['no interface {}'.format(interface)],\n parents=['router isis {}'.format(isis_instance)])\n return not self._has_cli_cmd_failed(out)\n","sub_path":"tests/common/devices/cisco.py","file_name":"cisco.py","file_ext":"py","file_size_in_byte":14492,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"227907602","text":"# -*- coding: utf-8 -*-\nimport time\nimport codecs\nimport math\nf_vectors = codecs.open('text_output_standard.txt', 'r', 'utf8')\nf_results = open('min_resWordsWord2vec.txt', 'w')\nf_group = codecs.open('phrases(200,250).txt', 'r', 'utf8')\ngroup = f_group.read().split()\ndict = {}\nfor i in range(len(group)//4):\n dict[group[4 * i + 1] + '_' + group[4 * i + 2]] = 0\n dict[group[4 * i + 1]] = 0\n dict[group[4 * i + 2]] = 0\nword_vector = f_vectors.readline().split()\nstart = time.time()\nwhile len(word_vector) != 0:\n if word_vector[0] in dict:\n dict[word_vector[0]] = word_vector[1:]\n word_vector = f_vectors.readline().split()\nprint (\"--- %s seconds ---\" % (time.time() - start))\nresults = {}\nfor i in range(len(group)//4):\n sq1, sq2, sq3 = 0.0, 0.0, 0.0\n #vector1 = dict[group[4 * i + 1] + '_' + group[i * 4 + 2]]\n vector2 = dict[group[4 * i + 1]]\n vector3 = dict[group[4 * i + 2]]\n if vector2 == 0 or vector3 == 0:\n results[group[4 * i] + ' ' + group[4 * i + 1] + ' ' + group[i * 4 + 2]] = 0.0\n continue\n for j in range(len(vector2)):\n #sq1 += float(vector1[j]) ** 2\n sq2 += float(vector2[j]) ** 2\n sq3 += float(vector3[j]) ** 2\n #sq1 = math.sqrt(sq1)\n sq2, sq3 = math.sqrt(sq2), math.sqrt(sq3)\n sum = 0.0\n sum_vector= 0.0\n for j in range(len(vector2)):\n sum += (float(vector2[j]) / sq2 + float(vector3[j]) / sq3) ** 2\n #sum_vector += float(vector1[j]) * (float(vector2[j]) / sq2 + float(vector3[j]) / sq3)\n sum_vector += float(vector2[j]) * float(vector3[j])\n #f_results.write(words_group[3 * i] + u' : ' + str(sum_vector/ (sq1 * math.sqrt(sum))) + '\\n')\n results[group[4 * i] + ' ' + group[4 * i + 1] + ' ' + group[i * 4 + 2]] = math.fabs(sum_vector/ (math.sqrt(sum)))\nfor i in sorted(results, key=results.__getitem__, reverse=False):\n f_results.write(str(i) + ' ' + str(results[i]) + '\\n')\nf_vectors.close()\nf_results.close()\nf_group.close()","sub_path":"src/2_step.py","file_name":"2_step.py","file_ext":"py","file_size_in_byte":1960,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"317682374","text":"import os\nimport requests\nfrom bs4 import BeautifulSoup\n\ndef dl_img(user, current, limit):\n images = []\n while current <= limit:\n url = 'http://%s.tumblr.com/mobile/page/%d' % (user, current)\n source_code = requests.get(url)\n raw_text = source_code.text\n soup = BeautifulSoup(raw_text)\n for image in soup.find_all('img'):\n src = image.get('src')\n images.append(src)\n if not os.path.exists(user):\n os.makedirs(user)\n for image in images:\n dl = requests.get(image)\n content = dl.content\n fo = open('%s/%s' % (user, image[-34:]), 'wb')\n fo.write(content)\n fo.close()\n current += 1\ndef dl_txt(user, current, limit):\n text = []\n while current <= limit:\n url = 'http://%s.tumblr.com/mobile/page/%d' % (user, current)\n source_code = requests.get(url)\n raw_text = source_code.text\n soup = BeautifulSoup(raw_text)\n if not os.path.exists(user):\n os.makedirs(user)\n content = soup.get_text()\n fo = open('%s/%d.txt' % (user, current), 'w', encoding = 'utf8')\n fo.write(content)\n fo.close()\n current += 1\ndef num_pages(user):\n num = 0\n page = 1\n while True:\n url = 'http://%s.tumblr.com/mobile/page/%d' % (user, page)\n source_code = requests.get(url)\n raw_text = source_code.text\n soup = BeautifulSoup(raw_text)\n if len(soup.find_all('img')) > 0:\n num += 1\n else:\n break\n page += 1\n return num\n \n \n","sub_path":"crawlblr.py","file_name":"crawlblr.py","file_ext":"py","file_size_in_byte":1615,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"620388897","text":"#!/usr/bin/env python\nimport sys\nimport os\nimport pandas as pd\nimport matplotlib\nimport getpass\nimport uuid\nimport argparse\nmatplotlib.use('agg')\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport matplotlib.colors as clr\nimport numpy as np\nfrom matplotlib import cm\nfrom matplotlib.colors import ListedColormap, LinearSegmentedColormap\nimport scipy\nimport glob\nimport datetime\n\n\"\"\"Plot heatmap given data frame\n\n\n\"\"\"\ncurrent_file_base_name = __file__.split(\"/\")[-1].split(\".\")[0]\n\ndef my_args():\n\tusername = getpass.getuser()\n\taddon_string = str(uuid.uuid4()).split(\"-\")[-1]\n\tmainParser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter,description=\"plot heatmap given dataframe.\")\n\tmainParser.add_argument('-f',\"--input\", help=\"data table input\",required=True)\n\tmainParser.add_argument(\"--sort_by\", help=\"usually this column should be logFC\",default=None)\n\tmainParser.add_argument('-c1',\"--value_cutoff_max\", help=\"subset genes by filtering out values above the cutoff, usually it should be p-value, input should be: column_name,value\",default=None)\n\tmainParser.add_argument('-c2',\"--value_cutoff_min\", help=\"subset genes by filtering out values below the cutoff, usually it should be log p-value, input should be: column_name,value\",default=None)\n\tmainParser.add_argument(\"--index_by\", help=\"if the first column is not the correct index, which one?\",default=None)\n\tmainParser.add_argument(\"--remove_cols\",default=\"\")\n\tmainParser.add_argument('-o',\"--output\", help=\"output table name\",default=\"gene_exp_heatmap.\"+username+\"_\"+str(datetime.date.today()))\n\tmainParser.add_argument(\"-pdf\", help=\"plot pdf instead of png, can be slower for large dataset\", action='store_true')\n\tmainParser.add_argument(\"-W\", \"--width\", help=\"Figure width, maximal use 200, usually 8 to 20\",type=int,required=True)\n\tmainParser.add_argument(\"-H\", \"--height\", help=\"Figure height, maximal use 200, usually 10 to 50\",type=int,required=True)\n\tmainParser.add_argument(\"--fontsize\", help=\"you can choose from 8 to 20 \",type=int,default=10)\n\tmainParser.add_argument(\"--linewidths\", help=\"you can choose from 0, 0.1 \",type=float,default=0.1)\n\tmainParser.add_argument(\"--log2_transform\", help=\"input values will be log2 transformed\", action='store_true')\n\tmainParser.add_argument(\"--show_name\", help=\"by default, >100 genes, name will not be shown, this option enforce to show name\", action='store_true')\n\t\n\t\n\n\t##------- add parameters above ---------------------\n\targs = mainParser.parse_args()\t\n\treturn args\ndef guess_sep(x):\n\twith open(x) as f:\n\t\tfor line in f:\n\t\t\ttmp1 = len(line.strip().split(\",\"))\n\t\t\ttmp2 = len(line.strip().split(\"\\t\"))\n\t\t\t# print (tmp1,tmp2)\n\t\t\tif tmp1 > tmp2:\n\t\t\t\treturn \",\"\n\t\t\tif tmp2 > tmp1: \n\t\t\t\treturn \"\\t\"\n\t\t\telse:\n\t\t\t\tprint (\"Can't determine the separator. Please input manually\")\n\t\t\t\texit()\n\t\t\t\t\n\ndef plot_n_save(df,args):\n\tdf_z = df.T\n\tfor c in df_z:\n\t\tdf_z[c] = (df_z[c] - df_z[c].mean())/df_z[c].std()\n\tplt.figure(figsize=(args.width,args.height))\n\tif df_z.T.shape[0]<=100 or args.show_name:\n\t\tg=sns.heatmap(df_z.T,vmin=-2,vmax=2,cmap=\"RdBu_r\",linewidths=args.linewidths,yticklabels=df_z.columns.tolist())\n\t\tg.set_yticklabels(g.get_ymajorticklabels(), fontsize = args.fontsize)\n\telse:\n\t\tg=sns.heatmap(df_z.T,vmin=-2,vmax=2,cmap=\"RdBu_r\",yticklabels=\"\")\n\tif args.pdf:\n\t\tplt.savefig(\"%s_heatmap.pdf\"%(args.output), bbox_inches='tight')\n\telse:\n\t\tplt.savefig(\"%s_heatmap.png\"%(args.output), bbox_inches='tight')\n\ndef main():\n\n\targs = my_args()\n\t\"\"\"below is the same for very dataframe scripts\n\t\n\tby default our input dataframe is bed format, which is \\t separated with no header no index \n\t\n\t\"\"\"\n\n\tdf = pd.read_csv(args.input,sep=guess_sep(args.input),index_col=0)\n\tif args.value_cutoff_max:\n\t\tname,value = args.value_cutoff_max.split(\",\")\n\t\tvalue = float(value)\n\t\tdf = df[df[name]<=value]\n\t\tdf = df.drop([name],axis=1)\n\tif args.value_cutoff_min:\n\t\tname,value = args.value_cutoff_min.split(\",\")\n\t\tvalue = float(value)\n\t\tdf = df[df[name]>=value]\n\t\tdf = df.drop([name],axis=1)\n\tif args.index_by:\n\t\tdf = df.set_index(args.index_by)\n\tif args.args.sort_by:\n\t\tdf = df.sort_values(args.sort_by)\n\t\tdf = df.drop([args.sort_by],axis=1)\n\t#-------------- pre-processing ----------------------\n\tremove_cols = str(args.remove_cols).split(\",\")\n\ttry:\n\t\tremove_cols.remove(\"\")\n\texcept:\n\t\tpass\n\tif len(remove_cols)>0:\n\t\tdf = df.drop(remove_cols,axis=1)\n\tif args.log2_transform:\n\t\tdf = df.transform(lambda x:np.log2(x+1))\n\tprint (\"ploting size:\",df.shape)\n\tplot_n_save(df,args)\n\n\nif __name__ == \"__main__\":\n\tmain()\n\n\n\n\n","sub_path":"bin/plot_gene_exp_heatmap.py","file_name":"plot_gene_exp_heatmap.py","file_ext":"py","file_size_in_byte":4553,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"401636093","text":"n = int(input())\na = list(map(int, input().split()))\ncash = 1000\nbill = 0\ntable = []\n\nfor i in range(n-1):\n if a[i] > a[i+1]:\n table.append(False)\n elif a[i] == a[i+1]:\n table.append(\"-\")\n else:\n table.append(True)\n# print(table)\nif table[0] == True:\n bill, cash = divmod(cash, a[0])\n# print(cash, bill)\ni = 0\nwhile i < n-1:\n if table[i] == True:\n while not table[i] == False:\n i += 1\n if i == n-1:\n break\n cash += bill * a[i]\n bill = 0\n # print(cash, bill)\n else:\n while not table[i] == True:\n i += 1\n if i == n-1:\n break\n bill, cash = divmod(cash, a[i])\n # print(cash, bill)\ncash += bill * a[-1]\nprint(cash)\n","sub_path":"ABC/M-SOLUTIONS2020/d.py","file_name":"d.py","file_ext":"py","file_size_in_byte":772,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"537540820","text":"#! python\n\n\"\"\"this module defines functions for parsing c++ enums\"\"\"\n\nimport urllib.request\nimport re\nimport datetime\nfrom tooling.enum_converter import CxxEnum\n\n__author__ = \"Sebastiaan Saarloos\"\n__date__ = datetime.date(2019, 6, 3)\n__status__ = \"Developing\"\n__maintainer__ = \"Isha Geurtsen\"\n\n\ndef get_enum_strings(file_contents: str) -> list:\n \"\"\"Scrapes all the enum strings from a given string\"\"\"\n # Remove all comments from the file.\n file_contents = re.sub(r\"//.*\\n\", \"\", file_contents)\n # Remove redundant spaces\n file_contents = re.sub(r\"\\s+\", \" \", file_contents)\n # enum class regular expression\n pattern = re.compile(\"enum class\\\\s*\\\\w+\\\\s*\\\\:\\\\s*\\\\w+\\\\s*\\\\{[^\\\\{\\\\}]+\\\\};\")\n # Iterates over all the matches, and will yield returned\n for item in pattern.finditer(file_contents):\n # Yield, means: give back a single item, go further a new item is requested\n yield item.group(0).replace(\"\\n\", \"\")\n\n\ndef get_enum_definition(enum_string: str) -> CxxEnum:\n \"\"\"Converts a enum definition represented as a string to a CXX enum object\"\"\"\n # Get the name of the enum\n name = re.findall(r\"(?<=enum class )\\w+\", enum_string)[0]\n # Get the defined type of the enum\n inner_type = re.findall(r\":\\s*(\\w+)\", enum_string)[0]\n # Get all the items of that are defined inside the enum class\n inner_items = re.findall(\n r\"\\{([^{}]+)\\}\", enum_string)[0].replace(\" \", \"\").split(\",\")\n return CxxEnum(name, inner_type, dict(zip(inner_items, [None]*len(inner_items))))\n\n\ndef get_github_file(repository: str, branch: str, file: str):\n \"\"\"Helper function to get a github file from a given repository\"\"\"\n url = (\n \"https://raw.githubusercontent.com/R2D2-2019/\" +\n f\"{repository}/{branch}/{file}\")\n return (\n urllib.request.urlopen(url)\n .read()\n .decode(\"utf-8\")\n )\n\n\ndef get_enum_definitions() -> list:\n \"\"\"collect enums defined in the frame_enums.hpp from the external communications file\"\"\"\n # Get the file from GitHub\n file_content = get_github_file(\n \"internal_communication\",\n \"master\",\n \"code/headers/frame_enums.hpp\"\n )\n # Seperate file in seperate enum strings\n enum_strings = get_enum_strings(file_content)\n # Loop over all enum strings\n for enum_string in enum_strings:\n yield get_enum_definition(enum_string)\n\n\nif __name__ == \"__main__\":\n for definition in get_enum_definitions():\n print(definition)\n","sub_path":"tooling/enum_parser.py","file_name":"enum_parser.py","file_ext":"py","file_size_in_byte":2474,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"473860626","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sat Feb 27 11:25:46 2021\r\n\r\n@author: ChrisLe\r\n\"\"\"\r\n\r\nimport sys\r\nfrom PyQt5.QtWidgets import QMainWindow, QAction, qApp, QApplication\r\nfrom PyQt5.QtGui import QIcon\r\n\r\nclass Example(QMainWindow):\r\n def __init__(self):\r\n super().__init__()\r\n self.initUI()\r\n \r\n def initUI(self):\r\n exitAct = QAction (QIcon ('exit.png'), '&Salir', self)\r\n exitAct.setShortcut('Ctrl+Q')\r\n exitAct.setStatusTip('Salir de la Aplicación')\r\n exitAct.triggered.connect(qApp.quit)\r\n \r\n self.statusBar()\r\n menubar = self.menuBar()\r\n fileMenu = menubar.addMenu('&File')\r\n fileMenu.addAction(exitAct)\r\n \r\n self.setGeometry(400, 400, 350, 250)\r\n self.setWindowTitle('Menu Simple')\r\n self.show()\r\n \r\ndef main():\r\n app = QApplication(sys.argv)\r\n ex = Example()\r\n sys.exit(app.exec_())\r\n\r\nif __name__=='__main__':\r\n main()\r\n \r\n ","sub_path":"PyQt5Tutorial/menus-barrasHerramientas/ej8.py","file_name":"ej8.py","file_ext":"py","file_size_in_byte":984,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"479238212","text":"from RemoveDuplicatesfromSortedList import ListNode\nclass SortLi:\n def sortList(self, head: ListNode) -> ListNode:\n lis=[]\n while head:\n lis.append(head.val)\n head=head.next\n lis.sort()\n p=ListNode(0)\n s=p\n for i in range(len(lis)):\n p.next=ListNode(lis[i])\n p=p.next\n p.next=None\n return s.next\n","sub_path":"SolutionsPY/SortList.py","file_name":"SortList.py","file_ext":"py","file_size_in_byte":399,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"22923430","text":"import random\n\ndef weighted_choice(weights):\n totals = []\n running_total = 0\n\n for w in weights:\n running_total += w\n totals.append(running_total)\n\n rnd = random.random() * running_total\n for i, total in enumerate(totals):\n if rnd < total:\n return i\n\nfor i in range(10):\n weights= [0.2,0.3,0.5]\n list = weighted_choice(weights)\n print (list)\n\n#zipcodef = open(\"zipcodes.txt\",\"r\")\n#for lines in zipcodef:\n# values = lines.split()\n# number = random.randrange(int(values[1]),int(values[2]))\n# print (values[0],number)\n","sub_path":"rand03.py","file_name":"rand03.py","file_ext":"py","file_size_in_byte":592,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"495253167","text":"#!/usr/bin/python\n\n# (1) http://rosalind.info/problems/iprb/\n# (2) https://en.wikipedia.org/wiki/Punnett_square\n\n\ndef punnett_square(genotype_a, genotype_b):\n grid = []\n for i in genotype_a:\n for j in genotype_b:\n grid.append(i+j)\n\n total = 4\n dominant = sum(1 for l in grid if l.isupper())\n recessive = sum(1 for l in grid if l.islower())\n heterozygous = total-dominant-recessive\n return {\"not-recessive\": (dominant/4.0)+(heterozygous/4.0), \"recessive\": recessive/4.0}\n\n\ndef comb(elements):\n return (elements * (elements - 1))/2\n\n\ndef mendels_first_law(k, m, n):\n \"\"\"\n k individuals are homozygous dominant, m are heterozygous, and n are homozygous recessive.\n\n >>> mendels_first_law(2,2,2)\n '0.78333'\n \"\"\"\n ps = punnett_square \n a = ps(\"AA\", \"AA\")['not-recessive'] * comb(k)\n b = ps(\"AA\", \"Bb\")['not-recessive'] * (k*m)\n c = ps(\"AA\", \"cc\")['not-recessive'] * (k*n)\n d = ps(\"Bb\", \"Bb\")['not-recessive'] * comb(m)\n e = ps(\"Bb\", \"cc\")['not-recessive'] * (m*n)\n f = ps(\"cc\", \"cc\")['not-recessive'] * comb(n)\n return \"%.5f\" % ((a+b+c+d+e+f)/comb(k+m+n))\n\n\nif __name__ == \"__main__\":\n print(\"Running test case...\")\n import doctest\n doctest.testmod()","sub_path":"iprb/__main__.py","file_name":"__main__.py","file_ext":"py","file_size_in_byte":1238,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"643438328","text":"#!/usr/bin/python\n\n# Basic Ship Class\n\nclass ship:\n def __init__(self):\n # What is the name of the ship\n self.name = \"\"\n # How much Exp does this ship have\n self.exp = \"\"\n # What level is the ship\n self.level = \"\"\n # Max Level\n self.max_level = \"\"\n # What id number is the ship\n self.ident = \"\"\n # Rarity\n self.rarity = \"\"\n # How Many HP the ship has\n self.shield_points = \"\"\n # How much HP is recovered when heal gems are matched\n self.repair_rate = \"\"\n # Percentage of incoming damage reduced\n self.damage_reduction = \"\"\n # Type of damage that gets reduced\n self.damage_reduction_type = \"\"\n # How much damage the ship does when damage gems are matched\n self.attack_value = \"\"\n # How many turns between skill deployment\n self.speed = \"\"\n # Skill ability the ship possesses\n self.skill = \"\"\n # How much damage is increased or decreased via modifications\n self.weapon_mod = \"\"\n # How much armor is increased or decreased via modifications\n self.armor_mod = \"\"\n # How much shield is increased or decreased via modifications\n self.shield_mod = \"\"\n # How much skills speed is increased or decreased via modifications\n self.engine_mod = \"\"\n # What type of damage does this ship do\n self.damage_type = \"\"\n # How many weapon hardpoints does this ship have\n self.weapon_hardpoint = \"\"\n # How many armor hardpoints does this ship have\n self.armor_hardpoint = \"\"\n # How many shield hardpoints does this ship have\n self.shield_hardpoint = \"\"\n # How many engine hardpoints does this have\n self.engine_hardpoint = \"\"\n # What skill does this ship confer when assigned as wing leader\n self.wing_leader_skill = \"\"","sub_path":"Classes/ships.py","file_name":"ships.py","file_ext":"py","file_size_in_byte":1922,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"362415939","text":"\n# -*- coding: utf-8 -*-\nfrom bs4 import BeautifulSoup\nimport requests\nimport json\n\nresults = []\nurl_list = []\nresponse = requests.get('https://www.borenius.com/references/')\nsoup = BeautifulSoup(response.content, \"html.parser\")\nurl_base = 'https://www.borenius.com'\nall_anchors = soup.find_all('a', href=True)\n\nfor single_anchor in all_anchors:\n link = single_anchor['href']\n if '/' == link[:1] and '//' != link[:2]:\n link = url_base + link\n if link not in url_list:\n url_list.append(link)\n elif 'www.borenius.com' in link[:len(url_base)]:\n if link not in url_list:\n url_list.append(link)\n\nfor single_url in url_list:\n try:\n response = requests.get(single_url)\n soup = BeautifulSoup(response.content, \"html.parser\")\n page_title = soup.find_all('h1')[0].text\n page_content = soup.find_all(class_='col-sm-12 col-md-6 content-container')[0].text\n\n parsed_results = {}\n parsed_results['title'] = page_title\n parsed_results['content'] = page_content\n parsed_results['link'] = single_url\n results.append(parsed_results)\n except:\n print ('Crawling of URL ' + single_url + ' has failed')\n continue\n\nprint (json.dumps(results, indent=4))\nprint (len(results))\n\n","sub_path":"automatic_crawler_for_url_www_borenius_com.py","file_name":"automatic_crawler_for_url_www_borenius_com.py","file_ext":"py","file_size_in_byte":1279,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"429764013","text":"# uncompyle6 version 3.7.4\n# Python bytecode 3.7 (3394)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.macosx-10.9-x86_64/egg/getml/hyperopt/load_hyperopt.py\n# Compiled at: 2020-03-16 07:21:38\n# Size of source mod 2**32: 10728 bytes\nimport json\nimport getml.communication as comm\nimport getml.models as models\nfrom .hyperopt import RandomSearch, LatinHypercubeSearch, GaussianHyperparameterSearch\n\ndef _decode_hyperopt(rawStr):\n \"\"\"A custom decoder function for\n :class:`~getml.hyperopt.RandomSearch`,\n :class:`~getml.hyperopt.LatinHypercubeSearch`, and\n :class:`~getml.hyperopt.GaussianHyperparameterSearch`.\n\n Args:\n rawStr (str): string containing a valid JSON message.\n\n Raises:\n KeyError: If not all required fields are present in `rawStr`\n to reconstruct a hyperparameter optimization search.\n ValueError: If not all keys in `rawStr` have a trailing\n underscore.\n TypeError: If `rawStr` is not of type :py:class:`dict`.\n\n Returns: \n Union[:class:`~getml.hyperopt.RandomSearch`,:class:`~getml.hyperopt.LatinHypercubeSearch`,:class:`~getml.hyperopt.GaussianHyperparameterSearch`]\n \"\"\"\n if type(rawStr) is not str:\n raise TypeError('_decode_hyperopt is expecting a str containing a valid JSON as input')\n else:\n rawDict = json.loads(rawStr)\n requiredFields = {\n 'name_', 'param_space_', 'session_name_',\n 'n_iter_', 'ratio_iter_',\n 'optimization_algorithm_',\n 'optimization_burn_in_algorithm_',\n 'optimization_burn_ins_',\n 'seed_',\n 'surrogate_burn_in_algorithm_',\n 'gaussian_kernel_',\n 'gaussian_optimization_algorithm_',\n 'gaussian_optimization_burn_in_algorithm_',\n 'gaussian_optimization_burn_ins_'}\n if set(rawDict.keys()).intersection(requiredFields) != requiredFields:\n raise KeyError('Not enough information contained in the response to reconstruct the hyperparameter optimization: ' + str(rawDict.keys()))\n decodingDict = dict()\n for kkey in rawDict:\n if kkey[(len(kkey) - 1)] != '_':\n raise ValueError('All keys in the JSON must have a trailing underscore.')\n elif kkey == 'name_':\n decodingDict['model'] = models.load_model(rawDict[kkey])\n elif kkey == 'param_space_':\n param_space = dict()\n for ddimension in rawDict[kkey]:\n param_space[ddimension[:len(ddimension) - 1]] = rawDict[kkey][ddimension]\n\n decodingDict['param_space'] = param_space\n else:\n if kkey in ('peripheral_names_', 'population_training_name_', 'population_validation_name_'):\n continue\n decodingDict[kkey[:len(kkey) - 1]] = rawDict[kkey]\n\n if decodingDict['ratio_iter'] == 1 and decodingDict['surrogate_burn_in_algorithm'] == 'latinHypercube':\n h = LatinHypercubeSearch(model=(decodingDict['model']), param_space=(decodingDict['param_space']),\n seed=(decodingDict['seed']),\n session_name=(decodingDict['session_name']),\n n_iter=(decodingDict['n_iter']))\n else:\n if decodingDict['ratio_iter'] == 1 and decodingDict['surrogate_burn_in_algorithm'] == 'random':\n h = RandomSearch(model=(decodingDict['model']), param_space=(decodingDict['param_space']),\n seed=(decodingDict['seed']),\n session_name=(decodingDict['session_name']),\n n_iter=(decodingDict['n_iter']))\n else:\n h = GaussianHyperparameterSearch(model=(decodingDict['model']), param_space=(decodingDict['param_space']),\n session_name=(decodingDict['session_name']),\n ratio_iter=(decodingDict['ratio_iter']),\n n_iter=(decodingDict['n_iter']),\n optimization_algorithm=(decodingDict['optimization_algorithm']),\n optimization_burn_in_algorithm=(decodingDict['optimization_burn_in_algorithm']),\n optimization_burn_ins=(decodingDict['optimization_burn_ins']),\n seed=(decodingDict['seed']),\n surrogate_burn_in_algorithm=(decodingDict['surrogate_burn_in_algorithm']),\n gaussian_kernel=(decodingDict['gaussian_kernel']),\n gaussian_optimization_algorithm=(decodingDict['gaussian_optimization_algorithm']),\n gaussian_optimization_burn_in_algorithm=(decodingDict['gaussian_optimization_burn_in_algorithm']),\n gaussian_optimization_burn_ins=(decodingDict['gaussian_optimization_burn_ins']))\n if 'score' in decodingDict:\n h.score = decodingDict['score']\n return h\n\n\ndef load_hyperopt(session_name):\n \"\"\"Loads a hyperparameter optimization run into the Python API.\n\n Args:\n session_name (string): Unique identifier of a particular\n hyperparameter optimization run.\n\n Returns:\n Union[:class:`~getml.hyperopt.RandomSearch`, :class:`~getml.hyperopt.LatinHypercubeSearch`, :class:`~getml.hyperopt.GaussianHyperparameterSearch`]\n\n Raises:\n IOError: If the messages received from the engine is not a\n valid JSON.\n TypeError: if `session_name` is not a string.\n ValueError: if `session_name` is an empty string.\n\n \"\"\"\n if type(session_name) is not string:\n raise TypeError('Only strings are allowed as session_name!')\n elif session_name == '':\n raise ValueError('The session_name must not be empty!')\n else:\n print('Not supported yet!')\n return\n h = _decode_hyperopt(msg)\n multithreaded = False\n if h.model.num_threads > 1:\n multithreaded = True\n if h.model.predictor is not None:\n if isinstance(h.model.predictor, predictors.XGBoostClassifier) or isinstance(h.model.predictor, predictors.XGBoostRegressor):\n if h.model.predictor.n_jobs > 1:\n multithreaded = True\n if not h.model.feature_selector is not None or isinstance(h.model.feature_selector, predictors.XGBoostClassifier) or isinstance(h.model.feature_selector, predictors.XGBoostRegressor):\n if h.model.feature_selector.n_jobs > 1:\n multithreaded = True\n if multithreaded:\n h.model.seed = None","sub_path":"pycfiles/getml-0.10.0-py3.7/load_hyperopt.cpython-37.py","file_name":"load_hyperopt.cpython-37.py","file_ext":"py","file_size_in_byte":6470,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"296379012","text":"count = 0\ntotal = 0\navg = 0\n\nwhile True:\n entry = input('Enter a number: ')\n if entry == 'done':\n break\n try:\n total = int(entry) + total\n except Exception as e:\n print('Invalid input')\n continue\n count = count + 1\n\nif count != 0:\n avg = total / count\n\nprint(total, count, avg)\n","sub_path":"course1/week5/ex_05_01.py","file_name":"ex_05_01.py","file_ext":"py","file_size_in_byte":324,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"92415597","text":"from base.basetestcase import BaseTestCase\nimport unittest\nfrom selenium.webdriver.support.ui import Select\n\n\nclass CookiesTest(BaseTestCase):\n URL = 'http://demo-store.seleniumacademy.com/'\n ENV = 'LINUX'\n\n def test_store_cookie(self):\n driver = self.driver\n # get the your language dropdown as instance of select class\n select_language = Select(self.driver.\n find_element_by_id('select-language'))\n\n # check default selected option is english\n self.assertEqual('English', select_language.first_selected_option.text)\n # store cookies should be none\n store_cookie = driver.get_cookie('store')\n self.assertEqual(None, store_cookie)\n\n # select an option using select_by_visible text\n select_language.select_by_visible_text('French')\n\n # store cookie should be populated with selected country\n store_cookie = driver.get_cookie('store')['value']\n self.assertEqual('french', store_cookie)\n\nif __name__ == '__main__':\n unittest.main(verbosity=2)\n","sub_path":"chapter_9/cookie_test.py","file_name":"cookie_test.py","file_ext":"py","file_size_in_byte":1075,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"519827698","text":"import os\n\nfrom flask import render_template, send_from_directory, url_for, flash, redirect, request, abort\nfrom flask_login import login_user, current_user, logout_user, login_required\nfrom Course import app, db, bcrypt, qrcode\nfrom Course.forms import LoginForm, CompanyRegistrationForm, UserRegistrationForm, QuestionForm, ClassForm\nfrom Course.models import User, Question, Clas, UserClas, Company\nimport random\n\n\n@app.route('/')\ndef index():\n return redirect(url_for('home'))\n\n\n@app.route('/home')\n@login_required\ndef home():\n c = Clas.query.filter_by(company_id=current_user.company_id)\n mental = c.filter_by(category='Mental').all()\n physical = c.filter_by(category='Physical').all()\n verbal = c.filter_by(category='Verbal').all()\n if current_user.role == 'admin':\n return redirect(url_for('admin'))\n return render_template('home.html', c=[mental, physical, verbal])\n\n\n@app.route('/class')\n@login_required\ndef classs():\n records = current_user.clases\n return render_template('class.html', title='Class', classes=records)\n\n\n@app.route('/intro/', methods=['GET', 'POST'])\n@login_required\ndef intro(class_id):\n course = Clas.query.filter_by(id=class_id).first()\n mark = -1\n enrolled = False\n\n for c in course.users:\n if c.user_id == current_user.id:\n enrolled = True\n mark = c.mark\n return render_template(\n 'intro.html',\n c=course,\n enrolled=enrolled,\n mark=mark,\n courses=Clas.query.all())\n\n\n@app.route('/login', methods=['GET', 'POST'])\ndef login():\n if current_user.is_authenticated:\n return redirect(url_for('home'))\n form = LoginForm()\n if form.validate_on_submit():\n user = User.query.filter_by(username=form.username.data).first()\n if user:\n if bcrypt.check_password_hash(user.password, form.password.data):\n login_user(user, remember=form.remember.data)\n next_page = request.args.get('next')\n if user.username == 'super_admin':\n return redirect(url_for('super_admin'))\n if next_page:\n return redirect(next_page)\n else:\n return redirect(url_for('home'))\n else:\n flash('Login Unsuccessful. Please check your password', 'danger')\n flash('Login Unsuccessful. Username does not exist', 'danger')\n return render_template('login.html', title='Login', form=form)\n\n\n@app.route('/logout')\ndef logout():\n logout_user()\n return redirect(url_for('login'))\n\n\n@app.route('/favicon.ico')\ndef favicon():\n return send_from_directory(\n os.path.join(\n app.root_path,\n 'static'),\n 'favicon.ico',\n mimetype='image/vnd.microsoft.icon')\n\n\n@app.route(\"/class/\", methods=['GET'])\n@login_required\ndef get_class(class_id):\n c = Clas.query.get_or_404(class_id)\n enrolled = False\n for user_class in c.users:\n if current_user.id == user_class.user_id:\n enrolled = True\n if not enrolled:\n user = User.query.filter_by(id=current_user.id).first()\n enroll = UserClas(mark=0, user=user)\n c.users.append(enroll)\n db.session.commit()\n flash(\n f'You have been successfully enrolled in {c.description}',\n 'success')\n return render_template('video.html', title='Class', c=c)\n\n\n@app.route(\"/admin\", methods=['GET', 'POST'])\n@login_required\ndef admin():\n if current_user.role != 'admin':\n abort(403)\n users = User.query.filter_by(company_id=current_user.company_id).all()\n form = UserRegistrationForm()\n if form.validate_on_submit():\n hashed_password = bcrypt.generate_password_hash(\n form.password.data).decode('utf8')\n user = User(\n username=form.username.data,\n password=hashed_password,\n role='student',\n company_id=current_user.company_id)\n db.session.add(user)\n db.session.commit()\n return redirect(url_for('admin'))\n return render_template(\n 'user_admin.html',\n title='admin',\n form=form,\n users=users)\n\n\n@app.route(\"/test_admin\", methods=['GET', 'POST'])\n@login_required\ndef test_admin():\n if current_user.role != 'admin':\n abort(403)\n tests = Clas.query.filter_by(company_id=current_user.company_id).all()\n form = ClassForm()\n if form.validate_on_submit():\n c = Clas(\n category=form.category.data,\n description=form.description.data,\n introduction=form.introduction.data,\n content=form.content.data,\n outcome=form.outcome.data,\n lecturer=form.lecturer.data,\n share=form.share.data,\n video_link=form.video_link.data,\n limit_time=30,\n company_id=current_user.company_id)\n db.session.add(c)\n db.session.commit()\n return redirect(url_for('test_admin'))\n return render_template(\n 'test_admin.html',\n title='admin',\n tests=tests,\n form=form)\n\n\n@app.route(\"/test\", methods=['POST', 'GET'])\n@login_required\ndef exam():\n class_id = request.args.get('class_id', default=1, type=int)\n typ = request.args.get('type', default='', type=str)\n current_year = 2021\n current_semester = 2\n q = Question.query.filter_by(\n clas_id=class_id,\n year=current_year,\n semester=current_semester).all()[:5]\n time_limit = Clas.query.filter_by(id=class_id).first().limit_time\n if typ == 'mock':\n q = Question.query.filter_by(clas_id=class_id).filter(\n Question.year != current_year and Question.semester != current_semester).all()\n # q = random.choices(q, k=5)\n q = q[:5]\n return render_template(\n 'test.html',\n title='Test',\n q=q,\n class_id=class_id,\n type=typ,\n time=time_limit)\n\n\n@app.route(\"/super_admin\", methods=['GET', 'POST'])\n@login_required\ndef super_admin():\n c = Company.query.all()\n if current_user.username != 'super_admin':\n abort(403)\n form = CompanyRegistrationForm()\n if form.validate_on_submit():\n hashed_password = bcrypt.generate_password_hash(\n form.password.data).decode('utf8')\n company = Company(name=form.username.data)\n db.session.add(company)\n db.session.commit()\n company_id = Company.query.filter_by(\n name=form.username.data).first().id\n user = User(\n username=form.username.data,\n password=hashed_password,\n role='admin',\n company_id=company_id)\n db.session.add(user)\n db.session.commit()\n return redirect(url_for('logout'))\n return render_template('register.html', title='Register', form=form, c=c)\n\n\n@app.route(\"/change_time//\", methods=['GET', 'POST'])\ndef change_time(clas_id, minutes):\n course = Clas.query.get_or_404(clas_id)\n if current_user.role != 'admin':\n abort(403)\n course.limit_time = minutes\n db.session.commit()\n return redirect(url_for('test_admin'))\n\n\n@app.route(\"/qrcode_login\")\ndef qrcode_login():\n abort(401)\n\n\n@app.route(\"/question_admin\", methods=['GET', 'POST'])\ndef question_admin():\n if current_user.role != 'admin':\n abort(403)\n clas_id = request.args.get('class_id', default=-1, type=int)\n semester = request.args.get('semester', default=-1, type=int)\n year = request.args.get('year', default=-1, type=int)\n questions = Question.query\n if clas_id != -1:\n questions = questions.filter_by(clas_id=clas_id)\n if year != -1:\n questions = questions.filter_by(year=year)\n if semester != -1:\n questions = questions.filter_by(semester=semester)\n\n questions = questions.all()\n form = QuestionForm()\n form.course.choices = list(\n map(lambda x: (x.id, x.description), current_user.owner.clases))\n if form.validate_on_submit():\n q = Question(\n question_type=form.question_type.data,\n year=form.year.data,\n semester=form.semester.data,\n question_text=form.question_text.data,\n clas_id=form.course.data)\n db.session.add(q)\n db.session.commit()\n flash('Question Added', 'success')\n return redirect(\n url_for(\n 'question_admin',\n class_id=clas_id,\n year=year,\n semester=semester))\n return render_template(\n 'question_admin.html',\n title='admin',\n form=form,\n questions=questions,\n dyear=year,\n dsemester=semester,\n dclas_id=clas_id,\n Q=Question)\n\n\n@app.route('/finish_course//')\n@login_required\ndef finish_course(student_id, course_id):\n user_class = UserClas.query.filter_by(\n user_id=student_id, clas_id=course_id).first()\n user_class.mark = random.randint(60, 100)\n db.session.commit()\n return redirect(url_for('classs'))\n","sub_path":"Course/routes.py","file_name":"routes.py","file_ext":"py","file_size_in_byte":9098,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"34960761","text":"def CheckForTrainStation(station):\r\n\r\n\t#f = open('stops.txt', 'r')\r\n\tstationLocation = []\r\n\t\r\n\twith open('stops.txt', 'r') as f:\r\n\t\tfor line in f:\r\n\t\t\twords = line.split(',')\r\n\r\n\t\t\tif station in words[1]:\r\n\t\t\t\tstationLocation.append(words[2])\r\n\t\t\t\tstationLocation.append(words[3])\r\n\t\t\t\t\r\n\treturn stationLocation","sub_path":"CheckForTrainStation.py","file_name":"CheckForTrainStation.py","file_ext":"py","file_size_in_byte":311,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"130900137","text":"from functools import reduce\nfrom collections import Counter\n\n\nimport itertools\nimport operator\n\ndef most_common(L):\n # get an iterable of (item, iterable) pairs\n SL = sorted((x, i) for i, x in enumerate(L))\n # print 'SL:', SL\n groups = itertools.groupby(SL, key=operator.itemgetter(0))\n # auxiliary function to get \"quality\" for an item\n def _auxfun(g):\n item, iterable = g\n count = 0\n min_index = len(L)\n for _, where in iterable:\n count += 1\n min_index = min(min_index, where)\n # print 'item %r, count %r, minind %r' % (item, count, min_index)\n return count, -min_index\n # pick the highest-count/earliest item\n return max(groups, key=_auxfun)[0]\n\nfd = open('data.txt', 'r')\nbook = fd.read().lower()\ncount=0\nfd.close()\n\n\n\nprint(\"Count of proletariat: \" + str(book.count('proletariat')))\nprint(\"Count of 'working class': \" + str(book.count('working class')))\n\nalphabet = \"abcdefghijklmnopqrstuvwxyz\"\nwords = []\ncurrent = \"\"\nfor char in book:\n if char in alphabet:\n current+=char\n elif current!=\"\":\n words.append(current)\n current=\"\"\n\n#print(words)\n\n# count=0\n# for x in words:\n# if x==\"proletariat\":\n# count+=1\n# print(count)\n\ncommon_words = [\"the\", \"and\", \"to\", \"by\", \"is\", \"in\", \"of\", \"a\", \"that\", \"it\", \"them\"]\nmax_word = \"\"\nmax_count = 0\nfor x in words:\n if words.count(x)>max_count and x not in common_words:\n max_count = words.count(x)\n max_word = x\n\n\nlist=[[x,words.count(x)] for x in words if x not in common_words]\n#print(most_common(list))\n\nword=[x for x in list if x[1]==max_count]\nw=word[0][0]\nwc=word[0][1]\nprint(\"Max uncommon word: \" + str(w) + \"; Count: \" + str(wc))\n","sub_path":"20_cw/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":1675,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"493169979","text":"from ipykernel.kernelbase import Kernel\nfrom pexpect import replwrap, EOF, spawn\n\nfrom subprocess import check_output\n\nfrom os import path\nimport re\nimport signal\nimport bisect\n\nfrom codecs import open\ndef readfile(filename):\n with open(filename, encoding='utf-8') as f:\n return f.read()\n__version__ = readfile(path.join(path.dirname(__file__), 'VERSION'));\nversion_pat = re.compile(r'version (\\d+(\\.\\d+)+)')\n\n\n\nclass MagmaKernel(Kernel):\n implementation = 'magma_kernel'\n implementation_version = __version__\n\n\n\n language_info = {'name': 'magma',\n 'codemirror_mode': 'magma',\n 'mimetype': 'text/x-sh',\n 'file_extension': '.mgm'}\n _banner = None\n\n @property\n def banner(self):\n if self._banner is None:\n self._banner = check_output('magma').decode('utf-8').split('\\n')[0]\n return self._banner\n\n def __init__(self, **kwargs):\n Kernel.__init__(self, **kwargs)\n self._start_magma()\n\n def _start_magma(self):\n # Signal handlers are inherited by forked processes, and we can't easily\n # reset it from the subprocess. Since kernelapp ignores SIGINT except in\n # message handlers, we need to temporarily reset the SIGINT handler here\n # so that bash and its children are interruptible.\n sig = signal.signal(signal.SIGINT, signal.SIG_DFL)\n try:\n magma = spawn('magma', echo=False, encoding='utf-8')\n magma.expect(u'> ')\n magma.sendline(u'SetLineEditor(false);')\n magma.expect(u'> ')\n magma.sendline(u'')\n self.magmawrapper = replwrap.REPLWrapper(magma,\n u'> ', u'SetPrompt(\"{}\");')\n finally:\n signal.signal(signal.SIGINT, sig)\n\n\n def do_execute(self, code, silent, store_history=True,\n user_expressions=None, allow_stdin=False):\n if not code.strip():\n return {'status': 'ok', 'execution_count': self.execution_count,\n 'payload': [], 'user_expressions': {}}\n\n interrupted = False\n try:\n output = self.magmawrapper.run_command(code.rstrip(), timeout=None)\n except KeyboardInterrupt:\n self.magmawrapper.child.sendintr()\n interrupted = True\n self.magmawrapper._expect_prompt()\n output = self.magmawrapper.child.before\n except EOF:\n output = self.magmawrapper.child.before + 'Restarting Magma'\n self._start_magma()\n\n if not silent:\n # Send standard output\n stream_content = {'name': 'stdout', 'text': output}\n self.send_response(self.iopub_socket, 'stream', stream_content)\n\n if interrupted:\n return {'status': 'abort', 'execution_count': self.execution_count}\n\n return {'status': 'ok', 'execution_count': self.execution_count,\n 'payload': [], 'user_expressions': {}}\n\n _magma_builtins = None;\n def do_complete(self, code, cursor_pos):\n code = code[:cursor_pos];\n default = {'matches': [], 'cursor_start': 0,\n 'cursor_end': cursor_pos, 'metadata': dict(),\n 'status': 'ok'};\n if not code or code[-1] == ' ':\n return default;\n\n if self._magma_builtins is None:\n # the list of builtins is generated by:\n # $ magmadoc=/path/to/magma/doc/html\n # $ cat $magmadoc/*.htm | sed -nr \"s/^.*NAME = \\\"([A-Z][[:alnum:]]*)\\\".*$/\\1/p\" | sort > magma-builtins\n from os import path\n F = open(path.join(path.dirname(__file__), \"magma-builtins\"), 'r');\n self._magma_builtins = F.read().split('\\n');\n F.close();\n\n tokens = code.replace(';', ' ').split();\n if not tokens:\n return default;\n token = tokens[-1];\n start = cursor_pos - len(token);\n\n low = bisect.bisect_left(self._magma_builtins, token);\n #very hacky\n high = bisect.bisect_right(self._magma_builtins, token+chr(127), low); \n matches = self._magma_builtins[low:high];\n\n #TODO add global variables\n\n if not matches:\n return default;\n\n return {'matches': matches, 'cursor_start': start,\n 'cursor_end': cursor_pos, 'metadata': dict(),\n 'status': 'ok'};\n\n","sub_path":"magma_kernel/kernel.py","file_name":"kernel.py","file_ext":"py","file_size_in_byte":4394,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"642451212","text":"import os\nimport time\n\nimport numpy as np\nimport pandas as pd\n\nEvaluation = pd.DataFrame(columns=['Date', 'StartTime', 'EndTime'])\ndata_list = []\nStartTime = []\nEndTime = []\n\n\ndef read_txt(file_path):\n i = 0\n i_list = []\n with open(file_path, 'r') as f:\n for line in f.readlines():\n line = line.strip('\\n')\n # print(line)\n if not 'PM1' in line:\n Starttime = line.split(' ')[0][:-3]\n Endtime = line.split(' ')[1][:-3]\n StartTime.append(Starttime)\n EndTime.append(Endtime)\n i += 1\n else:\n i_list.append(i)\n if len(i_list) > 1:\n for k in range(i_list[-1] - i_list[-2] - 1):\n data_list.append(data_list[-1])\n front = line.split('.')[0]\n Date = front[-10:].replace('_', '-')\n data_list.append(Date)\n\n data_list.append(data_list[-1])\n data_list.append(data_list[-1])\n data_list.append(data_list[-1])\n\n Evaluation['Date'] = data_list\n Evaluation['StartTime'] = StartTime\n Evaluation['EndTime'] = EndTime\n Evaluation['Value_Count'] = 1\n return Evaluation\n\n\ndef Evaluation_Process(PATH, Evaluation, analomy_data):\n Evaluation['Timestamp'] = Evaluation['Date'].apply(lambda x: int(time.mktime(time.strptime(x, '%Y-%m-%d'))))\n\n # Evaluation['Start'] = Evaluation['StartTime'].apply(lambda x: int((float(x.split(':')[0])*60)+float(x.split(':')[1])))\n # Evaluation['End'] = Evaluation['EndTime'].apply(lambda x: int((float(x.split(':')[0])*60)+float(x.split(':')[1])))\n\n Evaluation['Start'] = Evaluation['StartTime'].apply(\n lambda x: int((float(x.split(':')[0]) * 4) + float(x.split(':')[1]) // 15))\n Evaluation['End'] = Evaluation['EndTime'].apply(\n lambda x: int((float(x.split(':')[0]) * 4) + float(x.split(':')[1]) // 15))\n\n Evaluation['Start_End'] = Evaluation.apply(lambda row: StartEnd(row['Start'], row['End']), axis=1)\n Evaluation = Evaluation[['Timestamp', 'Start_End']]\n Evaluation = Evaluation.explode('Start_End')\n Evaluation.drop_duplicates(keep='first', inplace=True)\n\n anamaly_path = os.path.join(PATH, 'Analysis_PM1', 'Evaluation_Data.csv')\n # Evaluation.to_csv(anamaly_path)\n np.savetxt(anamaly_path, Evaluation, delimiter=',', fmt='%d')\n Common_Row = pd.merge(Evaluation, analomy_data, how='inner', on=['Timestamp', 'Start_End'])\n print('Common_Row', Common_Row.shape[0])\n print('Evaluation_Row', Evaluation.shape[0])\n print('analomy_data_Row', analomy_data.shape[0])\n # compare = datacompy.Compare(Evaluation,analomy_data,join_columns='Timestamp')\n # print(compare.report())\n print('DONE')\n\n\ndef StartEnd(start, end):\n item_list = []\n for item in range(start, end + 1):\n item_list.append(item)\n return item_list\n\n\nif __name__ == '__main__':\n PATH = os.path.abspath(os.path.dirname(os.getcwd()))\n folder = 'Analysis_PM1'\n file_name = 'CSVs_datalogger630094-CUBIC_IAQ_PM1_UpToAug22_GT.txt'\n analomy_file = 'anomaly_data(GMT)630094_Per15_Weekall_2021-02-28_2021-08-22.csv'\n analomy_data = pd.read_csv(os.path.join(PATH, folder, analomy_file), usecols=range(0, 2))\n file_path = os.path.join(PATH, folder, file_name)\n Evaluation = read_txt(file_path)\n Evaluation_Process(PATH, Evaluation, analomy_data)\n","sub_path":"pyfiles/Quantitative_Evaluation.py","file_name":"Quantitative_Evaluation.py","file_ext":"py","file_size_in_byte":3421,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"397770554","text":"from app.database.sqlalchemy_extension import db\nfrom app.database.models.user import UserModel\nfrom app.database.models.application import ApplicationModel\nfrom app.database.models.institution import InstitutionModel\n\nfrom app.database.models.application_donor import application_donor\nfrom app.database.models.application_moderator import application_moderator\nfrom datetime import date, timedelta\n\nclass ReservedApplicationModel(db.Model):\n \n # Specifying table\n __tablename__ = \"reserved_application\"\n __table_args__ = {\"extend_existing\": True}\n \n \n id = db.Column(db.Integer, primary_key=True)\n application_id = db.Column(db.Integer, db.ForeignKey(\"application.id\"))\n application = db.relationship(\n ApplicationModel,\n backref='reserved',\n primaryjoin=\"ReservedApplicationModel.application_id == ApplicationModel.id\",\n )\n donor_id = db.Column(db.Integer, db.ForeignKey(\"user.id\"))\n donor = db.relationship(\n UserModel,\n backref=\"reserved_as_donor\",\n primaryjoin=\"ReservedApplicationModel.donor_id == UserModel.id\",\n )\n moderator_id = db.Column(db.Integer, db.ForeignKey(\"user.id\"))\n moderator = db.relationship(\n UserModel,\n backref=\"reserved_as_moderator\",\n primaryjoin=\"ReservedApplicationModel.moderator_id == UserModel.id\",\n )\n \n is_active = db.Column(db.Boolean)\n verified = db.Column(db.Boolean)\n verification_date = db.Column(db.String(20))\n donation_date = db.Column(db.String(20))\n amount = db.Column(db.Integer)\n \n \n def save_to_db(self) -> None:\n '''Add application to database'''\n db.session.add(self)\n db.session.commit()\n\n def delete_from_db(self) -> None:\n '''Deletes application from the database.'''\n db.session.delete(self)\n db.session.commit()\n \n \n @classmethod \n def find_by_id(cls, id: int) -> 'ReservedApplicationModel':\n '''Returns reserved application of given id.'''\n return cls.query.filter_by(id= id).first()\n \n @classmethod \n def find_by_application_id(cls, application_id: int) -> 'ReservedApplicationModel':\n '''Returns reserved application of given donor id.'''\n return cls.query.filter_by(application_id= application_id).first()\n \n @classmethod \n def find_by_donor_id(cls, donor_id: int) -> 'ReservedApplicationModel':\n '''Returns reserved application of given donor id.'''\n return cls.query.filter_by(donor_id= donor_id).first()\n \n @classmethod \n def find_by_moderator_id(cls, moderator_id: int) -> 'ReservedApplicationModel':\n '''Returns reserved application of given moderator id.'''\n return cls.query.filter_by(moderator_id= moderator_id).first()\n \n @classmethod \n def find_reserved_application(cls, application_id: int, donor_id: int, moderator_id: int) -> 'ReservedApplicationModel':\n '''Returns reserved application of given moderator id.'''\n return cls.query.filter_by(application_id=application_id, donor_id=donor_id, moderator_id= moderator_id).first()\n ","sub_path":"app/database/models/reserved_application.py","file_name":"reserved_application.py","file_ext":"py","file_size_in_byte":3144,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"191081712","text":"import random\nfrom collections.abc import Iterable\nimport pickle\nimport numpy as np\nfrom tqdm import tqdm\nfrom pytorch_lightning.metrics import Accuracy\nimport torch\nimport os\nfrom p_tqdm import p_umap\nfrom pathos.multiprocessing import ProcessingPool as Pool\nimport multiprocessing\n\ndef generate_random_params(params, n_iter, seed=10):\n \n random_params = []\n random.seed(seed)\n\n # loop for n_iter\n for _iter in range(0,n_iter):\n random_param = {}\n # loop over params\n for key, values in params.items():\n # assert values is a list\n assert isinstance(values,Iterable)\n\n # select a random item out of values\n value = random.choice(values)\n\n random_param[key] = value\n\n random_params.append(random_param)\n\n return random_params\n\n\ndef create_chunks(l, n):\n for i in range(0, len(l), n):\n yield l[i:i + n]\n\ndef model_qualifier(model_type):\n return model_type[:2]\n\ndef save_ann_model(model,file_name):\n # save actual model\n import joblib\n import copy \n\n # copy the nn module\n torch_nn_module = copy.deepcopy(model.module_).to('cpu')\n\n # reset the net to cpu\n model.set_params(device=\"cpu\")\n model = model.initialize()\n \n # attach the old nn module\n model.module_ = torch_nn_module\n\n # dump to disk\n joblib.dump(model,file_name)\n \n\ndef load_ann_model(file_name):\n import joblib\n model = joblib.load(file_name)\n return model\n\n\ndef save_model(model, file_name):\n import joblib\n from sklearn.utils.validation import check_is_fitted\n check_is_fitted(model)\n joblib.dump(model,file_name)\n\ndef load_model(file_name):\n import joblib\n return joblib.load(file_name)\n\ndef save_scaler(scaler, file_name):\n import joblib\n from sklearn.utils.validation import check_is_fitted\n check_is_fitted(scaler)\n joblib.dump(scaler,file_name)\n \n\ndef load_scaler(file_name):\n import joblib\n return joblib.load(file_name)\n\ndef clean_proteins(protein):\n cleaned = protein.split(\"-\")[0]\n return cleaned\n\ndef get_embedding_vector_for_edge(edge,embeddings_pd,operator=\"l1\"):\n source_node = edge[0]\n target_node = edge[1]\n source_embedding = embeddings_pd[embeddings_pd[\"protein\"] == source_node].drop(columns=[\"protein\"]).to_numpy()\n\n target_embedding = embeddings_pd[embeddings_pd[\"protein\"] == target_node].drop(columns=[\"protein\"]).to_numpy()\n \n # if we get empty vectors then zero arrs\n if source_embedding.size == 0:\n source_embedding = np.zeros(target_embedding.shape)\n\n if target_embedding.size == 0:\n target_embedding = np.zeros(source_embedding.shape)\n\n if operator == \"l1\":\n embedding_np = operator_l1(source_embedding,target_embedding)\n elif operator == \"l2\":\n embedding_np = operator_l2(source_embedding,target_embedding)\n elif operator == \"hadamard\":\n embedding_np = operator_hadamard(source_embedding,target_embedding)\n elif operator == \"avg\":\n embedding_np = operator_avg(source_embedding,target_embedding)\n elif operator == \"cat\":\n embedding_np = operator_cat(source_embedding,target_embedding)\n else:\n raise Exception(f\"Unknown operator : {operator}\")\n \n return embedding_np\n\ndef get_embeddings_vector(edge_list,embeddings_pd,operator=\"l1\",show_progress=True):\n embeddings = []\n cpu_count = multiprocessing.cpu_count()\n pool = Pool(processes=cpu_count-1)\n do_work = lambda edge: get_embedding_vector_for_edge(edge,embeddings_pd,operator=operator)\n\n if show_progress == True:\n embeddings = p_umap(do_work,\n edge_list,\n num_cpus=cpu_count)\n else:\n embeddings = pool.map(do_work, edge_list)\n \n embeddings_vec = np.concatenate(embeddings,axis=0)\n return embeddings_vec\n\ndef calculate_accuracy(labels_predicted,labels_actual):\n accuracy = Accuracy()\n return accuracy(labels_predicted,labels_actual)\n\n\n# define the binary operators\ndef operator_hadamard(u, v):\n return u * v\n\ndef operator_l1(u, v):\n return np.abs(u - v)\n\ndef operator_l2(u, v):\n return (u - v) ** 2\n\ndef operator_avg(u, v):\n return (u + v) / 2.0\n\ndef operator_cat(u, v):\n return np.concatenate([u,v],axis=1)\n\nclass IKGDataSet(torch.utils.data.Dataset):\n\n def __init__(self,embeddings, labels):\n self.embedddings = embeddings\n self.labels = labels\n\n def __len__(self):\n return self.labels.shape[0]\n\n def __getitem__(self,index):\n # get embedding row\n embedding = self.embedddings[index]\n\n # get the label\n label = self.labels[index]\n\n return embedding, label\n\n\ndef get_best_experiment(results_dir):\n files = os.listdir(results_dir)\n \n # sort in descending ordert\n files.sort(reverse=True)\n\n best_result = files[0]\n\n # get the absolute path\n best_result = os.path.join(results_dir,best_result)\n\n return best_result\n\ndef extract_config(result_dict):\n config = {}\n for key,value in result_dict.items():\n # check if this is config key\n if \"config.\" in key:\n # if yes then extract the name\n key_name = key.replace(\"config.\",\"\")\n\n config[key_name] = value\n\n return config\n \n ","sub_path":"src/misc/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":5282,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"479573990","text":"import pickle\nimport numpy as np\nimport matplotlib.pyplot as plt\n\ntrain_balanced_data_file_preprocessed = \"traffic-signs-data/train_balanced_preprocessed.p\"\ntest_data_file_preprocessed = \"traffic-signs-data/test_preprocessed.p\"\nvalid_data_file_preprocessed = \"traffic-signs-data/valid_preprocessed.p\"\n\ndef load_data(path):\n with open(path, mode='rb') as f:\n data = pickle.load(f)\n X, y = data['features'], data['labels']\n return X, y\n\ndef normalize(data): \n data = (data - data.min()) / (data.max() - data.min())\n return data\n\nX_train, y_train = load_data(train_balanced_data_file_preprocessed)\nX_valid, y_valid = load_data(valid_data_file_preprocessed)\nX_test, y_test = load_data(test_data_file_preprocessed)\n \nX_train = normalize(X_train)\nX_valid = normalize(X_valid)\nX_test = normalize(X_test)\n\n# Implement LeNet-5\n# Implement the [LeNet-5](http://yann.lecun.com/exdb/lenet/) neural network architecture.\nimport tensorflow as tf\nfrom tensorflow.contrib.layers import flatten\n\ndef conv2d(x, W, b, strides=1):\n x = tf.nn.conv2d(x, W, strides=[1, strides, strides, 1], padding='VALID')\n x = tf.nn.bias_add(x, b)\n # Activation.\n return tf.nn.relu(x)\n\ndef maxpool2d(x, k=2):\n return tf.nn.max_pool(x, ksize=[1, k, k, 1], strides=[1, k, k, 1], padding='VALID')\n\ndef LeNet(x, is_training):\n # Arguments used for tf.truncated_normal, randomly defines variables for the weights and biases for each layer\n mu = 0\n sigma = 0.1\n output_size = 43\n \n # Layer 1: Convolutional. Input = 32x32x1. Output = 28x28x32.\n w1 = tf.Variable(tf.truncated_normal(shape=(5, 5, 1, 32), mean = mu, stddev = sigma))\n b1 = tf.Variable(tf.zeros(32))\n conv1 = conv2d(x, w1, b1)\n\n # Pooling. Input = 28x28x32. Output = 14x14x32.\n conv1 = maxpool2d(conv1, k=2)\n conv1 = tf.cond(is_training, lambda: tf.nn.dropout(conv1, keep_prob = 0.9), lambda: conv1)\n\n # Layer 2: Convolutional. Output = 10x10x64.\n w2 = tf.Variable(tf.truncated_normal(shape=(5, 5, 32, 64), mean = mu, stddev = sigma))\n b2 = tf.Variable(tf.zeros(64))\n conv2 = conv2d(conv1, w2, b2)\n\n # Pooling. Input = 10x10x64. Output = 5x5x64.\n conv2 = maxpool2d(conv2, k=2)\n conv2 = tf.cond(is_training, lambda: tf.nn.dropout(conv2, keep_prob = 0.75), lambda: conv2)\n\n # Flatten. Input = 5x5x64. Output = 1600.\n fc0 = flatten(conv2)\n \n # Layer 3: Fully Connected. Input = 1600. Output = 1024.\n wd1 = tf.Variable(tf.truncated_normal(shape=(1600, 1024), mean = mu, stddev = sigma))\n bd1 = tf.Variable(tf.zeros(1024))\n fc1 = tf.matmul(fc0, wd1) + bd1\n \n # Activation.\n fc1 = tf.nn.relu(fc1)\n\n # Layer 4: Fully Connected. Input = 1024. Output = 512.\n wd2 = tf.Variable(tf.truncated_normal(shape=(1024, 512), mean = mu, stddev = sigma))\n bd2 = tf.Variable(tf.zeros(512))\n fc2 = tf.matmul(fc1, wd2) + bd2\n \n # Activation.\n fc2 = tf.nn.relu(fc2)\n fc2 = tf.cond(is_training, lambda: tf.nn.dropout(fc2, keep_prob = 0.5), lambda: fc2)\n\n # Layer 5: Fully Connected. Input = 512. Output = 43.\n wd3 = tf.Variable(tf.truncated_normal(shape=(512, output_size), mean = mu, stddev = sigma))\n bd3 = tf.Variable(tf.zeros(output_size))\n logits = tf.matmul(fc2, wd3) + bd3\n weights = [wd1,wd2,wd3]\n \n return logits, weights\n\n\nfrom sklearn.utils import shuffle\n\nsave_file = './lenet'\nEPOCHS = 100\nBATCH_SIZE = 256\n\nx = tf.placeholder(tf.float32, (None, 32, 32, 1))\ny = tf.placeholder(tf.int32, (None))\nis_training = tf.placeholder(tf.bool)\none_hot_y = tf.one_hot(y, 43)\n\nrate = 0.001\n\nlogits, weights = LeNet(x, is_training)\ncross_entropy = tf.nn.softmax_cross_entropy_with_logits(labels=one_hot_y, logits=logits)\n\n# L2 Regularization \nregularization = 0.0\nfor w in weights:\n regularization += tf.nn.l2_loss(w)\n\n#http://docs.aws.amazon.com/machine-learning/latest/dg/training-parameters.html\nL2_rate = 1e-4\n\nloss_operation = tf.reduce_mean(cross_entropy) + L2_rate*regularization\noptimizer = tf.train.AdamOptimizer(learning_rate = rate)\ntraining_operation = optimizer.minimize(loss_operation)\n\ncorrect_prediction = tf.equal(tf.argmax(logits, 1), tf.argmax(one_hot_y, 1))\naccuracy_operation = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))\nsaver = tf.train.Saver()\n\ndef evaluate(X_data, y_data):\n num_examples = len(X_data)\n total_accuracy = 0\n sess = tf.get_default_session()\n for offset in range(0, num_examples, BATCH_SIZE):\n batch_x, batch_y = X_data[offset:offset+BATCH_SIZE], y_data[offset:offset+BATCH_SIZE]\n accuracy = sess.run(accuracy_operation, feed_dict={x: batch_x, y: batch_y, is_training: False})\n total_accuracy += (accuracy * len(batch_x))\n return total_accuracy / num_examples\n\n# Train the model\nwith tf.Session() as sess:\n sess.run(tf.global_variables_initializer())\n num_examples = len(X_train)\n \n print(\"Training...\")\n print()\n for i in range(EPOCHS):\n X_train, y_train = shuffle(X_train, y_train)\n for offset in range(0, num_examples, BATCH_SIZE):\n end = offset + BATCH_SIZE\n batch_x, batch_y = X_train[offset:end], y_train[offset:end]\n sess.run(training_operation, feed_dict={x: batch_x, y: batch_y, is_training: True})\n \n validation_accuracy = evaluate(X_valid, y_valid)\n if i%5 == 0:\n print(\"EPOCH {} ...\".format(i+1))\n print(\"Validation Accuracy = {:.3f}\".format(validation_accuracy))\n print()\n \n saver.save(sess, save_file)\n print(\"Model saved\")\n","sub_path":"train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":5530,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"123826350","text":"import serial\nimport traceback\nimport threading\nimport serial.threaded\n\n# arguments\n# port – Device name or None.\n# baudrate (int) – Baud rate such as 9600 or 115200 etc.\n# bytesize – Number of data bits. Possible values: FIVEBITS, SIXBITS, SEVENBITS, EIGHTBITS\n# parity – Enable parity checking. Possible values: PARITY_NONE, PARITY_EVEN, PARITY_ODD PARITY_MARK, PARITY_SPACE\n# stopbits – Number of stop bits. Possible values: STOPBITS_ONE, STOPBITS_ONE_POINT_FIVE, STOPBITS_TWO\n# timeout (float) – Set a read timeout value.\n# xonxoff (bool) – Enable software flow control.\n# rtscts (bool) – Enable hardware (RTS/CTS) flow control. ALWAYS TRUE\n# dsrdtr (bool) – Enable hardware (DSR/DTR) flow control. ALWAYS TRUE\n# write_timeout (float) – Set a write timeout value.\n# inter_byte_timeout (float) – Inter-character timeout, None to disable (default).\n\n# Raises:\n# ValueError – Will be raised when parameter are out of range, e.g. baud rate, data bits.\n# SerialException – In case the device can not be found or can not be configured.\n\n\nclass Connection:\n \"\"\"\n Физическое подключение\n \"\"\"\n class PrintLines(serial.threaded.LineReader):\n \"\"\"\n Класс для асинхронного чтения из порта\n \"\"\"\n on_received = None\n on_exception = None\n buffer = bytearray()\n receiving_now = False\n\n def connection_made(self, transport):\n super(Connection.PrintLines, self).connection_made(transport)\n\n def data_received(self, data):\n # Если порт не расчитан на прием, игнорируем данные\n if self.on_received is None:\n return\n for b in data:\n if b == Connection.start_byte():\n self.receiving_now = True\n self.buffer = bytearray()\n elif b == Connection.stop_byte():\n self.receiving_now = False\n self.on_received(self.buffer)\n self.buffer = bytearray()\n elif self.receiving_now:\n self.buffer.append(b)\n\n def handle_line(self, line):\n pass\n\n def connection_lost(self, exc):\n # exc - исключение, если что-то пошло не так\n if exc:\n if self.on_exception is not None:\n self.on_exception()\n else:\n traceback.print_exc(exc)\n\n def __init__(self):\n self.ser = None\n self.stop_event = threading.Event()\n self.thread = threading.Thread(target=self.reading, args=(self.stop_event, ))\n self.on_received = None\n self.on_exception = None\n\n @staticmethod\n def start_byte():\n return 0x02\n\n @staticmethod\n def stop_byte():\n return 0x04\n\n def connect(self, port_name, **kargs):\n self.ser = serial.Serial(port_name, rtscts=True, dsrdtr=True, **kargs)\n self.thread.start()\n\n def disconnect(self):\n # Останавливаем поток, в котором читаются данные\n self.stop_event.set()\n self.thread.join()\n # Закрываем соединение\n if self.ser is not None:\n self.ser.close()\n self.ser = None\n\n def is_connected(self):\n return self.ser is not None\n\n def reading(self, stop_event):\n # Запускаем чтение в новом потоке\n with serial.threaded.ReaderThread(self.ser, Connection.PrintLines) as protocol:\n protocol.on_received = self.on_received\n protocol.on_exception = self.on_exception\n while not stop_event.is_set():\n serial.time.sleep(0.05)\n\n def write(self, data):\n buffer = bytearray()\n buffer.append(Connection.start_byte())\n buffer.extend(data)\n buffer.append(Connection.stop_byte())\n try:\n self.ser.write(buffer)\n except serial.SerialException as e:\n if self.on_exception is not None:\n self.on_exception()\n else:\n traceback.print_exc(e)\n","sub_path":"physical_layer/connection.py","file_name":"connection.py","file_ext":"py","file_size_in_byte":4285,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"512207712","text":"import cv2\nimport numpy as np\nimport sysv_ipc\n\n# Read image data from shared memory.\nmemory = sysv_ipc.SharedMemory(123456)\ndata = memory.read()\n\n# Construct image using numpy.\nframe = np.frombuffer(data, 'uint8')\nwidth = 640\nheight = 424\nshape = (height, width, 3)\nstrides = (width * 3, 3, 1)\nframe = np.lib.stride_tricks.as_strided(frame, shape, strides)\n\n# Show the image using OpenCV.\ncv2.imshow('image from shared memory', frame)\ncv2.waitKey()\ncv2.destroyAllWindows()\n","sub_path":"imshow.py","file_name":"imshow.py","file_ext":"py","file_size_in_byte":473,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"473400803","text":"import os\n\n\ndef has_weights(root, path_to_check):\n \"\"\"Checks for model weight paths from weights folder\n\n Args:\n root (str): path of peekingduck root folder\n path_to_check (List[str]): list of files/directories to check\n to see if weights exists\n\n Returns:\n boolean: True is files/directories needed exists, else False\n \"\"\"\n\n # Check for whether weights dir even exist. If not make directory\n # Empty directory should then return False\n weights_dir = os.path.join(root, '..', 'weights')\n if not os.path.isdir(weights_dir):\n os.mkdir(weights_dir)\n return False\n\n for check in path_to_check:\n if not os.path.exists(os.path.join(root, check)):\n return False\n return True","sub_path":"peekingduck/weights_utils/checker.py","file_name":"checker.py","file_ext":"py","file_size_in_byte":761,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"451910979","text":"############################################################################\r\n# #\r\n# Copyright (c)2008, 2009, Digi International (Digi). All Rights Reserved. #\r\n# #\r\n# Permission to use, copy, modify, and distribute this software and its #\r\n# documentation, without fee and without a signed licensing agreement, is #\r\n# hereby granted, provided that the software is used on Digi products only #\r\n# and that the software contain this copyright notice, and the following #\r\n# two paragraphs appear in all copies, modifications, and distributions as #\r\n# well. Contact Product Management, Digi International, Inc., 11001 Bren #\r\n# Road East, Minnetonka, MN, +1 952-912-3444, for commercial licensing #\r\n# opportunities for non-Digi products. #\r\n# #\r\n# DIGI SPECIFICALLY DISCLAIMS ANY WARRANTIES, INCLUDING, BUT NOT LIMITED #\r\n# TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A #\r\n# PARTICULAR PURPOSE. THE SOFTWARE AND ACCOMPANYING DOCUMENTATION, IF ANY, #\r\n# PROVIDED HEREUNDER IS PROVIDED \"AS IS\" AND WITHOUT WARRANTY OF ANY KIND. #\r\n# DIGI HAS NO OBLIGATION TO PROVIDE MAINTENANCE, SUPPORT, UPDATES, #\r\n# ENHANCEMENTS, OR MODIFICATIONS. #\r\n# #\r\n# IN NO EVENT SHALL DIGI BE LIABLE TO ANY PARTY FOR DIRECT, INDIRECT, #\r\n# SPECIAL, INCIDENTAL, OR CONSEQUENTIAL DAMAGES, INCLUDING LOST PROFITS, #\r\n# ARISING OUT OF THE USE OF THIS SOFTWARE AND ITS DOCUMENTATION, EVEN IF #\r\n# DIGI HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. #\r\n# #\r\n############################################################################\r\n\r\n\"\"\"\\\r\n The Xbee Device Manager Event Specs base class, as well as\r\n a few various classes built upon the base class.\r\n\r\n\r\n An event spec is a tuple:\r\n (addr, endpoint, profile_id, cluster_id)\r\n\r\n The mask is an (inverted) Boolean, so True means check this field.\r\n\r\n (True, False, False, True) means addr and cluster_id must match.\r\n\r\n\"\"\"\r\n\r\n# imports\r\nimport string\r\n\r\nfrom devices.xbee.common.addressing import addresses_equal\r\n\r\n# constants\r\n\r\n# exception classes\r\n\r\n# interface functions\r\n\r\n# classes\r\n\r\nclass XBeeDeviceManagerEventSpec:\r\n \"\"\"\\\r\n The Event Spec base class.\r\n\r\n This class is the foundation class to build various different\r\n Event Spec classes around.\r\n\r\n \"\"\"\r\n def __init__(self):\r\n self.__cb = None\r\n\r\n def cb_set(self, cb):\r\n self.__cb = cb\r\n\r\n\r\n def cb_get(self):\r\n return self.__cb\r\n\r\n\r\n def match_spec_set(self, *args, **kwargs):\r\n \"\"\"\\\r\n This function's calling interface is allowed to change, it is\r\n never called by the device manager.\r\n\r\n \"\"\"\r\n raise NotImplementedError(\"virtual function\")\r\n\r\n\r\n def match_spec_test(self, candidate):\r\n raise NotImplementedError(\"virtual function\")\r\n\r\n\r\nclass XBeeDeviceManagerRxEventSpec(XBeeDeviceManagerEventSpec):\r\n \"\"\"\\\r\n The Rx Event Spec class.\r\n\r\n XBeeDeviceManagerRxEventSpec is called when data comes\r\n in from a specific XBee device that matches the given\r\n specs.\r\n\r\n \"\"\"\r\n def __init__(self):\r\n self._match_spec = None\r\n self._match_mask = None\r\n self._listen_endpoint = None\r\n\r\n XBeeDeviceManagerEventSpec.__init__(self)\r\n\r\n def get_listen_endpoint(self):\r\n if self._listen_endpoint != None:\r\n return self._listen_endpoint\r\n else:\r\n return self._match_spec[1]\r\n\r\n def _check_addr(self, addr, varname=\"addr\"):\r\n if not (isinstance(addr, tuple) and len(addr) >= 4):\r\n raise AttributeError(\r\n \"bad %s, must be tuple of >= length 4. (%s given)\" \\\r\n % (varname, addr))\r\n\r\n if not isinstance(addr[0], str) and addr[0] != None:\r\n raise AttributeError(\r\n \"bad %s, first item must be string or None\" % (varname))\r\n\r\n if not reduce(lambda p, q: p and isinstance(q, int), addr[1:4], True):\r\n raise AttributeError(\r\n \"bad %s, [1:4] must be integers.\" % (varname))\r\n\r\n def match_spec_set(self, match_spec, match_mask, listen_endpoint=None):\r\n\r\n self._check_addr(match_spec, \"match_spec\")\r\n\r\n # is our listen endopoint different than the one passed in\r\n # the match spec? (0 is a valid endpoint)\r\n if listen_endpoint != None:\r\n self._listen_endpoint = listen_endpoint\r\n\r\n if not (isinstance(match_mask, tuple) and len(match_mask) == 4):\r\n raise AttributeError(\"bad match_mask, must be tuple of length 4.\")\r\n\r\n if not reduce(lambda p, q: p and \\\r\n isinstance(q, bool), match_mask, True):\r\n raise AttributeError(\"bad match_spec, [0:4] must be booleans.\")\r\n\r\n if match_spec[0] != None:\r\n addr = match_spec[0].lower()\r\n else:\r\n addr = match_spec[0]\r\n final_match_spec = (addr, match_spec[1], match_spec[2], match_spec[3])\r\n\r\n self._match_spec = final_match_spec\r\n self._match_mask = match_mask\r\n\r\n\r\n def match_spec_get(self):\r\n return (self._match_spec, self._match_mask)\r\n\r\n\r\n def match_spec_test(self, candidate, mac_prematch=False):\r\n if not mac_prematch:\r\n self._check_addr(candidate, \"candidate\")\r\n\r\n # TODO: move stripping logic into common addressing helper lib:\r\n if self._match_mask[0] and \\\r\n not addresses_equal(self._match_spec[0], candidate[0]):\r\n return False\r\n\r\n ##Couple of points to make here. We only match elements 1-3, and allow\r\n ##Short circuiting the if comparison. spec and mask are moved into local\r\n ##scope for performance boost instead of being class variables.\r\n\r\n ##Return False if not matched,\r\n ##Return True if matched.\r\n\r\n spec = self._match_spec\r\n mask = self._match_mask\r\n\r\n if not ((not mask[1] or (candidate[1] == spec[1])) and\r\n (not mask[2] or (candidate[2] == spec[2])) and\r\n (not mask[3] or (candidate[3] == spec[3]))):\r\n return False\r\n return True\r\n\r\n\r\nclass XBeeDeviceManagerRxConfigEventSpec(XBeeDeviceManagerRxEventSpec):\r\n \"\"\"\r\n The Rx Config Event Spec class.\r\n\r\n XBeeDeviceManagerRxConfigEventSpec is a XBeeDeviceManagerRxEventSpec\r\n except that its callbacks get processed when a device is in the\r\n config state instead of the running state.\r\n\r\n \"\"\"\r\n pass\r\n\r\n\r\nclass XBeeDeviceManagerRunningEventSpec(XBeeDeviceManagerEventSpec):\r\n \"\"\"\r\n The Running Event Spec class.\r\n\r\n XBeeDeviceManagerRunningEventSpec is called when a device enters\r\n the running state.\r\n\r\n \"\"\"\r\n def match_spec_set(self):\r\n pass\r\n\r\n\r\n def match_spec_test(self):\r\n pass\r\n","sub_path":"src/devices/xbee/xbee_device_manager/xbee_device_manager_event_specs.py","file_name":"xbee_device_manager_event_specs.py","file_ext":"py","file_size_in_byte":7208,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"244484908","text":"import folium\nfrom parser import parse_file, convert_to_coordinates, neighbors\n\n\ndef create_map(dct, user_input):\n '''\n dict(), tuple -> ()\n Function for creating a map with all layers.\n '''\n map = folium.Map(location=user_input, zoom_start=3)\n tooltip = 'Click me!'\n\n fg1 = folium.FeatureGroup(name='Locations')\n fg2 = folium.FeatureGroup(name='Movies')\n fg3 = folium.FeatureGroup(name=\"Area\")\n for key in dct:\n folium.Marker(user_input(), popup=key, tooltip=tooltip).add_to(map)\n fg1.add_child(folium.CircleMarker(location=[key[0], key[1]],\n radius=10,\n popup=key,\n color='red',\n fill_opacity=0.5))\n\n fg2.add_child(folium.Marker(location=[key[0], key[1]],\n popup=dct[key],\n icon=folium.Icon()))\n\n fg3.add_child(folium.GeoJson(data=open('world.json', 'r',\n encoding='utf-8-sig').read(),\n style_function=lambda x: {'fillColor': 'white'\n if x['properties']['AREA'] < 1000\n else 'blue' if 1000 <= x['properties']['AREA'] < 100000\n else 'green' if 100000 <= x['properties']['AREA'] < 1000000\n else 'purple'}))\n map.add_child(fg1)\n map.add_child(fg2)\n map.add_child(fg3)\n map.add_child(folium.LayerControl())\n map.save(\"Map_movies.html\")\n\n\nif __name__ == \"__main__\":\n year = input(\"Please enter a year you would like to have a map for: \")\n coordinates = input(\"Please enter your location (format: lat, long): \")\n coordinates = eval(coordinates)\n inf = convert_to_coordinates(parse_file(year))\n loc = neighbors(coordinates, inf)\n create_map(loc, coordinates)\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2005,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"301292726","text":"from keras.models import Sequential\r\nimport cv2\r\nimport numpy as np\r\nfrom keras.preprocessing import image\r\nfrom keras.models import load_model\r\nfrom keras.layers import Convolution2D as Conv2D, MaxPooling2D, Dropout\r\nfrom keras.layers.convolutional import Deconv2D as Conv2DTranspose\r\nfrom keras.layers import Dense, Activation, Flatten\r\n\r\ndef nothing(x):\r\n pass\r\n\r\ndef create_model():\r\n # Initialising the CNN\r\n classifier = Sequential()\r\n\r\n# Adding first convolutional layer, followed by pooling, and dropout\r\n classifier.add(Conv2D(32, (3, 3), input_shape = (64, 64, 3), activation = 'relu'))\r\n classifier.add(MaxPooling2D(pool_size = (2, 2)))\r\n classifier.add(Dropout(0.25))\r\n\r\n# Adding second convolutional layer, followed by pooling, and dropout\r\n classifier.add(Conv2D(32, (3, 3), activation = 'relu'))\r\n classifier.add(MaxPooling2D(pool_size = (2, 2)))\r\n classifier.add(Dropout(0.25))\r\n\r\n# Adding third convolutional layer, followed by pooling, and dropout\r\n classifier.add(Conv2D(32, (3, 3), activation = 'relu'))\r\n classifier.add(MaxPooling2D(pool_size = (2, 2)))\r\n classifier.add(Dropout(0.25))\r\n\r\n# Flattening\r\n classifier.add(Flatten())\r\n\r\n# Full connection\r\n classifier.add(Dense(units = 128, activation = 'relu'))\r\n classifier.add(Dense(units = 10, activation = 'softmax'))\r\n\r\n# Compiling the CNN\r\n classifier.compile(optimizer = 'adam', loss = 'categorical_crossentropy', metrics = ['accuracy'])\r\n return classifier\r\n\r\n\r\n\r\n\r\nimage_x, image_y = 64,64\r\nmodel = create_model()\r\nmodel.load_weights('model.h5')\r\n\r\ndef predictor():\r\n\r\n img = image.load_img('c2.jpg', target_size=(64, 64))\r\n img = image.img_to_array(img)\r\n img = np.expand_dims(img, axis = 0)\r\n result = model(img)\r\n \r\n if result[0][0] == 1:\r\n return '0'\r\n elif result[0][1] == 1:\r\n return '1'\r\n elif result[0][2] == 1:\r\n return '2'\r\n elif result[0][3] == 1:\r\n return '3'\r\n elif result[0][4] == 1:\r\n return '4'\r\n elif result[0][5] == 1:\r\n return '5'\r\n elif result[0][6] == 1:\r\n return '6'\r\n elif result[0][7] == 1:\r\n return '7'\r\n elif result[0][8] == 1:\r\n return '8'\r\n elif result[0][9] == 1:\r\n return '9'\r\ncommand = predictor();\r\nprint(command);","sub_path":"NewModel.py","file_name":"NewModel.py","file_ext":"py","file_size_in_byte":2362,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"143023029","text":"# Copyright (c) 2014-2017, iocage\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted providing that the following conditions\n# are met:\n# 1. Redistributions of source code must retain the above copyright\n# notice, this list of conditions and the following disclaimer.\n# 2. Redistributions in binary form must reproduce the above copyright\n# notice, this list of conditions and the following disclaimer in the\n# documentation and/or other materials provided with the distribution.\n#\n# THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR\n# IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED\n# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE\n# ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY\n# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL\n# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS\n# OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)\n# HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,\n# STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING\n# IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE\n# POSSIBILITY OF SUCH DAMAGE.\nimport libiocage.lib.errors\n\n\nclass ZFSShareStorage:\n @property\n def zfs_datasets(self):\n return self._get_zfs_datasets(auto_create=self.auto_create)\n\n def mount_zfs_shares(self, auto_create=False):\n self.logger.log(\"Mounting ZFS shares\")\n self._mount_procfs()\n self._mount_jail_datasets(auto_create=auto_create)\n\n def _get_zfs_datasets(self, auto_create=None):\n dataset_names = self.jail.config[\"jail_zfs_dataset\"]\n\n auto_create = self.auto_create if auto_create is None else auto_create\n\n datasets = set()\n for name in dataset_names:\n\n zpool = None\n try:\n zpool = self._get_pool_from_dataset_name(name)\n except:\n pass\n\n pool_name = f\"{self.jail.zfs_pool_name}/{name}\"\n try:\n # legacy support (datasets not prefixed with pool/)\n zpool = self._get_pool_from_dataset_name(pool_name)\n name = f\"{self.jail.zfs_pool_name}/{name}\"\n except:\n pass\n\n try:\n if auto_create:\n zpool.create(name, {}, create_ancestors=True)\n except:\n pass\n\n try:\n dataset = self.zfs.get_dataset(name)\n datasets.add(dataset)\n except:\n raise libiocage.lib.errors.DatasetNotAvailable(\n dataset_name=name,\n logger=self.logger\n )\n\n return datasets\n\n def _mount_jail_dataset(self, dataset_name):\n self.jail.exec(['zfs', 'mount', dataset_name])\n\n def _mount_jail_datasets(self, auto_create=None):\n\n auto_create = self.auto_create if auto_create is None else (\n auto_create is True)\n\n if self.safe_mode:\n self._require_datasets_exist_and_jailed()\n\n for dataset in self.zfs_datasets:\n\n self._unmount_local(dataset)\n\n # ToDo: bake jail feature into py-libzfs\n libiocage.lib.helpers.exec(\n [\"zfs\", \"jail\", self.jail.identifier, dataset.name])\n\n if dataset.properties['mountpoint']:\n for child in list(dataset.children):\n self._ensure_dataset_exists(child)\n self._mount_jail_dataset(child.name)\n\n def _get_pool_name_from_dataset_name(self, dataset_name):\n return dataset_name.split(\"/\", maxsplit=1)[0]\n\n def _get_pool_from_dataset_name(self, dataset_name):\n target_pool_name = self._get_pool_name_from_dataset_name(dataset_name)\n for zpool in list(self.zfs.pools):\n if zpool.name == target_pool_name:\n return zpool\n\n # silent exception, no logger defined\n raise libiocage.lib.errors.ZFSPoolUnavailable(\n pool_name=target_pool_name\n )\n\n def _require_datasets_exist_and_jailed(self):\n existing_datasets = self.get_zfs_datasets(auto_create=False)\n for existing_dataset in existing_datasets:\n if existing_dataset.properties[\"jailed\"] != \"on\":\n raise libiocage.lib.errors.DatasetNotJailed(\n dataset=existing_dataset,\n logger=self.logger\n )\n","sub_path":"libiocage/lib/ZFSShareStorage.py","file_name":"ZFSShareStorage.py","file_ext":"py","file_size_in_byte":4556,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"361004516","text":"#!/usr/bin/python3 \n\nimport os, glob \n\n# getting file list\ndef get_file_list(path):\n return glob.glob(path)\n\n\ndef filter_file_list(path, exp, dir_filter):\n\n def is_dir(fo):\n return True if os.path.isdir(fo) else False\n\n flist = glob.glob(os.path.join(path, exp))\n\n if dir_filter is None:\n return flist\n elif dir_filter:\n return list(filter(os.path.isdir, flist))\n else:\n return list(filter(lambda x:not os.path.isdir(x), flist))\n\n return glob.glob(os.path.join(path, exp))\n\n\ndef main():\n print(\"all list\")\n flist = get_file_list(\"./*\")\n print(flist+\"\\n\")\n\n print(\"all list with filter\")\n flist = filter_file_list(\"./\", \"*.py\", None)\n print(flist+\"\\n\")\n\n print(\"all list\")\n flist = filter_file_list(\"./\", \"*\", None)\n print(flist+\"\\n\")\n\n print(\"all dir list\")\n flist = filter_file_list(\"./\", \"*\", True)\n print(flist+\"\\n\")\n\n print(\"all file list\")\n flist = filter_file_list(\"./\", \"*\", False)\n print(flist+\"\\n\")\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"fs/get_file_list.py","file_name":"get_file_list.py","file_ext":"py","file_size_in_byte":1042,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"116361730","text":"# -*- coding: utf-8 -*-\n\"\"\"\n@author:XuMing(xuming624@qq.com), Abtion(abtion@outlook.com)\n@description: \n\"\"\"\nimport operator\nfrom abc import ABC\n\nimport torch\nimport torch.nn as nn\nimport numpy as np\nimport pytorch_lightning as pl\nfrom pycorrector.macbert import lr_scheduler\nfrom pycorrector.macbert.evaluate_util import compute_corrector_prf, compute_sentence_level_prf\nfrom pycorrector.utils.logger import logger\n\n\nclass FocalLoss(nn.Module):\n \"\"\"\n Softmax and sigmoid focal loss.\n copy from https://github.com/lonePatient/TorchBlocks\n \"\"\"\n\n def __init__(self, num_labels, activation_type='softmax', gamma=2.0, alpha=0.25, epsilon=1.e-9):\n super().__init__()\n self.num_labels = num_labels\n self.gamma = gamma\n self.alpha = alpha\n self.epsilon = epsilon\n self.activation_type = activation_type\n\n def forward(self, input, target):\n \"\"\"\n Args:\n logits: model's output, shape of [batch_size, num_cls]\n target: ground truth labels, shape of [batch_size]\n Returns:\n shape of [batch_size]\n \"\"\"\n if self.activation_type == 'softmax':\n idx = target.view(-1, 1).long()\n one_hot_key = torch.zeros(idx.size(0), self.num_labels, dtype=torch.float32, device=idx.device)\n one_hot_key = one_hot_key.scatter_(1, idx, 1)\n logits = torch.softmax(input, dim=-1)\n loss = -self.alpha * one_hot_key * torch.pow((1 - logits), self.gamma) * (logits + self.epsilon).log()\n loss = loss.sum(1)\n elif self.activation_type == 'sigmoid':\n multi_hot_key = target\n logits = torch.sigmoid(input)\n zero_hot_key = 1 - multi_hot_key\n loss = -self.alpha * multi_hot_key * torch.pow((1 - logits), self.gamma) * (logits + self.epsilon).log()\n loss += -(1 - self.alpha) * zero_hot_key * torch.pow(logits, self.gamma) * (1 - logits + self.epsilon).log()\n return loss.mean()\n\n\ndef make_optimizer(lr, weight_decay, optimizer_name, param_dict):\n params = []\n for key, value in param_dict:\n if not value.requires_grad:\n continue\n if \"bias\" in key:\n lr = lr * 2\n weight_decay = 0\n params += [{\"params\": [value], \"lr\": lr, \"weight_decay\": weight_decay}]\n optimizer = getattr(torch.optim, optimizer_name)(params)\n return optimizer\n\n\ndef build_lr_scheduler(optimizer):\n scheduler_args = {\n \"optimizer\": optimizer,\n\n # warmup options\n \"warmup_factor\": 0.01,\n \"warmup_epochs\": 1024,\n \"warmup_method\": \"linear\",\n\n # multi-step lr scheduler options\n \"milestones\": (10,),\n \"gamma\": 0.9999,\n\n # cosine annealing lr scheduler options\n \"max_iters\": 10,\n \"delay_iters\": 0,\n \"eta_min_lr\": 3e-7,\n\n }\n scheduler = getattr(lr_scheduler, \"WarmupExponentialLR\")(**scheduler_args)\n return {'scheduler': scheduler, 'interval': 'step'}\n\n\nclass BaseTrainingEngine(pl.LightningModule):\n def __init__(self, lr=5e-5, weight_decay=0.01, optimizer_name='AdamW', *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.lr = lr\n self.weight_decay = weight_decay\n self.optimizer_name = optimizer_name\n\n def configure_optimizers(self):\n optimizer = make_optimizer(self.lr, self.weight_decay, self.optimizer_name, self.named_parameters())\n scheduler = build_lr_scheduler(optimizer)\n\n return [optimizer], [scheduler]\n\n def on_validation_epoch_start(self) -> None:\n logger.info('Valid.')\n\n def on_test_epoch_start(self) -> None:\n logger.info('Testing...')\n\n\nclass CscTrainingModel(BaseTrainingEngine, ABC):\n \"\"\"用于CSC的TrainingModel, 定义了训练及预测步骤\"\"\"\n\n def __init__(self, lr=5e-5, weight_decay=0.01, optimizer_name='AdamW',\n loss_coefficient=0.3, device=torch.device('cuda'), *args, **kwargs):\n super().__init__(lr=lr, weight_decay=weight_decay, optimizer_name=optimizer_name, *args, **kwargs)\n # loss weight\n self.w = loss_coefficient\n self._device = device\n\n def training_step(self, batch, batch_idx):\n ori_text, cor_text, det_labels = batch\n outputs = self.forward(ori_text, cor_text, det_labels)\n loss = self.w * outputs[1] + (1 - self.w) * outputs[0]\n return loss\n\n def validation_step(self, batch, batch_idx):\n ori_text, cor_text, det_labels = batch\n outputs = self.forward(ori_text, cor_text, det_labels)\n loss = self.w * outputs[1] + (1 - self.w) * outputs[0]\n det_y_hat = (outputs[2] > 0.5).long()\n cor_y_hat = torch.argmax((outputs[3]), dim=-1)\n encoded_x = self.tokenizer(cor_text, padding=True, return_tensors='pt')\n encoded_x.to(self._device)\n cor_y = encoded_x['input_ids']\n cor_y_hat *= encoded_x['attention_mask']\n\n results = []\n det_acc_labels = []\n cor_acc_labels = []\n for src, tgt, predict, det_predict, det_label in zip(ori_text, cor_y, cor_y_hat, det_y_hat, det_labels):\n _src = self.tokenizer(src, add_special_tokens=False)['input_ids']\n _tgt = tgt[1:len(_src) + 1].cpu().numpy().tolist()\n _predict = predict[1:len(_src) + 1].cpu().numpy().tolist()\n cor_acc_labels.append(1 if operator.eq(_tgt, _predict) else 0)\n det_acc_labels.append(det_predict[1:len(_src) + 1].equal(det_label[1:len(_src) + 1]))\n results.append((_src, _tgt, _predict,))\n\n return loss.cpu().item(), det_acc_labels, cor_acc_labels, results\n\n def validation_epoch_end(self, outputs) -> None:\n det_acc_labels = []\n cor_acc_labels = []\n results = []\n for out in outputs:\n det_acc_labels += out[1]\n cor_acc_labels += out[2]\n results += out[3]\n loss = np.mean([out[0] for out in outputs])\n self.log('val_loss', loss)\n logger.info(f'loss: {loss}')\n logger.info(f'Detection:\\n'\n f'acc: {np.mean(det_acc_labels):.4f}')\n logger.info(f'Correction:\\n'\n f'acc: {np.mean(cor_acc_labels):.4f}')\n compute_corrector_prf(results, logger)\n compute_sentence_level_prf(results, logger)\n\n def test_step(self, batch, batch_idx):\n return self.validation_step(batch, batch_idx)\n\n def test_epoch_end(self, outputs) -> None:\n logger.info('Test.')\n self.validation_epoch_end(outputs)\n\n def predict(self, texts):\n inputs = self.tokenizer(texts, padding=True, return_tensors='pt')\n inputs.to(self._device)\n with torch.no_grad():\n outputs = self.forward(texts)\n y_hat = torch.argmax(outputs[1], dim=-1)\n expand_text_lens = torch.sum(inputs['attention_mask'], dim=-1) - 1\n rst = []\n for t_len, _y_hat in zip(expand_text_lens, y_hat):\n rst.append(self.tokenizer.decode(_y_hat[1:t_len]).replace(' ', ''))\n return rst\n","sub_path":"pycorrector/macbert/base_model.py","file_name":"base_model.py","file_ext":"py","file_size_in_byte":7036,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"358471591","text":"#Displays the credits\r\n\r\nimport pygame\r\nimport constants\r\n\r\nfrom level_manager import *\r\nfrom art import *\r\n\r\nclass CreditsScreen():\r\n def __init__(self):\r\n \r\n title_font = pygame.font.SysFont('castellar', 48, True, False)\r\n credit_font = pygame.font.SysFont('mangal', 20, True, False)\r\n instruction_font = pygame.font.SysFont('mangal', 32, True, False)\r\n self._text = title_font.render(\"Credits\",True,constants.BLACK)\r\n self._instruct_text1 = instruction_font.render(\"Press escape to return to the title screen.\", True, constants.BLACK)\r\n self._credit_text1 = credit_font.render(\"Game designed by Monica Timmerman, Jack Olson,\"\r\n + \" Carrie Mannilla, Bryant Lennick, Camryn Roadley, & Dan Nygard\", True, constants.BLACK)\r\n self._credit_text2 = credit_font.render(\"Game background music is \\\"Awesome Call\\\" by Kevin MacLeod (incompetech.com) \",True, constants.BLACK)\r\n self._credit_text3 = credit_font.render(\"Licensed under Creative Commons: By Attribution 3.0 License http://creativecommons.org/licenses/by/3.0/\", True, constants.BLACK)\r\n self._credit_text4 = credit_font.render(\"Sound effects from free sounds library at https://www.partnersinrhyme.com/pir/PIRsfx.shtml:\", True, constants.BLACK)\r\n self._credit_text5 = credit_font.render(\"Rabbit attack: https://www.partnersinrhyme.com/soundfx/fight_sounds/fight_punch_wav.shtml\", True, constants.BLACK)\r\n self._credit_text6 = credit_font.render(\"Harvest sound: https://www.partnersinrhyme.com/soundfx/office_sounds/office_cash-register2_wav.shtml\", True, constants.BLACK)\r\n\r\n self._credit_text7 = credit_font.render(\"All game art by Bryant Lennick using Piskel (https://www.piskelapp.com/).\",True, constants.BLACK)\r\n self._credit_text8 = credit_font.render(\"Sprite animation created using PygAnim, Copyright 2014 Al Sweigart\", True, constants.BLACK)\r\n\r\n \r\n\r\n def handle_keyboard_event(self, event):\r\n if event.type == pygame.KEYDOWN:\r\n if event.key == pygame.K_ESCAPE:\r\n LevelManager().leave_level()\r\n\r\n #No need to do anything here, unless we've got some animation\r\n def update(self):\r\n pass\r\n \r\n def draw(self, screen):\r\n # Clear the screen\r\n screen.blit(pygame.image.load('images/background.png').convert_alpha(), [0, 0])\r\n \r\n # Draw my credits text!\r\n screen.blit(self._text, [380, 150])\r\n screen.blit(self._credit_text1, [210, 200])\r\n screen.blit(self._credit_text2, [210, 225])\r\n screen.blit(self._credit_text3, [225, 250])\r\n screen.blit(self._credit_text4, [210, 275])\r\n screen.blit(self._credit_text5, [225, 300])\r\n screen.blit(self._credit_text6, [225, 325])\r\n screen.blit(self._credit_text7, [210, 350])\r\n screen.blit(self._credit_text8, [210, 375])\r\n\r\n screen.blit(self._instruct_text1, [210, 450])\r\n \r\n","sub_path":"credits_screen.py","file_name":"credits_screen.py","file_ext":"py","file_size_in_byte":2994,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"640786965","text":"# -*- coding: utf-8 -*-\nfrom multiprocessing import Process, Manager\n\n# API CODE\napi_key = \"\"\napi_secret = \"\"\n\nimport time, base64, hmac, hashlib, urllib.parse, pycurl, json, io\n#빗썸 API 호출 클래스\nclass XCoinAPI:\n def __init__(self):\n self.api_url = \"https://api.bithumb.com\"\n self.api_key = api_key\n self.api_secret = api_secret\n\n def xcoinApiCall(self, EP, Params, nonce = str(time.time()*1000)):\n Params = dict({\"endpoint\" : EP},**Params)\n str_data = urllib.parse.urlencode(Params)\n contents = io.BytesIO()\n\n utf8_data = (EP + chr(0) + str_data + chr(0) + nonce).encode('utf8')\n utf8_key = self.api_secret.encode('utf8')\n utf8_api_sign = base64.b64encode(hmac.new(bytes(utf8_key), utf8_data, hashlib.sha512).hexdigest().encode('utf8')).decode('utf8')\n\n curl_handle = pycurl.Curl()\n curl_handle.setopt(pycurl.POST, 1)\n curl_handle.setopt(pycurl.POSTFIELDS, str_data)\n curl_handle.setopt(curl_handle.URL, self.api_url + EP)\n curl_handle.setopt(curl_handle.HTTPHEADER, ['Api-Key: ' + self.api_key, 'Api-Sign: ' + utf8_api_sign, 'Api-Nonce: ' + nonce])\n curl_handle.setopt(curl_handle.WRITEFUNCTION, contents.write)\n curl_handle.setopt(pycurl.SSL_VERIFYPEER, False)\n curl_handle.setopt(pycurl.SSL_VERIFYHOST, False)\n curl_handle.perform()\n curl_handle.close()\n return json.loads(contents.getvalue().decode())\n\n#apicall\nclass ApiCall() :\n def __init__(self,Coin_List) :\n self.api = XCoinAPI()\n self.Coin_List = Coin_List\n self.Checker = 0\n self.mydict = Manager().dict()\n\n\n def Call(self) :\n start = time.time()\n for self.Coin in self.Coin_List :\n self.mydict[self.Coin] = {}\n self.pro = []\n #for self.Coin in self.Coin_List :\n # self.pro.append(Process(target=self.Odb, args = (self.Coin,)))\n self.pro.append(Process(target=self.Odb_All, args = (self.mydict,)))\n self.pro.append(Process(target=self.Ticker_All, args = (self.mydict,)))\n if (self.Checker % 100) == 0 :\n self.pro.append(Process(target=self.Account, args = (self.mydict,)))\n if (self.Checker % 3) == 0 :\n self.pro.append(Process(target=self.Balance, args = (self.mydict,)))\n self.Ran = len(self.pro)\n for i in range(0,self.Ran) :\n self.pro[i].daemon = True\n self.pro[i].start()\n for i in range(0,self.Ran) :\n self.pro[i].join()\n self.pro[i].terminate()\n self.end = round(time.time() - start,2)\n self.mydict[\"end\"] = self.end\n self.Checker += 1\n return self.mydict\n\n def Ticker_All(self,mydict) :\n self.result = self.api.xcoinApiCall(\"/public/ticker/All\",{}, str(time.time()*1000))\n if self.result['status'] == \"0000\" :\n for self.Coin in self.Coin_List :\n self.mydata = self.result[\"data\"][self.Coin.upper()]\n if self.mydict[self.Coin] == {} :\n mydict[self.Coin] = {\"ticker\" : dict( self.mydata, **{\"Message\" : None})}\n else :\n mydict[self.Coin] = dict(mydict[self.Coin], **{\"ticker\" : dict( self.mydata, **{\"Message\" : None})})\n else :\n for self.Coin in self.Coin_List :\n mydict[self.Coin] = {\"ticker\" : \"Error\"}\n\n def Odb_All(self,mydict) :\n self.result = self.api.xcoinApiCall(\"/public/orderbook/All\",{}, str(time.time()*1000))\n if self.result['status'] == \"0000\" :\n mydict[\"timestamp\"] = self.result[\"data\"][\"timestamp\"]\n for self.Coin in self.Coin_List :\n self.mydata = self.result[\"data\"][self.Coin.upper()]\n if self.mydict[self.Coin] == {} :\n mydict[self.Coin] = {\"orderbook\" : { \"asks\" : self.mydata[\"asks\"], \"bids\" : self.mydata[\"bids\"], \"Message\" : None }}\n else :\n mydict[self.Coin] = dict(mydict[self.Coin], **{\"orderbook\" : { \"asks\" : self.mydata[\"asks\"], \"bids\" : self.mydata[\"bids\"], \"Message\" : None }})\n else :\n for self.Coin in self.Coin_List :\n mydict[self.Coin] = dict(mydict[self.Coin], **{\"orderbook\" : { \"asks\" : \"Error\", \"bids\" : \"Error\", \"Message\" : self.result[\"message\"] }})\n\n def Balance(self,mydict) :\n self.result = self.api.xcoinApiCall(\"/info/balance\",{\"currency\" : \"ALL\"}, str(time.time()*1000))\n if self.result['status'] == \"0000\" :\n mydict[\"balance\"] = dict(self.result[\"data\"],**{ \"Message\" : None })\n else :\n mydict[\"balance\"] = \"Error;\"+str(self.result['message'])\n\n def Account(self,mydict) :\n self.result = self.api.xcoinApiCall(\"/info/account\",{}, str(time.time()*1000))\n if self.result['status'] == \"0000\" :\n mydict[\"account\"] = dict(self.result[\"data\"],**{ \"Message\" : None })\n else :\n mydict[\"account\"] = \"Error;\"+str(self.result['message'])\n\nif __name__ == \"__main__\" :\n ApiCall(['btc', 'eth', 'dash', 'ltc', 'etc', 'xrp', 'bch', 'xmr', 'zec', 'qtum', 'btg', 'eos'])\n","sub_path":"Bithumb_Api.py","file_name":"Bithumb_Api.py","file_ext":"py","file_size_in_byte":5149,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"228737208","text":"# -*- coding: UTF-8 -*-\n\n\"\"\"\nThis file is part of Pondus, a personal weight manager.\nCopyright (C) 2007-11 Eike Nicklas \n\nThis program is free software licensed under the MIT license. For details\nsee LICENSE or http://www.opensource.org/licenses/mit-license.php\n\"\"\"\n\nimport pygtk\npygtk.require('2.0')\n\nimport gtk\n\nfrom pondus.core import parameters\n\n\ndef sort_function_weight(listmodel, iter1, iter2, data):\n \"\"\"Sorts the weight column correctly, i.e. interprets the weight\n data as floats instead of strings.\"\"\"\n weight1 = float(listmodel.get_value(iter1, 2))\n weight2 = float(listmodel.get_value(iter2, 2))\n return cmp(weight1, weight2)\n\n\ndef register_icons():\n \"\"\"Adds custom icons to the list of stock IDs.\"\"\"\n icon_info = {'pondus_plot': parameters.plot_button_path}\n iconfactory = gtk.IconFactory()\n stock_ids = gtk.stock_list_ids()\n for stock_id in icon_info:\n # only load image files when our stock_id is not present\n if stock_id not in stock_ids:\n icon_file = icon_info[stock_id]\n pixbuf = gtk.gdk.pixbuf_new_from_file(icon_file)\n iconset = gtk.IconSet(pixbuf)\n iconfactory.add(stock_id, iconset)\n iconfactory.add_default()\n\n\ndef get_tooltip(dataset):\n \"\"\"Returns a string containing the data of the optional parameters\n of the dataset.\"\"\"\n\n def append_to_tooltip(tooltip, text):\n if tooltip != '':\n tooltip += '\\n'\n return tooltip + text\n\n tooltip = ''\n if dataset.bodyfat is not None:\n text = _('Bodyfat') + ': ' + str(round(dataset.bodyfat, 1)) + '%'\n tooltip = append_to_tooltip(tooltip, text)\n if dataset.muscle is not None:\n text = _('Muscle') + ': ' + str(round(dataset.muscle, 1)) + '%'\n tooltip = append_to_tooltip(tooltip, text)\n if dataset.water is not None:\n text = _('Water') + ': ' + str(round(dataset.water, 1)) + '%'\n tooltip = append_to_tooltip(tooltip, text)\n if dataset.note is not None:\n text = _('Note') + ': ' + dataset.note\n tooltip = append_to_tooltip(tooltip, text)\n if tooltip == '':\n return None\n return tooltip\n","sub_path":"pondus/gui/guiutil.py","file_name":"guiutil.py","file_ext":"py","file_size_in_byte":2173,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"227796129","text":"# Default Imports\nfrom greyatomlib.python_intermediate.q05_read_csv_data.build import read_ipl_data_csv\nimport numpy as np\n\npath = 'data/ipl_matches_small.csv'\n\n# Enter Code Here\ndef get_total_extras():\n extras=0\n data=read_ipl_data_csv(path,\"|S50\")\n for i in range(0,data.shape[0]):\n extras+=int(data[i][17])\n\n return extras\n","sub_path":"q08_get_total_extras/build.py","file_name":"build.py","file_ext":"py","file_size_in_byte":345,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"562289580","text":"fin = open('short_house_list', 'r')\nlines = fin.readlines()\nname2info = dict()\nfor line in lines:\n\tls = line.split('\\t')\n\tif len(ls) == 7:\n\t\tname = line.split('\\t')[0]\n\t\tname2info[name] = line\nfin.close()\n\nfin = open('dict/house_id_match', 'r')\nfout = open('house_list_final', 'w')\nlines = fin.readlines()\nfor line in lines:\n\tname = line.split('\\t')[0]\n\ttid = line.split('\\t')[2]\n\ttid = tid.split('\\n')[0]\n\tif name in name2info:\n\t\tnewline = tid + '\\t' + name2info[name]\n\t\tfout.write(newline)\nfin.close()\nfout.close()\n","sub_path":"data/1.py","file_name":"1.py","file_ext":"py","file_size_in_byte":517,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"432667779","text":"'''\nImage Transformations\n1 Translate\n2 Rotate\n3 Resize\n4 Flip\n'''\nimport os \nimport cv2 as cv\nimport numpy as np \n\ndir_path = os.getcwd() + '/data/images/'\n\nimg=cv.imread(dir_path+'cat.jpeg')\ncv.imshow('Original cat',img)\n\n# x -->right\n# y -->down\n# -x -->left\n# -y -->up\n\n# Translation\ndef translate(img,x,y):\n\ttransMat=np.float32([[1,0,x],[0,1,y]])\n\tdimensions= (img.shape[1],img.shape[0])\n\treturn cv.warpAffine(img,transMat,dimensions)\n\n# Rotation\ndef rotate(img,angle,rotatePoint=None):\n\n\theight,width=img.shape[:2]\n\n\tif rotatePoint is None:\n\t\trotatePoint=(width/2,height/2)\n\n\trotMat=cv.getRotationMatrix2D(rotatePoint,angle,1.0)\n\tdimensions= (width,height)\n\treturn cv.warpAffine(img,rotMat,dimensions)\n\n\n# Translate\ntranslated=translate(img,100,100)\ncv.imshow('Translated',translated)\n\n# Rotate\nrotated=rotate(img,30)\ncv.imshow('Rotate',rotated)\n\n# Resize\nresized=cv.resize(img,(500,500),interpolation=cv.INTER_CUBIC)\ncv.imshow('Resized',resized)\n\n# Flip\n# 0 is vertical flip, 1 is horizontal flip, -1 is both\nflip=cv.flip(img,0)\ncv.imshow('flip',flip)\n\ncv.waitKey(0)","sub_path":"5 Transform.py","file_name":"5 Transform.py","file_ext":"py","file_size_in_byte":1073,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"462492359","text":"\"\"\"\nTests the re-creation of object, using the UUID feature\n\"\"\"\nimport zproc\n\n\ndef test_random_addr():\n ctx = zproc.Context()\n ctx.state[\"foo\"] = 42\n\n ctx = zproc.Context(ctx.server_address)\n assert ctx.state.copy() == {\"foo\": 42}\n\n state = zproc.State(ctx.server_address)\n assert state.copy() == {\"foo\": 42}\n\n\nADDRESS = \"tcp://127.0.0.1:50000\"\n\n\ndef test_static_addr():\n zproc.start_server(ADDRESS)\n\n ctx = zproc.Context(ADDRESS)\n ctx.state[\"foo\"] = 42\n\n assert ctx.state.copy() == {\"foo\": 42}\n\n state = zproc.State(ADDRESS)\n assert state.copy() == {\"foo\": 42}\n","sub_path":"tests/test_server_address.py","file_name":"test_server_address.py","file_ext":"py","file_size_in_byte":598,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"594972297","text":"#coding:utf8\n\nfrom _base.orm import Model, Field\nfrom time import localtime, strftime\nfrom config.constant import DEFAULT_ROWS, PUBLISH\n\nclass Article(Model):\n \n id = Field()\n title = Field()\n summary = Field()\n content = Field()\n publishtime = Field()\n labelid = Field()\n status = Field()\n\n @classmethod\n def post(cls, title, summary, content, label_id, status, id=None):\n ''' post the article\n '''\n # set publish time of the article \n publishtime = strftime(\"%Y-%m-%d %H:%M:%S\", localtime())\n\n params = {'title':title, 'summary':summary, 'content':content, 'labelid':int(label_id), \\\n 'status':status, 'publishtime':publishtime}\n if id:\n Article().where(id=id).update(**params)\n else:\n Article(**params).save() \n\n @classmethod\n def get_articles_total_page(cls, rows=DEFAULT_ROWS, status=None, labelid=None):\n '''get total articles page by status\n '''\n # print Article().where().count()[0]\n total_records = 0\n if not status and not labelid:\n total_records = Article().where().count()[0]\n elif not labelid:\n total_records = Article().where(status=status).count()[0]\n else:\n total_page = Article().where(status=status, labelid=labelid).count()[0]\n total_page = total_records / rows;\n total_page = total_page +1 if total_records % rows != 0 or total_records ==0 else total_page\n # return total_page + 1 #19 / 10 = 1 5/5 = 1 4/5 = 0\n return total_page \n\n\n @classmethod\n def get_articles_limit(cls, current_page, status=None, rows=DEFAULT_ROWS,labelid=None):\n offset = (current_page - 1) * rows\n if status != None:\n if labelid:\n return Article().where(labelid=labelid,status=status) \\\n .order_by(('publishtime',),('desc',)) \\\n .limit(rows, offset).select_many()\n else: \n return Article().where(status=status).order_by(('publishtime',),('desc',)) \\\n .limit(rows, offset).select_many()\n else:\n return Article().where().order_by(('publishtime',),('desc',)) \\\n .limit(rows, offset).select_many()\n\n @classmethod\n def get_article(cls, id):\n return Article().where(id=id).select_one()\n\n","sub_path":"model/article.py","file_name":"article.py","file_ext":"py","file_size_in_byte":2417,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"266131947","text":"import tensorflow as tf\r\nimport numpy as np\r\nimport Layer_class\r\nimport matplotlib.pyplot as plt\r\nfrom PIL import Image\r\n\r\nwith tf.Graph().as_default():\r\n BATCH_SIZE = 1\r\n N_CLASSES = 2\r\n val_src = 'images/validate/'\r\n filename = input('write file name that you wanna predict : ')\r\n\r\n image_file = Image.open(val_src + filename)\r\n image_file = image_file.resize([208, 208])\r\n plt.imshow(image_file)\r\n img_file1 = np.array(image_file)\r\n\r\n images = np.reshape(img_file1,newshape=[BATCH_SIZE,208,208,3])\r\n\r\n x = tf.placeholder(tf.float32, shape=[BATCH_SIZE, 208, 208, 3])\r\n keep_prob = tf.placeholder(tf.float32)\r\n\r\n logits = Layer_class._build_net(x, batch_size=BATCH_SIZE, n_classes=N_CLASSES, keep_prob=1)\r\n logits = tf.nn.softmax(logits)\r\n logs_train_dir = 'train/logs/'\r\n\r\n saver = tf.train.Saver()\r\n with tf.Session() as sess:\r\n print(\"Reading checkpoints...\")\r\n ckpt = tf.train.get_checkpoint_state(logs_train_dir)\r\n if ckpt and ckpt.model_checkpoint_path:\r\n global_step = ckpt.model_checkpoint_path.split('/')[-1].split('-')[-1]\r\n saver.restore(sess, ckpt.model_checkpoint_path)\r\n print('Loading success, global_step is %s' % global_step)\r\n else:\r\n print('No checkpoint file found')\r\n prediction = sess.run(logits, feed_dict={x: images, keep_prob : 1})\r\n max_index = np.argmax(prediction)\r\n\r\n if max_index == 1 :\r\n print(\"Prediction : White\")\r\n if max_index == 0 :\r\n print(\"Prediction : Asian\")\r\n if max_index == 2 :\r\n print(\"Prediction : Black\")","sub_path":"validate.py","file_name":"validate.py","file_ext":"py","file_size_in_byte":1639,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"519619702","text":"from django.db import models\nfrom django.db.models import fields\nfrom rest_framework import serializers\nfrom ..models import Products, Email\nfrom django.contrib.auth import get_user_model\nfrom ast import literal_eval\n\nclass ProductSerializer(serializers.ModelSerializer):\n class Meta:\n model = Products\n fields = ['name', 'price', 'count_stock', 'brand']\n\nclass ProductCreateSerializer(serializers.ModelSerializer):\n class Meta:\n model = Products\n fields = '__all__'\nclass ProductAddSerializer(serializers.ModelSerializer):\n userid = serializers.IntegerField()\n prodid = serializers.IntegerField()\n class Meta:\n model = Products\n fields = ['prodid', 'userid']\n def add_to_user(self):\n userid = self.validated_data['userid']\n prodid = self.validated_data['prodid']\n user = get_user_model().objects.get(id = userid)\n prod = Products.objects.get(id = prodid)\n count_check = prod.count_stock \n user.budget -= int(prod.price)\n if count_check < 1:\n res = \"Out of Stock\"\n return res\n elif user.budget < 0:\n res = \"Your Budget has been exceeded\"\n return res\n else:\n user.product_list.add(prod)\n prod.count_stock = count_check - 1\n prod.save()\n user.save()\n return True\n\nclass ProductReviewSerializer(serializers.ModelSerializer):\n review = serializers.CharField()\n prodid = serializers.IntegerField()\n class Meta:\n model = Products\n fields = ['prodid','review']\n \n def review_add(self):\n review = self.validated_data['review']\n prodid = self.validated_data['prodid']\n item = Products.objects.get(id=prodid)\n rlist = literal_eval(item.review_list)\n rlist.append(review)\n rstring = str(rlist)\n item.review_list = rstring\n item.save()\n\nclass AllReviewSerializer(serializers.ModelSerializer):\n class Meta:\n model = Products\n fields = ['review_list']\n\nclass ProductRateSerializer(serializers.ModelSerializer):\n rate = serializers.IntegerField()\n prodid = serializers.IntegerField()\n class Meta:\n model = Products\n fields = [ 'rate', 'prodid']\n\n def rate_prod(self):\n rate = self.validated_data['rate']\n prodid = self.validated_data['prodid']\n item = Products.objects.get(id=prodid)\n if rate > 0 and rate <= 5:\n new_rating = (item.rating + rate)/2\n item.rating = new_rating\n item.save()\n return True\n else:\n return False\n\n \nclass EmailSerializer(serializers.ModelSerializer):\n class Meta:\n model = Email\n fields = '__all__'","sub_path":"products/api/serializers.py","file_name":"serializers.py","file_ext":"py","file_size_in_byte":2774,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"434737143","text":"\"\"\"\n2019, day 5\n\"\"\"\nfrom day02 import read_intcode_into_dict\n\ndef update_pointer(pointer, opcode, new_pointer):\n if opcode in [1, 2, 7, 8]:\n pointer += 4\n elif opcode in [3, 4, 9]:\n pointer += 2\n elif opcode in [5, 6]:\n if pointer == new_pointer:\n pointer += 3\n else:\n pointer = new_pointer\n else:\n raise NotImplementedError\n return pointer\n\ndef retrieve_modes(raw_opcode):\n opcode = raw_opcode // 100\n return opcode % 10, (opcode // 10) % 10, opcode // 100\n\ndef read_opcode(raw_opcode):\n return raw_opcode % 100\n\ndef retrieve_destination(pointer, opcode, intcode):\n if opcode in [4, 5, 6]:\n return None\n elif opcode == 3:\n return intcode[pointer + 1]\n else:\n return intcode[pointer + 3]\n\ndef apply_parameter(pos, mode, intcode):\n #print(f'parameter: mode {mode} at position {pos}')\n if mode == 0:\n return intcode[intcode[pos]]\n else:\n return intcode[pos]\n\ndef retrieve_inputs(pointer, opcode, intcode):\n modes = retrieve_modes(intcode[pointer])\n inp1 = None\n inp2 = None\n if opcode != 3:\n inp1 = apply_parameter(pointer+1, modes[0], intcode)\n if opcode != 4:\n inp2 = apply_parameter(pointer+2, modes[1], intcode)\n return inp1, inp2\n\ndef run_program(intcode, inp0, pointer=0):\n opcode = read_opcode(intcode[pointer])\n out = None\n inp0_cnt = 0 # used for part7\n while opcode != 99:\n new_pointer = None\n to_store = None\n # eval inputs\n (inp1, inp2) = retrieve_inputs(pointer, opcode, intcode)\n # eval dest\n dest = retrieve_destination(pointer, opcode, intcode)\n # sum or factor two inputs\n if opcode in [1, 2]:\n to_store = inp1 + inp2 if opcode == 1 else inp1 * inp2\n # take input\n elif opcode == 3:\n if isinstance(inp0, int):\n to_store = inp0\n else:\n try:\n to_store = inp0[inp0_cnt]\n inp0_cnt += 1\n except IndexError:\n # we need a new input... exit loop, current pointer is returned\n break\n # give output\n elif opcode == 4:\n out = inp1\n # jump if true / jump if false\n elif opcode in [5, 6]:\n if opcode == 5 and inp1 != 0:\n new_pointer = inp2\n elif opcode == 6 and inp1 == 0:\n new_pointer = inp2\n else:\n new_pointer = pointer\n # less than or equal\n elif opcode in [7, 8]:\n cond_met = inp1 < inp2 if opcode == 7 else inp1 == inp2\n to_store = 1 if cond_met else 0\n else:\n raise NotImplementedError\n if to_store is not None:\n intcode[dest] = to_store\n\n pointer = update_pointer(pointer, opcode, new_pointer)\n opcode = read_opcode(intcode[pointer])\n #if out is not None and out != 0 and opcode != 99:\n # raise ValueError\n\n return out, pointer, opcode == 99\n\ndef part1(content):\n raw_intcode = read_intcode_into_dict(content)\n (out, _, _) = run_program(raw_intcode, 1)\n print(f'Part 1: Diagnostic code: {out}')\n\ndef part2(content):\n raw_intcode = read_intcode_into_dict(content)\n (out, _, _) = run_program(raw_intcode, 5)\n print(f'Part 2: Diagnostic code: {out}')\n","sub_path":"2019/day05.py","file_name":"day05.py","file_ext":"py","file_size_in_byte":3410,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"292069289","text":"LOWER = 33\nUPPER = 127\n\ndef main():\n for i in range(LOWER, UPPER):\n print(\"{:>3} {}\".format(i,chr(i)))\n entered_character = str(input(\"Enter a character: \"))\n print(\"The ASCII code for {0} is {1}\".format(entered_character,ord(entered_character)))\n entered_number = int(input(\"Enter a number between {0} and {1}: \".format(LOWER, UPPER)))\n print(\"The character for {0} is {1}\".format(entered_number,chr(entered_number)))\n\nmain()","sub_path":"prac_02/ascii_table/ascii_table.py","file_name":"ascii_table.py","file_ext":"py","file_size_in_byte":448,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"516212359","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# author: anne\n# datetime: 2020-12-02 09:39\n# filename: PythonProjectDemo/requests_user\n\n\nimport json\nimport logging\nimport time\n\nimport requests\nfrom requests.auth import HTTPBasicAuth\n\nFORMAT = '%(asctime)s--%(levelname)s: %(message)s'\nlogging.basicConfig(format=FORMAT, level=logging.INFO)\n\n\ndef decorator(func):\n def wrapper(*params):\n print('%s方法 is running...' % func.__name__)\n func(*params)\n\n return wrapper\n\n\n@decorator\ndef get(host, url, args=None):\n r = requests.get(host + url, params=args)\n if r.status_code == 200:\n logging.info(r.json())\n else:\n logging.info('接口返回错���')\n\n\n@decorator\ndef get_header(host, url, header):\n r = requests.get(host + url, headers=header)\n logging.info(r.text)\n\n\n# 加代理\n@decorator\ndef get_proxy(host, url, proxy):\n r = requests.get(host + url, proxies=proxy)\n logging.info(r.text)\n\n\n# 加鉴权\n@decorator\ndef get_auth(host, url):\n r = requests.get(host + url, auth=HTTPBasicAuth('user', 'password'))\n logging.info(r.text)\n\n\n@decorator\ndef post_data(host, url, params):\n if isinstance(params, dict):\n data = json.dumps(params) # 字典转换成json字符串\n r = requests.post(host + url, data=data)\n logging.info(r.text)\n else:\n r = requests.post(host + url, data=params)\n # logging.info(f'{params}不是字典')\n logging.info(r.text)\n\n\n@decorator\ndef post_json(host, url, params):\n before = time.time()\n r = requests.post(host + url, data=params, timeout=1) # 自动转换成json字符串\n after = time.time()\n if r.status_code == 200:\n logging.info(r.text)\n logging.info(f'接口耗时:{after - before}s')\n else:\n r.raise_for_status()\n\n\n@decorator\ndef put(host, url, data):\n r = requests.put(host + url, data=data)\n logging.info(r.text)\n print(r.headers, r.cookies.get_dict()) # 获取headers和cookies\n\n\n@decorator\ndef delete(host, url, data):\n r = requests.delete(host + url, data=data)\n logging.info(r.content) # 返回二进制的内容,一般是图片、视频、音频等数据流\n\n\ndef session():\n s = requests.Session() # 跨请求保持某些参数\n s.get(\"https://httpbin.org/cookies/set/sessioncookie/anneyang\")\n r = s.get(\"https://httpbin.org/cookies\")\n logging.info(r.text)\n\n\nif __name__ == '__main__':\n host = 'https://httpbin.org/'\n args = {'name': 'anne', 'age': 20, 'attributes': ['eat', 'sing', 'run']}\n args1 = [('key1', 'value1'), ('key2', 'value2')]\n header = {'user_agent': 'my-app/0.0.1'}\n proxies = {'http': 'http://127.0.0.1:8080',\n 'https': 'http://127.0.0.1:8080'}\n\n # get(host, 'ip')\n # get(host, 'get', args)\n # get_header(host, 'get', header)\n # # get_proxy(host, 'get', proxies)\n # get_auth(host, 'user')\n # post_data(host, 'post', args)\n # post_json(host, 'post', args1)\n # put(host, 'put', args)\n # delete(host, 'delete', args)\n session()\n","sub_path":"UnittestFramework/requests_use.py","file_name":"requests_use.py","file_ext":"py","file_size_in_byte":3021,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"298305051","text":"from twitter import api\nfrom urlShortener import urlShortener\nfrom logger import logger\nimport random \n\nclass Tweet():\n def __init__ (self):\n self.title = None\n self.link = None\n self.hashtags = None\n self.shortLink = None\n \n def _getTweetFormat(self):\n # 0 = title, 1 = hashtags, 3 = shortlik\n formatArray = [\n '{0} | {1} | {2}',\n '{0} {1} {2}',\n '{0} {1}\\n\\n{2}',\n '{0}\\n\\n{1}\\n\\n{2}',\n '{0}\\n\\n{1}\\n{2}'\n ]\n \n return random.choice(formatArray)\n \n def generate(self):\n self.shortlink = urlShortener.shorten(self.link)\n message = self._getTweetFormat().format(self.title, self.hashtags, self.shortlink)\n return message\n \n def getLength(self):\n return len('{} {}| '.format(self.title, self.hashtags)) + 20\n \n def tweet(self):\n api.tweet(self.generate())","sub_path":"twitter/Tweet.py","file_name":"Tweet.py","file_ext":"py","file_size_in_byte":949,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"574705992","text":"import os\nimport requests\n\nclass GiantbombAPI():\n \"\"\"Simple implementation of the Giantbomb API that only offers GET /platforms/ call as a generator.\"\"\"\n\n base_url = \"http://www.giantbomb.com/api\"\n\n def __init__(self, api_key):\n self.api_key = api_key\n\n def get_platforms(self, sort=None, filter=None, field_list=None):\n \"\"\"Generator yielding platforms matching given criteria. Returns all platforms if no limit is specified.\"\"\"\n\n #Convert dictionary of criteria into a comma-separated list of key:value pairs.\n params = {}\n if sort is not None:\n params['sort'] = sort\n if field_list is not None:\n params['field_list'] = field_list\n if filter is not None:\n params['filter'] = filter\n parsed_filters = []\n for key, value in filter.iteritems():\n parsed_filters.append('{0}:{1}', format(key, value))\n params['filter'] = ','.join(parsed_filters)\n\n #Finally, append our API key to the list of params and tell the API to transfer the data as JSON.\n params['api_key'] = self.api_key\n params['format'] = 'json'\n\n incomplete_result = True\n num_total_results = None\n num_fetched_results = 0\n counter = 0\n\n while incomplete_result:\n params['offset'] = num_fetched_results\n result = requests.get(self.base_url + '/platforms/', params=params)\n result = result.json()\n \n if num_total_results is None:\n num_total_results = int(result['number_of_total_results'])\n num_fetched_results += int(result['number_of_total_results'])\n\n if num_fetched_results >= num_total_results:\n incomplete_result = False\n\n for item in result['results']:\n logging.debug(\"Yielding platform {0} of {1}\".format(counter + 1, num_total_results))\n \n #Convert values into a more useful format where appropriate.\n if 'original_price' in item and item['original_price']:\n item['original_price'] = float(item['original_price'])\n\n yield item\n counter += 1","sub_path":"Python-API-Testing/GiantbombAPI.py","file_name":"GiantbombAPI.py","file_ext":"py","file_size_in_byte":2231,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"322161449","text":"class Solution:\n def uniquePaths(self, m, n):\n import pdb;pdb.set_trace()\n dp = [[1]*n] + [[1]+[0] * (n-1) for _ in range(m-1)]\n print(dp)\n for i in range(1, m):\n for j in range(1, n):\n dp[i][j] = dp[i-1][j]+dp[i][j-1]\n return dp[-1][-1]\n\n\nins = Solution()\nprint(ins.uniquePaths(5, 4))\n","sub_path":"62.py","file_name":"62.py","file_ext":"py","file_size_in_byte":350,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"113139332","text":"\n\nfrom xai.brain.wordbase.nouns._convertible import _CONVERTIBLE\n\n#calss header\nclass _CONVERTIBLES(_CONVERTIBLE, ):\n\tdef __init__(self,): \n\t\t_CONVERTIBLE.__init__(self)\n\t\tself.name = \"CONVERTIBLES\"\n\t\tself.specie = 'nouns'\n\t\tself.basic = \"convertible\"\n\t\tself.jsondata = {}\n","sub_path":"xai/brain/wordbase/nouns/_convertibles.py","file_name":"_convertibles.py","file_ext":"py","file_size_in_byte":273,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"121594020","text":"# 버블 정렬 알고리즘 구현-개선2(이미 정렬된 원소를 제외한 나머지만 비교,교환하도록 스캔 범위를 제한함)\n\nfrom typing import MutableSequence\n\n\ndef bubble_sort_verbose(a: MutableSequence) -> None:\n \"\"\"버블 정렬(스캔 범위를 제한)\"\"\"\n ccnt = 0 # 비교 횟수\n scnt = 0 # 교환 횟수\n n = len(a)\n k = 0\n while k < n - 1:\n last = n - 1\n exchng = 0\n print(f'패스 {k + 1}')\n for j in range(n - 1, k, -1):\n for m in range(0, n - 1):\n print(f'{a[m]:2}' + (' ' if m != j - 1 else\n ' +' if a[j - 1] > a[j] else ' -'), end='')\n print(f'{a[n - 1]:2}')\n ccnt += 1\n if a[j - 1] > a[j]:\n scnt += 1\n a[j - 1], a[j] = a[j], a[j - 1]\n last = j\n exchng += 1\n k = last\n if exchng == 0:\n break\n for m in range(0, n - 1):\n print(f'{a[m]:2}', end=' ')\n print(f'{a[n - 1]:2}')\n print(f'비교를 {ccnt}번 했습니다.')\n print(f'교환을 {scnt}번 했습니다.')\n\n\nif __name__ == '__main__':\n print('버블 정렬을 수행합니다.')\n num = int(input('원소 수를 입력하세요 : '))\n x = [None] * num # 원소 수가 num인 배열을 생성\n\n for i in range(num):\n x[i] = int(input(f'x[{i}]: '))\n\n bubble_sort_verbose(x) # 배열 x를 버블 정렬\n print('오름차순으로 정렬했습니다.')\n for i in range(num):\n print(f'x[{i}] = {x[i]}')\n","sub_path":"Do_it!/6.정렬 알고리즘/1.버블정렬/개선2-스캔범위제한.py","file_name":"개선2-스캔범위제한.py","file_ext":"py","file_size_in_byte":1591,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"301200759","text":"class GTCTelescope():\n\n __instrument_default = \"OSIRIS\"\n\n __params={\n \"OSIRIS\":{\n \"gain\":0.95, #e-/ADU\n \"readnoise\": 4.5, #15-,\n \"readtime\": 21.0, # 1.45 S(Non-destructive)\n \"pixelscale\": 0.127*2, #0.184 arcsec/pixel (unbinned)\n \"telescope\":\"gtc\"\n }\n }\n\n __key_band = {\"mag_u\": \"sdss-u\", \"mag_g\": \"sdss-g\", \"mag_r\": \"sdss-r\", \"mag_i\": \"sdss-i\", \"mag_z\": \"sdss-z\"}\n\n def getBand(self,band):\n band = str(band).lower()\n if band in self.__key_band.keys():\n return self.__key_band[band]\n elif band in self.__key_band.values():\n for mag, filter in self.__key_band.items():\n if filter == band:\n return mag\n else:\n return None\n\n @staticmethod\n def getParam(key):\n\n lv=GTCTelescope()\n\n if key in lv.__params:\n return lv.__params[key]\n else:\n raise KeyError(\"No found \"+key)\n\n @staticmethod\n def getParams(instrument=\"all\"):\n lv = GTCTelescope()\n if instrument == \"all\":\n return lv.__params\n else:\n instrument= instrument.replace(\":\",\"\")\n if instrument in lv.__params:\n return lv.__params[instrument]\n else:\n raise KeyError(\"No found \"+instrument)","sub_path":"telescopes/gtc.py","file_name":"gtc.py","file_ext":"py","file_size_in_byte":1367,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"630751117","text":"#!/usr/bin/python3\n# -*- encoding: utf-8 -*-\n\"\"\"\n\n跟卖维护\n\n如果同时有初始底价和最低底价的, 判断时,以最低底价为准,如果没有最低底价的,以初始底价为准。\n\n\"\"\"\n\nimport traceback\nfrom fun import *\n\nstime = int(time.time())\ncdir = os.path.dirname(os.path.abspath(__file__))\nlogfile = cdir + \"/0gmwh.log\"\n\nlogtxt = \"{}\\n{}\\n\\n\".format(__file__, now())\nfile_write(logfile, logtxt, 'w')\n\nUARGS = sys.argv\n\nDIC['pool'] = mysqlc.pool(CONFIG, 3)\ndb = DIC['pool'].get()\n\ndb.query('SET wait_timeout=7200')\nDIC['db'] = db\nDIC['url_id'] = 0\nDIC['fp'] = 0\nDIC['country_filter'] = 1 # 过滤发货地\nDIC['sform'] = {}\nDIC['referer'] = 'https://www.cdiscount.com'\nDIC['filter'] = ''\n\n# --------------------------------------------------------------------\n\ncond = 'where gm.status=2 and caiji=1 and p.addtime < {}'.format(days_before(1))\n# cond = \"where gm.id in(7749)\"\n\n# 单独采集P_ID\nif len(UARGS) == 2 and UARGS[1].isdigit():\n cond = \"where gm.id={}\".format(UARGS[1])\n\nsql0 = \"\"\"\nselect count(*) from t_gm gm\nLEFT JOIN `cd_product` p ON p.`id` = gm.`pro_id`\n{}\n\"\"\".format(cond)\n\nsql1 = \"\"\"\nselect gm.id, s.store_code, p.id pid, p.cd_url, p.addtime from t_gm gm\nLEFT JOIN cd_product p ON p.id = gm.pro_id\nLEFT JOIN t_store s ON s.id = gm.store\n{}\nlimit 10\n\"\"\".format(cond)\n\nprint(sql1)\n\nnum = db.field(sql0)\npagesize = 10\npages = math.ceil(num / pagesize)\nprint(\"总数: {}, pages: {}\\n\".format(num, pages))\n\nfor i in range(1, pages + 1):\n rows = db.fetch(sql1)\n if not rows: continue\n args = [ [a[3], {}, 0, 0, a[1]] for a in rows ]\n # print(i, args)\n print(rows)\n\n threads_run(parse_product_page, args) # 更新cd_product表的产品信息\n\n for row in rows:\n cond1 = \"where gm.id='{}' limit 1\".format(row[0])\n zds = \"gm.id, p.id, p.gm_price, gm.price_gm, gm.price_d, gm.min_dprice\"\n ret = db.one(\"select {} from t_gm gm LEFT JOIN `cd_product` p ON p.`id` = gm.`pro_id` {}\".format(zds, cond1), 1)\n # print(ret)\n\n cjgm_price = float(ret[\"gm_price\"]) # 采集到的最低跟卖价\n price_gm = float(ret[\"price_gm\"]) # 跟卖时填写的跟卖价\n price_d = float(ret[\"price_d\"]) # 跟卖时填写的底价\n min_dprice = float(ret[\"min_dprice\"]) # 填写的最低底价\n\n # if gm_price<=0: continue # 没有跟卖卖家时\n\n ubase_price = price_d\n if min_dprice>0: ubase_price = min_dprice\n\n gm_status = 1;\n if cjgm_price > 0:\n\n print(price_gm, cjgm_price)\n\n if ubase_price > cjgm_price: # 如果底价大于采集到的最低跟卖价,则状态为低于底价\n gm_status = 3;\n elif price_gm > cjgm_price: # 如果跟卖价大于采集到的最低跟卖价,则状态为无购物车\n gm_status = 2;\n\n dd = {}\n dd['gm_status'] = gm_status;\n cond2 = \"id='{}' limit 1\".format(row[0])\n db.update(\"t_gm\", dd, cond2, 0) # 更新跟卖表的记录\n\n print(\"主循环{} 休眠 {}s\".format(i, 10))\n time.sleep(10)\n\nDIC['pool'].empty()\n\nrun_time = int(time.time()) - stime\nif run_time > 3600: run_time = datetime.timedelta(seconds=run_time)\n\nprint(\"花费时间:{} \".format(run_time ))\n\nlogtxt = \"总耗时: {}, 内存: {} \\n\".format(run_time, mem() )\nlogtxt += now() + \"\\n\"\nfile_write(logfile, logtxt, 'a')\n\n\n","sub_path":"cdiscount/gmwh.py","file_name":"gmwh.py","file_ext":"py","file_size_in_byte":3491,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"312980357","text":"# -*- coding: utf-8 -*-\nimport cv2\nimport sys\n\ndataset_dir = \"./img/\"\n \ndef hog_func():\n argv = sys.argv\n argc = len(argv)\n if (argc != 2):\n #引数がちゃんとあるかチェック\n #正しくなければメッセージを出力して終了\n print('Usage: python %s arg1', argv[0])\n quit()\n # 画像の読み込み\n img_name = dataset_dir + argv[1]\n im = cv2.imread(img_name)\n # HoG特徴量の計算\n hog = cv2.HOGDescriptor()\n # SVMによる人検出\n hog.setSVMDetector(cv2.HOGDescriptor_getDefaultPeopleDetector())\n hogParams = {'winStride': (8, 8), 'padding': (32, 32), 'scale': 1.05}\n # 人を検出した座標\n human, r = hog.detectMultiScale(im, **hogParams)\n # 長方形で人を囲う\n for (x, y, w, h) in human:\n cv2.rectangle(im, (x, y),(x+w, y+h),(0,50,255), 3)\n # 人を検出した座標\n cv2.imshow(\"Human detection\",im)\n cv2.waitKey(0)\n # 画像保存\n out_file_name = \"after\" + argv[1]\n cv2.imwrite(out_file_name,im)\n print (\"saved\")\n\nif __name__ == '__main__':\n hog_func()\n\n","sub_path":"traincascade03.py","file_name":"traincascade03.py","file_ext":"py","file_size_in_byte":1102,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"442103410","text":"#!/usr/bin/env python\n# license removed for brevity\n\nimport rospy\nimport random\n# Brings in the SimpleActionClient\nimport actionlib\n# Brings in the .action file and messages used by the move base action\nfrom move_base_msgs.msg import MoveBaseAction, MoveBaseGoal\nfrom pedsim_msgs.msg import AgentStates\n\ndef movebase_client():\n # Create an action client called \"move_base\" with action definition file \"MoveBaseAction\"\n client = actionlib.SimpleActionClient('move_base',MoveBaseAction)\n\n # Waits until the action server has started up and started listening for goals.\n client.wait_for_server()\n\n # Select a pedsim pedestrian\n data = rospy.wait_for_message(\"/pedsim_simulator/simulated_agents\", AgentStates)\n n = len(data.agent_states)\n ped_n = random.randint(1,n)\n ped_x = data.agent_states[ped_n-1].pose.position.x\n ped_y = data.agent_states[ped_n-1].pose.position.y\n print(\"Pedestrian Info: x = %.2f, y = %.2f\" % (ped_x, ped_y))\n\n # Creates a new goal with the MoveBaseGoal constructor\n goal = MoveBaseGoal()\n goal.target_pose.header.frame_id = \"odom\"\n goal.target_pose.header.stamp = rospy.Time.now()\n # Move 0.5 meters forward along the x axis of the \"map\" coordinate frame\n goal.target_pose.pose.position.x = ped_x\n goal.target_pose.pose.position.y = ped_y\n # No rotation of the mobile base frame w.r.t. map frame\n goal.target_pose.pose.orientation.w = 1.0\n\n # Sends the goal to the action server.\n client.send_goal(goal)\n # Waits for the server to finish performing the action.\n wait = client.wait_for_result()\n # If the result doesn't arrive, assume the Server is not available\n if not wait:\n rospy.logerr(\"Action server not available!\")\n rospy.signal_shutdown(\"Action server not available!\")\n else:\n # Result of executing the action\n return client.get_result()\n\n# If the python node is executed as main process (sourced directly)\nif __name__ == '__main__':\n try:\n # Initializes a rospy node to let the SimpleActionClient publish and subscribe\n rospy.init_node('movebase_client_py')\n result = movebase_client()\n if result:\n rospy.loginfo(\"Goal execution done!\")\n except rospy.ROSInterruptException:\n rospy.loginfo(\"Navigation test finished.\")\n","sub_path":"src/development/scripts/old/ped_goal_v1.py","file_name":"ped_goal_v1.py","file_ext":"py","file_size_in_byte":2296,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"266251099","text":"from PIL import Image\nfrom PIL import ImageChops\nfrom sys import argv,exit\n \ndef compare_images(path_one, path_two):\n \"\"\"\n Compares to images and saves a diff image, if there\n is a difference\n \n @param: path_one: The path to the first image\n @param: path_two: The path to the second image\n \"\"\"\n image_one = Image.open(path_one)\n image_two = Image.open(path_two)\n \n diff = ImageChops.difference(image_one, image_two)\n pixels = list(diff.getdata())\n\n score = 0\n for p in pixels:\n r,g,b = p\n if r+g+b > 0:\n score += 1\n\n return score\n \n \nif __name__ == '__main__':\n if len(argv) < 3:\n exit()\n else:\n img1 = argv[1]\n img2 = argv[2]\n score = compare_images(img1,img2)\n print(score)","sub_path":"assignments/meme_generator/www/html/scripts/compare_image.py","file_name":"compare_image.py","file_ext":"py","file_size_in_byte":773,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"307504621","text":"#coding: utf-8\nimport sys\nimport re\nfrom collections import defaultdict\nmorpheme = []\nsentences = []\nfor line in open(\"neko.txt.mecab\"):\n if \"EOS\" not in line:\n line_s = re.split(\"\\s|,\", line)\n morpheme.append({\"surface\": line_s[0], \"base\":line_s[6], \"pos\":line_s[1], \"pos1\":line_s[2]})\n elif morpheme != []: \n sentences.append(morpheme)\n morpheme = []\ncount = defaultdict(int)\nfor sentence in sentences:\n for word in sentence:\n count[word[\"surface\"]] += 1\nfor k, v in sorted(count.items(), key = lambda x: x[1], reverse = True):\n print(\"{} {}\".format(k, v))\n","sub_path":"tomoya/chapter04/knock36.py","file_name":"knock36.py","file_ext":"py","file_size_in_byte":579,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"25711883","text":"from aoc_parser import Parser\n\n\ndef main():\n file_name = 'data'\n frequencies = Parser('{}.txt'.format(file_name)).int_lines()\n # Part 1: 540\n print('Part 1: {}'.format(sum(frequencies)))\n # Part 2: 73056\n print('Part 2: {}'.format(get_first_repeated(frequencies)))\n\n\ndef get_first_repeated(values):\n seen = set()\n result, i = 0, 0\n while True:\n if result in seen:\n return result\n seen.add(result)\n value = values[i % len(values)]\n result += value\n i += 1\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"2018/01/solver.py","file_name":"solver.py","file_ext":"py","file_size_in_byte":568,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"529806886","text":"#!/usr/bin/python3\n\"\"\"\n Exam from Week 2 (MITx: 6.00.1x): \n Introduction to Computer Science and Programming Using Python\n\n INFO:\n - When submitting the exercise, remove the lines that contains '###'\n (variables and prints). \n\n PROBLEM\n - Write a program to calculate the credit card balance \n after one year if a person only pays the minimum monthly payment\n required by the credit card company each month.\n\n FLOW\n - For each month, calculate statements on the monthly payment and\n remaining balance. At the end of 12 months, print out the remaining\n balance. Be sure to print out no more than two decimal digits of accuracy\n\n INPUTS\n - balance: the outstanding balance on the credit card\n - annualInterestRate: annual interest rate as a decimal\n - monthlyPaymentRate: minimum monthly payment rate as a decimal\n\n\"\"\"\n\nbalance = 42 ###\nannualInterestRate = 0.2 # annual interest ###\nmonthlyPaymentRate = 0.04 # % to pay, per month\nmonthlyInterestRate = annualInterestRate / 12.0 # annual interest, per month\n\nprevious_balance = balance\n\nprint(\"\\nINITIAL BALANCE:\", previous_balance) ###\n\nfor m in range(1, 13):\n print(\"MONTH\", m) ###\n minimum_payment = monthlyPaymentRate * previous_balance\n print(\" - minimum payment:\", minimum_payment) ###\n unpaid_balance = previous_balance - minimum_payment\n print(\" - unpaid balance:\", unpaid_balance) ###\n # calculating annual interest for the new balance... \n previous_balance = unpaid_balance + (monthlyInterestRate * unpaid_balance)\n print(\" - new balance\", previous_balance, \"\\n\") ###\n\nprint(\"Remaining balance: {:.2f}\".format(previous_balance))\n\n\n","sub_path":"edX/pset_02/week2_credit_card_balance.py","file_name":"week2_credit_card_balance.py","file_ext":"py","file_size_in_byte":1675,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"526402048","text":"import serial\nqte = input().split()\nif len(qte) > 4:\n qte = qte[:4]\nelif len(qte) < 4:\n while len(qte) < 4:\n qte.append('0')\n qte[-1] = '13'\nqte[0] = '0'\nser = serial.Serial('COM3', 9600)\ns = ser.read(100)\n\n\nfor i in range(len(qte)):\n ser.write(int(qte[i]))\n\n\n","sub_path":"controller.py","file_name":"controller.py","file_ext":"py","file_size_in_byte":279,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"158124684","text":"from copy import deepcopy\nfrom reparser.regex_parser import parse\nfrom query import Query, allQuery\n\n\ndef cross(s, t):\n return {x + y for x in s for y in t}\n\n\ndef get_trigrams(raw_str):\n prev = raw_str[:3]\n if len(prev) == 3:\n yield prev\n for j in range(3, len(raw_str)):\n prev = prev[1:] + raw_str[j]\n yield prev\n\n\n# print(list(get_trigrams(\"ab\")))\n\n\ndef freeQuery(tree):\n # if tree is None:\n # return []\n\n re_type = tree[\"type\"]\n if re_type == \"literal\":\n trigram_list = list(get_trigrams(tree[\"value\"]))\n\n if len(trigram_list):\n return Query(o=Query.QAnd, trigram=trigram_list)\n else:\n return deepcopy(allQuery)\n\n if re_type == \"concat\":\n q = freeQuery(tree[\"value\"][0])\n r = freeQuery(tree[\"value\"][1])\n return q.andOr(r, Query.QAnd)\n # return {\"AND\": [freeQuery(tree[\"value\"][0]), freeQuery(tree[\"value\"][1])]}\n\n if re_type == \"union\":\n q = freeQuery(tree[\"value\"][0])\n r = freeQuery(tree[\"value\"][1])\n return q.andOr(r, Query.QOr)\n # return {\"OR\": [freeQuery(tree[\"value\"][0]), freeQuery(tree[\"value\"][1])]}\n\n if re_type == \"repetition\":\n return deepcopy(allQuery)\n\n\nif __name__ == \"__main__\":\n tests = [\n (r\"(abcd|efgh)(ijklm|x*)\", '(\"abc\" \"bcd\")|(\"efg\" \"fgh\")'),\n (r\"(abc|cba)def\", '(\"abc\" | \"cba\") \"def\"'),\n (r\"abc+de\", \"+\"),\n (r\"(abc*)+de\", \"+\"),\n (r\"ab(cd)*ef\", \"+\"),\n (r\"abc|def\", '(\"abc\" | \"def\")')\n ]\n\n for test in tests:\n print(\"test:\", test[0])\n parse_tree = parse(test[0])\n # print(\"tree:\", parse_tree)\n print(\"actual:\", freeQuery(tree=parse_tree))\n print(\"expected:\", test[1])\n","sub_path":"requery/free.py","file_name":"free.py","file_ext":"py","file_size_in_byte":1745,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"408595014","text":"from django.shortcuts import render\nfrom django.http import Http404\nfrom .models import WallImage, Catlog, HomePage\nfrom django.core.paginator import Paginator\nfrom django.contrib import messages\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.db.models import Q\n\n\ndef home(request):\n # home page top image\n random_head_image = HomePage.objects.order_by('?')[:1]\n # end\n catlog_item = Catlog.objects.order_by('?')[:8]\n popular_item = Catlog.objects.order_by('?')[:7]\n latest_uploaded = WallImage.objects.order_by('?')[:30]\n images_data = WallImage.objects.order_by('-id')[:30]\n context = {\n 'popular_item':popular_item,\n 'catlog_item': catlog_item,\n 'random_head_image': random_head_image,\n 'latest_uploaded': latest_uploaded,\n 'images_data': images_data,\n }\n return render(request, 'wall/index.html', context)\n\n\ndef images(request):\n catlog_small = Catlog.objects.order_by('?')[:18]\n catlog_items = Catlog.objects.order_by('id')[:29]\n images_data = WallImage.objects.order_by('-id')\n paginator = Paginator(images_data, 30)\n page = request.GET.get('page')\n images_data = paginator.get_page(page)\n context = {\n 'catlog_items': catlog_items,\n 'catlog_small': catlog_small,\n 'images_data': images_data,\n }\n return render(request, 'wall/images.html', context)\n\n\ndef picture(request, pic_id):\n catlog_item = Catlog.objects.order_by('?')[:18]\n picture_data = WallImage.objects.get(id=pic_id)\n # remove after installing search similar items\n random_image = WallImage.objects.order_by('?')[:15]\n context = {\n 'catlog_item': catlog_item,\n 'picture_data': picture_data,\n 'random_image': random_image\n }\n return render(request, 'wall/pic.html', context)\n\n\ndef catlog(request):\n catlog_item = Catlog.objects.order_by('id')\n paginator = Paginator(catlog_item, 21)\n page = request.GET.get('page')\n catlog_item = paginator.get_page(page)\n context = {\n 'catlog_item': catlog_item\n }\n return render(request, 'wall/catlog.html', context)\n\n\ndef explore(request):\n catlog_small = Catlog.objects.order_by('?')[:18]\n catlog_items = Catlog.objects.order_by('?')[:29]\n images_data = WallImage.objects.order_by('?')\n paginator = Paginator(images_data, 30)\n page = request.GET.get('page')\n images_data = paginator.get_page(page)\n context = {\n 'catlog_items': catlog_items,\n 'catlog_small': catlog_small,\n 'images_data': images_data,\n }\n return render(request, 'wall/explore.html', context)\n\ndef catlog_search(request, cat_item):\n catlog_small = Catlog.objects.order_by('?')[:18]\n catlog_items = Catlog.objects.order_by('id')[:29]\n template = 'wall/catlog_search.html'\n queryset_list = WallImage.objects.all()\n query = cat_item\n if query == '':\n return HttpResponseRedirect('/')\n else:\n if query:\n queryset_list = queryset_list.filter(\n Q(image_tags__icontains=query) |\n Q(image_type__icontains=query)\n ).order_by('-id')\n random_search = WallImage.objects.order_by('?')[:1]\n paginator = Paginator(queryset_list, 30)\n page = request.GET.get('page')\n result = paginator.get_page(page)\n context = {\n 'catlog_items': catlog_items,\n 'catlog_small': catlog_small,\n 'result': result,\n 'random_search': random_search,\n }\n return render(request, template, context)\n else:\n messages.error(request, 'no results found')\n\n\ndef search(request):\n template = 'wall/search.html'\n queryset_list = WallImage.objects.all()\n query = request.GET.get('q')\n if query == '':\n return HttpResponseRedirect('/')\n else:\n if query:\n queryset_list = queryset_list.filter(\n Q(image_type__icontains=query) |\n Q(image_tags__icontains=query)\n\n ).order_by('-id')\n catlog_small = Catlog.objects.order_by('?')[:18]\n catlog_items = Catlog.objects.order_by('id')[:29]\n paginator = Paginator(queryset_list, 30)\n page = request.GET.get('page')\n result = paginator.get_page(page)\n context = {\n 'catlog_items': catlog_items,\n 'catlog_small': catlog_small,\n 'result': result,\n 'catlog_items': catlog_items,\n 'catlog_small': catlog_small,\n 'result': result,\n }\n return render(request, template, context)\n else:\n messages.error(request, 'no results found')\n","sub_path":"wall/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":4756,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"79629913","text":"import os\nimport numpy as np\nfrom PIL import Image\nfrom scipy.misc import imread\nfrom scipy.misc import imsave\nimport tensorflow as tf\nfrom tensorflow.contrib.slim.nets import inception\n\nfrom urllib.request import urlretrieve\nimport os\nimport json\nimport matplotlib.pyplot as plt\n\nfrom keras.models import Model\nfrom keras.models import load_model\nfrom keras.utils import np_utils\nfrom keras.datasets import mnist\n\nmodel = load_model('model_weights.h5')\n\n(x_train, y_train), (x_test, y_test) = mnist.load_data()\n\n\n# 重新定义数据格式,归一化\nx_train = x_train.reshape(x_train.shape[0], -1) / 255.0\nx_test = x_test.reshape(x_test.shape[0], -1) / 255.0\n\n# # 转one-hot标签\ny_train = np_utils.to_categorical(y_train, num_classes=10)\ny_test = np_utils.to_categorical(y_test, num_classes=10)\n\n\ndef fgm(x, preds, y=None, eps=0.3, ord=np.inf, clip_min=None, clip_max=None,targeted=False):\n \"\"\"\n TensorFlow implementation of the Fast Gradient Method.\n :param x: the input placeholder\n :param preds: the model's output tensor (the attack expects the\n probabilities, i.e., the output of the softmax)\n :param y: (optional) A placeholder for the model labels. If targeted\n is true, then provide the target label. Otherwise, only provide\n this parameter if you'd like to use true labels when crafting\n adversarial samples. Otherwise, model predictions are used as\n labels to avoid the \"label leaking\" effect (explained in this\n paper: https://arxiv.org/abs/1611.01236). Default is None.\n Labels should be one-hot-encoded.\n :param eps: the epsilon (input variation parameter) 邻域\n :param ord: (optional) Order of the norm (mimics NumPy).\n Possible values: np.inf, 1 or 2.\n :param clip_min: Minimum float value for adversarial example components\n :param clip_max: Maximum float value for adversarial example components\n :param targeted: Is the attack targeted or untargeted? Untargeted, the\n default, will try to make the label incorrect. Targeted\n will instead try to move in the direction of being more\n like y.\n :return: a tensor for the adversarial example\n \"\"\"\n\n # if y is None:\n # # Using model predictions as ground truth to avoid label leaking\n # preds_max = tf.reduce_max(preds, 1, keep_dims=True)\n # y = tf.to_float(tf.equal(preds, preds_max))\n # y = tf.stop_gradient(y)\n # y = y / tf.reduce_sum(y, 1, keep_dims=True)\n\n # Compute loss\n # loss = utils_tf.model_loss(y, preds, mean=False)\n loss = tf.nn.softmax_cross_entropy_with_logits(labels=y, logits=preds)\n # if targeted:\n # loss = -loss\n\n # Define gradient of loss wrt input\n grad, = tf.gradients(loss, x)\n\n if ord == np.inf:\n # Take sign of gradient\n normalized_grad = tf.sign(grad)\n # The following line should not change the numerical results.\n # It applies only because `normalized_grad` is the output of\n # a `sign` op, which has zero derivative anyway.\n # It should not be applied for the other norms, where the\n # perturbation has a non-zero derivative.\n normalized_grad = tf.stop_gradient(normalized_grad)\n elif ord == 1:\n red_ind = list(range(1, len(x.get_shape())))\n normalized_grad = grad / tf.reduce_sum(tf.abs(grad),\n reduction_indices=red_ind,\n keep_dims=True)\n elif ord == 2:\n red_ind = list(range(1, len(x.get_shape())))\n square = tf.reduce_sum(tf.square(grad),\n reduction_indices=red_ind,\n keep_dims=True)\n normalized_grad = grad / tf.sqrt(square)\n else:\n raise NotImplementedError(\"Only L-inf, L1 and L2 norms are \"\n \"currently implemented.\")\n\n # Multiply by constant epsilon\n scaled_grad = eps * normalized_grad\n\n # Add perturbation to original example to obtain adversarial example\n adv_x = x + scaled_grad\n\n # If clipping is needed, reset all values outside of [clip_min, clip_max]\n if (clip_min is not None) and (clip_max is not None):\n adv_x = tf.clip_by_value(adv_x, clip_min, clip_max)\n\n return adv_x\n\n# print(x_train[0])\n# print(y_train[0])\n#\n# print(len(x_train[0]))\n#\npred = model.predict(x_train[0:1])\n# print(pred)\n# print(y_train[0:1])\n# print(type(pred))\n# print(type(x_train[0]))\n# print(type(pred[0]))\n# print(type(y_train[0]))\n\nx = tf.convert_to_tensor(x_train[0])\npred = tf.convert_to_tensor(pred[0])\ny = tf.convert_to_tensor(y_train[0])\nprint(type(x), type(pred), type(y))\nadv_input = fgm(x=x, preds=pred, y=y, clip_min=-1.0, clip_max=1.0)\nprint(adv_input)\n\n\n","sub_path":"examples/mytest/fgsm.py","file_name":"fgsm.py","file_ext":"py","file_size_in_byte":4850,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"497357127","text":"import os\nimport numpy as np\nimport tensorflow as tf\n\nimport collections\nfrom nltk.corpus import stopwords\n\n########################################################################\n\nwrite_dir = 'models_dir'\nomnibus = 'omnibus.zip'\ndefault_embedding_size = 20\ndefault_epochs = 10\ninitial_learning_rate = 0.15\nnumber_negative_examples = 20\nbatch_size_default = 16\nwindow = 3\ndefault_min_count = 3\nnum_samples_per_word = 2\nthreads = 2\nsubsample_threashold =1e-3\ninteractive_mode = False\nstatistics_interval = 2\nsummary_interval = 5\ncheckpoint_interval = 600\n\n###############################################################################\n\n\n###############################################################################\n\nclass Options():\n\n ###########################################################################\n\n emb_dim = None\n num_samples = None\n learning_rate = None\n epochs_to_train = None\n concurrent_steps = None\n batch_size = None\n window_size = None\n min_count = None\n subsample = None\n statistics_interval = None\n summary_interval = None\n checkpoint_interval = None\n save_path = None\n #eval_data = None\n\n ###########################################################################\n\n def __init__(self):\n\n self.emb_dim = default_embedding_size\n self.num_samples = number_negative_examples\n self.learning_rate = initial_learning_rate\n self.epochs_to_train = default_epochs\n self.concurrent_steps = threads\n self.batch_size = batch_size_default\n self.window_size = window\n self.min_count = default_min_count\n self.samples_window = 2*window - 2\n self.subsample = subsample_threashold\n self.statistics_interval = statistics_interval\n self.summary_interval = summary_interval\n self.checkpoint_interval = checkpoint_interval\n self.save_path = write_dir\n if not os.path.exists(self.save_path):\n os.makedirs(self.save_path)\n #self.eval_data = fake_analogies\n\n ###########################################################################\n\n###############################################################################\n\n\n###############################################################################\n\nclass WordMapper():\n \n min_count = None\n null_words = None\n word_list = None\n unk_str = 'UNK'\n word_index = None\n vocab_size = None\n vocab_counts = None\n \n ###########################################################################\n \n def __init__(self, min_count):\n \n self.min_count = min_count\n \n ###########################################################################\n \n ###########################################################################\n \n def make(self, tokens):\n \n count = collections.Counter(tokens)\n count_common = {word: number for word, number\n in count.items() \n if number>=self.min_count}\n self.null_words = set(count.keys())-set(count_common.keys())\n sum_of_uncommon = sum(count[unk_word] \n for unk_word in self.null_words)\n\n self.word_list = np.array([word for word in count_common.keys()])\n \n self.word_index = {word: index for index, word\n in enumerate(self.word_list)}\n self.word_index.update({self.unk_str: len(self.word_list)})\n self.vocab_size = len(self.word_list) + 1\n self.vocab_counts = list(count_common.values()) + [sum_of_uncommon]\n return self\n \n ###########################################################################\n \n ###########################################################################\n \n def run(self, tokens):\n \n return [self.word_index[token] if token \n in self.word_list else self.word_index[self.unk_str]\n for token in tokens]\n \n ###########################################################################\n \n############################################################################### \n\n###############################################################################\n\nclass SkipNGramInput():\n\n ###########################################################################\n\n _options = None\n _wordmap = None\n data = None \n \n words_per_epoch = None\n\n ###########################################################################\n\n def __init__(self, options):\n\n self._options = options\n\n ###########################################################################\n \n ###########################################################################\n \n def make(self, tokenized_data):\n \n min_count = self._options.min_count\n self.length_corpus = len(tokenized_data) \n tokens = tokenized_data\n \n tokens = self.filter_stop_words(tokens)\n wm = WordMapper(min_count).make(tokens)\n cat_tokens = wm.run(tokens)\n self._wordmap = wm\n\n subsamples = self._options.samples_window\n self.words_per_epoch = subsamples*len(cat_tokens)\n words_per_epoch = self.words_per_epoch\n window = self._options.window_size\n \n pad_data = (cat_tokens[-window:] \n + cat_tokens + cat_tokens[:window])\n \n data_labels = np.array([subsamples*[datum] for\n datum in pad_data[window: -window]])\n data_window = np.array(list(map(lambda i: np.random.choice(\n np.hstack((pad_data[i-window:i], pad_data[i+1:i+1+window])),\n size=subsamples, replace=False),\n np.arange(window, len(pad_data)-window))))\n \n data_labels = data_labels.flatten()\n data_window = data_window.flatten()\n \n data_labels_tf = tf.data.Dataset.from_tensor_slices(\n tf.constant(data_labels,\n dtype=tf.int32, shape=(words_per_epoch, )))\n data_window_tf = tf.data.Dataset.from_tensor_slices(\n tf.constant(data_window,\n dtype=tf.int32, shape=(words_per_epoch, ))) \n self.data = tf.data.Dataset.zip(\n (data_labels_tf,\n data_window_tf)).batch(batch_size_default,\n True).repeat(default_epochs)\n\t# temporary\n self.data_labels = data_labels\n self.data_window = data_window\n\n return self\n\n ###########################################################################\n\n ###########################################################################\n \n def filter_stop_words(self, tokens):\n \n stop_words = set(stopwords.words('english'))\n stop_words = stop_words | set('.')\n # should enocde to bytes\n tokens = [word.lower() for word in tokens if word.lower()\n not in stop_words]\n return tokens\n \n ###########################################################################\n \n###############################################################################\n","sub_path":"word_embeddings_util.py","file_name":"word_embeddings_util.py","file_ext":"py","file_size_in_byte":7109,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"411573039","text":"# -*- coding: utf-8 -*-\n\n\"\"\"\n@author: Dylan Chen\n\nTo be continue (IV value)\n\n\"\"\"\n\nimport csv\nimport descriptive_statistics\nimport glob\nimport os\nfrom functools import partial\nfrom multiprocessing import Pool\n\nimport numpy as np\nfrom scipy.stats.stats import pearsonr\nfrom sklearn.feature_selection import chi2\n\n\nclass AdvancedEdd:\n \"\"\"\n\n split row in a vertical direction way, for low-memory computation\n ensure target y is in the last column\n\n header: boolean, default True [must be True :)]\n write each column to input_path directory\n\n \"\"\"\n\n def __init__(self, input_path, var_type, output_path, y_index):\n\n self.input_path = input_path\n self.var_type = var_type\n self.output_path = output_path\n self.y_index = y_index\n\n pass\n\n def row2fields(self, var_type=None, header=True):\n\n num_fields = 0\n tmp_list = []\n keys = []\n\n # ensure the input file containing header\n if not header:\n raise OSError('Ensure header exists in your file !!')\n\n if not var_type:\n raise OSError('Must provide each Variables type !!')\n\n with open(self.input_path, 'r') as csv_file:\n data = csv.reader(csv_file, delimiter=',')\n\n for row_num, line in enumerate(data):\n\n if row_num == 0 and header:\n\n num_fields = len(line)\n keys = line\n print('Label Y Column Name: %s' % keys[self.y_index])\n\n for i in range(num_fields - 1):\n # create columns temp file and write header (one column, y label) to each file\n tmp_list.append(open('%s/field' % os.path.dirname(self.input_path) + str(i + 1) + '.tmp', 'w'))\n tmp_list[i].write('{0},{1},{2}'.format(keys[i].replace(',', ','), 'variable_type', 'target_y') + \"\\n\")\n\n continue\n\n # write each column together with y to the corresponding tmp file\n for i in range(num_fields - 1):\n try:\n tmp_list[i].write(\n ','.join([line[i].replace(',', ','), var_type[keys[i]], line[self.y_index]]) + \"\\n\")\n except IndexError:\n tmp_list[i].write(\"\\n\")\n\n for i in range(num_fields - 1):\n tmp_list[i].close()\n\n def feature_explore(self, output_path, input_file):\n \"\"\"\n Notes:\n 1. ensure all missing values are represented by ''\n 2. the statistic score is implemented without nan values\n 3. this part is practised in parallel\n 4. refer to function (descriptive_statistics) in edd.py\n\n \"\"\"\n\n col_x = []\n tar_y = []\n key = ''\n var_type = ''\n\n type_algorithm = {'categorical': [chi2],\n 'numeric': [pearsonr]}\n\n with open(input_file) as csv_file:\n print(input_file)\n file = csv.reader(csv_file, delimiter=',')\n\n # collect column x and target y into list\n for row_num, line in enumerate(file):\n x, var_type, y = line\n\n if row_num == 0:\n key = x\n continue\n\n col_x.append(x)\n tar_y.append(float(y))\n var_type = var_type\n\n # get each variable type's corresponding statistic score\n des_dict, header = descriptive_statistics.desc_stat(col_x)\n\n # convert categorical columns to numeric\n if var_type == 'categorical':\n col_x = [col_x.index(x) if x != '' else '' for x in col_x]\n\n value_mask = [i for i, x in enumerate(col_x) if x != '']\n\n array_x = np.array(col_x)[value_mask].astype(np.float64)\n array_y = np.array(tar_y)[value_mask].astype(np.float64)\n\n for num, method in enumerate(type_algorithm[var_type]):\n if method == chi2:\n array_x = array_x.reshape(len(array_x), 1)\n\n try:\n statistic, p_value = method(array_x, array_y)\n except ValueError:\n statistic = ' '\n p_value = ' '\n\n if method == chi2:\n statistic = statistic[0]\n p_value = p_value[0]\n des_dict.update({'statistic_%s' % num: statistic, 'p_value_%s' % num: p_value})\n header += ['statistic_%s' % num, 'p_value_%s' % num]\n\n header = ['var_name'] + header\n des_dict.update({'var_name': key})\n\n output = []\n with open(output_path, 'a') as out_file:\n for name in header:\n output.append(str(des_dict[name]))\n out_file.write(','.join(output) + '\\n')\n\n def run(self):\n\n print(\"Getting each variables' type ...\")\n type_dict = {}\n with open(self.var_type) as csv_file:\n\n dict_file = csv.reader(csv_file, delimiter=',')\n\n for row_index, row_data in enumerate(dict_file):\n if row_index == 0:\n continue\n k, v = row_data\n type_dict.update({k: v})\n\n print(\"Splitting TXT file in a vertical direction way ...\")\n self.row2fields(var_type=type_dict)\n\n output_header = ['var_name', 'mean', 'std', 'No_obs', 'No_uniq', 'sum', 'missrate', 'misscnt', 'mod1', 'mod2',\n 'mod3', 'max', 'min', 'p1', 'p5', 'p10', 'p25', 'p33', 'p50', 'p66', 'p75', 'p90', 'p95',\n 'p99',\n 'statistic_0', 'p_value_0', 'statistic_1', 'p_value_1']\n\n print(\"Creating output header ...\")\n with open(self.output_path, 'w') as header_file:\n header_file.write(','.join(output_header) + '\\n')\n\n file_list = glob.glob(os.path.join(os.path.dirname(self.input_path), 'field*.tmp'))\n\n print(\"Computing each Column's descriptive statistics ... \")\n func = partial(self.feature_explore, self.output_path)\n pool = Pool(processes=10)\n p = pool.map(func, file_list)\n\n print(\"Removing temporary files ...\")\n for remove_file in file_list:\n os.remove(remove_file)\n\n print('Done')\n","sub_path":"advanced_edd.py","file_name":"advanced_edd.py","file_ext":"py","file_size_in_byte":6247,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"131359626","text":"# d[i] 값에 (현재 값)과 (d[i - 1] + 현재 값) 중 큰 값을 대입(음수가 포함돼 있을 수도 있기 때문에)\n\nn = int(input())\nnums = list(map(int, input().split()))\n\nd = [0] * n\nfor ix, num in enumerate(nums):\n if ix == 0:\n d[0] = num\n continue\n\n if num > num + d[ix - 1]:\n d[ix] = num\n else:\n d[ix] = num + d[ix - 1]\n\nprint(max(d))\n","sub_path":"BJ/DP/BJ_1912_continuous_sum.py","file_name":"BJ_1912_continuous_sum.py","file_ext":"py","file_size_in_byte":390,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"578252411","text":"from collections import OrderedDict\n\nfrom sqlalchemy.sql import text as sql_text\n\nfrom service_lib.database import DatabaseEngine\n\n\nclass PermissionsQuery:\n \"\"\"PermissionsQuery base class\n\n Query permissions for GDI resources.\n \"\"\"\n\n # name of public iam.role\n PUBLIC_ROLE_NAME = 'public'\n\n def __init__(self, config_models, logger):\n \"\"\"Constructor\n\n :param ConfigModels config_models: Helper for ORM models\n :param Logger logger: Application logger\n \"\"\"\n self.config_models = config_models\n self.logger = logger\n self.db_engine = DatabaseEngine()\n\n def public_role(self):\n \"\"\"Return public role name.\"\"\"\n return self.PUBLIC_ROLE_NAME\n\n def resource_ids(self, table_names, role, session):\n \"\"\"Query permissions for multiple GDI resource types.\n\n Return set of permitted GDI resource IDs.\n\n :param list[str] table_names: List of table names for GDI resource types\n :param str role: Role name\n :param Session session: DB session\n \"\"\"\n ResourcePermission = self.config_models.model('resource_permission')\n GDIResource = self.config_models.model('gdi_resource')\n\n # base query for all permissions of user\n query = self.role_permissions_query(role, session)\n\n # filter permissions by GDI resource types\n query = query.join(ResourcePermission.resource) \\\n .filter(GDIResource.table_name.in_(table_names))\n\n # execute query and pluck resource IDs\n resource_ids = [p.gdi_oid_resource for p in query.all()]\n\n return set(resource_ids)\n\n def resource_permissions(self, table_name, resource_name, role, session):\n \"\"\"Query permissions for a GDI resource type and optional name.\n\n Return resource permissions sorted by priority.\n\n :param str table_name: Table name for a GDI resource type\n :param str resource_name: optional GDI resource name (None for all)\n :param str role: Role name\n :param Session session: DB session\n \"\"\"\n ResourcePermission = self.config_models.model('resource_permission')\n GDIResource = self.config_models.model('gdi_resource')\n\n # base query for all permissions of user\n query = self.role_permissions_query(role, session)\n\n # filter permissions by GDI resource types\n query = query.join(ResourcePermission.resource) \\\n .filter(GDIResource.table_name == table_name)\n\n if resource_name is not None:\n # filter by resource name\n query = query.filter(GDIResource.name == resource_name)\n\n # order by priority\n query = query.order_by(ResourcePermission.priority.desc())\n\n # execute query and return results\n return query.all()\n\n def role_permissions_query(self, role, session):\n \"\"\"Create base query for all permissions of a role.\n\n :param str role: Role name\n :param Session session: DB session\n \"\"\"\n ResourcePermission = self.config_models.model('resource_permission')\n Role = self.config_models.model('role')\n\n # create query for permissions of role\n query = session.query(ResourcePermission). \\\n join(ResourcePermission.role). \\\n filter(Role.name == role)\n\n return query\n\n def all_resource_ids(self, table_names, session):\n \"\"\"Query multiple GDI resource types.\n\n Return set of GDI resource IDs.\n\n :param list[str] table_names: List of table names for GDI resource types\n :param Session session: DB session\n \"\"\"\n ResourcePermission = self.config_models.model('resource_permission')\n GDIResource = self.config_models.model('gdi_resource')\n\n # create query with any permissions\n query = session.query(ResourcePermission) \\\n .join(ResourcePermission.resource) \\\n .filter(GDIResource.table_name.in_(table_names))\n\n # execute query and pluck resource IDs\n resource_ids = [p.gdi_oid_resource for p in query.all()]\n\n return set(resource_ids)\n\n def postgis_metadata(self, conn_str, schema, table_name, geometry_column):\n \"\"\"Return primary key, geometry column, type and srid from a PostGIS\n table.\n\n :param str conn_str: DB connection URL\n :param str schema: DB schema name\n :param str table_name: DB table name\n :param str geometry_column: optional geometry column name if not unique\n \"\"\"\n info = {}\n\n conn = None\n try:\n # connect to GeoDB\n engine = self.db_engine.db_engine(conn_str)\n conn = engine.connect()\n\n # get primary key\n\n # build query SQL\n sql = sql_text(\"\"\"\n SELECT a.attname\n FROM pg_index i\n JOIN pg_attribute a ON a.attrelid = i.indrelid\n AND a.attnum = ANY(i.indkey)\n WHERE i.indrelid = '{schema}.{table}'::regclass\n AND i.indisprimary;\n \"\"\".format(schema=schema, table=table_name))\n\n # execute query\n result = conn.execute(sql)\n for row in result:\n info['primary_key'] = row['attname']\n\n # get geometry column and srid\n\n # build query SQL\n where_clause = \"\"\n if geometry_column:\n # select specific geometry column\n where_clause = sql_text(\"\"\"\n AND f_geometry_column = '{geom_column}'\n \"\"\".format(geom_column=geometry_column))\n\n sql = sql_text(\"\"\"\n SELECT f_geometry_column, srid, type\n FROM geometry_columns\n WHERE f_table_schema = '{schema}' AND f_table_name = '{table}'\n {where_clause};\n \"\"\".format(\n schema=schema, table=table_name, where_clause=where_clause)\n )\n\n # execute query\n result = conn.execute(sql)\n for row in result:\n info['geometry_column'] = row['f_geometry_column']\n info['geometry_type'] = row['type']\n info['srid'] = row['srid']\n\n # close database connection\n conn.close()\n except Exception as e:\n self.logger.error(\n \"Could not get PostGIS metadata for table '%s.%s':\\n%s\" %\n (schema, table_name, e)\n )\n if conn:\n conn.close()\n\n return info\n\n def attribute_metadata(self, conn_str, schema, table_name, column):\n \"\"\"Return data type and any constraints for a table column.\n\n :param str conn_str: DB connection URL\n :param str schema: DB schema name\n :param str table_name: DB table name\n :param str column: Column name\n \"\"\"\n data_type = 'text'\n # NOTE: use ordered keys\n constraints = OrderedDict()\n\n try:\n # connect to GeoDB\n geo_db = self.db_engine.db_engine(conn_str)\n conn = geo_db.connect()\n\n # build query SQL\n sql = sql_text(\"\"\"\n SELECT data_type, character_maximum_length, numeric_precision,\n numeric_scale\n FROM information_schema.columns\n WHERE table_schema = '{schema}' AND table_name = '{table}'\n AND column_name = '{column}'\n ORDER BY ordinal_position;\n \"\"\".format(schema=schema, table=table_name, column=column))\n\n # execute query\n result = conn.execute(sql)\n for row in result:\n data_type = row['data_type']\n\n # constraints from data type\n if (data_type in ['character', 'character varying'] and\n row['character_maximum_length']):\n constraints['maxlength'] = row['character_maximum_length']\n elif data_type in ['double precision', 'real']:\n # NOTE: use text field with pattern for floats\n constraints['pattern'] = '[0-9]+([\\\\.,][0-9]+)?'\n elif data_type == 'numeric' and row['numeric_precision']:\n step = pow(10, -row['numeric_scale'])\n max_value = pow(\n 10, row['numeric_precision'] - row['numeric_scale']\n ) - step\n constraints['numeric_precision'] = row['numeric_precision']\n constraints['numeric_scale'] = row['numeric_scale']\n constraints['min'] = -max_value\n constraints['max'] = max_value\n constraints['step'] = step\n elif data_type == 'smallint':\n constraints['min'] = -32768\n constraints['max'] = 32767\n elif data_type == 'integer':\n constraints['min'] = -2147483648\n constraints['max'] = 2147483647\n elif data_type == 'bigint':\n # NOTE: JSON/Javascript may reduce precision\n constraints['min'] = -9223372036854775808\n constraints['max'] = 9223372036854775807\n\n if not data_type:\n self.logger.warning(\n \"Could not find data type of column '%s' \"\n \"of table '%s.%s'\" % (column, schema, table_name)\n )\n data_type = 'text'\n\n # close database connection\n conn.close()\n except Exception as e:\n self.logger.error(\n \"Could not get data type of column '%s' \"\n \"of table '%s.%s':\\n%s\" % (column, schema, table_name, e)\n )\n data_type = 'text'\n\n return {\n 'data_type': data_type,\n 'constraints': constraints\n }\n","sub_path":"permissions_query.py","file_name":"permissions_query.py","file_ext":"py","file_size_in_byte":9943,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"181940904","text":"from pymongo import MongoClient\nimport pprint\nclient=MongoClient()\nsets={\"name\":\"monika\",\"surname\":\"singh\",\"id\":2314026,\"year\":[2012,2017],\"layout\":[\"good\",\"bad\"],\"designer\":{\"first\":\"tata\",\"last\":\"motor\"},\"work\":\"student\"}\ndb=client.examples #examples is the database\ndb.sets_table.insert(sets)#sets_table is the collection(table)\nfor a in db.sets_table.find():\n pprint.pprint(a)\n''' \ndef find():\n sets_table=db.sets_table.find({\"surname\":\"singh\"})\n for a in sets_table:\n pprint.pprint(a)\n\nif __name__==\"__main__\":\n find()\n'''\n","sub_path":"insert1_mongo.py","file_name":"insert1_mongo.py","file_ext":"py","file_size_in_byte":548,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"431609185","text":"import os\nimport re\nfrom inspect import getmembers, getmodule, ismethod\nfrom typing import Callable, Union\n\nfrom .connection import get_connection\n\n\nclass ImportNotFoundException(Exception):\n pass\n\n\nclass ModuleNotfoundException(Exception):\n pass\n\n\ndef sql_table_name(queue_table: str) -> str:\n return ('.').join([f'\"{x}\"' for x in queue_table.split('.')]) \n\n\nasync def get_fn_name(func: Union[str, Callable]) -> str:\n try:\n if isinstance(func, str):\n return func\n if ismethod(func):\n module_name = get_fn_name(dict(getmembers(func))['__self__'])\n else:\n module_name = getmodule(func).__name__\n name = func.__name__\n return '.'.join([module_name, name])\n except AttributeError as e:\n raise ModuleNotfoundException(e)\n\n\nasync def import_fn(fn_name) -> Callable:\n access_path = fn_name.split(\".\")\n module = None\n try:\n for index in range(1, len(access_path)):\n try:\n # import top level module\n module_name = \".\".join(access_path[:-index])\n module = __import__(module_name)\n except ImportError:\n continue\n else:\n for step in access_path[1:-1]: # walk down it\n module = getattr(module, step)\n break\n if module:\n return getattr(module, access_path[-1])\n else:\n return globals()[\"__builtins__\"][fn_name]\n except KeyError as e:\n raise ImportNotFoundException(e)\n\n\nasync def create_table(name: str):\n if not name:\n name = os.getenv('MOD_NGARN_TABLE', 'modngarn_job')\n print(f\"Creating table {name}...\")\n cnx = await get_connection()\n async with cnx.transaction():\n await cnx.execute(\n \"\"\"CREATE TABLE IF NOT EXISTS {queue_table} (\n id TEXT NOT NULL CHECK (id !~ '\\\\|/|\\u2044|\\u2215|\\u29f5|\\u29f8|\\u29f9|\\ufe68|\\uff0f|\\uff3c'),\n fn_name TEXT NOT NULL,\n args JSON DEFAULT '[]',\n kwargs JSON DEFAULT '{{}}',\n priority INTEGER DEFAULT 0,\n created TIMESTAMP WITH TIME ZONE DEFAULT NOW(),\n scheduled TIMESTAMP WITH TIME ZONE,\n executed TIMESTAMP WITH TIME ZONE,\n canceled TIMESTAMP WITH TIME ZONE,\n result JSON,\n reason TEXT,\n processed_time TEXT,\n PRIMARY KEY (id)\n );\n \"\"\".format(\n queue_table=name\n )\n )\n\n await cnx.execute(\n f\"\"\"CREATE INDEX IF NOT EXISTS idx_pending_jobs ON {name} (executed) WHERE executed IS NULL;\"\"\"\n )\n print(f\"Done\")\n","sub_path":"mod_ngarn/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":2798,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"82428226","text":"\"\"\"\nTic-Tac-Toe Game\n\"\"\"\n\nimport random\n\ngame_board = ['#', 'X', 'O', 'X', 'O', 'X', 'O', 'X', 'O', 'X']\n\n\"\"\"\nDisplay board to user.\n\"\"\"\n\n\ndef display_board(board):\n\n print(board[1]+'|'+board[2]+'|'+board[3])\n print(board[4]+'|'+board[5]+'|'+board[6])\n print(board[7]+'|'+board[8]+'|'+board[9])\n\n\n\"\"\"\nCheck to see what marker the players will have\n\"\"\"\n\n\ndef player_input():\n\n marker = ''\n\n while not (marker == 'X' or marker == 'O'):\n marker = input('Player 1: Choose X or O: ').upper()\n\n if marker == 'X':\n return ('X', 'O')\n else:\n return ('O', 'X')\n\n\n\"\"\"\nTakes in board list object, marker and assigns it to the board\n\"\"\"\n\n\ndef place_marker(board, marker, position):\n board[position] = marker\n\n\n\"\"\"\nTake in board list and checks to see if the marker has won\n\"\"\"\n\n\"\"\"\nChecking board to see if there is a winner\n\"\"\"\n\n\ndef win_check(board, mark):\n return ((board[1] == mark and board[2] == mark and board[3] == mark) or\n (board[4] == mark and board[5] == mark and board[6] == mark) or\n (board[7] == mark and board[8] == mark and board[9] == mark) or\n (board[1] == mark and board[4] == mark and board[7] == mark) or\n (board[2] == mark and board[5] == mark and board[8] == mark) or\n (board[3] == mark and board[6] == mark and board[9] == mark) or\n (board[1] == mark and board[5] == mark and board[9] == mark) or\n (board[3] == mark and board[5] == mark and board[7] == mark))\n\n\n\"\"\"\nPicking the first player to make a move\n\"\"\"\n\n\ndef choose_first():\n\n flip = random.randint(0, 1)\n\n if flip == 0:\n return 'Player 1'\n else:\n return 'Player 2'\n\n\n\"\"\"\nCheck to see if the position on a board is free to place\n\"\"\"\n\n\ndef space_check(board, position):\n return board[position] == ' '\n\n\n\"\"\"\nCheck if board is full and return a boolean value\n\"\"\"\n\n\ndef full_board_check(board):\n\n for i in range(1, 10):\n if space_check(board, i):\n return False\n return True\n\n\n\"\"\"\nAsk player for next position as a number and use previous function to check if\nthat play can be made\n\"\"\"\n\n\ndef player_choice(board):\n\n position = 0\n\n while position not in range(1, 10) or not space_check(board, position):\n position = int(input('Choose a position: 1-9: '))\n\n return position\n\n\n\"\"\"\nWrite a function that asks the player if they want to play again\n\"\"\"\n\n\ndef replay():\n\n choice = input('Play again? Enter Yes or No ')\n\n return choice == 'Yes'\n\n\n\"\"\"\nLogic\n\"\"\"\n\nprint('Welcome to Tic Tac Toe')\n\nwhile True:\n\n the_board = [' ']*10\n player1_marker, player2_marker = player_input()\n\n turn = choose_first()\n print(turn + ' will play first.')\n\n play_game = input('Ready to play? y or n?')\n\n if play_game == 'y':\n game_on = True\n else:\n game_on = False\n\n while game_on:\n\n if turn == 'Player 1':\n display_board(the_board)\n position = player_choice(the_board)\n place_marker(the_board, player1_marker, position)\n\n if win_check(the_board, player1_marker):\n display_board(the_board)\n print('Player 1 has won the game.')\n game_on = False\n else:\n if full_board_check(the_board):\n display_board(the_board)\n print('The game is a draw.')\n game_on = False\n else:\n turn = 'Player 2'\n\n else:\n display_board(the_board)\n position = player_choice(the_board)\n place_marker(the_board, player2_marker, position)\n\n if win_check(the_board, player2_marker):\n display_board(the_board)\n print('Player 2 has won the game.')\n game_on = False\n else:\n if full_board_check(the_board):\n display_board(the_board)\n print('The game is a draw.')\n game_on = False\n else:\n turn = 'Player 1'\n\n if not replay():\n print('Thanks for playing Tic Tac Toe.')\n break\n","sub_path":"Syllabus/06-Python-Milestone-Project-1/project.py","file_name":"project.py","file_ext":"py","file_size_in_byte":4152,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"648285008","text":"import unittest\nimport os\nfrom selenium.webdriver import DesiredCapabilities, Remote\nfrom pages.auth_page import AuthPage\nfrom pages.userinfo_page import UserinfoPage\n\nclass WrongTownTest(unittest.TestCase):\n WRONG_TOWN_NAME = 'qwertyuiop'\n TOP_MESSAGE = 'Некоторые поля заполнены неверно'\n TOWN_ERROR = 'Проверьте название города'\n\n def setUp(self):\n browser = os.environ.get('BROWSER', 'CHROME')\n\n self.driver = Remote(\n command_executor='http://127.0.0.1:4444/wd/hub',\n desired_capabilities=getattr(DesiredCapabilities, browser).copy()\n )\n\n def tearDown(self):\n self.driver.quit()\n\n def test(self):\n auth_page = AuthPage(self.driver)\n auth_page.open()\n auth_page.authorize()\n\n userinfo_page = UserinfoPage(self.driver)\n userinfo_page.open()\n userinfo_form = userinfo_page.form\n userinfo_form.set_town(self.WRONG_TOWN_NAME)\n userinfo_form.wait_for_suggests_invisible()\n userinfo_form.save()\n self.assertEqual(self.TOP_MESSAGE, userinfo_form.get_top_message())\n self.assertEqual(self.TOWN_ERROR, userinfo_form.get_town_message())","sub_path":"tests/userinfo/wrongtown_test.py","file_name":"wrongtown_test.py","file_ext":"py","file_size_in_byte":1227,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"321620945","text":"'''Desenvolva um programa que leia o comprimento de 3 retas e diga ao usuário se elas podem ou não formar um triangulo'''\r\n\r\nreta1 = float(input('Informe o valor da primeira reta: '))\r\nreta2 = float(input('Informe o valor da segunda reta: '))\r\nreta3 = float(input('Informe o valor da terceira reta: '))\r\n\r\nif reta1 < reta2 + reta3 and reta2 < reta1 + reta3 and reta3 < reta1 + reta2:\r\n print('Os valores informados formam um triângulo!')\r\nelse:\r\n print('Com os valores informados não é possível formar retângulos')","sub_path":"Desafio35.py","file_name":"Desafio35.py","file_ext":"py","file_size_in_byte":527,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"618469380","text":"# -*- coding: utf-8 -*-\n\"\"\"\nDate created: Tue Mar 7 15:38:34 2017\nAuthor: S. Hutchins\n\nScript description: Testing different Clustal Omega parameters.\n\n\"\"\"\nimport os\nfrom ClustalParams import COP1, COP2, COP3, COP4, COP5, COP6, COP7\n\n# Assess current working directories and establishing home and output directories.\n\na = '/work5/r2295/bin/ClustalParamsTest/' # Home directory\nhome = a\nos.chdir(home) # Directory change to output directory\n\nparamnames = [\"params1\", \"params2\", \"params3\", \"params4\", \"params5\", \"params6\", \"params7\"]\nparameters = [COP1, COP2, COP3, COP4, COP5, COP6, COP7]\n\nfor name, param in zip(paramnames, parameters):\n # Create directories for alignment files.\n b = a + name + \"-output\"\n os.makedirs('%s' % b, exist_ok=True) # Create a directory or don't if it exists.\n\n # Go to the CDS directory\n c = a + \"InputCDS\"\n\n os.chdir(c) # Change to cds.fasta file directory\n os.listdir(c) # Make a list of the files in the current directory\n print(\"➜ Current CDS/Gene directory: \"+ os.getcwd() + \"\\n\") # Print current working directory\n\n # Echos all commands in the current shell.\n os.system(\"set -x\")\n\n # Copies the profile.fasta and concatenated cds.fasta file to output dir.\n os.system(\"cp {APOL1_cds_nucl.fasta,profile.fasta} \" + b + \"/\")\n\n # Directory change to output directory\n os.chdir(b)\n\n # Run Clustal Omega Commandline tool\n param(in_file=\"APOL1_cds_nucl.fasta\", out_file=name + \".fasta\", outfmt=\"fasta\", logfile=name + \".log\")","sub_path":"{{ cookiecutter.project_slug }}/misc/Tests and Examples/ClustalOmega/ParametersTest/ClustalParamsTest1/params_test.py","file_name":"params_test.py","file_ext":"py","file_size_in_byte":1504,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"132240001","text":"from itertools import groupby\nimport operator\nimport time\nfrom datetime import datetime\n\ndef repeat_func(func, times, *args):\n for i in range(times):\n yield func(*args)\n\ndef to_range_str(iterable):\n '''\n Convert an iterable object of integers to a string in with consecutive\n numbers combined into a range.\n '''\n result = ''\n sorted_list = sorted(iterable)\n for (_, g) in groupby(enumerate(sorted_list), lambda t : t[0] - t[1]):\n rlist = list(map(operator.itemgetter(1), g))\n if len(rlist) > 1:\n single = '{}-{}'.format(rlist[0], rlist[-1])\n else:\n single = '{}'.format(rlist[0])\n if not result:\n result = single\n else:\n result += ' {}'.format(single)\n\n return result\n\ndef range_str_to_list(s):\n ret = []\n for f in s.strip(' ,\\n\\t').replace(',', ' ').split():\n r = list(map(int, f.split('-')))\n ret.extend(range(r[0], r[-1] + 1))\n\n return ret\n\ndef get_milliseconds():\n return int(time.time()) * 1000 + (datetime.now().microsecond) // 1000\n\ndef is_trade_time():\n now = datetime.now()\n return now.isoweekday() in (1, 2, 3, 4, 5) and now.hour in range(9, 16)\n\n# This must be put at the last line!\n__all__ = [x for x in locals() if not x.startswith('__')]\n","sub_path":"python/tools/binary-message/ouou/util/functions.py","file_name":"functions.py","file_ext":"py","file_size_in_byte":1297,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"212639598","text":"from threading import Thread\nimport time\nimport threading\n\n\ndef test1():\n g_num = 100\n name = threading.current_thread().name\n if name == \"Thread-1\":\n g_num += 1\n else:\n time.sleep(2)\n print(\"--Thread is %s--g_num=%d\" % (name, g_num))\n\n\nt1 = Thread(target=test1)\nt1.start()\n\nt2 = Thread(target=test1)\nt2.start()","sub_path":"PycharmProjects/PythonCodes/02-Senior/11-线程/08-多进程使用非共享变量.py","file_name":"08-多进程使用非共享变量.py","file_ext":"py","file_size_in_byte":340,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"331874288","text":"# -*- coding: utf-8 -*-\n# perimeter form tests\nimport datetime\n\nfrom django.core.exceptions import ValidationError\nfrom django.test import TestCase, RequestFactory, override_settings\nfrom django.utils.timezone import now\n\nfrom perimeter.forms import GatewayForm\nfrom perimeter.models import AccessToken, AccessTokenUse\n\nYESTERDAY = now().date() - datetime.timedelta(days=1)\n\nclass GatewayFormTests(TestCase):\n\n def setUp(self):\n self.factory = RequestFactory()\n self.payload = {\n 'token': \"test\",\n 'email': \"hugo@yunojuno.com\",\n 'name': \"Hugo Rodger-Brown\"\n }\n self.request = self.factory.post('/',\n data=self.payload,\n REMOTE_ADDR='127.0.0.1',\n HTTP_USER_AGENT='test_agent'\n )\n self.token = AccessToken(token=\"test\").save()\n self.form = GatewayForm(self.request.POST)\n\n def test_post_valid_token(self):\n self.assertTrue(self.form.is_valid())\n self.assertEqual(self.form.token, self.token)\n\n def test_clean_inactive_token(self):\n self.token.is_active = False\n self.token.save(update_fields=['is_active'])\n self.assertFalse(self.form.is_valid())\n self.assertRaises(ValidationError, self.form.clean_token)\n\n def test_clean_expired_token(self):\n self.token.expires_on = YESTERDAY\n self.token.save(update_fields=['expires_on'])\n self.assertFalse(self.form.is_valid())\n self.assertRaises(ValidationError, self.form.clean_token)\n\n def test_no_matching_token(self):\n AccessToken.objects.all().delete()\n self.assertFalse(self.form.is_valid())\n self.assertRaises(ValidationError, self.form.clean_token)\n\n def test_save(self):\n self.request.session = {}\n self.assertTrue(self.form.is_valid())\n au = self.form.save(self.request)\n self.assertTrue(AccessTokenUse.objects.get(), au)\n self.assertEqual(au.user_email, self.payload['email'])\n self.assertEqual(au.user_name, self.payload['name'])\n self.assertEqual(au.token, self.token)\n self.assertEqual(au.client_ip, '127.0.0.1')\n self.assertEqual(au.client_user_agent, 'test_agent')\n","sub_path":"perimeter/tests/test_forms.py","file_name":"test_forms.py","file_ext":"py","file_size_in_byte":2205,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"614363516","text":"import pygame;\nimport random;\nimport sys;\nimport pickle;\nimport numpy as np;\nfrom pygame.locals import *;\n\n### define game logic methods ####\n\ndef collide(x1, x2, y1, y2, w1, w2, h1, h2):\n\tif x1+w1>x2 and x1y2 and y1 -1 and ys[i]/obj_size > -1):\n\t\tgrid[xs[i]/obj_size][ys[i]/obj_size] = 0;\n\t\n\twhile i >= 1:\n\t\txs[i] = xs[i-1];\n\t\tys[i] = ys[i-1];\n\t\ti -= 1\n\n\tif dirs==0:\n\t\tys[0] += obj_size;\n\telif dirs==1:\n\t\txs[0] += obj_size;\n\telif dirs==2:\n\t\tys[0] -= obj_size;\n\telif dirs==3:\n\t\txs[0] -= obj_size;\n\n\tif(flag and xs[i]/obj_size < grid_size and ys[i]/obj_size < grid_size and xs[i]/obj_size > -1 and ys[i]/obj_size > -1):\n\t\tgrid[xs[i]/obj_size][ys[i]/obj_size] = 1;\n\ndef check_boundries(xs, ys):\n\tif xs[0] < 0 or xs[0] > (grid_size-1)*obj_size or ys[0] < 0 or ys[0] > (grid_size-1)*obj_size: \n\t\treturn True;\n\treturn False;\n\ndef check_self(xs,ys):\n\ti = len(xs)-1\n\twhile i >= 2:\n\t\tif collide(xs[0], xs[i], ys[0], ys[i], obj_size, obj_size, obj_size, obj_size):\n\t\t\treturn True;\n\t\ti-= 1\n\treturn False;\n\ndef check_apple(xs,ys):\n\treturn collide(xs[0], applepos[0], ys[0], applepos[1], obj_size, obj_size, obj_size, obj_size);\n\n### define init and main game loop ###\ndef init_game():\n\tcurr_it = 0;\n\t\n\tif TRAIN:\n\t\tglobal Q, Q_eps, Q_alpha, Q_gamma, Q_count, Q_runs;\n\t\t\n\t\ttrain_file = open(\"train\",\"rb+\");\n\n\t\tdump_limit = 500;\n\t\tdump_time = 0;\n\t\tcurr_it = 0;\n\n\t\tobj = pickle.load(train_file);\n\t\tcurr_it = obj[\"curr_it\"];\n\t\tQ_eps = obj[\"eps\"];\n\t\tQ_alpha = obj[\"alpha\"];\n\t\tQ_gamma = obj[\"gamma\"];\n\t\tQ_count = obj[\"count\"];\n\t\tQ_runs = obj[\"runs\"];\n\t\tQ = obj[curr_it];\n\n\t\tprint(\"loaded training at \" + str(curr_it));\n\n\tfor i in range(curr_it,100000):\n\t\tcurr_it += 1;\n\n\t\tif TRAIN:\n\t\t\tdump_time += 1;\n\n\t\t\tif(dump_time >= dump_limit):\n\t\t\t\tdump_time = 0;\n\n\t\t\t\tobj[\"runs\"] = Q_runs;\n\t\t\t\tobj[\"curr_it\"] = curr_it;\n\t\t\t\tobj[\"eps\"] = Q_eps;\n\t\t\t\tobj[\"alpha\"] = Q_alpha;\n\t\t\t\tobj[\"gamma\"] = Q_gamma;\n\t\t\t\tobj[\"count\"] = Q_count;\n\t\t\t\tobj[curr_it] = Q;\n\n\t\t\t\ttrain_file.seek(0);\n\t\t\t\tpickle.dump(obj,train_file,pickle.HIGHEST_PROTOCOL);\n\n\t\t\t\tprint(\"wrote training until now\");\n\n\t\t## reset varaibles\n\t\tglobal grid, xs, ys, dirs, score, applepos;\n\n\t\txs = np.copy(xs_orig);\n\t\tys = np.copy(ys_orig);\n\t\tdirs = 0;\n\t\tscore = snake_init_length;\n\t\tapplepos = rand_apple_pos();\n\t\tgrid = np.copy(grid_orig);\n\n\t\trun();\n\n\t\tQ_runs = np.append(Q_runs,score);\n\n\t\tprint(str(curr_it) + \" : \" + str(score));\n\ndef run():\n\tglobal xs, ys, applepos, score, dirs;\n\n\twhile True:\n\t\tif not DEBUG:\n\t\t\tif USER:\n\t\t\t\tclock.tick(20);\n\t\t\telse:\n\t\t\t\tclock.tick(FPS);\n\t\t\tfor e in pygame.event.get():\n\t\t\t\tif e.type == QUIT:\n\t\t\t\t\ttrain_file.close();\n\t\t\t\t\tsys.exit(0)\n\t\t\t\telif e.type == KEYDOWN and USER:\n\t\t\t\t\tif e.key == K_UP and dirs != 0:\n\t\t\t\t\t\tdirs = 2\n\t\t\t\t\telif e.key == K_DOWN and dirs != 2:\n\t\t\t\t\t\tdirs = 0\n\t\t\t\t\telif e.key == K_LEFT and dirs != 1:\n\t\t\t\t\t\tdirs = 3;\n\t\t\t\t\telif e.key == K_RIGHT and dirs != 3:\n\t\t\t\t\t\tdirs = 1\n\n\t\t## ai here\n\t\tif(ALGO_PARAMS==None):\n\t\t\tdirs = ALGO();\n\t\telse:\n\t\t\tdirs = ALGO(ALGO_PARAMS);\n\n\t\tmove(xs,ys,dirs,True);\n\n\t\t## chekc collisions (self, apple, boundries)\n\t\tif(check_self(xs,ys)):\n\t\t\treturn;\n\t\tif check_apple(xs,ys):\n\t\t\tscore+=1;\n\t\t\txs = np.append(xs,grid_size*obj_size + 100);\n\t\t\tys = np.append(ys,grid_size*obj_size + 100);\n\t\t\tapplepos=rand_apple_pos();\n\t\tif check_boundries(xs, ys): \n\t\t\treturn;\n\n\t\tif not DEBUG:\n\t\t\ts.fill((255, 255, 255));\n\t\t\ts.blit(img_head, (xs[0],ys[0]));\n\t\t\tfor i in range(1, len(xs)):\n\t\t\t\ts.blit(img_body, (xs[i], ys[i]))\n\t\t\ts.blit(appleimage, applepos);\n\t\t\tt=f.render(\"length: \" + str(score), True, (0, 0, 0));\n\t\t\ts.blit(t, (10, 10));\n\t\t\tt=f.render(\"eps: \" + str(Q_eps), True, (0,0,0));\n\t\t\ts.blit(t, (10, 30));\n\t\t\tpygame.display.update()\n\n### define ai algos ###\n\ndef ai_weighted_rand_walk(params):\n\t\n\tw = params[0]; # weights\n\ts = params[1]; # smart or not (crash into self/walls ok?)\n\t\n\ttries = [(forward,w[0]), (left_turn,w[1]), (right_turn,w[2])] # forward, left, right\n\n\twhile(len(tries)>0):\n\t\tfake_xs = np.copy(xs);\n\t\tfake_ys = np.copy(ys);\n\t\tfake_dirs = dirs;\n\t\t\n\t\t## weighted choice of what to do next\n\t\ttotal = sum(w for c,w in tries);\n\t\tr = random.uniform(0, total);\n\t\tupto = 0;\n\t\tfor c,w in tries:\n\t\t\t\tif upto + w >= r:\n\t\t\t\t\t\tfunc = c;\n\t\t\t\t\t\tweight = w;\n\t\t\t\t\t\tbreak;\n\t\t\t\tupto += w;\n\n\t\tfake_dirs = func(fake_dirs);\n\t\tmove(fake_xs,fake_ys,fake_dirs,False);\n\t\t\n\t\tif not s:\n\t\t\treturn fake_dirs;\n\n\t\tif not (check_boundries(fake_xs,fake_ys) or check_self(fake_xs,fake_ys)):\n\t\t\treturn fake_dirs;\n\t\t\n\t\ttries.remove((func,weight));\n\n\t## at this point we are in a corner and whatever choice will lead us to die\n\t## might as well just return the original direction\n\treturn dirs;\n\ndef ai_smart():\n\ttries = [forward, left_turn, right_turn];\n\twork = [];\n\n\twhile(len(tries)>0):\n\t\tfake_xs = np.copy(xs);\n\t\tfake_ys = np.copy(ys);\n\t\tfake_dirs = dirs;\n\n\t\tfunc = random.choice(tries);\n\t\tfake_dirs = func(fake_dirs);\n\n\t\tmove(fake_xs, fake_ys, fake_dirs, False);\n\n\t\tif(check_boundries(fake_xs,fake_ys)):\n\t\t\ttries.remove(func);\n\t\t\tcontinue;\n\n\t\tif(check_self(fake_xs,fake_ys)):\n\t\t\ttries.remove(func);\n\t\t\tcontinue;\n\n\t\tdist = np.sqrt(pow(fake_xs[0]-applepos[0],2) + pow(fake_ys[0]-applepos[1],2));\n\t\twork.append([dist,fake_dirs]);\n\t\ttries.remove(func);\n\n\tif(len(work)>0):\n\t\tind = 0;\n\t\tmin_dist = 10000000;\n\t\t\n\t\tfor i in range(len(work)):\n\t\t\tif work[i][0] < min_dist:\n\t\t\t\tind = i;\n\t\t\t\tmin_dist = work[i][0];\n\n\t\treturn work[ind][1];\n\n\t## at this we are in a corner and whatever choice will lead us to die \n\t## might as well just return the original direction\n\treturn dirs;\n\n# Q learning AI\nQ = {};\nQ_runs = np.array([]);\nQ_actions = [forward, left_turn, right_turn];\nQ_eps = .3;\nQ_alpha = .3;\nQ_gamma = .7;\nQ_count = 0;\n\n### find apple quadrant relative to the snake's head\ndef Q_apple_quad(dirs):\n\tdx = 1 if applepos[0] >= xs[0] else -1\n\tdy = 1 if grid_size*obj_size - applepos[1] >= grid_size*obj_size - ys[0] else -1;\n\n\tif dirs == 0: #down\n\t\tqx = -dx;\n\t\tqy = -dy;\n\telif dirs == 1: #right\n\t\tqx = -dy;\n\t\tqy = dx;\n\telif dirs == 2: #up (does nothing)\n\t\tqx = dx;\n\t\tqy = dy;\n\telif dirs == 3: #left\n\t\tqx = dy;\n\t\tqy = -dx;\n\n\treturn (qx,qy);\n\ndef Q_get_state(dirs):\n\tspc = [0,0,0] # forward, left, right\n\tspc_dist = [10000000, 10000000, 10000000];\n\tfake_dirs = None;\n\n\tfor j in range(3):\n\t\tfake_xs = np.copy(xs);\n\t\tfake_ys = np.copy(ys);\n\t\tfake_dirs = dirs;\n\t\n\t\tfake_dirs = Q_actions[j](fake_dirs);\n\t\tmove(fake_xs, fake_ys, fake_dirs, False);\n\t\t\n\t\tif(check_boundries(fake_xs, fake_ys) or check_self(fake_xs, fake_ys)):\n\t\t\tspc[j] = -1;\n\n\t\tif check_apple(fake_xs, fake_ys):\n\t\t\tspc[j] = 1;\n\n\t\tspc_dist[j] = np.sqrt(pow(fake_xs[0]-applepos[0],2) + pow(fake_ys[0]-applepos[1],2));\n\n\t(qx, qy) = Q_apple_quad(dirs);\n\tstate = (spc[0],spc[1],spc[2],qx,qy);\n\n\treturn { \"spc\" : spc, \"spc_dist\" : spc_dist, \"state\" : state};\n\ndef ai_qlearn():\n\tglobal Q, Q_count, Q_eps;\n\t\n\t## due to artifacts of implementation, instead of having a current state\n\t## and previous state, like the Q-learn algorithm requires, we instead have\n\t## a current state and a next state, where the current state effectivley \n\t## acts like the previous state, and the next state acts as the current state\n\t## in terms of the Q-Learn Algo.\n\n\t## epsilon greedy exploration\n\tQ_count += 1;\n\tif Q_count == 10000:\n\t\tQ_eps = (.5)*Q_eps;\n\t\tQ_count = 0;\n\n\t## init some variables\n\tepisode_ended = False;\n\treward = 0;\n\tdistance = 10000000;\n\taction = -1;\n\n\t## get current state\n\tstate_space = Q_get_state(dirs);\n\tstate = state_space[\"state\"];\n\tspc = state_space[\"spc\"];\n\tspc_dist = state_space[\"spc_dist\"];\n\n\t# calculate distance to apple\n\tdist = np.sqrt(pow(xs[0]-applepos[0],2) + pow(ys[0]-applepos[1],2));\n\n\t## take best action (or explore)\n\tif state not in Q:\n\t\tQ[state] = [10,10,10]; # assume optimistic actions\n\n\tq = Q[state];\n\tmaxQ = max(q);\n\n\tif random.random() < Q_eps:\n\t\taction = random.choice([0,1,2]);\n\telse:\n\t\tbest = [i for i,x in enumerate(q) if x == maxQ];\n\t\taction = random.choice(best);\n\n\t# get next action\n\tfunc = Q_actions[action];\n\tfake_dirs = func(dirs);\n\n\t## get reward from next state\n\t## action corresponds to the action the snake will take\n\t## spc[action] corresponds to the outcome in next state\n\tif spc[action] == -1: # crashed\n\t\treward = -100;\n\t\tepisode_ended = True;\n\telif spc[action] == 1: # ate apple\n\t\treward = 500;\n\telif spc_dist[action] < dist: # closer to apple\n\t\treward = 20;\n\telse: # anything else\n\t\treward = -10;\n\n\t## update q of current state\n\tif episode_ended:\n\t\tQ[state][action] += Q_alpha * (reward - Q[state][action]);\n\telse:\n\t\tstate_next = Q_get_state(fake_dirs)[\"state\"];\n\t\tif state_next not in Q:\n\t\t\tQ[state_next] = [10,10,10];\n\t\t\tQ[state][action] += Q_alpha * (reward + Q_gamma * max(Q[state_next]) - Q[state][action]);\n\t\n\treturn fake_dirs;\n\n## 'ai' which lets the user play, basically identity function\ndef ai_user():\n\treturn dirs;\n\n### set environment ###\nDEBUG = False;\nUSER = False;\nTRAIN = False;\nALGO = ai_weighted_rand_walk;\nALGO_PARAMS = [1/3,1/3,1/3]; # completley random walk\nFPS = 0;\n\nfor i in range(1,len(sys.argv)):\n\tcurr = sys.argv[i];\n\tif curr == \"-d\":\n\t\tDEBUG = True;\n\telif curr == \"-r\":\n\t\tALGO = ai_weighted_rand_walk;\n\t\tALGO_PARAMS = [[1/3,1/3,1/3], False];\n\telif curr == \"-wr\":\n\t\tALGO = ai_weighted_rand_walk;\n\t\tALGO_PARAMS = [[.8,.1,.1], False];\n\telif curr == '-sr':\n\t\tALGO = ai_weighted_rand_walk;\n\t\tALGO_PARAMS = [[1/3,1/3,1/3], True];\n\telif curr == '-swr':\n\t\tALGO = ai_weighted_rand_walk;\n\t\tALGO_PARAMS = [[.8,.1,.1], True];\n\telif curr == \"-s\":\n\t\tALGO = ai_smart;\n\t\tALGO_PARAMS = None;\n\telif curr == \"-q\":\n\t\tALGO = ai_qlearn;\n\t\tALGO_PARAMS = None;\n\telif curr == \"-usr\":\n\t\tALGO = ai_user;\n\t\tALGO_PARAMS = None;\n\t\tUSER = True;\n\telif curr == \"-t\":\n\t\tTRAIN = True;\n\telif curr.isdigit():\n\t\tFPS = float(curr);\n\telse:\n\t\traise Exception(\"Unrecognized parameter \" + curr)\n\n### set up game variables ###\n\nobj_size = 10;\ngrid_size = 30;\ngrid_orig = np.zeros((grid_size,grid_size));\n\nsnake_head_start = (np.floor(grid_size/2)*obj_size,np.floor(grid_size/2)*obj_size);\nsnake_init_length = 5;\n\nxs_orig = np.array([snake_head_start[0]]);\nys_orig = np.array([snake_head_start[1]]);\n\ngrid_orig[xs_orig[0]/obj_size][ys_orig[0]/obj_size] = 1;\n\nfor i in range(1,snake_init_length):\n\txs_orig = np.append(xs_orig, snake_head_start[0]);\n\tys_orig = np.append(ys_orig, snake_head_start[1] - i * obj_size);\n\tgrid_orig[xs_orig[0]/obj_size][ys_orig[i-1]/obj_size] = 1;\n\nxs = np.array([]);\nys = np.array([]);\ngrid = np.zeros((grid_size,grid_size));\n\ndirs = 0;\nscore = 0;\napplepos = (0,0);\n\n### set up non debug environment ###\n\nif not DEBUG:\n\tpygame.init();\n\ts=pygame.display.set_mode((grid_size*obj_size, grid_size*obj_size));\n\tpygame.display.set_caption('SnakeAI');\n\tappleimage = pygame.Surface((obj_size, obj_size));\n\tappleimage.fill((0, 255, 0));\n\timg_body = pygame.Surface((obj_size, obj_size));\n\timg_body.fill((255, 0, 0));\n\timg_head = pygame.Surface((obj_size, obj_size));\n\timg_head.fill((125, 0, 0));\n\tf = pygame.font.SysFont('Arial', 20);\n\tclock = pygame.time.Clock()\n\n### run game\ninit_game();\n","sub_path":"Machine Learning/Final Project - Snake AI (Python)/code/snake_old/snake.py","file_name":"snake.py","file_ext":"py","file_size_in_byte":11485,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"370892015","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Dec 13 19:02:40 2011\n\n@author: Dai\n\"\"\"\n\nimport pylab as plt\nimport mpl_toolkits.mplot3d.axes3d as p3\nimport numpy as np\nimport time\nimport pickle\n\n#from instruments import AWG81180A\nfrom slab.instruments import InstrumentManager\nfrom .spectrum_analyzer import *\nfrom .sa_calibration_manager import *\n\nfrom .single_sideband_mod import *\n\n\ndef int2amp(intensity):\n \"\"\"convert intensity (in dBm) to amplitude (or more precisely sqrt(intensity))\"\"\"\n return 10.0**(intensity/20.0)\n\nif __name__ == '__main__':\n #awg = AWG81180A(name='awg', address='GPIB::04::INSTR')\n im = InstrumentManager(\"ssm_optimization.cfg\" ) \n awg = im['AWG']\n #sa = SpectrumAnalyzer(protocol='serial', port=2)\n #Agilent analog waveform generator \n #sacm = SACalibrationManager(pickle.load(open('10dBm_cali.data')))\n \n sa = im['SA']\n rf = im['RF'] \n lo = im['LO']\n sacm = SACalibrationManager(pickle.load(open('10dBm_cali.data')))\n \n \"\"\"arbitrary setup parameters\"\"\"\n c_freq = 7e9\n m_freq = 0.1e9\n # amplitude in mV\n m_amp = 200.0\n \n \"\"\"setup analog waveform generators\"\"\"\n rf.set_frequency(c_freq)\n # +13dBm LO to drive IQB0618\n rf.set_power(13)\n lo.set_frequency(c_freq+sa.lo_offset)\n lo.set_power(10)\n \n \"\"\"awg global setup\"\"\"\n \"\"\"synchronize the two channels\"\"\"\n awg.write('inst:coup:stat on')\n \"\"\"turn on both channels\"\"\"\n set_output(awg, 1)\n set_output(awg, 2)\n\n \"\"\"sweep DC offsets and collect carrier amplitudes\"\"\"\n# s_step = 0.001\n# s_start = -0.5\n# s_end = 0.5\n# \n# d1s, d2s = np.meshgrid(np.arange(s_start, s_end, s_step), np.arange(s_start, s_end, s_step))\n# \n# t1 = []\n# cs = []\n# for d2 in d2s:\n# set_sin(awg, 2, m_amp, m_freq, d2[0])\n# for d1 in d1s[0]:\n# set_sin(awg, 1, m_amp, m_freq, d1, 90.0)\n# time.sleep(0.003)\n# \"\"\"carrier frequency\"\"\"\n# try:\n# c = int2amp(sacm.get_rf_power(c_freq-m_freq, get_power_at(sa, lo, c_freq-m_freq)))\n# except OutputOutOfRangeError as e:\n# c = int2amp(e.lower_bound_pwr)\n# \n# t1.append(c)\n# print 'd1, d2, c = '+repr((d1, d2[0], c))\n# cs.append(t1)\n# t1 = []\n# \n \n \"\"\"sweep phase and collect sidebands amplitudes\"\"\" \n s_step = 0.1\n s_start = -20.0\n s_end = 20.0\n\n ps = np.arange(s_start, s_end, s_step)\n\n plsbs = []\n pusbs = []\n i = 0\n for p in ps:\n set_sin(awg, 2, m_amp, m_freq, 0.0, 0.0)\n set_sin(awg, 1, m_amp, m_freq, 0.0, 90.0+p)\n time.sleep(0.1)\n \"\"\"lower side band\"\"\"\n #print 'raw # = '+repr(get_power_at(sa, lo, c_freq-m_freq))\n try:\n plsbs.append(sacm.get_rf_power(c_freq-m_freq, \n get_power_at(sa, lo, c_freq-m_freq)))\n except OutputOutOfRangeError as e:\n plsbs.append(e.lower_bound_pwr)\n\n \"\"\"upper side band\"\"\"\n #print 'raw # = '+repr(get_power_at(sa, lo, c_freq+m_freq))\n try:\n pusbs.append(sacm.get_rf_power(c_freq+m_freq, \n get_power_at(sa, lo, c_freq+m_freq)))\n except OutputOutOfRangeError as e:\n pusbs.append(e.lower_bound_pwr)\n \n print('p, lsb, usb = '+repr((p, plsbs[i], pusbs[i])))\n i += 1\n\n# \"\"\"sweep I, Q and collect sidebands amplitudes\"\"\" \n# s_step = 5.0\n# s_start = 50.0\n# s_end = 200.0\n#\n# iis, qqs = np.meshgrid(np.arange(s_start, s_end, s_step), np.arange(s_start, s_end, s_step))\n#\n# t1 = []\n# t2 = []\n# lsbs = []\n# usbs = []\n# for q in qqs:\n# set_sin(awg, 2, q[0], m_freq, 0.0, 0.0)\n# for i in iis[0]:\n# set_sin(awg, 1, i, m_freq, 0.0, 90.0)\n# time.sleep(0.01)\n# \"\"\"lower side band\"\"\"\n# #print 'raw # = '+repr(get_power_at(sa, lo, c_freq-m_freq))\n# try:\n# lsb = int2amp(sacm.get_rf_power(c_freq-m_freq, \n# get_power_at(sa, lo, c_freq-m_freq)))\n# except OutputOutOfRangeError as e:\n# lsb = int2amp(e.lower_bound_pwr)\n# t1.append(lsb)\n# \n# \"\"\"upper side band\"\"\"\n# #print 'raw # = '+repr(get_power_at(sa, lo, c_freq+m_freq))\n# try:\n# usb = int2amp(sacm.get_rf_power(c_freq+m_freq, \n# get_power_at(sa, lo, c_freq+m_freq)))\n# except OutputOutOfRangeError as e:\n# usb = int2amp(e.lower_bound_pwr)\n# t2.append(usb) \n# \n# print 'i, q, lsb, usb = '+repr((i, q[0], lsb, usb))\n# \n# lsbs.append(t1)\n# usbs.append(t2)\n# t1 = []\n# t2 = [] \n\n\n \"\"\"numpified for plotting\"\"\"\n# lsbs = np.array(lsbs) \n# usbs = np.array(usbs)\n# cs = np.array(cs)\n plsbs = np.array(plsbs)\n pusbs = np.array(pusbs)\n \n \"\"\"plots\"\"\" \n \"\"\"carrier amplitude over offsets\"\"\"\n# fig = plt.figure()\n# ax = p3.Axes3D(fig)\n# ax.set_xlabel('Q offset (V)')\n# ax.set_ylabel('I offset (V)')\n# ax.set_zlabel('Carrier Amplitude')\n# #ax.scatter(d1, d2, cs)\n# ax.plot_wireframe(d1s, d2s, cs)\n# plt.show()\n \n# \"\"\"lsb over IQ\"\"\"\n# fig = plt.figure()\n# ax = p3.Axes3D(fig)\n# ax.set_xlabel('Q amplitude')\n# ax.set_ylabel('I amplitude')\n# ax.set_zlabel('LSB Amplitude')\n# #ax.scatter(d1, d2, cs)\n# ax.plot_wireframe(iis, qqs, lsbs)\n# plt.show()\n# \n# \"\"\"usb over IQ\"\"\"\n# fig = plt.figure()\n# ax = p3.Axes3D(fig)\n# ax.set_xlabel('Q amplitude')\n# ax.set_ylabel('I amplitude')\n# ax.set_zlabel('USB Amplitude')\n# #ax.scatter(d1, d2, cs)\n# ax.plot_wireframe(iis, qqs, usbs)\n# plt.show()\n\n \"\"\"lsb over phase\"\"\"\n# ax.set_xlabel('phase (deg)')\n# ax.set_ylabel('LSB amplitude')\n plt.plot(ps, plsbs)\n plt.show()\n \n \"\"\"usb over phase\"\"\"\n# ax.set_xlabel('phase (deg)')\n# ax.set_ylabel('LSB amplitude')\n plt.plot(ps, pusbs)\n plt.show()","sub_path":"slab/instruments/spec_analyzer/mixer_std_calibration.py","file_name":"mixer_std_calibration.py","file_ext":"py","file_size_in_byte":6160,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"283553922","text":"import asyncio\nimport contextlib\nimport socket\n\nimport pytest\n\nimport asyncio_dgram\n\n\n@contextlib.contextmanager\ndef loop_exception_handler():\n \"\"\"\n Replace the current event loop exception handler with one that\n simply stores exceptions in the returned dictionary.\n\n @return - dictionary that is updated with the last loop exception\n \"\"\"\n context = {}\n\n def handler(self, c):\n context.update(c)\n\n loop = asyncio.get_event_loop()\n orig_handler = loop.get_exception_handler()\n loop.set_exception_handler(handler)\n yield context\n\n loop.set_exception_handler(orig_handler)\n\n\n@pytest.mark.asyncio\n@pytest.mark.parametrize(\n \"addr,family\",\n [((\"127.0.0.1\", 0), socket.AF_INET), ((\"::1\", 0), socket.AF_INET6)],\n ids=[\"INET\", \"INET6\"],\n)\nasync def test_connect_sync(addr, family):\n # Bind a regular socket, asyncio_dgram connect, then check asyncio send and\n # receive.\n with socket.socket(family, socket.SOCK_DGRAM) as sock:\n sock.bind(addr)\n client = await asyncio_dgram.connect(sock.getsockname()[:2])\n\n assert client.peername == sock.getsockname()\n\n await client.send(b\"hi\")\n got, client_addr = sock.recvfrom(4)\n assert got == b\"hi\"\n assert client_addr == client.sockname\n assert client.peername == sock.getsockname()\n\n sock.sendto(b\"bye\", client.sockname)\n got, server_addr = await client.recv()\n assert got == b\"bye\"\n assert server_addr == sock.getsockname()\n\n with pytest.raises(asyncio.TimeoutError):\n await asyncio.wait_for(client.recv(), 0.05)\n\n # Same as above but reversing the flow. Bind a regular socket, asyncio_dgram\n # connect, then check asyncio receive and send.\n with socket.socket(family, socket.SOCK_DGRAM) as sock:\n sock.bind(addr)\n client = await asyncio_dgram.connect(sock.getsockname()[:2])\n\n with pytest.raises(asyncio.TimeoutError):\n await asyncio.wait_for(client.recv(), 0.05)\n\n assert client.peername == sock.getsockname()\n\n sock.sendto(b\"hi\", client.sockname)\n got, server_addr = await client.recv()\n assert got == b\"hi\"\n assert server_addr == sock.getsockname()\n\n await client.send(b\"bye\")\n got, client_addr = sock.recvfrom(4)\n assert got == b\"bye\"\n assert client_addr == client.sockname\n\n\n@pytest.mark.asyncio\n@pytest.mark.parametrize(\n \"addr,family\",\n [((\"127.0.0.1\", 0), socket.AF_INET), ((\"::1\", 0), socket.AF_INET6)],\n ids=[\"INET\", \"INET6\"],\n)\nasync def test_bind_sync(addr, family):\n # Bind an asyncio_dgram, regular socket connect, then check asyncio send and\n # receive.\n with socket.socket(family, socket.SOCK_DGRAM) as sock:\n server = await asyncio_dgram.bind(addr)\n sock.connect(server.sockname)\n\n assert server.peername is None\n\n await server.send(b\"hi\", sock.getsockname())\n got, server_addr = sock.recvfrom(4)\n assert got == b\"hi\"\n assert server_addr == server.sockname\n\n sock.sendto(b\"bye\", server.sockname)\n got, client_addr = await server.recv()\n assert got == b\"bye\"\n assert client_addr == sock.getsockname()\n\n with pytest.raises(asyncio.TimeoutError):\n await asyncio.wait_for(server.recv(), 0.05)\n\n # Same as above but reversing the flow. Bind an asyncio_dgram, regular\n # socket connect, then check asyncio receive and send.\n with socket.socket(family, socket.SOCK_DGRAM) as sock:\n server = await asyncio_dgram.bind(addr)\n sock.connect(server.sockname)\n\n assert server.peername is None\n\n with pytest.raises(asyncio.TimeoutError):\n await asyncio.wait_for(server.recv(), 0.05)\n\n sock.sendto(b\"hi\", server.sockname)\n got, client_addr = await server.recv()\n assert got == b\"hi\"\n assert client_addr == sock.getsockname()\n\n await server.send(b\"bye\", sock.getsockname())\n got, server_addr = sock.recvfrom(4)\n assert got == b\"bye\"\n assert server_addr == server.sockname\n\n\n@pytest.mark.asyncio\n@pytest.mark.parametrize(\n \"addr,family\",\n [((\"127.0.0.1\", 0), socket.AF_INET), ((\"::1\", 0), socket.AF_INET6)],\n ids=[\"INET\", \"INET6\"],\n)\nasync def test_from_socket_streamtype(addr, family):\n with socket.socket(family, socket.SOCK_DGRAM) as sock:\n sock.bind(addr)\n stream = await asyncio_dgram.from_socket(sock)\n\n assert stream.sockname is not None\n assert sock.getsockname() == stream.sockname\n assert stream.peername is None\n assert stream.socket.fileno() == sock.fileno()\n assert isinstance(stream, asyncio_dgram.aio.DatagramServer)\n\n with socket.socket(family, socket.SOCK_DGRAM) as sock:\n sock.bind(addr)\n\n with socket.socket(family, socket.SOCK_DGRAM) as tsock:\n tsock.connect(sock.getsockname())\n stream = await asyncio_dgram.from_socket(tsock)\n\n assert stream.sockname is not None\n assert tsock.getsockname() == stream.sockname\n assert isinstance(stream, asyncio_dgram.aio.DatagramClient)\n assert stream.peername == sock.getsockname()\n assert stream.socket.fileno() == tsock.fileno()\n\n # Make sure that the transport stored the peername\n with loop_exception_handler() as context:\n await stream.send(b\"abc\")\n assert context == {}\n\n\n@pytest.mark.asyncio\nasync def test_from_socket_bad_socket():\n with socket.socket(socket.AF_UNIX, socket.SOCK_DGRAM) as sock:\n with pytest.raises(TypeError, match=\"either AddressFamily.AF\"):\n await asyncio_dgram.from_socket(sock)\n\n with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as sock:\n with pytest.raises(TypeError, match=\"must be SocketKind.SOCK_DGRAM\"):\n await asyncio_dgram.from_socket(sock)\n\n\n@pytest.mark.asyncio\n@pytest.mark.parametrize(\n \"addr,family\",\n [((\"127.0.0.1\", 0), socket.AF_INET), ((\"::1\", 0), socket.AF_INET6)],\n ids=[\"INET\", \"INET6\"],\n)\nasync def test_no_server(addr, family):\n with socket.socket(family, socket.SOCK_DGRAM) as sock:\n sock.bind(addr)\n free_addr = sock.getsockname()\n\n client = await asyncio_dgram.connect(free_addr[:2])\n await client.send(b\"hi\")\n\n for _ in range(20):\n try:\n await client.send(b\"hi\")\n except ConnectionRefusedError:\n break\n await asyncio.sleep(0.01)\n else:\n pytest.fail(\"ConnectionRefusedError not raised\")\n\n assert client.peername == free_addr\n client.close()\n\n\n@pytest.mark.asyncio\n@pytest.mark.parametrize(\"addr\", [(\"127.0.0.1\", 0), (\"::1\", 0)], ids=[\"INET\", \"INET6\"])\nasync def test_echo(addr):\n server = await asyncio_dgram.bind(addr)\n client = await asyncio_dgram.connect(server.sockname[:2])\n\n await client.send(b\"hi\")\n data, client_addr = await server.recv()\n assert data == b\"hi\"\n assert client_addr == client.sockname\n\n await server.send(b\"bye\", client_addr)\n data, server_addr = await client.recv()\n assert data == b\"bye\"\n assert server_addr == server.sockname\n\n assert server.peername is None\n assert client.peername == server.sockname\n\n\n@pytest.mark.asyncio\n@pytest.mark.parametrize(\"addr\", [(\"127.0.0.1\", 0), (\"::1\", 0)], ids=[\"INET\", \"INET6\"])\nasync def test_echo_bind(addr):\n server = await asyncio_dgram.bind(addr)\n client = await asyncio_dgram.bind(addr)\n\n await client.send(b\"hi\", server.sockname)\n data, client_addr = await server.recv()\n assert data == b\"hi\"\n assert client_addr == client.sockname\n\n await server.send(b\"bye\", client_addr)\n data, server_addr = await client.recv()\n assert data == b\"bye\"\n assert server_addr == server.sockname\n\n assert server.peername is None\n assert client.peername is None\n\n\n@pytest.mark.asyncio\n@pytest.mark.parametrize(\"addr\", [(\"127.0.0.1\", 0), (\"::1\", 0)], ids=[\"INET\", \"INET6\"])\nasync def test_unconnected_sender(addr):\n # Bind two endpoints and connect to one. Ensure that only the endpoint\n # that was connected to can send.\n ep1 = await asyncio_dgram.bind(addr)\n ep2 = await asyncio_dgram.bind(addr)\n connected = await asyncio_dgram.connect(ep1.sockname[:2])\n\n await ep1.send(b\"from-ep1\", connected.sockname)\n await ep2.send(b\"from-ep2\", connected.sockname)\n data, server_addr = await connected.recv()\n assert data == b\"from-ep1\"\n assert server_addr == ep1.sockname\n\n with pytest.raises(asyncio.TimeoutError):\n await asyncio.wait_for(connected.recv(), 0.05)\n","sub_path":"test/test_aio.py","file_name":"test_aio.py","file_ext":"py","file_size_in_byte":8593,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"82843608","text":"\"\"\"\nRelationship between array indexes and tree element\nA complete binary tree has an interesting property that we can use to find the children and parents of any node.\n\nIf the index of any element in the array is i, the element in the index 2i+1 will become the left child and element in 2i+2 index will become the right child. Also, the parent of any element at index i is given by the lower bound of (i-1)/2.\"\"\"\nclass Node:\n\n def __init__(self, key):\n self.key = key\n self.left = None\n self.right = None\n\n\ndef countNodes(root):\n if root is None:\n return 0\n return (1 + countNodes(root.left) + countNodes(root.right))\n\n\ndef isComplete(root, index, numberNodes):\n\n if root is None:\n return True\n\n if index >= numberNodes:\n return False\n\n return (isComplete(root.left, 2 * index + 1, numberNodes)\n and isComplete(root.right, 2 * index + 2, numberNodes))\n\n\nroot = Node(1)\nroot.left = Node(2)\nroot.right = Node(3)\nroot.left.left = Node(4)\nroot.left.right = Node(5)\nroot.right.left = Node(6)\n\nnode_count = countNodes(root)\nindex = 0\n\nif isComplete(root, index, node_count):\n print(\"The tree is a complete binary tree\")\nelse:\n print(\"The tree is not a complete binary tree\")\n","sub_path":"Trees/complete binary tree1.py","file_name":"complete binary tree1.py","file_ext":"py","file_size_in_byte":1248,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"41582687","text":"import ply.lex as lex\nimport ply.yacc as yacc\nimport sys\n\n# Create a list to hold all of the token names\ntokens = [\n 'BOOL',\n 'NAME',\n 'LPAREN',\n 'RPAREN',\n 'NEG',\n 'AND',\n 'OR',\n 'COND',\n 'BICOND',\n 'EQUALS'\n]\n\n# Use regular expressions to define what each token is\nt_LPAREN = r'\\('\nt_RPAREN = r'\\)'\nt_NEG = r'\\~'\nt_AND = r'\\&'\nt_OR = r'\\|'\nt_COND = r'\\->'\nt_BICOND = r'\\<->'\nt_EQUALS = r'\\='\n\n# Ply's special t_ignore variable allows us to define characters the lexer will ignore.\n# We're ignoring spaces.\nt_ignore = r' '\n\n# More complicated tokens, such as tokens that are more than 1 character in length\n# are defined using functions.\n\n# An int is 1 or more numbers.\ndef t_BOOL(t):\n r'(F|T)'\n if t.value == 'F':\n t.value = False\n else:\n t.value = True\n return t\n\n# A NAME is a variable name. A variable can be 1 or more characters in length.\n# The first character must be in the ranges a-z A-Z or be an underscore.\n# Any character following the first character can be a-z A-Z 0-9 or an underscore.\ndef t_NAME(t):\n r'(?!(F|T))[a-zA-Z_][a-zA-Z_0-9]*'\n t.type = 'NAME'\n return t\n\n# Skip the current token and output 'Illegal characters' using the special Ply t_error function.\ndef t_error(t):\n #print(\"Illegal character '%s'\" % t.value[0])\n t.lexer.skip(1)\n raise CalcError(\"Illegal character '%s'\" % t.value[0])\n\n# Build the lexer\nlexer = lex.lex()\n\n# Ensure our parser understands the correct order of operations.\n# The precedence variable is a special Ply variable.\nprecedence = (\n ('left', 'BICOND'),\n ('left', 'COND'),\n ('left', 'OR'),\n ('left', 'AND'),\n ('left', 'NEG'),\n ('left', 'LPAREN','RPAREN')\n)\n\n# Convert a boolean element to their char representation\ndef boolToChar(boolean):\n if boolean == True:\n return 'T'\n else:\n return 'F'\n\n# Define our grammar. We allow expressions, var_assign's and empty's.\ndef p_calc(p):\n '''\n calc : expression\n | var_assign\n | empty\n '''\n print(boolToChar(run(p[1])))\n\ndef p_var_assign(p):\n '''\n var_assign : NAME EQUALS expression\n '''\n # Build our tree\n p[0] = ('=', p[1], p[3])\n\ndef p_expression_negation(p):\n '''\n expression : NEG expression\n '''\n p[0] = not p[2]\n\ndef p_expression_parentesis(p):\n '''\n expression : LPAREN expression RPAREN\n '''\n p[0] = p[2]\n\n# Expressions are recursive.\ndef p_expression(p):\n '''\n expression : expression BICOND expression\n | expression COND expression\n | expression OR expression\n | expression AND expression\n '''\n # Build our tree.\n p[0] = (p[2], p[1], p[3])\n\ndef p_expression_bool(p):\n '''\n expression : BOOL\n '''\n p[0] = p[1]\n\ndef p_expression_var(p):\n '''\n expression : NAME\n '''\n p[0] = ('var', p[1])\n\n# Output to the user that there is an error in the input as it doesn't conform to our grammar.\n# p_error is another special Ply function.\ndef p_error(p):\n #print(\"Syntax error found!\")\n raise CalcError(\"Syntax error found!\")\n\ndef p_empty(p):\n '''\n empty :\n '''\n p[0] = None\n\n# Build the parser\nparser = yacc.yacc(debug=0, write_tables=0)\n# Create the environment upon which we will store and retreive variables from.\nenv = {}\n# The run function is our recursive function that 'walks' the tree generated by our parser.\ndef run(p):\n global env\n if type(p) == tuple:\n if p[0] == '&':\n return run(p[1]) and run(p[2])\n elif p[0] == '|':\n return run(p[1]) or run(p[2])\n elif p[0] == '->':\n return not run(p[1]) or run(p[2])\n elif p[0] == '<->':\n return (not run(p[1]) or run(p[2])) and (not run(p[2]) or run(p[1]))\n elif p[0] == '=':\n env[p[1]] = run(p[2])\n return ''\n elif p[0] == 'var':\n if p[1] not in env:\n return 'Undeclared variable found!'\n else:\n return env[p[1]]\n else:\n return p\n\n# Create a REPL to provide a way to interface with our interpreter.\n'''\nwhile True:\n try:\n s = input('>> ')\n \n except EOFError:\n break\n parser.parse(s)\n'''\n\nclass CalcError(Exception):\n def __init__(self, message):\n self.message = message\n\ntry:\n while True:\n s = input('>>> ')\n try:\n answer = parser.parse(s)\n if answer is not None:\n print(answer)\n except CalcError as e:\n print(e.message)\nexcept EOFError:\n print(\"XD\")","sub_path":"prologic.py","file_name":"prologic.py","file_ext":"py","file_size_in_byte":4563,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"512773401","text":"# -*- coding: utf-8 -*-\n\"\"\"\n pyrseas.dbobject.rule\n ~~~~~~~~~~~~~~~~~~~~~\n\n This defines two classes, Rule and RuleDict, derived from\n DbSchemaObject and DbObjectDict, respectively.\n\"\"\"\nfrom pyrseas.dbobject import DbObjectDict, DbSchemaObject\nfrom pyrseas.dbobject import quote_id, commentable\n\n\nclass Rule(DbSchemaObject):\n \"\"\"A rewrite rule definition\"\"\"\n\n keylist = ['schema', 'table', 'name']\n objtype = \"RULE\"\n\n def identifier(self):\n \"\"\"Return a full identifier for a rule object\n\n :return: string\n \"\"\"\n return \"%s ON %s\" % (quote_id(self.name), self._table.qualname())\n\n def to_map(self):\n \"\"\"Convert rule to a YAML-suitable format\n\n :return: dictionary\n \"\"\"\n dct = self._base_map()\n del dct['_table']\n return {self.name: dct}\n\n @commentable\n def create(self):\n \"\"\"Return SQL statements to CREATE the rule\n\n :return: SQL statements\n \"\"\"\n where = instead = ''\n if hasattr(self, 'condition'):\n where = ' WHERE %s' % self.condition\n if hasattr(self, 'instead'):\n instead = 'INSTEAD '\n return [\"CREATE RULE %s AS ON %s\\n TO %s%s\\n DO %s%s\" % (\n quote_id(self.name), self.event.upper(),\n self._table.qualname(), where, instead, self.actions)]\n\n\nclass RuleDict(DbObjectDict):\n \"The collection of rewrite rules in a database.\"\n\n cls = Rule\n query = \\\n \"\"\"SELECT nspname AS schema, relname AS table, rulename AS name,\n split_part('select,update,insert,delete', ',',\n ev_type::int - 48) AS event, is_instead AS instead,\n pg_get_ruledef(r.oid) AS definition,\n obj_description(r.oid, 'pg_rewrite') AS description\n FROM pg_rewrite r JOIN pg_class c ON (ev_class = c.oid)\n JOIN pg_namespace n ON (relnamespace = n.oid)\n WHERE relkind = 'r'\n AND (nspname != 'pg_catalog' AND nspname != 'information_schema')\n ORDER BY nspname, relname, rulename\"\"\"\n\n def _from_catalog(self):\n \"\"\"Initialize the dictionary of rules by querying the catalogs\"\"\"\n for rule in self.fetch():\n do_loc = rule.definition.index(' DO ')\n if 'WHERE' in rule.definition:\n rule.condition = rule.definition[rule.definition.index(\n ' WHERE ') + 7:do_loc]\n if hasattr(rule, 'instead') and rule.instead:\n do_loc += 8\n rule.actions = rule.definition[do_loc + 4:-1]\n del rule.definition\n self[rule.key()] = rule\n\n def from_map(self, table, inmap):\n \"\"\"Initialize the dictionary of rules by examining the input map\n\n :param inmap: the input YAML map defining the rules\n \"\"\"\n for rul in inmap:\n inrule = inmap[rul]\n rule = Rule(table=table.name, schema=table.schema, name=rul,\n **inrule)\n if inrule:\n if 'oldname' in inrule:\n rule.oldname = inrule['oldname']\n del inrule['oldname']\n if 'description' in inrule:\n rule.description = inrule['description']\n self[(table.schema, table.name, rul)] = rule\n\n def diff_map(self, inrules):\n \"\"\"Generate SQL to transform existing rules\n\n :param input_map: a YAML map defining the new rules\n :return: list of SQL statements\n\n Compares the existing rule definitions, as fetched from the\n catalogs, to the input map and generates SQL statements to\n transform the rules accordingly.\n \"\"\"\n stmts = []\n # check input rules\n for rul in inrules:\n inrul = inrules[rul]\n # does it exist in the database?\n if rul in self:\n stmts.append(self[rul].diff_map(inrul))\n else:\n # check for possible RENAME\n if hasattr(inrul, 'oldname'):\n oldname = inrul.oldname\n try:\n stmts.append(self[oldname].rename(inrul.name))\n del self[oldname]\n except KeyError as exc:\n exc.args = (\"Previous name '%s' for rule '%s' \"\n \"not found\" % (oldname, inrul.name), )\n raise\n else:\n # create new rule\n stmts.append(inrul.create())\n # check database rules\n for (sch, tbl, rul) in self:\n # if missing, drop it\n if (sch, tbl, rul) not in inrules:\n stmts.append(self[(sch, tbl, rul)].drop())\n\n return stmts\n","sub_path":"pyrseas/dbobject/rule.py","file_name":"rule.py","file_ext":"py","file_size_in_byte":4789,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"592807597","text":"import json\nimport requests\n\nclass RestAgent:\n\n def __init__(self, url, key):\n self.__session = requests.Session()\n self.base_url = url\n self.client_list = {}\n self.session = requests.Session()\n self.api_secret = key\n\n def send_request(self, request, param=None, method='POST', private=False):\n url = self.base_url + request\n headers = {}\n if private:\n headers = {\n 'Content-Type': \"application/json\",\n 'Authorization': 'Bearer {}'.format(self.api_secret)\n }\n\n try:\n if method == 'GET':\n response = self.session.get(url, headers=headers)\n elif method == 'DELETE':\n response = self.session.delete(url, headers=headers)\n elif method == 'POST' and param is not None:\n response = self.session.post(url, headers=headers, data=json.dumps(param))\n elif method == 'POST' and param is None:\n response = self.session.post(url, headers=headers)\n return response.json()\n except IndexError as e:\n print(e)\n\n\nif __name__ == \"__main__\":\n settings = {\n 'key': '',\n 'secret': ''\n }\n rest_wrapper = RestAgent('https://api.exchange.ripio.com/api/v1', 'a963ae2fccf59bbaae607b1a65b3ca2d3305378b2dc59a0659a02b3b675a6513')\n print('Depth test:')\n resp = rest_wrapper.send_request(request='/orderbook/btc_usdc/', param=None, method='GET')\n print(resp)\n\n print('Ticker test:')\n resp = rest_wrapper.send_request(request='/rate/btc_usdc', param=None, method='GET')\n print(resp)\n\n print('Pairs test:')\n resp = rest_wrapper.send_request(request='/pair/', param=None, method='GET')\n print(resp)\n\n print('Balance test:')\n resp = rest_wrapper.send_request(request='/balances/exchange_balances/', param=None, method='GET', private=True)\n print(resp)\n\n while True:\n pass","sub_path":"test/ripio_tests/test_agent.py","file_name":"test_agent.py","file_ext":"py","file_size_in_byte":1958,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"84019066","text":"from __future__ import unicode_literals, division, print_function\n\nimport numpy as np\nimport re\nimport os\nimport csv\n\nfrom vgg16 import Vgg16\nfrom utils import save_array, load_array, get_data, get_batches, onehot\nfrom toolbox import extract_images_and_classes, MODELS_PATH\nfrom keras.models import Sequential\nfrom keras.layers import Input\nfrom keras.layers.core import Flatten, Dense, Dropout, Lambda\nfrom keras.optimizers import SGD, RMSprop\nfrom keras import backend as K\n\n\ndef write_submission_csv(data, columns):\n \"\"\"\n Write data according to the Kaggle submission format.\n \"\"\"\n with open('submission.csv', 'wb') as f:\n w = csv.writer(f)\n w.writerow(columns)\n for key in data.keys():\n w.writerow([key[8:]] + data[key])\n\n\ndef main():\n path = 'sample/' # In order to use the big dataset set it to empty string.\n batch_size = 1 # If you are using a GPU, please change it to 64\n # We are going to use the 1000 categories from the IMAGENET model,\n # so we need to re-use an instance of it.\n vgg = Vgg16()\n # First, we check if we have saved arrays with the data and classes.\n train_data, val_data, batches_classes, val_classes = extract_images_and_classes(\n path, batch_size=batch_size\n )\n\n # We need the labels in \"hot encoding form\". e.g [1, 0] or [0, 1]\n train_labels = onehot(batches_classes)\n val_labels = onehot(val_classes)\n\n # It's always good to see an overview of the whole model.\n vgg.model.summary()\n # Now we need to create the model.\n # --------------------------------------------------------\n # 0) Replace and train the last Dense layer as it doesn't really\n # fit our purposes.\n print('Let\\'s remove the last layer as it\\'s already DENSE...')\n vgg.model.pop()\n # Here we say that we only want to train the last layer\n for layer in vgg.model.layers: layer.trainable = False\n print('And train a new layer that outputs 10 classes (c0..c9)')\n vgg.model.add(Dense(10, activation='softmax'))\n opt = RMSprop(lr=0.1)\n vgg.model.compile(optimizer=opt,\n loss='categorical_crossentropy',\n metrics=['accuracy'])\n\n vgg.model.fit(train_data, train_labels, nb_epoch=1,\n validation_data=(val_data, val_labels), batch_size=batch_size)\n # --------------------------------------------------------\n # 1) Backpropagate and train the previous layers.\n print('Now let\\'s train the previous layers using backpropagation')\n layers = vgg.model.layers\n # Get the index of the first dense layer...\n first_dense_idx = [index for index, layer in enumerate(layers) if type(layer) is Dense][0]\n # ...and set this and all subsequent layers to trainable\n for layer in layers[first_dense_idx:]: layer.trainable = True\n K.set_value(opt.lr, 0.01)\n vgg.model.fit(train_data, train_labels, nb_epoch=1,\n validation_data=(val_data, val_labels), batch_size=batch_size)\n\n print('Saving weights')\n vgg.model.save_weights(MODELS_PATH + 'weights.bc')\n #vgg.model.load_weights(MODEL_PATH + 'weights.bc')\n\n test_batches, predictions = vgg.test(path+'test', batch_size=batch_size)\n\n d = {}\n for idx, filename in enumerate(test_batches.filenames):\n # We use a trick to never show 0 or 1, but 0.05 and 0.95.\n # This is required becase log loss penalizes predictions that are confident and wrong.\n d[filename] = [prediction.clip(min=0.05, max=0.95) for prediction in predictions[idx]]\n write_submission_csv(d, ['img', 'c0', 'c1', 'c2', 'c3', 'c4', 'c5', 'c6', 'c7', 'c8', 'c9'])\n\nif __name__ == '__main__':\n main()\n","sub_path":"state-farm-distracted-driver-detection/generate_submission_statefarm.py","file_name":"generate_submission_statefarm.py","file_ext":"py","file_size_in_byte":3661,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"481397604","text":"# ----------------------------------------------------------------\n# Copyright 2016 Cisco Systems\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ------------------------------------------------------------------\n\n\"\"\"test_sanity_rpc.py\nsanity test for netconf\n\"\"\"\nfrom __future__ import absolute_import\n\nimport sys\nimport unittest\n\nfrom ydk.errors import YPYModelError, YPYError, YPYServiceError\nfrom ydk.models.ydktest import ydktest_sanity as ysanity\nfrom ydk.providers import NetconfServiceProvider\nfrom ydk.services import NetconfService, Datastore\n\nfrom test_utils import ParametrizedTestCase\nfrom test_utils import get_device_info\n\n\nclass SanityNetconf(ParametrizedTestCase):\n\n @classmethod\n def setUpClass(cls):\n cls.ncc = NetconfServiceProvider(\n cls.hostname,\n cls.username,\n cls.password,\n cls.port,\n cls.protocol,\n not cls.on_demand,\n cls.common_cache,\n cls.timeout)\n cls.netconf_service = NetconfService()\n\n def setUp(self):\n from ydk.services import CRUDService\n crud = CRUDService()\n runner = ysanity.Runner()\n crud.delete(self.ncc, runner)\n\n def tearDown(self):\n pass\n\n def test_edit_commit_get(self):\n runner = ysanity.Runner()\n runner.one.number = 1\n runner.one.name = 'runner:one:name'\n\n get_filter = ysanity.Runner()\n\n op = self.netconf_service.edit_config(self.ncc, Datastore.candidate, runner)\n self.assertEqual(True, op)\n\n result = self.netconf_service.get_config(self.ncc, Datastore.candidate, get_filter)\n self.assertEqual(runner, result)\n\n op = self.netconf_service.commit(self.ncc)\n self.assertEqual(True, op)\n\n result = self.netconf_service.get(self.ncc, get_filter)\n self.assertEqual(runner, result)\n\n def test_lock_unlock(self):\n op = self.netconf_service.lock(self.ncc, Datastore.running)\n self.assertEqual(True, op)\n\n op = self.netconf_service.unlock(self.ncc, Datastore.running)\n self.assertEqual(True, op)\n\n # Failing - NetconfService glue code needed\n def test_lock_unlock_fail(self):\n op = self.netconf_service.lock(self.ncc, Datastore.candidate)\n self.assertEqual(True, op)\n\n try:\n op = self.netconf_service.unlock(self.ncc, Datastore.running)\n except Exception as e:\n self.assertIsInstance(e, YPYError)\n\n def test_validate(self):\n op = self.netconf_service.validate(self.ncc, source=Datastore.candidate)\n self.assertEqual(True, op)\n\n runner = ysanity.Runner()\n runner.one.number = 1\n runner.one.name = 'runner:one:name'\n op = self.netconf_service.validate(self.ncc, source=runner)\n self.assertEqual(True, op)\n\n def test_validate_fail(self):\n # should have been handled by YDK local validation\n pass\n\n def test_commit_discard(self):\n runner = ysanity.Runner()\n runner.two.number = 2\n runner.two.name = 'runner:two:name'\n get_filter = ysanity.Runner()\n\n op = self.netconf_service.edit_config(self.ncc, Datastore.candidate, runner)\n self.assertEqual(True, op)\n\n op = self.netconf_service.discard_changes(self.ncc)\n self.assertEqual(True, op)\n\n op = self.netconf_service.edit_config(self.ncc, Datastore.candidate, runner)\n self.assertEqual(True, op)\n\n op = self.netconf_service.commit(self.ncc)\n self.assertEqual(True, op)\n\n result = self.netconf_service.get(self.ncc, get_filter)\n self.assertEqual(runner, result)\n\n @unittest.skip('No message id in cancel commit payload')\n def test_confirmed_commit(self):\n runner = ysanity.Runner()\n runner.two.number = 2\n runner.two.name = 'runner:two:name'\n get_filter = ysanity.Runner()\n\n op = self.netconf_service.edit_config(self.ncc, Datastore.candidate, runner)\n self.assertEqual(True, op)\n\n op = self.netconf_service.commit(self.ncc, confirmed=True, confirm_timeout=120)\n self.assertEqual(True, op)\n\n result = self.netconf_service.get(self.ncc, get_filter)\n self.assertEqual(runner, result)\n\n op = self.netconf_service.cancel_commit(self.ncc)\n self.assertEqual(True, op)\n\n def test_copy_config(self):\n op = self.netconf_service.copy_config(self.ncc, Datastore.candidate, Datastore.running)\n self.assertEqual(True, op)\n\n runner = ysanity.Runner()\n runner.two.number = 2\n runner.two.name = 'runner:two:name'\n get_filter = ysanity.Runner()\n\n op = self.netconf_service.edit_config(self.ncc, Datastore.candidate, runner)\n self.assertEqual(True, op)\n\n op = self.netconf_service.copy_config(self.ncc, Datastore.running, Datastore.candidate)\n self.assertEqual(True, op)\n\n result = self.netconf_service.get_config(self.ncc, Datastore.running, get_filter)\n self.assertEqual(runner, result)\n\n runner.two.name = '%smodified' % runner.two.name\n\n op = self.netconf_service.copy_config(self.ncc, Datastore.running, runner)\n self.assertEqual(True, op)\n\n result = self.netconf_service.get_config(self.ncc, Datastore.running, get_filter)\n self.assertEqual(runner, result)\n\n def test_delete_config(self):\n pass\n # startup and candidate cannot be both enabled in ConfD\n # op = self.netconf_service.delete_config(self.ncc, Datastore.startup)\n # self.assertEqual(True, op)\n\n # Error not thrown by TCP client, YPYError is populated instead\n def test_delete_config_fail(self):\n found = False\n try:\n self.netconf_service.delete_config(self.ncc, Datastore.running)\n except (YPYError, YPYModelError):\n found = True\n self.assertEqual(found, True)\n\n # Failing - NetconfService glue code needed\n def test_copy_config_fail(self):\n self.assertRaises(YPYServiceError,\n self.netconf_service.copy_config,\n self.ncc,\n target=123,\n source=456)\n\n # Failing - NetconfService glue code needed\n def test_edit_config_fail(self):\n self.assertRaises(YPYServiceError,\n self.netconf_service.edit_config,\n self.ncc,\n Datastore.startup,\n Datastore.candidate)\n\n # Failing - NetconfService glue code needed\n def test_get_config_fail(self):\n runner = ysanity.Runner()\n self.assertRaises(YPYServiceError,\n self.netconf_service.get_config,\n self.ncc,\n \"invalid-input\",\n runner)\n\n # Failing - NetconfService glue code needed\n def test_lock_fail(self):\n self.assertRaises(YPYServiceError,\n self.netconf_service.lock,\n self.ncc,\n \"invalid-input\")\n\n # Failing - NetconfService glue code needed\n def test_unlock_fail(self):\n self.assertRaises(YPYServiceError,\n self.netconf_service.unlock,\n self.ncc,\n \"invalid-input\")\n\n\nif __name__ == '__main__':\n device, non_demand, common_cache, timeout = get_device_info()\n\n suite = unittest.TestSuite()\n suite.addTest(ParametrizedTestCase.parametrize(\n SanityNetconf,\n device=device,\n non_demand=non_demand,\n common_cache=common_cache,\n timeout=timeout))\n ret = not unittest.TextTestRunner(verbosity=2).run(suite).wasSuccessful()\n sys.exit(ret)\n","sub_path":"sdk/python/core/tests/test_sanity_netconf.py","file_name":"test_sanity_netconf.py","file_ext":"py","file_size_in_byte":8267,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"340534587","text":"import time #pour le chrono\n\n# Definition fonction\ndef f(x):\n return x**2 -10\n\n## Sans recursivité\n\n\n\ndef regula_falsi(f, a, b, epsilon):\n nIter = 0 #nombre ditérations\n\n if f(a) * f(b) > 0:\n return None #TVI pas applicable ici\n u, v = float(a), float(b)\n while (abs(f(u)) > epsilon):\n i = i+1 #incrémentation compteur boucle\n w = u -f(u)*(v-u)/(f(v)-f(u))\n if f(u) * f(w) <= 0: #w remplace v\n v = w\n else: #w remplace u\n u = w\n print(\"Nb d'Itérations':\", nIter)\n return w\n\n\n\n# Lecture de l'intervalle\na = 0\nb = 10\n\n\nepsilon=1e-12\n\n\n# Résultats\nstart_time = time.perf_counter_ns() #début du chrono\nprint(\"La valeur du zéro de cette fonction sur (a,b) est:\",regula_falsi(f,a,b,epsilon))\n\ninterval = time.perf_counter_ns() - start_time #calcul du temps\n\nprint(\"Temps d'execution de la fonction:\", interval*10**(-6))\n\n## Avec recursivité\n\n\ndef regu_rec(f,a, b,prec):\n a, b = float(a), float(b) #convertion bornes en réels\n if (abs(f(a))<=prec): #alors a est racine\n return a -f(a)*(b-a)/(f(b)-f(a))\n else:\n c = a -f(a)*(b-a)/(f(b)-f(a)) #calcul nouvelle borne\n\n if f(a)*f(c) <= 0: #Si le TVI est applicable entre a et c, c remplace b\n return regu_rec(f,a,c,prec)\n else: #sinon c remplace b\n return regu_rec(f,c,b,prec)\n\n\n\n# Résultats\nstart_time = time.perf_counter_ns() #début du chrono\nprint(\"La valeur du zéro de cette fonction sur (a,b) est:\", regu_rec(f,a,b,epsilon))\n\ninterval = time.perf_counter_ns() - start_time #calcul du temps\n\nprint(\"Temps d'execution de la fonction:\", interval*10**(-6))","sub_path":"Grandes méthode d'optimisation/Regula Falsi/regula.py","file_name":"regula.py","file_ext":"py","file_size_in_byte":1629,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"387582885","text":"# read data file\nf = open('d16-input.txt', 'r')\ndata = [line.rstrip() for line in f.readlines()]\n# Test data\n# data = [\"class: 1-3 or 5-7\",\"row: 6-11 or 33-44\",\"seat: 13-40 or 45-50\",\"your ticket:\",\"7,1,14\",\"\",\"nearby tickets:\",\"7,3,47\",\"40,4,50\",\"55,2,20\",\"38,6,12\"]\n\n# Split data into areas - fields, my ticket, nearby tickets\nfields, myTicket, nearbyTickets = {}, [], []\n# split fields into dictionary with class type, min and max values\nfor idx, line in enumerate(data):\n if line == \"\": continue\n if line[:11] == \"your ticket\":\n break\n # split the line into text before the colon to form the key, then the value pairs afer the colon seperated by ' or ' to form value pairs\n fields[line.split(\": \")[0]] = (line.split(\": \")[1].split(\" or \")[0], line.split(\": \")[1].split(\" or \")[1])\n #the min of the min value\n min_min = int(fields[line.split(\": \")[0]][0].split(\"-\")[0])\n # the max of the min value\n min_max = int(fields[line.split(\": \")[0]][0].split(\"-\")[1])\n #the min of the max value\n max_min = int(fields[line.split(\": \")[0]][1].split(\"-\")[0])\n #the max of the max value\n max_max = int(fields[line.split(\": \")[0]][1].split(\"-\")[1])\n # redefine the dictionary entry with new min/max values\n fields[line.split(\": \")[0]] = (min_min, min_max, max_min, max_max)\n# split myTicket data into list\nfor line in data[idx+1:]:\n if line == \"\": continue\n if line[:14] == \"nearby tickets\":\n break\n myTicket.append(line)\n# split nearbyTicket data into list\nfor line in data[idx+4:]:\n if line == \"\": continue\n nearbyTickets.append(line)\n\n# ignore my ticket\n# loop though each field, and compare ticket values to given parameters\n# store sum of invalid parameters\n# can combine all parameters into one set to compare with each line of list\nvalidParams = set() # define empty set\nticketScanningErrorRate = 0\n# add paramteters to set as union of sets of values from fields\nfor keys in fields:\n for idx in range(0,len(fields[keys])-1,2):\n validParams = validParams | set(range(fields[keys][idx],fields[keys][idx+1]+1))\n# loop though nearbyTickets and sum values not in parameters\nfor line in nearbyTickets:\n values = line.split(\",\")\n for value in values:\n ticketScanningErrorRate += int(not(int(value) in validParams)) * int(value)\nprint(\"ticket scanning error rate =\",ticketScanningErrorRate)","sub_path":"d16p1-solution.py","file_name":"d16p1-solution.py","file_ext":"py","file_size_in_byte":2372,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"237020033","text":"from unittest import TestCase\n\nfrom app import app\nfrom models import db, Pet\n\n# Use test database and don't clutter tests with SQL\napp.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql:///pet_shop_test'\napp.config['SQLALCHEMY_ECHO'] = False\n\n# Make Flask errors be real errors, rather than HTML pages with error info\napp.config['TESTING'] = True\n\n# This is a bit of hack, but don't use Flask DebugToolbar\napp.config['DEBUG_TB_HOSTS'] = ['dont-show-debug-toolbar']\n\ndb.drop_all()\ndb.create_all()\n\n\nclass PetViewsTestCase(TestCase):\n \"\"\"Tests for views for Pets.\"\"\"\n\n def setUp(self):\n \"\"\"Add sample pet.\"\"\"\n\n Pet.query.delete()\n\n pet = Pet(name=\"TestPet\", species=\"dog\", hunger=10)\n db.session.add(pet)\n db.session.commit()\n\n self.pet_id = pet.id\n self.pet = pet\n\n def tearDown(self):\n \"\"\"Clean up any fouled transaction.\"\"\"\n\n db.session.rollback()\n\n def test_list_pets(self):\n with app.test_client() as client:\n resp = client.get(\"/\")\n html = resp.get_data(as_text=True)\n\n self.assertEqual(resp.status_code, 200)\n self.assertIn('TestPet', html)\n\n def test_show_pet(self):\n with app.test_client() as client:\n resp = client.get(f\"/{self.pet_id}\")\n html = resp.get_data(as_text=True)\n\n self.assertEqual(resp.status_code, 200)\n self.assertIn('TestPet
', html)\n self.assertIn(self.pet.species, html)\n \n\n def test_add_pet(self):\n with app.test_client() as client:\n d = {\"name\": \"TestPet2\", \"species\": \"cat\", \"hunger\": 20}\n resp = client.post(\"/\", data=d, follow_redirects=True)\n html = resp.get_data(as_text=True)\n\n self.assertEqual(resp.status_code, 200)\n self.assertIn(\"TestPet2
\", html)\n","sub_path":"test_flask.py","file_name":"test_flask.py","file_ext":"py","file_size_in_byte":1866,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"22165861","text":"# Lint as: python2, python3\n# Copyright 2019 Google LLC. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Python source file include Iris pipeline functions and necessary utils.\n\nThe utilities in this file are used to build a model with native Keras.\nThis module file will be used in Transform and generic Trainer.\n\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nfrom typing import Text\n\nimport absl\nimport tensorflow as tf\nfrom tensorflow import keras\nimport tensorflow_transform as tft\n\nfrom tfx.components.trainer.executor import TrainerFnArgs\n\n_FEATURE_KEYS = ['sepal_length', 'sepal_width', 'petal_length', 'petal_width']\n_LABEL_KEY = 'variety'\n\n\ndef _transformed_name(key):\n return key + '_xf'\n\n\ndef _gzip_reader_fn(filenames):\n \"\"\"Small utility returning a record reader that can read gzip'ed files.\"\"\"\n return tf.data.TFRecordDataset(filenames, compression_type='GZIP')\n\n\ndef _get_serve_tf_examples_fn(model, tf_transform_output):\n \"\"\"Returns a function that parses a serialized tf.Example.\"\"\"\n\n @tf.function\n def serve_tf_examples_fn(serialized_tf_examples):\n \"\"\"Returns the output to be used in the serving signature.\"\"\"\n feature_spec = tf_transform_output.raw_feature_spec()\n feature_spec.pop(_LABEL_KEY)\n parsed_features = tf.io.parse_example(serialized_tf_examples, feature_spec)\n\n transformed_features = tf_transform_output.transform_raw_features(\n parsed_features)\n transformed_features.pop(_transformed_name(_LABEL_KEY))\n\n outputs = model(transformed_features)\n return {'outputs': outputs}\n\n return serve_tf_examples_fn\n\n\ndef _input_fn(file_pattern: Text,\n tf_transform_output: tft.TFTransformOutput,\n batch_size: int = 200) -> tf.data.Dataset:\n \"\"\"Generates features and label for tuning/training.\n\n Args:\n file_pattern: input tfrecord file pattern.\n tf_transform_output: A TFTransformOutput.\n batch_size: representing the number of consecutive elements of returned\n dataset to combine in a single batch\n\n Returns:\n A dataset that contains (features, indices) tuple where features is a\n dictionary of Tensors, and indices is a single Tensor of label indices.\n \"\"\"\n transformed_feature_spec = (\n tf_transform_output.transformed_feature_spec().copy())\n\n dataset = tf.data.experimental.make_batched_features_dataset(\n file_pattern=file_pattern,\n batch_size=batch_size,\n features=transformed_feature_spec,\n reader=_gzip_reader_fn,\n label_key=_transformed_name(_LABEL_KEY))\n\n return dataset\n\n\ndef _build_keras_model() -> tf.keras.Model:\n \"\"\"Creates a DNN Keras model for classifying iris data.\n\n Returns:\n A Keras Model.\n \"\"\"\n # The model below is built with Functional API, please refer to\n # https://www.tensorflow.org/guide/keras/overview for all API options.\n inputs = [\n keras.layers.Input(shape=(1,), name=_transformed_name(f))\n for f in _FEATURE_KEYS\n ]\n d = keras.layers.concatenate(inputs)\n for _ in range(3):\n d = keras.layers.Dense(8, activation='relu')(d)\n output = keras.layers.Dense(3, activation='softmax')(d)\n\n model = keras.Model(inputs=inputs, outputs=output)\n model.compile(\n optimizer=keras.optimizers.Adam(lr=0.001),\n loss='sparse_categorical_crossentropy',\n metrics=[keras.metrics.SparseCategoricalAccuracy(name='accuracy')])\n\n absl.logging.info(model.summary())\n return model\n\n\n# TFX Transform will call this function.\ndef preprocessing_fn(inputs):\n \"\"\"tf.transform's callback function for preprocessing inputs.\n\n Args:\n inputs: map from feature keys to raw not-yet-transformed features.\n\n Returns:\n Map from string feature key to transformed feature operations.\n \"\"\"\n outputs = {}\n\n for key in _FEATURE_KEYS:\n outputs[_transformed_name(key)] = tft.scale_to_z_score(inputs[key])\n outputs[_transformed_name(_LABEL_KEY)] = inputs[_LABEL_KEY]\n\n return outputs\n\n\n# TFX Trainer will call this function.\ndef run_fn(fn_args: TrainerFnArgs):\n \"\"\"Train the model based on given args.\n\n Args:\n fn_args: Holds args used to train the model as name/value pairs.\n \"\"\"\n tf_transform_output = tft.TFTransformOutput(fn_args.transform_output)\n\n train_dataset = _input_fn(fn_args.train_files, tf_transform_output, 40)\n eval_dataset = _input_fn(fn_args.eval_files, tf_transform_output, 40)\n\n # To use distribution strategy, create an appropriate tf.distribute.Strategy\n # and move the creation and compiling of Keras model inside `strategy.scope`.\n #\n # For example, replace `model = _build_keras_model()` with:\n # mirrored_strategy = tf.distribute.MirroredStrategy()\n # with mirrored_strategy.scope():\n # model = _build_keras_model()\n model = _build_keras_model()\n\n model.fit(\n train_dataset,\n steps_per_epoch=fn_args.train_steps,\n validation_data=eval_dataset,\n validation_steps=fn_args.eval_steps)\n\n signatures = {\n 'serving_default':\n _get_serve_tf_examples_fn(model,\n tf_transform_output).get_concrete_function(\n tf.TensorSpec(\n shape=[None],\n dtype=tf.string,\n name='examples')),\n }\n model.save(fn_args.serving_model_dir, save_format='tf', signatures=signatures)\n","sub_path":"tfx/examples/iris/iris_utils_native_keras.py","file_name":"iris_utils_native_keras.py","file_ext":"py","file_size_in_byte":5924,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"165715591","text":"#!/usr/bin/env python\n\nfrom Bio import SeqIO\nimport click\nfrom copy import deepcopy\nimport numpy as np\nimport pandas as pd\nimport torch\nfrom torch.utils.data import DataLoader, TensorDataset\n\nfrom models.danq import DanQ\nfrom utils.io import parse_fasta_file, write\nfrom utils.data import one_hot_encode, reverse_complement\n\nCONTEXT_SETTINGS = {\n \"help_option_names\": [\"-h\", \"--help\"],\n}\n\n@click.command(context_settings=CONTEXT_SETTINGS, no_args_is_help=True)\n@click.option(\n \"-f\", \"--fasta-file\",\n help=\"FASTA file with sequences.\",\n metavar=\"FILENAME\",\n required=True\n)\n@click.option(\n \"-o\", \"--out-file\",\n help=\"Output file.\",\n metavar=\"FILENAME\",\n required=True\n)\n@click.option(\n \"-r\", \"--rev-complement\",\n help=\"Predict on reverse complement sequences.\",\n is_flag=True,\n default=False\n)\n@click.option(\n \"-s\", \"--state-dict\",\n help=\"Model state dict to use.\",\n metavar=\"FILENAME\",\n required=True\n)\n@click.option(\n \"-t\", \"--threads\",\n default=1,\n help=\"Number of CPU threads to use.\",\n show_default=True\n)\n\ndef predict(fasta_file, out_file, state_dict, rev_complement=False, threads=1):\n\n _predict(fasta_file, out_file, state_dict, rev_complement, threads)\n\ndef _predict(fasta_file, out_file, state_dict, rev_complement=False, threads=1):\n\n # Sequences\n sequences = []\n for seq_record in parse_fasta_file(fasta_file):\n sequences.append((seq_record.id, str(seq_record.seq).upper()))\n df = pd.DataFrame(sequences, columns=[\"Id\", \"Sequence\"])\n\n # One-hot encode\n encoded_sequences = []\n for seq in df[\"Sequence\"]:\n encoded_sequences.append(one_hot_encode(seq))\n encoded_sequences = np.array(encoded_sequences)\n\n # TensorDataset\n ix = np.array([[i] for i in range(len(sequences))])\n dataset = TensorDataset(torch.Tensor(encoded_sequences), torch.Tensor(ix))\n if rev_complement:\n encoded_sequences_rc = np.array(reverse_complement(encoded_sequences))\n dataset_rc = TensorDataset(\n torch.Tensor(encoded_sequences_rc), torch.Tensor(ix)\n )\n\n # DataLoader\n parameters = dict(batch_size=64, num_workers=threads)\n dataloader = DataLoader(dataset, **parameters)\n if rev_complement:\n dataloader_rc = DataLoader(dataset_rc, **parameters)\n\n # Predict\n sequence_length = len(sequences[0][1])\n predictions = __predict(sequence_length, 1, dataloader, state_dict)\n if rev_complement:\n predictions_rc = __predict(\n sequence_length, 1, dataloader_rc, state_dict\n )\n else:\n predictions_rc = np.empty((len(predictions)))\n predictions_rc[:] = np.NaN\n\n # Save predictions\n zipped_predictions = np.array(\n list(zip(df[\"Id\"].to_list(), predictions, predictions_rc[:]))\n )\n df = pd.DataFrame(zipped_predictions, columns=[\"Id\", \"Fwd\", \"Rev\"])\n df.to_csv(out_file, compression=\"gzip\", index=False)\n\ndef __predict(sequence_length, n_features, dataloader, state_dict):\n\n predictions = None\n\n device = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n model = DanQ(sequence_length, n_features).to(device)\n model.load_state_dict(torch.load(state_dict))\n model.eval() # set the model in evaluation mode\n\n for seqs, labels in dataloader:\n x = seqs.to(device) # shape = (batch_size, 4, 200)\n labels = labels.to(device)\n with torch.no_grad():\n # Forward pass\n outputs = model(x)\n # Save predictions\n if predictions is None:\n predictions = outputs.data.cpu().numpy()\n else:\n predictions = np.append(\n predictions, outputs.data.cpu().numpy(), axis=0\n )\n\n return(predictions.flatten())\n\nif __name__ == \"__main__\":\n predict()","sub_path":"DanQ/predict.py","file_name":"predict.py","file_ext":"py","file_size_in_byte":3808,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"614973472","text":"import requests\nfrom lxml import html\n\n\ndef user_info(user_id):\n url = \"https://www.shiyanlou.com/user/{}/\".format(user_id)\n response = requests.get(url)\n rep = response.content.decode()\n tree = html.fromstring(rep, 'lxml')\n user_name = tree.xpath(\"//div/span[@class='username']/text()\")\n user_level = tree.xpath(\"//div/span[@class='user-level']/text()\")\n join_date = tree.xpath(\"//div/span[@class='join-date']/text()\")\n\n try:\n user_name = user_name[0]\n user_level = int(user_level[0][1:])\n join_date = join_date[0].split()[0]\n except:\n user_name = None\n user_level = None\n join_date = None\n return user_name, user_level, join_date\n\n\nuser_info(\"214893\")","sub_path":"test/test4/user.py","file_name":"user.py","file_ext":"py","file_size_in_byte":725,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"503796354","text":"import tweepy\nimport sys\n\n\nconsumer_key = \"ZCYgdNKycYTWUB6IReeqmqish\"\nconsumer_secret = \"xm4gKvCaQVxTnAHlbXmH46MhqJbw0FCVAq72hsayMSo06rdndP\"\naccess_key=\"1426498003-PMr1Fds4lCJulrjRMd3wsYZfKLXXeIez0uOveqt\"\naccess_secret=\"3JmD7iedgX2auVOsQQPdEF5E0mlZsLp6mtPsJZ8nWxwar\"\nauth = tweepy.OAuthHandler(consumer_key, consumer_secret)\nauth.set_access_token(access_key, access_secret)\napi=tweepy.API(auth)\n\ndef gnm(name):\n try:\n user =api.get_user(name)\n foll=(user.friends_count)\n stuff = api.user_timeline(id = name, count = 10, include_rts = False,tweet_mode='extended')\n sum_tweet=0\n for tweet in stuff:\n text=(tweet.full_text).split()#Spao to kathe post se lekseis \n sum_tweet+=len(text)\n return(sum_tweet*foll)\n except:\n print('The username \"%s\" is not valid.'%(name))\n return(1)\n \n\n\nname1=input(\"First user: \")\nname2=input(\"Second user: \")\nginomeno1=gnm(name1)\nginomeno2=gnm(name2)\nif((ginomeno1!=1) or (ginomeno2!=1)): #Elegxo an iparxei to username\n if(ginomeno1>ginomeno2):\n print(\"Ο χρήστης %s έχει μεγαλύτερο γινόμενο από τον χρήστη %s.\"%(name1,name2))\n elif(ginomeno1==ginomeno2):\n print(\"Οι δύο χρήστες έχουν το ίδιο γινόμενο.\")\n else:\n print(\"Ο χρήστης %s έχει μεγαλύτερο γινόμενο από τον χρήστη %s.\"%(name2,name1))\n","sub_path":"ERGASIES_PYTHON/Askisi8.py","file_name":"Askisi8.py","file_ext":"py","file_size_in_byte":1456,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"242552584","text":"# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.macosx-10.7-intel/egg/etk/docking/util.py\n# Compiled at: 2011-01-15 06:50:51\nimport gtk\n\ndef rect_contains(rect, x, y):\n \"\"\"\n The rect_contains function checks if a point, defined by x and y falls\n within the gdk.Rectangle defined by rect.\n\n Note: Unlike rect_overlaps defined below, this function ignores a 1 pixel border.\n \"\"\"\n if x > rect.x and x < rect.x + rect.width and y > rect.y and y < rect.y + rect.height:\n return True\n else:\n return False\n\n\ndef rect_overlaps(rect, x, y):\n \"\"\"\n The rect_overlaps function checks if a point, defined by x and y overlaps\n the gdk.Rectangle defined by rect.\n\n Note: Unlike rect_contains defined above, this function does not ignore a 1 pixel border.\n \"\"\"\n if x >= rect.x and x <= rect.x + rect.width and y >= rect.y and y <= rect.y + rect.height:\n return True\n else:\n return False\n\n\ndef load_icon(icon_name, size):\n icontheme = gtk.icon_theme_get_default()\n if not icontheme.has_icon(icon_name):\n icon_name = 'gtk-missing-image'\n return icontheme.load_icon(icon_name, size, gtk.ICON_LOOKUP_USE_BUILTIN)\n\n\ndef load_icon_image(icon_name, size):\n icontheme = gtk.icon_theme_get_default()\n if not icontheme.has_icon(icon_name):\n icon_name = 'gtk-missing-image'\n return gtk.image_new_from_icon_name(icon_name, size)\n\n\ndef flatten(w, child_getter=gtk.Container.get_children):\n \"\"\"\n Generator function that returns all items in a hierarchy.\n Default `child_getter` returns children in a GTK+ widget hierarchy.\n \"\"\"\n yield w\n try:\n for c in child_getter(w):\n for d in flatten(c, child_getter):\n yield d\n\n except TypeError:\n pass","sub_path":"pycfiles/etk.docking-0.2-py2.7/util.py","file_name":"util.py","file_ext":"py","file_size_in_byte":1896,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"226061573","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nimport datetime\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('learn', '0001_initial'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='contract',\n name='create_time',\n field=models.DateTimeField(default=datetime.datetime.now),\n preserve_default=True,\n ),\n migrations.AlterField(\n model_name='contract',\n name='message',\n field=models.CharField(max_length=5000),\n preserve_default=True,\n ),\n ]\n","sub_path":"learn/migrations/0002_auto_20150925_0431.py","file_name":"0002_auto_20150925_0431.py","file_ext":"py","file_size_in_byte":658,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"162058732","text":"#!/usr/bin/python3\n\nimport sys\n\nfrom os import system as cmd\nfrom os import name as name\nfrom string import Template\n\nclass Updater(object):\n def __init__(self):\n # Init to inherit classes\n super(Updater, self).__init__()\n\n def generate(self):\n\n form=\"html\"\n source=\"docs/source/\"\n target=\"classes/\"\n build=\"docs/build/\"\n\n # Generate documentation\n print(\"\\n - Building documentation\")\n cmd_doc = Template(\"sphinx-apidoc -f -o $source $target\")\n cmd(cmd_doc.substitute(source=source, target=target))\n\n # Generate documentation\n print(\"\\n - Generating html documentation\")\n cmd_build = Template(\"sphinx-build -b $form $source $build$form\")\n cmd(cmd_build.substitute(form=form, source=source, build=build))\n\n def sphinx(self):\n \"\"\"Initialize sphinx documentation.\n\n \"\"\"\n print(\"\\n - Initializing sphinx\")\n cmd(\"./docs/sphinx-quickstart\")\n\n def ghpages(self):\n \"\"\"push to github gh-pages branch\n\n \"\"\"\n print(\"\\n - Uploading documentation to gh-pages branch\")\n cmd(\"git subtree push --prefix docs/build/html origin gh-pages\")\n\n def master(self, msg):\n print(\"\\n - Uploading to git master branch: \" + sys.argv[1])\n cmd(\"git add -A\")\n cmd('git commit -m \"' + msg +'\"')\n cmd(\"git pull origin master\")\n cmd(\"git push origin master\")","sub_path":"classes/app/updater.py","file_name":"updater.py","file_ext":"py","file_size_in_byte":1428,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"76988310","text":"from builtins import str\nimport six\n\n\ndef dict_to_query_params(d):\n if d is None or len(d) == 0:\n return ''\n\n param_list = [param + '=' + (str(value).lower() if type(value) == bool else str(value))\n for param, value in six.iteritems(d) if value is not None]\n return '?' + \"&\".join(param_list)","sub_path":"calendar_event_app/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":325,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"69722401","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Oct 16 20:38:07 2018\n\n@author: ztw1e12\n\"\"\"\n\nfrom sklearn import model_selection, preprocessing, linear_model, naive_bayes, metrics, svm\nfrom sklearn.feature_extraction.text import TfidfVectorizer, CountVectorizer\nfrom sklearn import decomposition \n# from textblob import TextBlob \nfrom keras.models import Sequential\nfrom keras.preprocessing.text import Tokenizer\n\nimport pandas, numpy, string\nfrom keras import layers, models, optimizers\nfrom sklearn.linear_model import LogisticRegression\n\n\nfrom sklearn.linear_model import Ridge\nimport random\n\n\n\n# load the dataset\n\ndf = pandas.read_csv(\"sentiment labelled sentences/yelp_labelled.txt\", names=['sentence', 'label'], sep=\"\\t\")\n\n\n# Splitting the data (sentence and labbels) into training and testing data \n\nsentences = df['sentence'].values\ny = df['label'].values\n\n\n# create a dataframe using texts and lables\ntrainDF = pandas.DataFrame()\ntrainDF['text'] = sentences\ntrainDF['label'] = y\n\n\n\n# split the dataset into training and validation datasets \ntrain_x, valid_x, train_y, valid_y = model_selection.train_test_split(trainDF['text'], trainDF['label'])\n\n\n\n# Will be used to train models\n\ntrainDF['char_count'] = trainDF['text'].apply(len)\ntrainDF['word_count'] = trainDF['text'].apply(lambda x: len(x.split()))\ntrainDF['word_density'] = trainDF['char_count'] / (trainDF['word_count']+1)\ntrainDF['punctuation_count'] = trainDF['text'].apply(lambda x: len(\"\".join(_ for _ in x if _ in string.punctuation))) \ntrainDF['title_word_count'] = trainDF['text'].apply(lambda x: len([wrd for wrd in x.split() if wrd.istitle()]))\ntrainDF['upper_case_word_count'] = trainDF['text'].apply(lambda x: len([wrd for wrd in x.split() if wrd.isupper()]))\n\n\n'''\npos_family = {\n 'noun' : ['NN','NNS','NNP','NNPS'],\n 'pron' : ['PRP','PRP$','WP','WP$'],\n 'verb' : ['VB','VBD','VBG','VBN','VBP','VBZ'],\n 'adj' : ['JJ','JJR','JJS'],\n 'adv' : ['RB','RBR','RBS','WRB']\n}\n\n\n# function to check and get the part of speech tag count of a words in a given sentence\ndef check_pos_tag(x, flag):\n cnt = 0\n try:\n wiki = TextBlob(x)\n for tup in wiki.tags:\n ppo = list(tup)[1]\n if ppo in pos_family[flag]:\n cnt += 1\n except:\n pass\n return cnt\n\ntrainDF['noun_count'] = trainDF['text'].apply(lambda x: check_pos_tag(x, 'noun'))\ntrainDF['verb_count'] = trainDF['text'].apply(lambda x: check_pos_tag(x, 'verb'))\ntrainDF['adj_count'] = trainDF['text'].apply(lambda x: check_pos_tag(x, 'adj'))\ntrainDF['adv_count'] = trainDF['text'].apply(lambda x: check_pos_tag(x, 'adv'))\ntrainDF['pron_count'] = trainDF['text'].apply(lambda x: check_pos_tag(x, 'pron'))\n'''\n\n\n\n# Our list of functions to apply.\ntransform_functions = [\n trainDF['char_count'],\n trainDF['word_count'],\n trainDF['word_density'],\n trainDF['punctuation_count'],\n trainDF['title_word_count'],\n trainDF['upper_case_word_count']\n]\n\n# Apply each function and put the results into a list.\ncolumns = []\nfor func in transform_functions:\n columns.append(func)\n \n# Convert the meta features to a numpy array.\nmeta = numpy.asarray(columns).T\n\nprint (meta)\n\n\ntrain_rows = 750\n# Set a seed to get the same \"random\" shuffle every time.\nrandom.seed(1)\n\n# Shuffle the indices for the matrix.\nindices = list(range(meta.shape[0]))\nrandom.shuffle(indices)\n\n# Create train and test sets.\ntrain = meta[indices[:train_rows], :]\ntest = meta[indices[train_rows:], :]\n# train = numpy.nan_to_num(train)\n'''\n# Run the regression and generate predictions for the test set.\nreg = Ridge(alpha=.1)\nreg.fit(train,train_y)\npredictions = reg.predict(test)\n'''\n\nclassifier = LogisticRegression()\nclassifier.fit(train, train_y)\nscore = classifier.score(test, valid_y)\n\nprint (\"Test acurracy: \", score)\n\n\n# =============================================================================\n# building sequential model layer by layer \n# =============================================================================\n\ninput_dim = train.shape[1]\nmodel = Sequential ()\nmodel.add(layers.Dense(10, input_dim=input_dim, activation='relu'))\nmodel.add(layers.Dense(1, activation='sigmoid'))\n\nmodel.compile(loss='binary_crossentropy', optimizer='adam', metrics=['accuracy'])\nmodel.summary()\n\n\nhistory = model.fit(train, train_y, epochs=20, verbose=False, validation_data=(test, valid_y), batch_size=10)\n\n\n\nloss, accuracy = model.evaluate(train, train_y, verbose=False)\nprint(\"Training Accuracy: {:.4f}\".format(accuracy))\nloss, accuracy = model.evaluate(test, valid_y, verbose=False)\nprint(\"Testing Accuracy: {:.4f}\".format(accuracy))\n\n\ntokenizer = Tokenizer(num_words=5000)\ntokenizer.fit_on_texts(train_x)\n\nX_train = tokenizer.texts_to_sequences(train_x)\nX_test = tokenizer.texts_to_sequences(valid_x)\n\nvocab_size = len(tokenizer.word_index) + 1 # Adding 1 because of reserved 0 index\n\n\nembedding_dim = 6\nmaxlen = 6\n\nmodel = Sequential()\nmodel.add(layers.Embedding(vocab_size, embedding_dim, input_length=maxlen))\nmodel.add(layers.Conv1D(128, 5, activation='relu'))\nmodel.add(layers.GlobalMaxPooling1D())\nmodel.add(layers.Dense(10, activation='relu'))\nmodel.add(layers.Dense(1, activation='sigmoid'))\nmodel.compile(optimizer='adam',\n loss='binary_crossentropy',\n metrics=['accuracy'])\nmodel.summary()\n\nhistory = model.fit(train, train_y, epochs=10, verbose=False,validation_data=(test, valid_y),batch_size=10)\nloss, accuracy = model.evaluate(train, train_y, verbose=False)\nprint(\"Training Accuracy: {:.4f}\".format(accuracy))\nloss, accuracy = model.evaluate(test, valid_y, verbose=False)\nprint(\"Testing Accuracy: {:.4f}\".format(accuracy))\n\n\n\n","sub_path":"nlp-features.py","file_name":"nlp-features.py","file_ext":"py","file_size_in_byte":5691,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"546368860","text":"#!/opt/libreoffice5.2/program/python\n# -*- coding: utf-8 -*-\nIMPLE_NAME = \"pq.Tcu\"\nSERVICE_NAME = \"pq.Tcu\"\ndef create(ctx, *args): \n from inoxt import component\n return component.create(ctx, *args, imple_name=IMPLE_NAME, service_name=SERVICE_NAME)\n# Registration\nimport unohelper\ng_ImplementationHelper = unohelper.ImplementationHelper()\ng_ImplementationHelper.addImplementation(create, IMPLE_NAME, (SERVICE_NAME,),)\n","sub_path":"TCU/src/pyunocomponent.py","file_name":"pyunocomponent.py","file_ext":"py","file_size_in_byte":434,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"643663971","text":"# coding=UTF-8\n# **********************************************************************\n# Copyright (c) 2013-2020 Cisco Systems, Inc. All rights reserved\n# written by zen warriors, do not modify!\n# **********************************************************************\n\n\nfrom cobra.mit.meta import ClassMeta\nfrom cobra.mit.meta import StatsClassMeta\nfrom cobra.mit.meta import CounterMeta\nfrom cobra.mit.meta import PropMeta\nfrom cobra.mit.meta import Category\nfrom cobra.mit.meta import SourceRelationMeta\nfrom cobra.mit.meta import NamedSourceRelationMeta\nfrom cobra.mit.meta import TargetRelationMeta\nfrom cobra.mit.meta import DeploymentPathMeta, DeploymentCategory\nfrom cobra.model.category import MoCategory, PropCategory, CounterCategory\nfrom cobra.mit.mo import Mo\n\n\n# ##################################################\nclass Route(Mo):\n \"\"\"\n The EIGRP route information.\n\n \"\"\"\n\n meta = ClassMeta(\"cobra.model.eigrp.Route\")\n\n meta.moClassName = \"eigrpRoute\"\n meta.rnFormat = \"rt-[%(pfx)s]\"\n meta.category = MoCategory.REGULAR\n meta.label = \"EIGRP Route\"\n meta.writeAccessMask = 0x8008020040001\n meta.readAccessMask = 0x8008020040001\n meta.isDomainable = False\n meta.isReadOnly = True\n meta.isConfigurable = False\n meta.isDeletable = False\n meta.isContextRoot = False\n\n meta.childClasses.add(\"cobra.model.eigrp.Nexthop\")\n\n meta.childNamesAndRnPrefix.append((\"cobra.model.eigrp.Nexthop\", \"nh-\"))\n\n meta.parentClasses.add(\"cobra.model.eigrp.Db\")\n\n meta.superClasses.add(\"cobra.model.nw.DbRec\")\n meta.superClasses.add(\"cobra.model.l3.DbRec\")\n meta.superClasses.add(\"cobra.model.eigrp.DbRec\")\n meta.superClasses.add(\"cobra.model.nw.Conn\")\n meta.superClasses.add(\"cobra.model.nw.Item\")\n meta.superClasses.add(\"cobra.model.nw.GEp\")\n\n meta.rnPrefixes = [\n ('rt-', True),\n ]\n\n prop = PropMeta(\"str\", \"actStQual\", \"actStQual\", 17384, PropCategory.REGULAR)\n prop.label = \"Query Origin Qualifier\"\n prop.isOper = True\n prop.range = [(0, 512)]\n meta.props.add(\"actStQual\", prop)\n\n prop = PropMeta(\"str\", \"chgQual\", \"chgQual\", 17385, PropCategory.REGULAR)\n prop.label = \"Last route state transition Qualifier\"\n prop.isOper = True\n prop.range = [(0, 512)]\n meta.props.add(\"chgQual\", prop)\n\n prop = PropMeta(\"str\", \"childAction\", \"childAction\", 4, PropCategory.CHILD_ACTION)\n prop.label = \"None\"\n prop.isImplicit = True\n prop.isAdmin = True\n prop._addConstant(\"deleteAll\", \"deleteall\", 16384)\n prop._addConstant(\"deleteNonPresent\", \"deletenonpresent\", 8192)\n prop._addConstant(\"ignore\", \"ignore\", 4096)\n meta.props.add(\"childAction\", prop)\n\n prop = PropMeta(\"str\", \"dn\", \"dn\", 1, PropCategory.DN)\n prop.label = \"None\"\n prop.isDn = True\n prop.isImplicit = True\n prop.isAdmin = True\n prop.isCreateOnly = True\n meta.props.add(\"dn\", prop)\n\n prop = PropMeta(\"str\", \"fDist\", \"fDist\", 17383, PropCategory.REGULAR)\n prop.label = \"Feasible Distance\"\n prop.isOper = True\n meta.props.add(\"fDist\", prop)\n\n prop = PropMeta(\"str\", \"flags\", \"flags\", 17388, PropCategory.REGULAR)\n prop.label = \"Route Flags\"\n prop.isOper = True\n prop._addConstant(\"stuck-act\", \"stuck-in-active\", 1)\n meta.props.add(\"flags\", prop)\n\n prop = PropMeta(\"str\", \"lastActTs\", \"lastActTs\", 17386, PropCategory.REGULAR)\n prop.label = \"Last active state timestamp\"\n prop.isOper = True\n meta.props.add(\"lastActTs\", prop)\n\n prop = PropMeta(\"str\", \"modTs\", \"modTs\", 7, PropCategory.REGULAR)\n prop.label = \"None\"\n prop.isImplicit = True\n prop.isAdmin = True\n prop.defaultValue = 0\n prop.defaultValueStr = \"never\"\n prop._addConstant(\"never\", \"never\", 0)\n meta.props.add(\"modTs\", prop)\n\n prop = PropMeta(\"str\", \"name\", \"name\", 16437, PropCategory.REGULAR)\n prop.label = \"Name\"\n prop.isConfig = True\n prop.isAdmin = True\n prop.range = [(1, 128)]\n meta.props.add(\"name\", prop)\n\n prop = PropMeta(\"str\", \"operSt\", \"operSt\", 17382, PropCategory.REGULAR)\n prop.label = \"Operational State\"\n prop.isOper = True\n prop.defaultValue = 0\n prop.defaultValueStr = \"passive\"\n prop._addConstant(\"active\", \"active\", 1)\n prop._addConstant(\"passive\", \"passive\", 0)\n meta.props.add(\"operSt\", prop)\n\n prop = PropMeta(\"str\", \"pfx\", \"pfx\", 17381, PropCategory.REGULAR)\n prop.label = \"Address\"\n prop.isConfig = True\n prop.isAdmin = True\n prop.isCreateOnly = True\n prop.isNaming = True\n meta.props.add(\"pfx\", prop)\n\n prop = PropMeta(\"str\", \"rn\", \"rn\", 2, PropCategory.RN)\n prop.label = \"None\"\n prop.isRn = True\n prop.isImplicit = True\n prop.isAdmin = True\n prop.isCreateOnly = True\n meta.props.add(\"rn\", prop)\n\n prop = PropMeta(\"str\", \"siaQueryCnt\", \"siaQueryCnt\", 17387, PropCategory.REGULAR)\n prop.label = \"Number of Stuck-in-active queries sent for the active route\"\n prop.isOper = True\n meta.props.add(\"siaQueryCnt\", prop)\n\n prop = PropMeta(\"str\", \"status\", \"status\", 3, PropCategory.STATUS)\n prop.label = \"None\"\n prop.isImplicit = True\n prop.isAdmin = True\n prop._addConstant(\"created\", \"created\", 2)\n prop._addConstant(\"deleted\", \"deleted\", 8)\n prop._addConstant(\"modified\", \"modified\", 4)\n meta.props.add(\"status\", prop)\n\n meta.namingProps.append(getattr(meta.props, \"pfx\"))\n getattr(meta.props, \"pfx\").needDelimiter = True\n\n def __init__(self, parentMoOrDn, pfx, markDirty=True, **creationProps):\n namingVals = [pfx]\n Mo.__init__(self, parentMoOrDn, markDirty, *namingVals, **creationProps)\n\n\n\n# End of package file\n# ##################################################\n","sub_path":"venv/Lib/site-packages/cobra/modelimpl/eigrp/route.py","file_name":"route.py","file_ext":"py","file_size_in_byte":5660,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"594751408","text":"from django.urls import path\nfrom .views import (NewsListView , NewsDetailView , \n\t\t\t\t\t#NewsCreateT, \n\t\t\t\t\tNewsCreateView , \n\t\t\t\t\tNewsUpdateView , NewsDeleteView)\n#app_name = 'news' ##\nurlpatterns = [ \n\t\n\tpath('////delete', NewsDeleteView.as_view(), name='sub_news_delete'),\n\tpath('////edit/', NewsUpdateView.as_view(), name='sub_news_edit'),\n\tpath('create/', NewsCreateView.as_view(), name='news_create'),\n\t#path('salam/', NewsCreateT.as_view() ,name='news_test'),\n\tpath('////', NewsDetailView.as_view(), name='sub_news_url'),\n\tpath('', NewsListView.as_view(), name='main_news'),\n\t ] \n","sub_path":"news/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":709,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"200514612","text":"from PIL import Image\nfrom numpy import *\n\ndef varianceEdge(im,n):\n\tfor i in range(0,len(im),1):\n\t\tfor j in range(0,len(im[0]),1):\n\t\t\tsub = im[i:i+n+1,j:j+n+1]\n\t\t\tim[i,j] = std(sub)\n\treturn im\n\ndef main():\n\tname = 'parking'\n\tim = array(Image.open(name + '.jpg').convert('L'))\n\tim2 = Image.fromarray(varianceEdge(im,2))\n\tim2.convert('RGB').save('var' + name + '.png','png')\n\nmain()\n\n\n\n","sub_path":"Exercises/Exercise_3/ex3q2_variance.py","file_name":"ex3q2_variance.py","file_ext":"py","file_size_in_byte":385,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"367146526","text":"from enum import Enum\n\nimport numpy as np\n\nimport hiv_model_econ_eval.input_data as data\n\n\nclass Therapies(Enum):\n \"\"\" mono vs. combination therapy \"\"\"\n MONO = 0\n COMBO = 1\n\n\nclass Parameters:\n def __init__(self, therapy):\n\n # selected therapy\n self.therapy = therapy\n\n # initial health state\n self.initialHealthState = data.HealthStates.CD4_200to500\n\n # annual treatment cost\n if self.therapy == Therapies.MONO:\n self.annualTreatmentCost = data.Zidovudine_COST\n else:\n self.annualTreatmentCost = data.Zidovudine_COST + data.Lamivudine_COST\n\n # transition probability matrix of the selected therapy\n self.probMatrix = []\n\n # calculate transition probabilities between hiv states\n if self.therapy == Therapies.MONO:\n # calculate transition probability matrix for the mono therapy\n self.probMatrix = get_prob_matrix_mono(trans_matrix=data.TRANS_MATRIX)\n\n elif self.therapy == Therapies.COMBO:\n # calculate transition probability matrix for the combination therapy\n self.probMatrix = get_prob_matrix_combo(\n prob_matrix_mono=get_prob_matrix_mono(trans_matrix=data.TRANS_MATRIX),\n combo_rr=data.TREATMENT_RR)\n\n # annual state costs and utilities\n self.annualStateCosts = data.ANNUAL_STATE_COST\n self.annualStateUtilities = data.ANNUAL_STATE_UTILITY\n\n # discount rate\n self.discountRate = data.DISCOUNT\n\n\ndef get_prob_matrix_mono(trans_matrix):\n \"\"\"\n :param trans_matrix: transition matrix containing counts of transitions between states\n :return: transition probability matrix\n \"\"\"\n\n # initialize transition probability matrix\n trans_prob_matrix = []\n\n # for each row in the transition matrix\n for row in trans_matrix:\n # calculate the transition probabilities\n prob_row = np.array(row)/sum(row)\n # add this row of transition probabilities to the transition probability matrix\n trans_prob_matrix.append(prob_row)\n\n return trans_prob_matrix\n\n\ndef get_prob_matrix_combo(prob_matrix_mono, combo_rr):\n \"\"\"\n :param prob_matrix_mono: (list of lists) transition probability matrix under mono therapy\n :param combo_rr: relative risk of the combination treatment\n :returns (list of lists) transition probability matrix under combination therapy \"\"\"\n\n # create an empty list of lists\n matrix_combo = []\n for row in prob_matrix_mono:\n matrix_combo.append(np.zeros(len(row))) # adding a row [0, 0, 0, 0]\n\n # populate the combo matrix\n # calculate the effect of combo-therapy on non-diagonal elements\n for s in range(len(matrix_combo)):\n for next_s in range(s + 1, len(prob_matrix_mono[s])):\n matrix_combo[s][next_s] = combo_rr * prob_matrix_mono[s][next_s]\n\n # diagonal elements are calculated to make sure the sum of each row is 1\n for s in range(len(matrix_combo)):\n matrix_combo[s][s] = 1 - sum(matrix_combo[s][s+1:])\n\n return matrix_combo\n\n\n# tests\nif __name__ == '__main__':\n matrix_mono = get_prob_matrix_mono(data.TRANS_MATRIX)\n matrix_combo = get_prob_matrix_combo(matrix_mono, data.TREATMENT_RR)\n\n print(matrix_mono)\n print(matrix_combo)\n","sub_path":"hiv_model_econ_eval/param_classes.py","file_name":"param_classes.py","file_ext":"py","file_size_in_byte":3297,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"651684861","text":"from random import randrange\nfrom mininet.log import info\n\nclass Placer( object ):\n \"The base object of the placer classes.\"\n\n def __init__( self, servers=None, nodes=None, hosts=None,\n switches=None, controllers=None, links=None ):\n \"\"\"Initialize necessary info,\n They are optional.\n If not feed, they will be empty.\n \"\"\"\n self.servers = servers or []\n self.nodes = nodes or []\n self.hosts = hosts or []\n self.switches = switches or []\n self.controllers = controllers or []\n self.links = links or []\n\n def place( self, node ):\n \"This is going to be overridden. It should return the server to place the node.\"\n return None\n\n\nclass RandomPlacer( Placer ):\n \"Random placement\"\n def place( self, nodename ):\n \"\"\"\n Random placement function\n Nodename is the name of node.\n \"\"\"\n # This may be slow with lots of servers\n return self.servers[ randrange( 0, len( self.servers ) ) ]\n\n\nclass RoundRobinPlacer( Placer ):\n \"\"\"Round-robin placement\n Note this will usually result in cross-server links between\n hosts and switches\"\"\"\n\n def __init__( self, *args, **kwargs ):\n Placer.__init__( self, *args, **kwargs )\n self.next = 0\n\n def place( self, nodename ):\n \"\"\"\n Return the server to place the node.\n Place nodes circularly.\n Nodename is the name of node.\n \"\"\"\n server = self.servers[ self.next ]\n self.next = ( self.next + 1 ) % len( self.servers )\n return server\n\n\nclass SwitchBinPlacer( Placer ):\n \"\"\"Place switches (and controllers) into evenly-sized bins,\n and attempt to co-locate hosts and switches\"\"\"\n\n def __init__( self, *args, **kwargs ):\n Placer.__init__( self, *args, **kwargs )\n self.servdict = dict( enumerate( self.servers ) )\n self.hset = frozenset( self.hosts )\n self.sset = frozenset( self.switches )\n self.cset = frozenset( self.controllers )\n self.placement = self.calculatePlacement()\n\n @staticmethod\n def bin( nodes, servers ):\n \"\"\"Distribute nodes evenly over servers.\"\"\"\n # Calculate base bin size\n nlen = len( nodes )\n slen = len( servers )\n quotient = int( nlen / slen )\n binsizes = { server: quotient for server in servers }\n # Distribute remainder\n remainder = nlen % slen\n for server in servers[ 0 : remainder ]:\n binsizes[ server ] += 1\n # Create binsize[ server ] tickets for each server\n tickets = sum( [ binsizes[ server ] * [ server ]\n for server in servers ], [] )\n # And assign one ticket to each node\n return { node: ticket for node, ticket in zip( nodes, tickets ) }\n\n def calculatePlacement( self ):\n \"Pre-calculate node placement\"\n placement = {}\n # Create host-switch connectivity map,\n # associating host with last switch that it's\n # connected to\n switchFor = {}\n for src, dst in self.links:\n if src in self.hset and dst in self.sset:\n switchFor[ src ] = dst\n if dst in self.hset and src in self.sset:\n switchFor[ dst ] = src\n # Place switches\n placement = self.bin( self.switches, self.servers )\n # Place controllers and merge into placement dict\n placement.update( self.bin( self.controllers, self.servers ) )\n # Co-locate hosts with their switches\n for h in self.hosts:\n if h in placement:\n # Host is already placed - leave it there\n continue\n if h in switchFor:\n placement[ h ] = placement[ switchFor[ h ] ]\n else:\n raise Exception(\n \"SwitchBinPlacer: cannot place isolated host \" + h )\n return placement\n\n def place( self, node ):\n \"\"\"Simple placement algorithm:\n place switches into evenly sized bins,\n and place hosts near their switches\"\"\"\n return self.placement[ node ]\n\n\nclass HostSwitchBinPlacer( Placer ):\n \"\"\"Place switches *and hosts* into evenly-sized bins\n Note that this will usually result in cross-server\n links between hosts and switches\"\"\"\n\n def __init__( self, *args, **kwargs ):\n Placer.__init__( self, *args, **kwargs )\n # Calculate bin sizes\n scount = len( self.servers )\n self.hbin = max( int( len( self.hosts ) / scount ), 1 )\n self.sbin = max( int( len( self.switches ) / scount ), 1 )\n self.cbin = max( int( len( self.controllers ) / scount ) , 1 )\n info( 'scount:', scount )\n info( 'bins:', self.hbin, self.sbin, self.cbin, '\\n' )\n self.servdict = dict( enumerate( self.servers ) )\n self.hset = frozenset( self.hosts )\n self.sset = frozenset( self.switches )\n self.cset = frozenset( self.controllers )\n self.hind, self.sind, self.cind = 0, 0, 0\n\n def place( self, nodename ):\n \"\"\"Simple placement algorithm:\n place nodes into evenly sized bins\"\"\"\n # Place nodes into bins\n if nodename in self.hset:\n server = self.servdict[ self.hind / self.hbin ]\n self.hind += 1\n elif nodename in self.sset:\n server = self.servdict[ self.sind / self.sbin ]\n self.sind += 1\n elif nodename in self.cset:\n server = self.servdict[ self.cind / self.cbin ]\n self.cind += 1\n else:\n info( 'warning: unknown node', nodename )\n server = self.servdict[ 0 ]\n return server\n\n","sub_path":"scalablemininet/scalabletopo.py","file_name":"scalabletopo.py","file_ext":"py","file_size_in_byte":5720,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"213205340","text":"######################################################\n# RTRATProd.py\n# ---------\n# Author: Matt Mottram\n# \n#\n# Description:\n# Ganga runtime handlers for SNO+ RATProd application.\n#\n# Prepares RATUser application for given backend.\n#\n# - RTHandler: handles submission to local/batch backends\n# - WGRTHandler: handles submission to WestGrid backend\n# - LCGRTHandler: handles submission to LCG backend\n#\n# Revision History:\n# - 26/08/14: M. Mottram - moved from RATProd (as with RTRATUser)\n#\n######################################################\n\nimport os\nimport shutil\n\nfrom Ganga.GPIDev.Adapters.IRuntimeHandler import IRuntimeHandler\n\nfrom Ganga.GPIDev.Lib.File import *\n\n\n# Assume that the applications should come from the GangaSNOplus/Lib/Applications directory\n_app_directory = os.path.abspath(os.path.join(os.path.dirname(__file__), \"../Applications\"))\n\nclass RTHandler(IRuntimeHandler):\n '''Standard RTHandler (for all batch/local submission).\n '''\n def prepare(self,app,appconfig,appmasterconfig,jobmasterconfig):\n\n logger.debug('RAT::RTHandler prepare ...')\n from Ganga.GPIDev.Adapters.StandardJobConfig import StandardJobConfig\n\n #create the backend wrapper script\n job=app.getJobObject()\n\n # Check that all is OK in the application\n app.verify()\n\n #Check whether we're looking for a non-default sw dir\n if app.softwareEnvironment is None:\n logger.error('Must specify a RAT environment')\n raise Exception\n\n #we need to know the name of the file to run\n macroFile = None\n prodFile = None\n if app.ratMacro!='':\n decimated=app.ratMacro.split('/')\n macroFile=decimated[len(decimated)-1]\n else:\n decimated=app.prodScript.split('/')\n prodFile=decimated[len(decimated)-1]\n\n foutList='[%s]' % (','.join(f for f in app.outputFiles))\n finList='[%s]' % (','.join(f for f in app.inputFiles))\n\n if app.environment==None:\n args = []\n args += ['-e',app.softwareEnvironment]\n args += ['-d',app.outputDir]\n args += ['-o',foutList]\n args += ['-i',finList]\n if app.ratMacro!='':\n args += ['-m',macroFile]\n else:\n args += ['-k','-m',prodFile]\n if app.useDB:\n args += ['--dbuser',app.rat_db_user]\n args += ['--dbpassword',app.rat_db_pswd]\n args += ['--dbname',app.rat_db_name]\n args += ['--dbprotocol',app.rat_db_protocol]\n args += ['--dburl',app.rat_db_url]\n if app.discardOutput:\n args += ['--nostore']\n \n app._getParent().inputsandbox.append('%s/job_tools.py' % _app_directory)\n return StandardJobConfig(File('%s/ratProdRunner.py' % _app_directory),\n inputbox = app._getParent().inputsandbox,\n outputbox = app._getParent().outputsandbox,\n args = args)\n else:#need a specific environment setup \n #can either use a specific file or a list of strings. the latter needs to be converted to a temp file and shipped.\n envFile=None\n if type(app.environment)==list:\n tempname = 'tempRATProdEnv_%s'%os.getlogin()\n tempf = file('/tmp/%s'%(tempname),'w')\n for line in app.environment:\n tempf.write('%s \\n' % line)\n tempf.close()\n app._getParent().inputsandbox.append('/tmp/%s'%(tempname))\n envFile=tempname\n else:\n app._getParent().inputsandbox.append(app.environment)\n envFile=os.path.basename(app.environment)\n args = ''\n args += '-e %s '%(app.softwareEnvironment)\n args += '-d %s '%(app.outputDir)\n args += '-o %s '%(foutList)\n args += '-i %s '%(finList)\n if app.ratMacro!='':\n args += '-m %s '%(macroFile)\n else:\n args += '-k -m %s '%(prodFile)\n if app.useDB:\n args += '--dbuser %s '%(app.rat_db_user)\n args += '--dbpassword %s '%(app.rat_db_pswd)\n args += '--dbname %s '%(app.rat_db_name)\n args += '--dbprotocol %s '%(app.rat_db_protocol)\n args += '--dburl %s '%(app.rat_db_url)\n if app.discardOutput:\n args += '--nostore '\n\n wrapperArgs = ['-f', envFile, '-a', '%s' % args]\n wrapperArgs += ['ratProdRunner.py', 'misc']\n\n app._getParent().inputsandbox.append('%s/ratProdRunner.py' % _app_directory)\n app._getParent().inputsandbox.append('%s/job_tools.py' % _app_directory)\n\n return StandardJobConfig(File('%s/sillyPythonWrapper.py' % _app_directory),\n inputbox = app._getParent().inputsandbox,\n outputbox = app._getParent().outputsandbox,\n args = wrapperArgs)\n\n###################################################################\n\nclass WGRTHandler(IRuntimeHandler):\n '''WGRTHandler for WestGrid submission.\n '''\n def prepare(self,app,appconfig,appmasterconfig,jobmasterconfig):\n\n logger.debug('RAT::RTHandler prepare ...')\n from Ganga.GPIDev.Adapters.StandardJobConfig import StandardJobConfig\n\n #create the backend wrapper script\n job=app.getJobObject()\n\n # Check that all is OK in the application\n app.verify()\n\n #Check whether we're looking for a non-default sw dir\n if app.softwareEnvironment is None:\n logger.error('Must specify a RAT directory')\n raise Exception\n\n voproxy = job.backend.voproxy\n if voproxy==None:\n #use the proxy from the environment (default behaviour) \n try:\n voproxy = os.environ[\"X509_USER_PROXY\"]\n if not voproxy.startswith(os.environ[\"HOME\"]):\n # If the proxy is not on the home directory, copy if there (for jobs on worker nodes)\n cacheDir = os.path.join(os.environ[\"HOME\"], \"gaspCache\")\n if not os.path.exists(os.path.join(cacheDir)):\n os.makedirs(cacheDir)\n proxy_path = os.path.join(cacheDir, os.path.basename(voproxy))\n shutil.copy(voproxy, proxy_path) # Maintains the permissions: important!\n logger.warn(\"Proxy not stored in home directory, transferring across to %s\" % proxy_path)\n logger.warn(\"Note that this proxy may expire during the job lifetime!\")\n voproxy = proxy_path\n except:\n logger.error('Cannot run without voproxy either in environment (X509_USER_PROXY) or specified for WG backend')\n raise Exception\n if not os.path.exists(voproxy): \n logger.error('Valid WestGrid backend voproxy location MUST be specified: %s'%(voproxy))\n raise Exception\n\n #we need to know the name of the file to run\n macroFile = None\n prodFile = None\n if app.ratMacro!='':\n decimated=app.ratMacro.split('/')\n macroFile=decimated[len(decimated)-1]\n else:\n decimated=app.prodScript.split('/')\n prodFile=decimated[len(decimated)-1]\n\n foutList='[%s]' % (','.join(f for f in app.outputFiles))\n finList='[%s]' % (','.join(f for f in app.inputFiles))\n\n args = ''\n args += '-g srm '\n args += '-e %s '%(app.softwareEnvironment)\n args += '-d %s '%(app.outputDir)\n args += '-o %s '%(foutList)\n args += '-i %s '%(finList)\n if app.ratMacro!='':\n args += '-m %s '%(macroFile)\n else:\n args += '-k -m %s '%(prodFile)\n args += '--voproxy %s '%(voproxy)\n if app.useDB:\n args += '--dbuser %s '%(app.rat_db_user)\n args += '--dbpassword %s '%(app.rat_db_pswd)\n args += '--dbname %s '%(app.rat_db_name)\n args += '--dbprotocol %s '%(app.rat_db_protocol)\n args += '--dburl %s '%(app.rat_db_url)\n if app.discardOutput:\n args += '--nostore '\n args += '--se %s ' % job.backend.se\n args += '--surl %s ' % job.backend.baseSURL\n\n\n wrapperArgs = ['-a', '%s' % args]\n wrapperArgs += ['ratProdRunner.py', 'wg']\n\n app._getParent().inputsandbox.append('%s/ratProdRunner.py' % _app_directory)\n app._getParent().inputsandbox.append('%s/job_tools.py' % _app_directory)\n\n return StandardJobConfig(File('%s/sillyPythonWrapper.py' % _app_directory),\n inputbox = app._getParent().inputsandbox,\n outputbox = app._getParent().outputsandbox,\n args = wrapperArgs)\n\n###################################################################\n\nclass LCGRTHandler(IRuntimeHandler):\n '''RTHandler for Grid submission.\n Could include CE options and tags here.\n '''\n def prepare(self,app,appconfig,appmasterconfig,jobmasterconfig):\n\n logger.debug('RAT::LCGRTHandler prepare ...')\n from Ganga.Lib.LCG import LCGJobConfig\n\n #create the backend wrapper script\n job=app.getJobObject()\n\n # Check that all is OK in the application\n app.verify()\n\n # Check the current worker node permissions\n grid_config = RATUtil.GridConfig.get_instance()\n for ce in grid_config.get_excluded_worker_nodes():\n ganga_job.backend.requirements.excludedCEs += ' %s ' % ce\n\n #Check whether we're looking for a non-default sw dir\n if app.softwareEnvironment is None:\n # The relative path for CVMFS, SNOPLUS_CVMFS_DIR will be set at the backend\n # Note the extra \\ to escape the dollar in the initial python wrapper\n app.softwareEnvironment = '\\$SNOPLUS_CVMFS_DIR/sw/%s/env_rat-%s.sh' % (app.ratVersion, app.ratVersion)\n\n #we need to know the name of the file to run\n macroFile = None\n prodFile = None\n if app.ratMacro!='':\n decimated=app.ratMacro.split('/')\n macroFile=decimated[len(decimated)-1]\n else:\n decimated=app.prodScript.split('/')\n prodFile=decimated[len(decimated)-1]\n\n foutList='[%s]' % (','.join(f for f in app.outputFiles))\n finList='[%s]' % (','.join(f for f in app.inputFiles))\n\n args = ''\n args += '-g lcg '\n args += '-e %s '%(app.softwareEnvironment)\n args += '-d %s '%(app.outputDir)\n args += '-o %s '%(foutList)\n args += '-i %s '%(finList)\n if app.ratMacro!='':\n args += '-m %s '%(macroFile)\n else:\n args += '-k -m %s '%(prodFile)\n if app.useDB:\n args += '--dbuser %s '%(app.rat_db_user)\n args += '--dbpassword %s '%(app.rat_db_pswd)\n args += '--dbname %s '%(app.rat_db_name)\n args += '--dbprotocol %s '%(app.rat_db_protocol)\n args += '--dburl %s '%(app.rat_db_url)\n if app.discardOutput:\n args += '--nostore '\n\n wrapperArgs = ['-a', '\"%s\"' % (args)]\n wrapperArgs += ['ratProdRunner.py', 'lcg']\n\n app._getParent().inputsandbox.append('%s/ratProdRunner.py' % _app_directory)\n app._getParent().inputsandbox.append('%s/job_tools.py' % _app_directory)\n\n return LCGJobConfig(File('%s/sillyPythonWrapper.py' % _app_directory),\n inputbox = app._getParent().inputsandbox,\n outputbox = app._getParent().outputsandbox,\n args = wrapperArgs)\n\n###################################################################\n","sub_path":"gasp/GangaSNOplus/Lib/RTHandlers/RTRATProd.py","file_name":"RTRATProd.py","file_ext":"py","file_size_in_byte":12025,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"364691607","text":"def take_in(num): return [float(input()) for _ in range(num)] \r\nsample=int(input())\r\nx=int(input())\r\ninstance=take_in(x)\r\ny=int(input())\r\nprices=take_in(y)\r\ndef interpolate(instance,prices,sample):\r\n try:\r\n position=instance.index(list(filter(lambda x:x>sample,instance))[0])\r\n except:\r\n position= len(instance)-1\r\n if position<1:\r\n xs=instance[0:2]\r\n ys=prices[0:2]\r\n else:\r\n xs=instance[position-1:position+1]\r\n ys=prices[position-1:position+1]\r\n ans=((ys[1]-ys[0])*(sample-xs[1])/(xs[1]-xs[0]) +ys[1])*100\r\n #round and format failed to approx 6.125 to 6.13\r\n print(ans)\r\n return (ans//1)/100 if ans%1 <0.5 else ((ans+1)//1)/100\r\nprint(interpolate(instance,prices,sample))\r\n","sub_path":"alg.py","file_name":"alg.py","file_ext":"py","file_size_in_byte":743,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"612077587","text":"from __future__ import print_function\nfrom matplotlib import pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\nimport numpy as np\nimport Read_Data as RD\n\n#dir = \"wine-5-fold/wine-5-1tra.dat\"\ndir = \"yeast4.dat\"\n\nRD.Initialize_Data(dir)\n\n#Features_Attribute = [0, 13 , 14, 15, 16, 17, 18]\nFeatures_Attribute = np.arange(0, RD.Num_Features, 1)\n\nl = len(Features_Attribute)\n\nfor i in range(0, l):\n for j in range(i+1, l):\n for k in range(j+1, l):\n X_index = Features_Attribute[i]\n Y_index = Features_Attribute[j]\n Z_index = Features_Attribute[k]\n print(X_index, Y_index, Z_index)\n ax = plt.subplot(111, projection='3d')\n ax.scatter(RD.Negative_Feature[:,X_index], RD.Negative_Feature[:,Y_index], RD.Negative_Feature[:,Z_index], marker = 'o', color = '#539caf', label='1', s = 30, alpha=0.3)\n ax.scatter(RD.Positive_Feature[:,X_index], RD.Positive_Feature[:,Y_index], RD.Positive_Feature[:,Z_index], marker = '+', color = 'r', label='2', s = 50)\n plt.show()","sub_path":"3D_Scatter_Plot.py","file_name":"3D_Scatter_Plot.py","file_ext":"py","file_size_in_byte":1054,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"475880618","text":"# Given some integer, find the maximal number you can obtain by deleting exactly one digit of the given number.\n#\n# Example\n#\n# For n = 152, the output should be\n# deleteDigit(n) = 52;\n# For n = 1001, the output should be\n# deleteDigit(n) = 101.\n\n\ndef deleteDigit(n):\n num = 0\n\n s = str(n)\n for i in range(len(s)):\n if int(s[:i] + s[i + 1:]) > num:\n num = int(s[:i] + s[i + 1:])\n\n return num\n\n\nprint(deleteDigit(1001))\nprint(deleteDigit(152))\n\n","sub_path":"arcade/deleteDigit.py","file_name":"deleteDigit.py","file_ext":"py","file_size_in_byte":474,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"38458274","text":"# Copyright (c) 2015, Frappe Technologies Pvt. Ltd. and Contributors\n# License: GNU General Public License v3. See license.txt\n\n\nfrom __future__ import unicode_literals\nimport unittest\nimport frappe\nfrom erpnext.selling.doctype.sales_order.test_sales_order import make_sales_order\nfrom erpnext.accounts.doctype.sales_invoice.test_sales_invoice import create_sales_invoice\nfrom erpnext.stock.get_item_details import get_item_details\nfrom frappe import MandatoryError\n\nclass TestPricingRule(unittest.TestCase):\n\tdef test_pricing_rule_for_discount(self):\n\t\tfrom erpnext.stock.get_item_details import get_item_details\n\t\tfrom frappe import MandatoryError\n\n\t\tfrappe.db.sql(\"delete from `tabPricing Rule`\")\n\n\t\ttest_record = {\n\t\t\t\"doctype\": \"Pricing Rule\",\n\t\t\t\"title\": \"_Test Pricing Rule\",\n\t\t\t\"apply_on\": \"Item Code\",\n\t\t\t\"item_code\": \"_Test Item\",\n\t\t\t\"selling\": 1,\n\t\t\t\"price_or_discount\": \"Discount Percentage\",\n\t\t\t\"price\": 0,\n\t\t\t\"discount_percentage\": 10,\n\t\t\t\"company\": \"_Test Company\"\n\t\t}\n\t\tfrappe.get_doc(test_record.copy()).insert()\n\n\t\targs = frappe._dict({\n\t\t\t\"item_code\": \"_Test Item\",\n\t\t\t\"company\": \"_Test Company\",\n\t\t\t\"price_list\": \"_Test Price List\",\n\t\t\t\"currency\": \"_Test Currency\",\n\t\t\t\"doctype\": \"Sales Order\",\n\t\t\t\"conversion_rate\": 1,\n\t\t\t\"price_list_currency\": \"_Test Currency\",\n\t\t\t\"plc_conversion_rate\": 1,\n\t\t\t\"order_type\": \"Sales\",\n\t\t\t\"customer\": \"_Test Customer\",\n\t\t\t\"name\": None\n\t\t})\n\t\tdetails = get_item_details(args)\n\t\tself.assertEquals(details.get(\"discount_percentage\"), 10)\n\n\t\tprule = frappe.get_doc(test_record.copy())\n\t\tprule.applicable_for = \"Customer\"\n\t\tprule.title = \"_Test Pricing Rule for Customer\"\n\t\tself.assertRaises(MandatoryError, prule.insert)\n\n\t\tprule.customer = \"_Test Customer\"\n\t\tprule.discount_percentage = 20\n\t\tprule.insert()\n\t\tdetails = get_item_details(args)\n\t\tself.assertEquals(details.get(\"discount_percentage\"), 20)\n\n\t\tprule = frappe.get_doc(test_record.copy())\n\t\tprule.apply_on = \"Item Group\"\n\t\tprule.item_group = \"All Item Groups\"\n\t\tprule.title = \"_Test Pricing Rule for Item Group\"\n\t\tprule.discount_percentage = 15\n\t\tprule.insert()\n\n\t\targs.customer = \"_Test Customer 1\"\n\t\tdetails = get_item_details(args)\n\t\tself.assertEquals(details.get(\"discount_percentage\"), 10)\n\n\t\tprule = frappe.get_doc(test_record.copy())\n\t\tprule.applicable_for = \"Campaign\"\n\t\tprule.campaign = \"_Test Campaign\"\n\t\tprule.title = \"_Test Pricing Rule for Campaign\"\n\t\tprule.discount_percentage = 5\n\t\tprule.priority = 8\n\t\tprule.insert()\n\n\t\targs.campaign = \"_Test Campaign\"\n\t\tdetails = get_item_details(args)\n\t\tself.assertEquals(details.get(\"discount_percentage\"), 5)\n\n\t\tfrappe.db.sql(\"update `tabPricing Rule` set priority=NULL where campaign='_Test Campaign'\")\n\t\tfrom erpnext.accounts.doctype.pricing_rule.pricing_rule\timport MultiplePricingRuleConflict\n\t\tself.assertRaises(MultiplePricingRuleConflict, get_item_details, args)\n\n\t\targs.item_code = \"_Test Item 2\"\n\t\tdetails = get_item_details(args)\n\t\tself.assertEquals(details.get(\"discount_percentage\"), 15)\n\n\t\tfrappe.db.sql(\"delete from `tabPricing Rule`\")\n\n\tdef test_pricing_rule_for_margin(self):\n\t\tfrom erpnext.stock.get_item_details import get_item_details\n\t\tfrom frappe import MandatoryError\n\n\t\tfrappe.db.sql(\"delete from `tabPricing Rule`\")\n\n\t\ttest_record = {\n\t\t\t\"doctype\": \"Pricing Rule\",\n\t\t\t\"title\": \"_Test Pricing Rule\",\n\t\t\t\"apply_on\": \"Item Code\",\n\t\t\t\"item_code\": \"_Test FG Item 2\",\n\t\t\t\"selling\": 1,\n\t\t\t\"price_or_discount\": \"Discount Percentage\",\n\t\t\t\"price\": 0,\n\t\t\t\"margin_type\": \"Percentage\",\n\t\t\t\"margin_rate_or_amount\": 10,\n\t\t\t\"company\": \"_Test Company\"\n\t\t}\n\t\tfrappe.get_doc(test_record.copy()).insert()\n\t\t\n\t\titem_price = frappe.get_doc({\n\t\t\t\"doctype\": \"Item Price\",\n\t\t\t\"price_list\": \"_Test Price List 2\",\n\t\t\t\"item_code\": \"_Test FG Item 2\",\n\t\t\t\"price_list_rate\": 100\n\t\t})\n\t\t\n\t\titem_price.insert(ignore_permissions=True)\n\n\t\targs = frappe._dict({\n\t\t\t\"item_code\": \"_Test FG Item 2\",\n\t\t\t\"company\": \"_Test Company\",\n\t\t\t\"price_list\": \"_Test Price List\",\n\t\t\t\"currency\": \"_Test Currency\",\n\t\t\t\"doctype\": \"Sales Order\",\n\t\t\t\"conversion_rate\": 1,\n\t\t\t\"price_list_currency\": \"_Test Currency\",\n\t\t\t\"plc_conversion_rate\": 1,\n\t\t\t\"order_type\": \"Sales\",\n\t\t\t\"customer\": \"_Test Customer\",\n\t\t\t\"name\": None\n\t\t})\n\t\tdetails = get_item_details(args)\n\t\tself.assertEquals(details.get(\"margin_type\"), \"Percentage\")\n\t\tself.assertEquals(details.get(\"margin_rate_or_amount\"), 10)\n\n\t\tfrappe.db.sql(\"delete from `tabPricing Rule`\")\n\n\tdef test_pricing_rule_for_variants(self):\n\t\tfrom erpnext.stock.get_item_details import get_item_details\n\t\tfrom frappe import MandatoryError\n\n\t\tfrappe.db.sql(\"delete from `tabPricing Rule`\")\n\n\t\tif not frappe.db.exists(\"Item\", \"Test Variant PRT\"):\n\t\t\tfrappe.get_doc({\n\t\t\t\t\"doctype\": \"Item\",\n\t\t\t\t\"item_code\": \"Test Variant PRT\",\n\t\t\t\t\"item_name\": \"Test Variant PRT\",\n\t\t\t\t\"description\": \"Test Variant PRT\",\n\t\t\t\t\"item_group\": \"_Test Item Group\",\n\t\t\t\t\"is_stock_item\": 1,\n\t\t\t\t\"variant_of\": \"_Test Variant Item\",\n\t\t\t\t\"default_warehouse\": \"_Test Warehouse - _TC\",\n\t\t\t\t\"stock_uom\": \"_Test UOM\",\n\t\t\t\t\"attributes\": [\n\t\t\t\t\t{\n\t\t\t\t\t \"attribute\": \"Test Size\",\n\t\t\t\t\t \"attribute_value\": \"Medium\"\n\t\t\t\t\t}\n\t\t\t\t],\n\t\t\t}).insert()\n\n\t\tfrappe.get_doc({\n\t\t\t\"doctype\": \"Pricing Rule\",\n\t\t\t\"title\": \"_Test Pricing Rule 1\",\n\t\t\t\"apply_on\": \"Item Code\",\n\t\t\t\"item_code\": \"_Test Variant Item\",\n\t\t\t\"selling\": 1,\n\t\t\t\"price_or_discount\": \"Discount Percentage\",\n\t\t\t\"price\": 0,\n\t\t\t\"discount_percentage\": 7.5,\n\t\t\t\"company\": \"_Test Company\"\n\t\t}).insert()\n\n\t\targs = frappe._dict({\n\t\t\t\"item_code\": \"Test Variant PRT\",\n\t\t\t\"company\": \"_Test Company\",\n\t\t\t\"price_list\": \"_Test Price List\",\n\t\t\t\"currency\": \"_Test Currency\",\n\t\t\t\"doctype\": \"Sales Order\",\n\t\t\t\"conversion_rate\": 1,\n\t\t\t\"price_list_currency\": \"_Test Currency\",\n\t\t\t\"plc_conversion_rate\": 1,\n\t\t\t\"order_type\": \"Sales\",\n\t\t\t\"customer\": \"_Test Customer\",\n\t\t\t\"name\": None\n\t\t})\n\n\t\tdetails = get_item_details(args)\n\t\tself.assertEquals(details.get(\"discount_percentage\"), 7.5)\n\n\t\t# add a new pricing rule for that item code, it should take priority\n\t\tfrappe.get_doc({\n\t\t\t\"doctype\": \"Pricing Rule\",\n\t\t\t\"title\": \"_Test Pricing Rule 2\",\n\t\t\t\"apply_on\": \"Item Code\",\n\t\t\t\"item_code\": \"Test Variant PRT\",\n\t\t\t\"selling\": 1,\n\t\t\t\"price_or_discount\": \"Discount Percentage\",\n\t\t\t\"price\": 0,\n\t\t\t\"discount_percentage\": 17.5,\n\t\t\t\"company\": \"_Test Company\"\n\t\t}).insert()\n\n\t\tdetails = get_item_details(args)\n\t\tself.assertEquals(details.get(\"discount_percentage\"), 17.5)\n\n\tdef test_pricing_rule_for_stock_qty(self):\n\t\tfrappe.db.sql(\"delete from `tabPricing Rule`\")\n\n\t\ttest_record = {\n\t\t\t\"doctype\": \"Pricing Rule\",\n\t\t\t\"title\": \"_Test Pricing Rule\",\n\t\t\t\"apply_on\": \"Item Code\",\n\t\t\t\"item_code\": \"_Test Item\",\n\t\t\t\"selling\": 1,\n\t\t\t\"price_or_discount\": \"Discount Percentage\",\n\t\t\t\"price\": 0,\n\t\t\t\"min_qty\": 5,\n\t\t\t\"max_qty\": 7,\n\t\t\t\"discount_percentage\": 17.5,\n\t\t\t\"company\": \"_Test Company\"\n\t\t}\n\t\tfrappe.get_doc(test_record.copy()).insert()\n\n\t\tif not frappe.db.get_value('UOM Conversion Detail',\n\t\t\t{'parent': '_Test Item', 'uom': 'box'}):\n\t\t\titem = frappe.get_doc('Item', '_Test Item')\n\t\t\titem.append('uoms', {\n\t\t\t\t'uom': 'Box',\n\t\t\t\t'conversion_factor': 5\n\t\t\t})\n\t\t\titem.save(ignore_permissions=True)\n\n\t\t# With pricing rule\n\t\tso = make_sales_order(item_code=\"_Test Item\", qty=1, uom=\"Box\", do_not_submit=True)\n\t\tso.items[0].price_list_rate = 100\n\t\tso.submit()\n\t\tso = frappe.get_doc('Sales Order', so.name)\n\t\tself.assertEquals(so.items[0].discount_percentage, 17.5)\n\t\tself.assertEquals(so.items[0].rate, 82.5)\n\n\t\t# Without pricing rule\n\t\tso = make_sales_order(item_code=\"_Test Item\", qty=2, uom=\"Box\", do_not_submit=True)\n\t\tso.items[0].price_list_rate = 100\n\t\tso.submit()\n\t\tso = frappe.get_doc('Sales Order', so.name)\n\t\tself.assertEquals(so.items[0].discount_percentage, 0)\n\t\tself.assertEquals(so.items[0].rate, 100)\n\n\tdef test_pricing_rule_with_margin_and_discount(self):\n\t\tfrappe.delete_doc_if_exists('Pricing Rule', '_Test Pricing Rule')\n\t\tmake_pricing_rule(selling=1, margin_type=\"Percentage\", margin_rate_or_amount=10)\n\t\tsi = create_sales_invoice(do_not_save=True)\n\t\tsi.items[0].price_list_rate = 1000\n\t\tsi.insert(ignore_permissions=True)\n\n\t\titem = si.items[0]\n\t\tself.assertEquals(item.rate, 1100)\n\t\tself.assertEquals(item.margin_rate_or_amount, 10)\n\n\t\t# With discount\n\t\titem.discount_percentage = 10\n\t\tsi.save()\n\t\titem = si.items[0]\n\t\tself.assertEquals(item.rate, 990)\n\t\tself.assertEquals(item.discount_percentage, 10)\n\t\tfrappe.db.sql(\"delete from `tabPricing Rule`\")\n\ndef make_pricing_rule(**args):\n\targs = frappe._dict(args)\n\n\tdoc = frappe.get_doc({\n\t\t\"doctype\": \"Pricing Rule\",\n\t\t\"title\": args.title or \"_Test Pricing Rule\",\n\t\t\"company\": args.company or \"_Test Company\",\n\t\t\"apply_on\": args.apply_on or \"Item Code\",\n\t\t\"item_code\": args.item_code or \"_Test Item\",\n\t\t\"applicable_for\": args.applicable_for,\n\t\t\"selling\": args.selling or 0,\n\t\t\"buying\": args.buying or 0,\n\t\t\"min_qty\": args.min_qty or 0.0,\n\t\t\"max_qty\": args.max_qty or 0.0,\n\t\t\"price_or_discount\": args.price_or_discount or \"Discount Percentage\",\n\t\t\"discount_percentage\": args.discount_percentage or 0.0,\n\t\t\"price\": args.price or 0.0,\n\t\t\"margin_type\": args.margin_type,\n\t\t\"margin_rate_or_amount\": args.margin_rate_or_amount or 0.0\n\t}).insert(ignore_permissions=True)\n\n\tapply_on = doc.apply_on.replace(' ', '_').lower()\n\tif args.get(apply_on) and apply_on != \"item_code\":\n\t\tdoc.db_set(apply_on, args.get(apply_on))\n\n\tapplicable_for = doc.applicable_for.replace(' ', '_').lower()\n\tif args.get(applicable_for):\n\t\tdoc.db_set(applicable_for, args.get(applicable_for))","sub_path":"python/erpnext/2017/8/test_pricing_rule.py","file_name":"test_pricing_rule.py","file_ext":"py","file_size_in_byte":9368,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"257167137","text":"\nimport os\nimport datetime as dt\n\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\n\n\"\"\"\nmy_assets_selection: according to OI_chg or Vol_chg or fundamental change\nsymbols_selection: according to the most traded/OI contract\nmethods:\n get_backtest_cc_info(asset)\n get_backtest_data_cc(asset)\n get_backtest_data_adjusted_cc(asset)\n \n\"\"\"\ndict_asset_exchange = {\n 'cu': 'SHFE', 'al': 'SHFE', 'zn': 'SHFE', 'pb': 'SHFE',\n 'ni': 'SHFE', 'sn': 'SHFE', 'au': 'SHFE', 'ag': 'SHFE',\n 'rb': 'SHFE', 'wr': 'SHFE', 'hc': 'SHFE', 'sc': 'SHFE',\n 'fu': 'SHFE', 'bu': 'SHFE', 'ru': 'SHFE', 'sp': 'SHFE',\n 'm': 'DCE', 'y': 'DCE', 'a': 'DCE', 'b': 'DCE', 'p': 'DCE',\n 'c': 'DCE', 'cs': 'DCE', 'jd': 'DCE', 'bb': 'DCE', 'fb': 'DCE',\n 'l': 'DCE', 'v': 'DCE', 'eg': 'DCE', 'pp': 'DCE', 'j': 'DCE',\n 'jm': 'DCE', 'i': 'DCE', 'SR': 'CZCE', 'CF': 'CZCE',\n 'CY': 'CZCE', 'ZC': 'CZCE', 'FG': 'CZCE', 'TA': 'CZCE',\n 'MA': 'CZCE', 'WH': 'CZCE', 'PM': 'CZCE', 'RI': 'CZCE',\n 'LR': 'CZCE', 'JR': 'CZCE', 'RS': 'CZCE', 'OI': 'CZCE',\n 'RM': 'CZCE', 'SF': 'CZCE', 'SM': 'CZCE', 'AP': 'CZCE',\n 'CJ': 'CZCE'\n }\n\ndict_contract_size = {\n 'cu': 5, 'al': 5, 'zn': 5, 'pb': 5, 'ni': 1,\n 'sn': 1, 'au': 1000, 'ag': 15, 'rb': 10,\n 'wr': 10, 'hc': 10, 'sc': 1000, 'fu': 10,\n 'bu': 10, 'ru': 10, 'sp': 10, 'm': 10,\n 'y': 10, 'a': 10, 'b': 10, 'p': 10, 'c': 10,\n 'cs': 10, 'jd': 10, 'bb': 500, 'fb': 500,\n 'l': 5, 'v': 5, 'eg': 10, 'pp': 5, 'j': 100,\n 'jm': 60, 'i': 100, 'SR': 10, 'CF': 5, 'CY': 5,\n 'ZC': 100, 'FG': 20, 'TA': 5, 'MA': 10, 'WH': 20,\n 'PM': 50, 'RI': 20, 'LR': 20, 'JR': 20, 'RS': 10,\n 'OI': 10, 'RM': 10, 'SF': 5, 'SM': 5, 'AP': 10,\n 'CJ': 5\n }\n\n# data_path = 'D://BaiduNetdiskDownload//Data'\ndata_path = 'C://Users//nealz//Documents//Data'\n\n\ndef my_assets_selection():\n \"\"\"\n +-为多空方向?加入基本面信息?\n :return:\n \"\"\"\n dict_assets = {\n 'long': ['rb'],\n 'short': ['MA']\n }\n return dict_assets\n\n\ndef symbols_selection(dict_assets):\n \"\"\"\n\n :param dict_assets:\n :return:\n \"\"\"\n dict_symbols = {\n 'long': ['SHFE.rb1905'],\n 'short': ['CZCE.MA1905']\n }\n return dict_symbols\n\n\ndef get_backtest_cc_info(asset):\n str_exchange = dict_asset_exchange[asset]\n df_cc = pd.read_excel(data_path + '//' + str_exchange + '_Hist//' + asset + '.xlsx')\n df_cc.reset_index(inplace=True)\n if '日期' not in df_cc.columns.tolist():\n df_cc.drop(columns='index', inplace=True)\n df_cc.rename(columns={'Unnamed: 0': 'index'}, inplace=True)\n df_cc.loc[:, '日期'] = df_cc.loc[:, 'index'].apply(lambda x: str(x))\n df_cc.set_index(keys='日期', inplace=True)\n ls_idx = df_cc.index.tolist()\n ls_periods = [np.nan] * len(ls_idx)\n for num in range(1, len(ls_idx)):\n if df_cc.cc0[num - 1] != df_cc.cc0[num]:\n ls_periods[num] = ls_idx[num]\n\n ls_periods_start = [x for x in ls_periods if type(x) == str]\n ls_periods_end = [ls_idx[ls_idx.index(x)-1] for x in ls_periods_start]\n ls_periods_end = ls_periods_end[1:]\n ls_periods_end.append(ls_idx[-1])\n\n ls_contracts = list(df_cc.cc0.dropna().unique())\n dict_info = {\n 'contracts': ls_contracts,\n 'periods_start': ls_periods_start,\n 'periods_end': ls_periods_end,\n 'df_cc': df_cc\n }\n\n return dict_info\n\n\ndef get_backtest_data_cc(asset):\n str_exchange = dict_asset_exchange[asset]\n dict_info = get_backtest_cc_info(asset)\n df_cc0 = dict_info['df_cc'].loc[:, 'cc0']\n\n if str_exchange == 'DCE':\n df_tmp = pd.read_csv(data_path + '//' + str_exchange +'_Hist//' + asset + '_hist.csv',\n encoding='gbk')\n df_tmp.loc[:, '日期'] = df_tmp.loc[:, '日期'].apply(lambda x: str(int(x)))\n df_tmp.set_index(['日期', '合约'], inplace=True)\n list_tuple_cc = [(x, df_cc0[x]) for x in df_cc0.index.tolist()]\n df_data_cc = df_tmp.reindex(list_tuple_cc)\n else:\n if str_exchange == 'SHFE':\n df_tmp = pd.read_excel(data_path + '//' + str_exchange + '_Hist//' + 'SHFE_2009_2018.xlsx')\n df_tmp.loc[:, '日期'] = df_tmp.loc[:, '日期'].apply(lambda x: str(x))\n df_tmp.set_index(['日期', '合约'], inplace=True)\n list_tuple_cc = [(x, df_cc0[x]) for x in df_cc0.index.tolist()]\n df_data_cc = df_tmp.reindex(list_tuple_cc)\n elif str_exchange == 'CZCE':\n df_tmp = pd.read_excel(data_path + '//' + str_exchange + '_Hist//' + 'CZCE_2015_2018.xlsx')\n df_tmp.loc[:, '交易日期'] = df_tmp.loc[:, '交易日期'].apply(lambda x: str(x))\n df_tmp.set_index(['交易日期', '品种代码'], inplace=True)\n list_tuple_cc = [(x, df_cc0[x]) for x in df_cc0.index.tolist()]\n df_data_cc = df_tmp.reindex(list_tuple_cc)\n\n return df_data_cc\n\n\ndef get_backtest_data_adjusted_cc(asset):\n # TODO\n df_data_adjusted_cc = pd.DataFrame()\n return df_data_adjusted_cc\n\n\nif __name__ == '__main__':\n asset = 'rb'\n # asset = 'SR'\n # asset = 'j'\n dict_info = get_backtest_cc_info(asset)\n df_data = get_backtest_data_cc(asset)\n print(df_data.head(20))\n","sub_path":"Version2/trading_selection.py","file_name":"trading_selection.py","file_ext":"py","file_size_in_byte":5342,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"92075193","text":"\nfrom PIL import Image\n\nimport os.path\n\ndef findSecret(file):\n # 10 x 10 pixel grosse felder\n os.chdir(\"Images\")\n message = \"Nein, gar nicht. Es war ganz leicht.\"\n im = Image.open(file, 'r')\n\n width, height = im.size\n pixels = []\n for x in range(int(width/2)):\n\n pixels.append((im.getpixel((x, height/2))))\n\n print(pixels)\n for pixel in pixels:\n count = 0\n for toTest in pixels:\n if pixel == toTest:\n count +=1\n if count < 10:\n pixels.remove(pixel)\n\n print(pixels) #geeeht alles noch nich\n\n\nif __name__ == '__main__':\n filename = \"mi.png\"\n findSecret(filename)","sub_path":"pyprog/UEB_9/aufg1.py","file_name":"aufg1.py","file_ext":"py","file_size_in_byte":661,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"612268704","text":"import sys\nsys.stdin = open('input.txt','r')\n\nN = int(input())\nspec = []\nfor _ in range(N):\n spec.append(list(map(int,input().split())))\n# for a in spec : print(*a)\n\ntarget = N // 2\nnumli = [a for a in range(1,N+1)]\ncombset = set()\nQ = [(0,set())]\n\nwhile Q :\n cnt , tempset = Q.pop(0)\n if cnt == target :\n combset.add(tuple(tempset))\n continue\n else :\n for num in numli :\n if num not in tempset:\n Q.append((cnt+1, tempset | {num}))\n# print(combset)\n\ndef permutation(memli):\n result = 0\n for i in memli:\n for j in memli:\n result += spec[i-1][j-1]\n return result\n\nminvs = 999999999\n# teamset = set()\nwhile combset and minvs:\n tA = combset.pop()\n # if tuple(tA) not in teamset :\n tB = tuple(set(numli) - set(tA))\n # teamset.add(tuple(tA))\n combset.remove(tuple(tB))\n Ateamworks = permutation(tA)\n Bteamworks = permutation(tB)\n # print(Ateamworks, ' vs ', Bteamworks)\n vs = abs(Bteamworks - Ateamworks)\n # print(vs)\n if vs < minvs : minvs = vs\nprint(minvs)\n","sub_path":"ForNovTest/BeakJoon/14889.py","file_name":"14889.py","file_ext":"py","file_size_in_byte":1076,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"597209616","text":"# -*- coding: utf8 -*-\nfrom django.db import models\n\n\nclass SignUp(models.Model):\n first_name = models.CharField(max_length=120, null=True, blank=True)\n last_name = models.CharField(max_length=120, null=True, blank=True)\n email = models.EmailField()\n timestramp = models.DateTimeField(auto_now_add=True, auto_now=False)\n updated = models.DateTimeField(auto_now_add=False, auto_now=True)\n boobs = models.BooleanField(default=True, verbose_name=\"В лесу родилась елочка, в лесу она росла.\")\n\n def __unicode__(self):\n return self.email\n","sub_path":"home/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":592,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"569095130","text":"# -*- coding:utf-8 -*-\nfrom bs4 import BeautifulSoup\nimport requests\nfrom sqlalchemy import Column, String, Integer, CHAR\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy.orm import sessionmaker\nBase = declarative_base()\nclass Baike(Base):\n __tablename__=\"baike\"\n id = Column(Integer, primary_key=True)\n bai_ke_name = Column(CHAR(255))\n bai_ke_content = Column(String(1024))\n\nclass Crawl(Baike):\n def __init__(self):\n self.root_url = \"http://baike.baidu.com/subview/14806/8904138.html\"\n self.db_url = \"mysql://root@localhost:3306/test?charset=utf8\"\n\n def get_content(self, url):\n r = requests.get(url)\n if r.status_code != 200:\n return -1\n else:\n html = r.content\n return html\n\n\n def parse_content(self,content):\n html_doc = content\n Soup = BeautifulSoup(html_doc, \"lxml\", from_encoding=\"utf-8\")\n all_content = Soup.find(\"h2\")\n name_1 = all_content.get_text()\n describe = Soup.find(\"div\", class_=\"para\").get_text()\n return name_1, describe\n\n def mysql_insert(self, id, name, content):\n db = self.db_url\n engine = create_engine(db, echo=True)\n DBSession = sessionmaker(bind=engine)\n session = DBSession()\n new_Baike = Baike(id=id, bai_ke_name=name, bai_ke_content=content)\n session.add(new_Baike)\n session.commit()\n session.close()\n\n\nif __name__ == \"__main__\":\n Crawler = Crawl()\n url = Crawler.root_url\n html = Crawler.get_content(url)\n content = Crawler.parse_content(html)\n id = 1\n name = content[0]\n content = content[1]\n print(name)\n print(content)\n\n","sub_path":"001/001Baike.py","file_name":"001Baike.py","file_ext":"py","file_size_in_byte":1731,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"31057577","text":"# -*- coding: utf-8 -*-\n'''\n\tUsed both Class based Views and Function based views\n\tfor demonstration purpose.\n\t\n\tPlease note that, the code is written with optimization \n\tfor time & effort as this is a demonstration\n\twork only.\n'''\nfrom __future__ import unicode_literals\n\nfrom django.shortcuts import render, redirect, get_object_or_404\nfrom django.core.paginator import Paginator, EmptyPage, PageNotAnInteger\nfrom django.http import HttpResponse\nfrom django.views.generic import TemplateView,ListView\nfrom django.views.generic.detail import DetailView\nfrom django.views.generic.edit import CreateView, UpdateView, DeleteView\nfrom django.core.urlresolvers import reverse_lazy, reverse\nfrom django.views.decorators.csrf import csrf_protect\nfrom models import Genre as GenreModel\nfrom models import Track as TrackModel\nfrom models import UserRating\nfrom .forms import GenreForm, TrackForm\nimport json\nimport simplejson\n\n\nfrom django.contrib.auth.decorators import login_required\n\n\n@login_required(login_url='/login/')\ndef index(request):\n\tcontext = {}\n\treturn render(request, 'music/index.html', context)\n\n\n##############################\n# #\n# Class based views #\n# #\n##############################\n\n#create genre\nclass GenreAdd(CreateView):\n\ttemplate_name = \"music/genre_form.html\"\n\tform_class = GenreForm\n\n\tdef form_valid(self, form):\n\t\treturn super(GenreAdd, self).form_valid(form)\n\n\tdef get_success_url(self):\n\t\treturn reverse('genres')\n\n#read genre\nclass Genre(ListView):\n\tmodel = GenreModel\n\tpaginate_by = 3\n\ttemplate_name='music/genres.html'\n\n\tdef get_queryset(self):\n\t\tqs = super(Genre, self).get_queryset().order_by('created_at')\n\t\tqs = qs.reverse()\n\t\treturn qs\n\n#update genre\nclass GenreUpdate(UpdateView):\n\tmodel = GenreModel\n\tform_class = GenreForm\n\ttemplate_name = \"music/genre_update.html\"\n\n\tdef get_success_url(self):\n\t\treturn reverse('genre_detail', kwargs={\n\t\t\t'pk': self.object.pk,\n\t\t})\n\n#delete genre\nclass GenreDelete(DeleteView):\n\tmodel = GenreModel\n\ttemplate_name = \"music/confirm_delete.html\"\n\tsuccess_url = reverse_lazy('genres')\n\nclass GenreDetail(DetailView):\n\tmodel = GenreModel\n\n###################################\n# #\n#\tFunction base views #\n# #\n###################################\n\n@csrf_protect\n@login_required\ndef add_tracks(request):\n\tcontext = {}\n\tif request.method == \"POST\":\n\t\tdata = json.loads(request.body)\n\t\ttitle, genre_string, rating = data.get('title', ''), data.get('genre', ''), data.get('rating', '')\n\t\tgenre_names = genre_string.split(',')\n\t\ttrack = TrackModel(title=title, rating=rating)\n\t\ttrack.save()\n\t\tfor genre_name in genre_names:\n\t\t\tgenre_obj = GenreModel.objects.get(name=genre_name)\n\t\t\ttrack.genre.add(genre_obj)\n\t\tcontext['status'] = True\n\treturn HttpResponse(simplejson.dumps(context), content_type=\"application/json\")\n\n@login_required\ndef get_tracks(request):\n\tcontext = {}\n\ttotal_tracks = TrackModel.objects.all().order_by('created_at').reverse()\n\tpaginator = Paginator(total_tracks, 3)\n\tpage = request.GET.get('page', None)\n\ttracks = None\n\ttry:\n\t\ttracks = paginator.page(page)\n\texcept PageNotAnInteger:\n\t\ttracks = paginator.page(1)\n\texcept EmptyPage:\n\t\ttracks = paginator.page(paginator.num_pages)\n\tgenres = GenreModel.objects.all()\n\tcontext['genres'] = genres\n\tcontext['tracks'] = tracks\n\tcontext['is_paginated'] = True\n\treturn render(request, 'music/tracks.html', context)\n\ndef track_detail(request, pk):\n\tcontext = {}\n\tif request.method == \"GET\":\n\t\ttrack = TrackModel.objects.get(id=pk)\n\t\tcontext['track'] = track\n\t\ttry:\n\t\t\tuserrating = UserRating.objects.get(user=request.user, track=track)\n\t\t\tcontext['user_rating'] = userrating.rating\n\t\texcept UserRating.DoesNotExist:\n\t\t\tcontext['user_rating'] = \"NA\"\n\t\treturn render(request, 'music/track_detail.html', context)\n\n@csrf_protect\n@login_required\ndef update_track(request, id):\n\tcontext = {}\n\tcurrent_user = request.user\n\ttrack = TrackModel.objects.get(id=id)\n\tgenres = GenreModel.objects.all()\n\trating, title, genre_string = request.POST.get('rating', ''), request.POST.get('title', ''), request.POST.get('updated_genre', '')\n\ttry:\n\t\ttrack.title = title\n\t\tuserrating = UserRating.objects.get_or_create(user=current_user, track=track)\n\n\t\tif userrating[1]:\n\t\t\ttrack.rating = (track.rating + int(rating)) / (track.votes + 1)\n\t\t\ttrack.votes = track.votes + 1\n\t\telse:\n\t\t\tnew_rating = (track.rating * track.votes) - float(userrating[0].rating) + float(rating)\n\t\t\ttrack.rating = new_rating / track.votes\n\n\t\ttrack.save()\n\n\t\tuserrating[0].rating = int(rating)\n\t\tuserrating[0].save()\n\n\t\told_genres = track.genre.all()\n\t\told_genres_list = [genre.name for genre in old_genres]\n\t\tgenres_list = genre_string.split(',')\n\t\tadd_diff = list(set(genres_list) - set(old_genres_list))\n\n\t\tfor genre in add_diff:\n\t\t\tgenre_obj = GenreModel.objects.get(name=genre)\n\t\t\ttrack.genre.add(genre_obj)\n\t\trem_diff = list(set(old_genres_list) - set(genres_list))\n\n\t\tfor genre in rem_diff:\n\t\t\tgenre_obj = GenreModel.objects.get(name=genre)\n\t\t\ttrack.genre.remove(genre_obj)\n\t\treturn redirect('/tracks/')\n\texcept Exception as e:\n\t\tcontext = {'track': track, 'genres':genres}\n\t\treturn render(request, 'music/edit_track.html', context)\n\n@login_required\ndef delete_track(request, id, template_name='music/confirm_delete.html'):\n\ttrack = get_object_or_404(TrackModel, id=id) \n\tif request.method=='POST':\n\t\ttrack.delete()\n\t\treturn redirect('tracks')\n\treturn render(request, template_name, {'object':track})\n\n@login_required\ndef edit_track(request, track_id):\n\ttrack = TrackModel.objects.get(id=track_id)\n\tgenres = GenreModel.objects.all()\n\tcontext = {'track': track, 'genres':genres}\n\ttry:\n\t\tuserrating = UserRating.objects.get(user=request.user, track=track)\n\t\tcontext['user_rating'] = userrating.rating\n\texcept UserRating.DoesNotExist:\n\t\tcontext['user_rating'] = 5\n\treturn render(request, 'music/edit_track.html', context)\n","sub_path":"music/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":5936,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"267171562","text":"from __future__ import print_function\r\nfrom __future__ import absolute_import\r\nfrom __future__ import division\r\n\r\nimport copy\r\n\r\nfrom compas.utilities import pairwise\r\n\r\nfrom compas.utilities import geometric_key\r\nfrom compas.utilities import reverse_geometric_key\r\n\r\n__all__ = [\r\n 'mesh_unweld_vertices',\r\n 'mesh_weld',\r\n 'meshes_join',\r\n 'meshes_join_and_weld',\r\n]\r\n\r\ndef mesh_unweld_vertices(mesh, fkey, where=None):\r\n \"\"\"Unweld a face of the mesh.\r\n\r\n Parameters\r\n ----------\r\n mesh : Mesh\r\n A mesh object.\r\n fkey : hashable\r\n The identifier of a face.\r\n where : list (None)\r\n A list of vertices to unweld.\r\n Default is to unweld all vertices of the face.\r\n\r\n Examples\r\n --------\r\n .. plot::\r\n :include-source:\r\n\r\n import compas\r\n\r\n from compas.datastructures import Mesh\r\n from compas.plotters import MeshPlotter\r\n from compas.geometry import subtract_vectors\r\n\r\n mesh = Mesh.from_obj(compas.get('faces.obj'))\r\n\r\n vertices = set(mesh.vertices())\r\n\r\n fkey = 12\r\n where = mesh.face_vertices(fkey)[0:1]\r\n centroid = mesh.face_centroid(fkey)\r\n\r\n face = mesh.unweld_vertices(fkey, where)\r\n\r\n for key in face:\r\n if key in vertices:\r\n continue\r\n xyz = mesh.vertex_coordinates(key)\r\n v = subtract_vectors(centroid, xyz)\r\n mesh.vertex[key]['x'] += 0.3 * v[0]\r\n mesh.vertex[key]['y'] += 0.3 * v[1]\r\n mesh.vertex[key]['z'] += 0.3 * v[2]\r\n\r\n plotter = MeshPlotter(mesh, figsize=(10, 7))\r\n\r\n plotter.draw_vertices()\r\n plotter.draw_faces(text={fkey: fkey for fkey in mesh.faces()})\r\n\r\n plotter.show()\r\n\r\n \"\"\"\r\n face = []\r\n vertices = mesh.face_vertices(fkey)\r\n\r\n if not where:\r\n where = vertices\r\n\r\n for u, v in pairwise(vertices + vertices[0:1]):\r\n if u in where:\r\n x, y, z = mesh.vertex_coordinates(u)\r\n u = mesh.add_vertex(x=x, y=y, z=z)\r\n if u in where or v in where:\r\n mesh.halfedge[v][u] = None\r\n face.append(u)\r\n\r\n mesh.add_face(face, fkey=fkey)\r\n\r\n return face\r\n\r\ndef mesh_weld(mesh, precision = None, cls = None):\r\n \"\"\"Weld vertices of a mesh within some precision distance.\r\n\r\n Parameters\r\n ----------\r\n mesh : Mesh\r\n A mesh.\r\n\r\n precision: str\r\n Tolerance distance for welding.\r\n\r\n Returns\r\n -------\r\n mesh\r\n The welded mesh.\r\n\r\n \"\"\"\r\n\r\n if cls is None:\r\n cls = type(mesh)\r\n\r\n # create vertex map based on geometric keys in dictionary without duplicates\r\n vertex_map = {geometric_key(mesh.vertex_coordinates(vkey), precision): vkey for vkey in mesh.vertices()}\r\n # list vertices with coordinates\r\n vertices = [reverse_geometric_key(geom_key) for geom_key in vertex_map.keys()]\r\n # reorder vertex keys in vertex map\r\n vertex_map = {geom_key: i for i, geom_key in enumerate(vertex_map.keys())}\r\n # modify vertex indices in the faces\r\n faces = [ [vertex_map[geometric_key(mesh.vertex_coordinates(vkey), precision)] for vkey in mesh.face_vertices(fkey)] for fkey in mesh.faces()]\r\n\r\n return cls.from_vertices_and_faces(vertices, faces)\r\n\r\ndef meshes_join_and_weld(meshes, precision = None, cls = None):\r\n \"\"\"Join and and weld meshes within some precision distance.\r\n\r\n Parameters\r\n ----------\r\n meshes : list\r\n A list of meshes.\r\n\r\n precision: str\r\n Tolerance distance for welding.\r\n\r\n Returns\r\n -------\r\n mesh\r\n The joined and welded mesh.\r\n\r\n \"\"\"\r\n\r\n if cls is None:\r\n cls = type(meshes[0])\r\n\r\n # create vertex map based on geometric keys in dictionary without duplicates\r\n vertex_map = {geometric_key(mesh.vertex_coordinates(vkey), precision): vkey for mesh in meshes for vkey in mesh.vertices()}\r\n # list vertices with coordinates\r\n vertices = [reverse_geometric_key(geom_key) for geom_key in vertex_map.keys()]\r\n # reorder vertex keys in vertex map\r\n vertex_map = {geom_key: i for i, geom_key in enumerate(vertex_map.keys())}\r\n # modify vertex indices in the faces\r\n faces = [ [vertex_map[geometric_key(mesh.vertex_coordinates(vkey), precision)] for vkey in mesh.face_vertices(fkey)] for mesh in meshes for fkey in mesh.faces()]\r\n\r\n return cls.from_vertices_and_faces(vertices, faces)\r\n\r\ndef meshes_join(meshes, cls = None):\r\n \"\"\"Join meshes without welding.\r\n\r\n Parameters\r\n ----------\r\n meshes : list\r\n A list of meshes.\r\n\r\n Returns\r\n -------\r\n mesh\r\n The joined mesh.\r\n\r\n \"\"\"\r\n\r\n if cls is None:\r\n cls = type(meshes[0])\r\n\r\n vertices = []\r\n faces = []\r\n\r\n for mesh in meshes:\r\n # create vertex map based on geometric keys in dictionary with duplicates\r\n vertex_map = ({vkey: len(vertices) + i for i, vkey in enumerate(mesh.vertices())})\r\n # list vertices with coordinates\r\n vertices += [mesh.vertex_coordinates(vkey) for vkey in mesh.vertices()]\r\n # modify vertex indices in the faces\r\n faces += [ [vertex_map[vkey] for vkey in mesh.face_vertices(fkey)] for fkey in mesh.faces()]\r\n\r\n return cls.from_vertices_and_faces(vertices, faces)\r\n\r\n# ==============================================================================\r\n# Main\r\n# ==============================================================================\r\n\r\nif __name__ == \"__main__\":\r\n\r\n import compas\r\n\r\n from compas.datastructures import Mesh\r\n from compas.plotters import MeshPlotter\r\n from compas.geometry import subtract_vectors\r\n\r\n mesh = Mesh.from_obj(compas.get('faces.obj'))\r\n\r\n vertices = set(mesh.vertices())\r\n\r\n fkey = 12\r\n where = mesh.face_vertices(fkey)[0:2]\r\n centroid = mesh.face_centroid(fkey)\r\n\r\n face = mesh.unweld_vertices(fkey, where)\r\n\r\n for key in face:\r\n if key in vertices:\r\n continue\r\n xyz = mesh.vertex_coordinates(key)\r\n v = subtract_vectors(centroid, xyz)\r\n mesh.vertex[key]['x'] += 0.3 * v[0]\r\n mesh.vertex[key]['y'] += 0.3 * v[1]\r\n mesh.vertex[key]['z'] += 0.3 * v[2]\r\n\r\n plotter = MeshPlotter(mesh, figsize=(10, 7))\r\n\r\n plotter.draw_vertices()\r\n plotter.draw_faces(text={fkey: fkey for fkey in mesh.faces()})\r\n\r\n plotter.show()","sub_path":"src/compas/datastructures/mesh/operations/weld.py","file_name":"weld.py","file_ext":"py","file_size_in_byte":6335,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"392342376","text":"# -*- coding: utf-8 -*-\nimport math\nn=int(input('digite n: '))\nnum=1\nden=n\nS=num/den\nfor i in range(1,n-1,1):\n num=num+i\n den=den-i\n S=S+(num/den)\nprint(S)","sub_path":"moodledata/vpl_data/92/usersdata/203/37711/submittedfiles/atividade.py","file_name":"atividade.py","file_ext":"py","file_size_in_byte":164,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"616469274","text":"#!/usr/bin/env python\nimport os\nimport sys\n\nif __name__ == \"__main__\":\n if \"--all\" in sys.argv:\n u_dir = \"unixbench-5.1.2\"\n if not os.path.isdir(u_dir):\n u = \"https://byte-unixbench.googlecode.com/files/unixbench-5.1.2.tar.gz\"\n cmd = \"wget %s && tar xzf %s.tar.gz && \" + \\\n \"patch -p0 -i %s.patch && cd %s && make all\"\n os.system(cmd % (u, u_dir, u_dir, u_dir))\n\n g_dir = \"GChartWrapper\"\n if not os.path.isdir(g_dir):\n u = \"http://google-chartwrapper.googlecode.com/svn/trunk/%s\" % g_dir\n os.system(\"svn co %s\" % u)\n","sub_path":"bootstrap.py","file_name":"bootstrap.py","file_ext":"py","file_size_in_byte":603,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"157310245","text":"from collections import OrderedDict\nfrom enum import Enum\nfrom typing import Tuple\n\nimport torch\nfrom torch.nn import Sequential, Linear, Softmax, ReLU\n\nfrom .general import Endianness, DataTypesSize, int_to_bytes, string_to_bytes, list_to_bytes, to_int, to_string, to_list_int, to_list_float\n\n\ndef state_dict_to_bytes(state_dict: OrderedDict, endianness: Endianness = Endianness.Native):\n result = int_to_bytes(len(state_dict), endianness)\n for key in state_dict:\n result += string_to_bytes(key, endianness)\n result += tensor_to_bytes(state_dict[key], endianness)\n return result\n\n\ndef to_state_dict(value: bytes, start_index: int = 0, endianness: Endianness = Endianness.Native) -> Tuple[OrderedDict, int]:\n entries_length = to_int(value, start_index, endianness)\n total_read = DataTypesSize.Int\n\n d = OrderedDict()\n for i in range(entries_length):\n key, read_bytes = to_string(value, start_index + total_read, endianness)\n total_read += read_bytes\n tensor, read_bytes = to_tensor_float(value, start_index + total_read, endianness)\n total_read += read_bytes\n d[key] = tensor\n\n return d, total_read\n\n\ndef tensor_to_bytes(tensor_value: torch.Tensor, endianness: Endianness = Endianness.Native) -> bytes:\n return list_to_bytes(list(tensor_value.shape), endianness) + list_to_bytes(tensor_value.flatten().tolist(), endianness)\n\n\ndef to_tensor_float(value: bytes, start_index: int = 0, endianness: Endianness = Endianness.Native) -> Tuple[torch.Tensor, int]:\n shape, bytes_read = to_list_int(value, start_index, endianness)\n total_read = bytes_read\n\n items, bytes_read = to_list_float(value, start_index + total_read, endianness)\n total_read += bytes_read\n\n return torch.tensor(items).reshape(shape), total_read\n\n\ndef to_tensor_int(value: bytes, start_index: int = 0, endianness: Endianness = Endianness.Native) -> Tuple[torch.Tensor, int]:\n shape, bytes_read = to_list_int(value, start_index, endianness)\n total_read = bytes_read\n\n items, bytes_read = to_list_int(value, start_index + total_read, endianness)\n total_read += bytes_read\n\n return torch.tensor(items).reshape(shape), total_read\n\n\nclass LayerName(int, Enum):\n Linear = 0\n ReLU = 1\n Softmax = 2\n Sequential = 3\n\n\ndef layout_to_bytes(model: Sequential):\n if not isinstance(model, Sequential):\n raise RuntimeError(\"Works only with Sequential models\")\n\n result = int_to_bytes(len(model))\n\n for child in model.children():\n if isinstance(child, Linear):\n result += int_to_bytes(LayerName.Linear)\n result += int_to_bytes(child.in_features)\n result += int_to_bytes(child.out_features)\n elif isinstance(child, Softmax):\n result += int_to_bytes(LayerName.Softmax)\n elif isinstance(child, ReLU):\n result += int_to_bytes(LayerName.ReLU)\n else:\n raise RuntimeError(f\"Layer is not supported yet: {type(child)}\")\n\n return result\n","sub_path":"python/src/serialization/torch_serialization.py","file_name":"torch_serialization.py","file_ext":"py","file_size_in_byte":3004,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"574160352","text":"import tensorflow as tf\nfrom tqdm import tqdm\nimport pandas as pd\nimport pickle\nimport numpy as np\nimport spacy\nfrom tensorflow.keras.models import Sequential\nfrom tensorflow.keras.layers import Dense, Conv1D, MaxPooling1D, Flatten, Embedding, LSTM\nfrom tensorflow.keras.layers import Dense\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.preprocessing import LabelEncoder\nfrom tensorflow.keras.preprocessing.sequence import pad_sequences\nfrom tensorflow.keras.preprocessing.text import Tokenizer\nimport string\n\nprint(\"OKAY LETS STAAAART!!!\")\n\ndef encodeY(Y):\n '''create one-hot (dummies) for output, see also https://machinelearningmastery.com/multi-class-classification-tutorial-keras-deep-learning-library/\n encode class values as integers\n '''\n encoder = LabelEncoder()\n encoder.fit(Y)\n encoded_Y = encoder.transform(Y)\n dummy_y = tf.keras.utils.to_categorical(encoded_Y)\n return dummy_y\n\n\nAEM = '../tmpanne/fullsample/w2v_model_nr_7_window_10_size_300_negsample_15.txt'\nPATH_TO_DATA = \"../tmpanne/AEM_data/dataset_vermeer.pkl\"\n\ndf = pd.read_pickle(PATH_TO_DATA)\n\nX_train, X_test, y_train, y_test = train_test_split([t.translate(str.maketrans('', '', string.punctuation)) for t in df['text']], encodeY(df['topic'].map(int)), test_size = 0.2)\n\n# TOKENIZATION + PAD\n\n# create the tokenizer\ntokenizer = Tokenizer()\n# fit the tokenizer on the documents\ntokenizer.fit_on_texts(X_train)\n\n# sequence encode\nencoded_docs = tokenizer.texts_to_sequences(X_train)\n# pad sequences\nmax_length = max([len(s.split()) for s in X_train])\nXtrain = pad_sequences(encoded_docs, maxlen=max_length, padding='post')\n# HIER OVER NADENKEN\nencoded_docs = tokenizer.texts_to_sequences(X_test)\n# pad sequences\nXtest = pad_sequences(encoded_docs, maxlen=max_length, padding='post')\n\nembeddings_index = {}\nwith open(AEM) as f:\n numberofwordvectors, dimensions = [int(e) for e in next(f).split()]\n for line in tqdm(f):\n values = line.split()\n embeddings_index[values[0]] = np.asarray(values[1:], dtype='float32')\n\n # word = values[0]\n # coefs = np.asarray(values[1:], dtype='float32')\n # embeddings_index[word] = coefs\n\nprint('Found %s word vectors.' % len(embeddings_index))\nprint('Should be {} vectors with {} dimensions'.format(numberofwordvectors, dimensions))\n\n\n# create a weight matrix for the Embedding layer from a loaded embedding\n\ndef get_weight_matrix(embedding, vocab):\n words_not_found = 0\n total_words = 0\n DEBUG_lijstmetwoorden = []\n # total vocabulary size plus 0 for unknown words\n vocab_size = len(vocab) + 1\n # define weight matrix dimensions with all 0\n weight_matrix = np.zeros((vocab_size, 300))\n # step vocab, store vectors using the Tokenizer's integer mapping\n for word, i in tqdm(vocab.items()):\n e = embedding.get(word, None)\n if e is not None: # if we do not find the word, we do not want to replace anything but leave the zero's\n weight_matrix[i] = e\n total_words+=1\n else:\n words_not_found+=1\n DEBUG_lijstmetwoorden.append(word)\n print('Weight matrix created. For {} out of {} words, we did not have any embedding.'.format(words_not_found, total_words))\n return DEBUG_lijstmetwoorden, weight_matrix\n\n\nmissingwords, embedding_vectors = get_weight_matrix(embeddings_index, tokenizer.word_index)\n\nembedding_layer = Embedding(len(tokenizer.word_index)+1, 300, weights=[embedding_vectors], input_length=max_length, trainable=False)\nprint(\"created embedding layer\")\n\n\n\n# alternatief model\nnumberoflabels = 4\nmodel = Sequential()\nmodel.add(embedding_layer)\nmodel.add(Conv1D(128, 4, activation='relu'))\nmodel.add(MaxPooling1D(4))\nmodel.add(MaxPooling1D(4))\nmodel.add(Flatten())\nmodel.add(Dense(units=64, activation='relu'))\nmodel.add(Dense(units=numberoflabels, activation='softmax')) # voor twee categorien sigmoid, voor 1 tanh\n\nmodel.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])\nprint(model.summary())\n\n\nVALIDATION_SIZE=1000\n\nmodel.fit(Xtrain[:-VALIDATION_SIZE], y_train[:-VALIDATION_SIZE],\n epochs=30, verbose=True,\n validation_data=(Xtrain[-VALIDATION_SIZE:], y_train[-VALIDATION_SIZE:]))\n\nloss, acc = model.evaluate(Xtest, y_test, verbose=True)\nprint('Test Accuracy: %f' % (acc*100))\n","sub_path":"deeplearning.py","file_name":"deeplearning.py","file_ext":"py","file_size_in_byte":4380,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"648788260","text":"from flask_restful import Resource, reqparse\nfrom models.Hotel import HotelModel\nfrom flask_jwt_extended import jwt_required\n\n\nclass Hoteis(Resource):\n def get(self):\n return {'hoteis': [hotel.json() for hotel in HotelModel.query.all()]}\n\nclass Hotel(Resource):\n argumentos = reqparse.RequestParser()\n argumentos.add_argument('nome', type = str, required = True, help = 'The field \"nome\" cannot be blank.')\n argumentos.add_argument('estrelas', type = int, required = True, help = 'The field \"estrelas\" cannot be blank.')\n argumentos.add_argument('diaria', type = float, required = True, help = 'The field \"diaria\" cannot be blank.')\n argumentos.add_argument('cidade', type = str, required = True, help = 'The field \"cidade\" cannot be blank.')\n\n def findHotel(id):\n for hotel in hoteis:\n if hotel['id'] == id:\n return hotel\n return None\n\n def get(self, id):\n hotel = HotelModel.findHotel(id)\n if hotel:\n return hotel.json()\n \n return {'message': 'Hotel não encontrado!'}, 404 #Not Found.\n\n @jwt_required\n def post(self, id):\n if HotelModel.findHotel(id):\n return {\"message\": \"ID '{}', ja existe.\".format(id)}, 400 #Not Found.\n\n dados = Hotel.argumentos.parse_args()\n hotel = HotelModel(id, **dados)\n try:\n hotel.saveHotel()\n except:\n return {'message': 'Ocorreu um erro interno ao tentar gravar o registro.'}, 500\n\n return hotel.json()\n\n @jwt_required\n def put(self, id):\n dados = Hotel.argumentos.parse_args()\n hotelFind = HotelModel.findHotel(id)\n if hotelFind:\n hotelFind.updateHotel(**dados) #Alterando hotel\n hotelFind.saveHotel()\n return hotelFind.json(), 200 #Ok\n hotel = HotelModel(id, **dados)\n try:\n hotel.saveHotel()\n except:\n return {'message': 'Ocorreu um erro interno ao tentar gravar o registro.'}, 500 \n return hotel.json(), 201\n\n @jwt_required\n def delete(self, id):\n hotel = HotelModel.findHotel(id)\n if hotel:\n try:\n hotel.deleteHotel()\n except:\n return {'message': 'Ocorreu um erro interno ao tentar deletar o registro.'}, 500\n return {'message': 'Hotel deletado!'}\n return {'message': 'Hotel não encontrado'}, 404 \n","sub_path":"resources/hotel.py","file_name":"hotel.py","file_ext":"py","file_size_in_byte":2324,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"166993894","text":"import numpy as np\nimport tensorflow as tf\nfrom tensorflow import keras\nfrom tensorflow.keras import layers\nfrom tensorflow.keras.optimizers import SGD , Adam, RMSprop\nfrom tensorflow.keras.layers import Concatenate, Dense, LSTM, Input, concatenate\nfrom tensorflow.keras.layers import LeakyReLU \nfrom matplotlib import pyplot\nimport matplotlib.pyplot as plt\n\nimport pickle \nimport copy\nimport os\nimport gym\nfrom feedback import *\nimport time\nfrom collections import deque\nimport random\nfrom rl_dqn import DeepQNetwork\nfrom kinematic_model import Kinematic_Model\n\ntrial_no = '1'\nENV = \"CartPole-v0\"\nenv = gym.make(ENV)\n\nstate_dim = env.observation_space.shape[0]\naction_dim = env.action_space.n\n\n########################################################\n# RL Agent for Oracle\nweights_file = 'rldqn_cartpole.h5'\noracle = DeepQNetwork(state_dim, action_dim, 0.001, 0.95, 1, 0.001, 0.995 )\noracle.load_weights(weights_file)\n'''\nTEST_Episodes=10\nrewards = [] #Store rewards for graphing\n#Test the agent that was trained\n# In this section we ALWAYS use exploit don't train any more\nfor e_test in range(TEST_Episodes):\n state,done = env.reset(), False\n state = np.reshape(state, [1, state_dim])\n tot_rewards = 0\n while not done:\n env.render()\n action = oracle.test_action(state)\n KM_nstate = Kinematic_Model(state[0],action)\n nstate, reward, done, _ = env.step(action)\n nstate = np.reshape( nstate, [1, state_dim])\n print(\"KM_NS : {}\\t \\t true_NS: {} Diff {}\".format(KM_nstate, nstate[0],KM_nstate-nstate[0]))\n\n tot_rewards += reward\n #DON'T STORE ANYTHING DURING TESTING\n state = nstate\n #done: CartPole fell. \n #t_test == 209: CartPole stayed upright\n rewards.append(tot_rewards)\n print(\"episode: {}/{}, score: {}, e: {}\".format(e_test, TEST_Episodes, tot_rewards, 0))\n\nenv.close()\n\n'''\n\n\n########################################################\npause = 0 \nwhile pause == 1:\n pause = 1\n\n# This model maps an input to its next state\n# Input state\nAE_state = keras.Input(shape=(state_dim,),name=\"AE_state\")\n# 2layer neural network to predict the next state\nencoded = Dense(32,name=\"dense1_NS\")(AE_state)\nencoded = LeakyReLU(alpha=0.2,name=\"LeakyRelu1_NS\")(encoded)\nencoded = Dense(32,name=\"dense2_NS\")(encoded)\nencoded = LeakyReLU(alpha=0.2,name=\"LeakyRelu2_NS\")(encoded)\nn_state = layers.Dense(state_dim,name=\"dense3_NS\")(encoded)\nAE = keras.Model(inputs=AE_state, outputs=n_state,name=\"AE\")\n\n#print(AE.summary())\n#tf.keras.utils.plot_model(AE, to_file='AE_model_plot.png', show_shapes=True, show_layer_names=True)\n\nopt_AE = tf.keras.optimizers.RMSprop(learning_rate=0.00015)\nAE.compile(loss='mean_squared_error', optimizer=opt_AE, metrics=['mse'])\n\n\n\n\n\n# This model maps an input & action to its next state\n# Input state\ncurr_state = keras.Input(shape=(state_dim,),name=\"curr_state\")\ncurr_action = keras.Input(shape=(action_dim,),name=\"curr_action\")\n# FDM model\ncurr_state_action = concatenate([curr_state, curr_action])\nfdm_h1 = Dense(16,name=\"dense1_FDM\")(curr_state_action)\nfdm_h1 = LeakyReLU(alpha=0.2,name=\"LeakyRelu1_FDM\")(fdm_h1)\nfdm_h2 = Dense(16,name=\"dense2_FDM\")(fdm_h1)\nfdm_h2 = LeakyReLU(alpha=0.2,name=\"LeakyRelu2_FDM\")(fdm_h2)\nfdm_pred_state = layers.Dense(state_dim,name=\"dense3_FDM\")(fdm_h2)\nFDM = keras.Model(inputs=[curr_state,curr_action], outputs=fdm_pred_state,name=\"FDM\")\n\n#print(FDM.summary())\n#tf.keras.utils.plot_model(FDM, to_file='FDM_model_plot.png', show_shapes=True, show_layer_names=True)\n\nopt_FDM = tf.keras.optimizers.RMSprop(learning_rate=0.00015)\nFDM.compile(loss='mean_squared_error', optimizer=opt_FDM, metrics=['mse'])\n\npause = 0 \nwhile pause == 1:\n pause = 1\n\np_state = np.zeros((1,state_dim))\n# Action for FDM to sample from\nleft = np.zeros((1,action_dim))\nleft[0][0] = 1\nright = np.zeros((1,action_dim))\nright[0][1] = 1\n\nverbose = False\nimpove_FDM = True\n#Buffer for FMD online improvement\nFDM_buff_s =[]\nFDM_buff_a =[]\nFDM_buff_ns =[]\nFDM_loss = 1000.0\n\nAE_buff_s = []\nAE_buff_ns = []\nAE_loss = 1000.0\n\nfeedback_dict = {\n H_NULL: 0,\n H_UP: 0,\n H_DOWN: 0,\n H_LEFT: 1,\n H_RIGHT: 1,\n H_HOLD: 0,\n DO_NOTHING: 0\n }\n\n\n# set which game to play\nenv = gym.make('CartPole-v0')\nenv.reset()\nenv.render() # Make the environment visible\n\n# Initialise Human feedback (call render before this)\nhuman_feedback = Feedback(env)\n\nnum_steps = 3000\nsteps = 0\nEpisode = 1\ntotal_reward = []\nfeedback_rate = []\n\nstate = env.reset()\n#state = np.reshape(state, [-1, state_dim])\n\nprev_s = state\na = np.random.uniform(-1,1,action_dim)\n\nprint(\"3\")\ntime.sleep(1)\nprint(\"2\")\ntime.sleep(1)\nprint(\"1\")\ntime.sleep(1)\n\n\n\nwhile Episode < 50:\n obs, terminal = env.reset(), False\n prev_state = obs\n print(\"Episode# \",Episode)\n \n episode_rew = 0\n h_counter = 0\n t_counter = 0\n\n # Iterate over the episode\n while((not terminal) and (not human_feedback.ask_for_done()) ):\n \n env.render() # Make the environment visible\n time.sleep(0.1)\n \n # Get feedback signal\n h_fb = human_feedback.get_h()\n if h_fb == 1:\n print(\"Oracle\")\n oracle_action = oracle.test_action(np.reshape(obs, [1, state_dim]))\n h_fb = oracle_action + 3\n \n \n if (feedback_dict.get(h_fb) != 0): # if feedback is not zero i.e. is valid\n # Update policy\n #oracle_action = oracle.test_action(np.reshape(obs, [1, state_dim]))\n #h_fb = oracle_action + 3\n #print(\"Feedback\", h_fb)\n\n h_counter += 1\n\n # Get new state transition label using feedback\n state_corrected = copy.deepcopy(obs)\n if (h_fb == H_LEFT): # PUSH CART TO LEFT\n print(\"Feedback left\\t\",end =\" \")\n state_corrected = Kinematic_Model(state,0)\n #state_corrected[0] -= 0.01 # correction in pos\n #state_corrected[1] -= 0.2 # correction in vel\n #state_corrected[2] += 0.01 # correction in angle\n #state_corrected[3] += 0.27 # correction in anglar vel\n elif (h_fb == H_RIGHT):# PUSH CART TO RIGHT\n print(\"Feedback right\\t\",end =\" \")\n state_corrected = Kinematic_Model(state,1)\n #state_corrected[0] += 0.01 # correction in pos\n #state_corrected[1] += 0.2 # correction in vel\n #state_corrected[2] -= 0.01 # correction in angle\n #state_corrected[3] -= 0.27 # correction in anglar vel\n \n # Add state transition pair to demo buffer\n AE_buff_s.append(obs)\n AE_buff_ns.append(state_corrected)\n\n #print(\"State \",obs) \n #print(\"state_corrected \",state_corrected)\n pred_ns = np.reshape(state_corrected, [-1, state_dim])\n\n '''\n # Update policy (immediate)\n temp_s = np.array(p_state)\n temp_ns = np.array(state_corrected)\n history_AE = AE.fit(x=temp_s, y=temp_ns, verbose=1) \n\n\n # Train with batch from Demo buffer (if enough entries exist)\n num = len(AE_buff_s)\n print(\"num\",num)\n if(num >= 64): # batch size 64\n print(\"Training AE\")\n temp_s = np.array(AE_buff_s)\n temp_ns = np.array(AE_buff_ns)\n # do you need to run for 5 epochs\n history_AE = AE.fit(x=temp_s, y=temp_ns,batch_size=64,shuffle=True, verbose=1)\n '''\n\n else:\n # Use current policy\n print(\"Policy \\t\\t\",end =\" \")\n p_state = np.reshape(obs, [-1, state_dim])\n pred_ns = AE.predict(p_state)\n\n # Get action from ifdm\n FDM_ns_l = np.squeeze(FDM.predict([p_state,left]))\n FDM_ns_r = np.squeeze(FDM.predict([p_state,right]))\n FDM_ns_both = np.array([FDM_ns_l,FDM_ns_r])\n state_diff = np.abs(FDM_ns_both-pred_ns)\n cost = np.sum(state_diff,axis=1)\n action_from_IDM = np.argmin(cost, axis=0)\n action = action_from_IDM\n\n prev_state = obs\n obs, reward, terminal, _ = env.step(action)\n episode_rew += reward\n\n # Add state transition pair to demo buffer\n FDM_buff_s.append(prev_state)\n FDM_buff_a.append(action)\n FDM_buff_ns.append(obs)\n\n if verbose ==True:\n print(\"Curr state \",prev_state)\n print(\"True next state \",obs)\n print(\"AE pred Nstate \",pred_ns)\n print(\"FDM both next state \",FDM_ns_both)\n print(\"cost \",cost)\n print(\"partial cost \",par_cost)\n print(\"action from IDM full cost \",action_from_IDM)\n if action == 0:\n print(\"FDM left\") #0 Push cart to the left\n else:\n print(\"FDM right\") #1 Push cart to the right\n print(\"\")\n \n steps += 1\n t_counter+=1\n\n \n \n feedback_rate.append(h_counter/t_counter)\n total_reward.append(episode_rew)\n\n #print('Episode #%d Reward %d' % (Episode, episode_rew))\n print(\"## episode: {}, Reward: {}, h_counter: {}, t_counter: {}, feedback_rate: {} \".format(Episode, episode_rew,h_counter,t_counter,h_counter/t_counter))\n Episode +=1\n\n #Train Next State predictor\n # Train with batch from Demo buffer (if enough entries exist)\n num = len(AE_buff_s)\n if(num >= 64) and AE_loss > 0.0001: # batch size 64\n print(\"Training AE\")\n # do you need to run for 5 epochs\n history_AE = AE.fit(x=np.array(AE_buff_s), y=np.array(AE_buff_ns),batch_size=64,epochs=5,shuffle=True, verbose=False)\n AE_loss = history_AE.history['loss'][-1]\n print(\"AE loss\",AE_loss)\n\n\n if FDM_loss > 0.0001:\n print(\"Training FDM\")\n #Train FDM every episode,\n temp_s = np.array(FDM_buff_s)\n temp_a = np.zeros((len(FDM_buff_a),action_dim))\n for i in range(len(FDM_buff_a)):\n temp_a[i][FDM_buff_a[i]] =1\n history_FDM=FDM.fit(x=[temp_s,temp_a], y=np.array(FDM_buff_ns),epochs=5,batch_size=32, shuffle=True,verbose=False)\n FDM_loss = history_FDM.history['loss'][-1]\n print(\"FDM loss\",FDM_loss)\n\nfor i in range(Episode):\n print(\"episode: {}, Reward: {}\".format(i, total_reward[i-1]))\n\ntotal_reward =np.array(total_reward)\nrolling_average = np.convolve(total_reward, np.ones(100)/100)\nfeedback_rate = np.array(feedback_rate)\n\nplt.plot(total_reward)\nplt.plot(rolling_average, color='black')\nplt.plot(feedback_rate, color='green')\nplt.axhline(y=195, color='r', linestyle='-') #Solved Line\nplt.xlim( (0,Episode) )\nplt.ylim( (0,220) )\nplt.show()\n\n\nprint(\"Saving Reward\")\nfilename = 'log_files/'+trial_no+'/total_reward.npy'\npickle.dump(total_reward, open(filename, 'wb'))\nfilename = 'log_files/'+trial_no+'/feedback_rate.npy'\npickle.dump(feedback_rate, open(filename, 'wb'))\n\nAE.save('log_files/'+trial_no+'/AE')\nFDM.save('log_files/'+trial_no+'/FDM')\n\n'''\n\nelse: \n # save state, action, nstate\n filename = 'Data/State.npy'\n pickle.dump(s, open(filename, 'wb'))\n filename = 'Data/Action.npy'\n pickle.dump(a, open(filename, 'wb'))\n filename = 'Data/NState.npy'\n pickle.dump(ns, open(filename, 'wb'))\n filename = 'Data/Diff.npy'\n pickle.dump(d,open(filename,'wb'))\n'''","sub_path":"Cartpole/live_improve_v8a.py","file_name":"live_improve_v8a.py","file_ext":"py","file_size_in_byte":11460,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"577774077","text":"from apps.eur_basketball_spider.eur_playbyplay import *\nimport queue\nimport traceback\nfrom apps.send_error_msg import dingding_alter\nfrom apps.eur_basketball_spider.tools import *\nfrom orm_connection.eur_basketball import *\nfrom orm_connection.orm_session import MysqlSvr\nfrom common.libs.log import LogMgr\n\nlogger = LogMgr.get('eur_basketball_boxscore_live')\n\n\nclass EurLeagueSpider_boxscore(object):\n def __init__(self):\n self.data_queue_svr = queue.Queue()\n self.headers = {\n 'user_agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/78.0.3904.70 Safari/537.36',\n }\n self.get_team_id = get_team_id_box()\n self.get_player_id = get_player_id_key()\n self.get_player_shirt_number = get_player_shirt_number()\n\n def start_requests(self, data_queue, gamecode):\n while True:\n time.sleep(10)\n try:\n seasoncode = 2019\n url = 'https://live.euroleague.net/api/Boxscore?gamecode=%s&seasoncode=E%s&disp=' % (\n gamecode, seasoncode)\n logger.info(url)\n boxscore_api_res = requests.get(url, headers=self.headers)\n if boxscore_api_res.text == '':\n logger.info('box比赛未开赛。。。 %s' % str(gamecode))\n else:\n boxscore_json_dict = json.loads(boxscore_api_res.text)\n boxscore_player = boxscore_json_dict['Stats']\n team_stats_list = []\n player_stats_list = []\n belong = 1\n for index in boxscore_player:\n BkMatchTeamStats = {}\n BkMatchTeamStats['belong'] = belong\n name_en = index['Team']\n BkMatchTeamStats['team_id'] = self.get_team_id[name_en.lower()]\n BkMatchTeamStats['team_name'] = name_en\n BkMatchTeamStats['three_point_goals'] = safe_get(index,'totr.FieldGoalsMade3')\n BkMatchTeamStats['three_point_field'] = safe_get(index,'totr.three_point_field')\n BkMatchTeamStats['goals'] = safe_get(index,'totr.FieldGoalsMade2') + BkMatchTeamStats['three_point_goals']\n BkMatchTeamStats['field'] = safe_get(index,'totr.FieldGoalsAttempted2') + BkMatchTeamStats['three_point_field']\n BkMatchTeamStats['free_throw_goals'] = safe_get(index,'totr.FreeThrowsMade')\n BkMatchTeamStats['free_throw_field'] = safe_get(index,'totr.FreeThrowsAttempted')\n BkMatchTeamStats['offensive_rebounds'] = safe_get(index,'totr.OffensiveRebounds')\n BkMatchTeamStats['defensive_rebounds'] = safe_get(index,'totr.DefensiveRebounds')\n BkMatchTeamStats['rebounds'] = safe_get(index,'totr.TotalRebounds')\n BkMatchTeamStats['assists'] = safe_get(index,'totr.Assistances')\n BkMatchTeamStats['steals'] = safe_get(index,'totr.Steals')\n BkMatchTeamStats['blocks'] = safe_get(index,'totr.BlocksFavour')\n BkMatchTeamStats['turnovers'] = safe_get(index,'totr.Turnovers')\n BkMatchTeamStats['personal_fouls'] = safe_get(index,'totr.FoulsCommited')\n BkMatchTeamStats['point'] = safe_get(index,'totr.Points')\n BkMatchTeamStats['score_difference'] = safe_get(index,'totr.Valuation')\n team_stats_list.append(BkMatchTeamStats)\n belong += 1\n for i in range(2):\n for home_player in boxscore_json_dict['Stats'][i]['PlayersStats']:\n player = {}\n player['belong'] = i + 1\n player['player_name'] = translate_player_name(home_player['Player'])\n player_key = home_player['Player_ID'][1:]\n id = safe_get(self.get_player_id,player_key.strip())\n if id == 0:\n url = 'https://www.euroleague.net/competition/players/showplayer?pcode=%s&seasoncode=E2019' % player_key\n player_res = requests.get(url, headers=self.headers)\n player_tree = tree_parse(player_res)\n player['sport_id'] = 2\n try:\n player['name_en'] = player_tree.xpath('//div[@class=\"name\"]/text()')[0]\n except:\n player['name_en'] = ''\n player['key'] = player_key\n try:\n player['logo'] = player_tree.xpath('//div[@class=\"player_img-img\"]/img/@src')[0]\n except:\n player['logo'] = ''\n logger.info('没有该球员的图片...')\n try:\n player['shirt_number'] = player_tree.xpath('//span[@class=\"dorsal\"]/text()')[0]\n except:\n player['shirt_number'] = 0\n try:\n position = \\\n player_tree.xpath(\n '//div[@class=\"summary-first\"]/span[last()]/span[last()]/text()')[\n 0]\n player['position'] = position.encode('utf-8').decode('utf-8')[0]\n except:\n player['position'] = ''\n if 'Height' in \\\n player_tree.xpath('//div[@class=\"summary-second\"]/span[1]/text()')[0].split(\n ':')[0]:\n player['height'] = float(\n player_tree.xpath('//div[@class=\"summary-second\"]/span[1]/text()')[0].split(\n ':')[-1]) * 100\n time_birthday = player_tree.xpath('//div[@class=\"summary-second\"]/span[2]/text()')[\n 0]\n player['birthday'], player['age'] = time_stamp(time_birthday)\n player['nationality'] = \\\n player_tree.xpath('//div[@class=\"summary-second\"]/span[last()]/text()')[\n 0].split(\n ':')[-1]\n else:\n player['height'] = 0\n time_birthday = player_tree.xpath('//div[@class=\"summary-second\"]/span[1]/text()')[\n 0]\n player['birthday'], player['age'] = time_stamp(time_birthday)\n player['nationality'] = \\\n player_tree.xpath('//div[@class=\"summary-second\"]/span[last()]/text()')[\n 0].split(':')[-1]\n try:\n player['name_zh'] = translate_dict[player['name_en']]\n except:\n player['name_zh'] = ''\n data = {\n 'key': player['key'],\n 'name_en': player['name_en'],\n 'name_zh': player['name_zh'],\n 'sport_id': player['sport_id'],\n 'age': player['age'],\n 'birthday': player['birthday'],\n 'nationality': player['nationality'],\n 'height': player['height'],\n 'shirt_number': player['shirt_number'],\n 'position': player['position'],\n }\n spx_dev_session = MysqlSvr.get('spider_zl')\n BleagueEurBasketballPlayer.upsert(\n spx_dev_session,\n 'key',\n data\n )\n player['player_id'] = self.get_player_id[player_key.strip()]\n player['shirt_number'] = self.get_player_shirt_number[player_key.strip()]\n times = home_player['Minutes']\n if times != 'DNP':\n player['enter_ground'] = 1\n minutes = times.split(':')[0]\n seconds = times.split(':')[-1]\n if int(seconds) >= 30:\n player['minutes'] = int(minutes) + 1\n else:\n player['minutes'] = int(minutes)\n else:\n player['enter_ground'] = 0\n player['minutes'] = 0\n player['two_points_goals'] = home_player['FieldGoalsMade2']\n player['two_points_goals'] = safe_get(home_player,'FieldGoalsMade2')\n player['two_points_total'] = safe_get(home_player,'FieldGoalsAttempted2')\n player['three_point_goals'] = safe_get(home_player,'FieldGoalsMade3')\n player['three_point_field'] = safe_get(home_player,'FieldGoalsAttempted3')\n player['goals'] = int(player['two_points_goals']) + int(player['three_point_goals'])\n player['field'] = int(player['two_points_total']) + int(player['three_point_field'])\n player['free_throw_goals'] = safe_get(home_player,'FreeThrowsMade')\n player['free_throw_field'] = safe_get(home_player,'FreeThrowsAttempted')\n player['offensive_rebounds'] = safe_get(home_player,'OffensiveRebounds')\n player['defensive_rebounds'] = safe_get(home_player,'DefensiveRebounds')\n player['rebounds'] = safe_get(home_player,'TotalRebounds')\n player['assists'] = safe_get(home_player,'Assistances')\n player['steals'] = safe_get(home_player,'Steals')\n player['blocks'] = safe_get(home_player,'BlocksFavour')\n player['turnovers'] = safe_get(home_player,'Turnovers')\n player['personal_fouls'] = safe_get(home_player,'FoulsCommited')\n player['score_difference'] = safe_get(home_player,'Valuation')\n player['point'] = safe_get(home_player,'Points')\n player['first_publish'] = safe_get(home_player,'IsStarter')\n first = safe_get(home_player,'IsPlaying')\n if first == 1:\n player['on_ground'] = 0\n elif first == 0:\n player['on_ground'] = 1\n else:\n player['on_ground'] = 0\n player_data = {\n 'belong': int(player['belong']),\n 'player_id': int(player['player_id']),\n 'player_name': player['player_name'],\n 'minutes': int(player['minutes']),\n 'goals': int(player['goals']),\n 'field': int(player['field']),\n 'three_point_goals': int(player['three_point_goals']),\n 'three_point_field': int(player['three_point_field']),\n 'free_throw_goals': int(player['free_throw_goals']),\n 'free_throw_field': int(player['free_throw_field']),\n 'offensive_rebounds': int(player['offensive_rebounds']),\n 'defensive_rebounds': int(player['defensive_rebounds']),\n 'rebounds': int(player['rebounds']),\n 'assists': int(player['assists']),\n 'steals': int(player['steals']),\n 'blocks': int(player['blocks']),\n 'turnovers': int(player['turnovers']),\n 'personal_fouls': int(player['personal_fouls']),\n 'score_difference': int(player['score_difference']),\n 'point': int(player['point']),\n 'first_publish': int(player['first_publish']),\n 'enter_ground': int(player['enter_ground']),\n 'on_ground': int(player['on_ground']),\n 'shirt_number': int(player['shirt_number']),\n }\n player_stats_list.append(player_data)\n match_id = int(str(13) + '0000') + int(gamecode)\n match_data_boxscore = {'match': {'id': int(match_id),\n 'basketball_items': {\n 'player_stat': {\n 'items': player_stats_list},\n 'team_stat': {'items': team_stats_list}\n }}}\n if player_stats_list:\n data_queue.put(match_data_boxscore)\n logger.info('球员技术统计推送完成。。。 %s' % str(gamecode))\n minutes_team = boxscore_json_dict['Stats'][0]['totr']['Minutes']\n if minutes_team == '225:00' or minutes_team == '200:00':\n break\n except:\n dingding_alter(traceback.format_exc())\n logger.error(traceback.format_exc())\n continue\n","sub_path":"apps/eur_basketball_spider/eur_boxscore.py","file_name":"eur_boxscore.py","file_ext":"py","file_size_in_byte":14940,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"91162491","text":"import numpy as np\n\ndef getTestFold(data, idx):\n ends = [1273, 2225, 2501, 3482, 3888, 4387, 5343, 6051, 6606, 7702, 8215, 9527, 10068, 11265, 11346]\n\n start = 0\n if (idx > 0):\n start = ends[idx - 1]\n end = ends[idx]\n\n return data[start:end]\n\n\ndef getTrainFold(data, idx):\n ends = [1273, 2225, 2501, 3482, 3888, 4387, 5343, 6051, 6606, 7702, 8215, 9527, 10068, 11265, 11346]\n\n start = 0\n if (idx > 0):\n start = ends[idx - 1]\n end = ends[idx]\n\n take = data[:]\n take = np.delete(take, list(range(start, end)), axis=0)\n\n return take\n\n\ndef getFold(data, idx):\n ends = [1273, 2225, 2501, 3482, 3888, 4387, 5343, 6051, 6606, 7702, 8215, 9527, 10068, 11265, 11346]\n\n start = 0\n if (idx > 0):\n start = ends[idx - 1]\n end = ends[idx]\n\n return data[start:end]\n\n\ndef getFolds(data, foldsIndices):\n take = getFold(data, foldsIndices[0])\n\n for i in range(1, len(foldsIndices)):\n idx = foldsIndices[i]\n take = np.concatenate((take, getFold(data, idx)), axis=0)\n\n return take\n\n\ndef getSubTrainFold(data, idx):\n ends = [1273, 2225, 2501, 3482, 3888, 4387, 5343, 6051, 6606, 7702, 8215, 9527, 10068, 11265, 11346]\n\n remove = idx\n remove2 = idx - 1\n if (remove2 < 0):\n remove2 = 14\n\n indices = []\n for i in range(15):\n if (i != remove and i != remove2):\n indices.append(i)\n\n take = getFold(data, indices[0])\n\n for i in indices[1:]:\n take = np.concatenate((take, getFold(data, i)), axis=0)\n\n return take\n\n\ndef getSubTestFold(data, idx):\n ends = [1273, 2225, 2501, 3482, 3888, 4387, 5343, 6051, 6606, 7702, 8215, 9527, 10068, 11265, 11346]\n\n idx -= 1\n if (idx < 0):\n idx = 14\n\n take = getFold(data, idx)\n\n return take\n\n\ndef loadLines(path):\n with open(path, \"r\") as f:\n lines = [l.strip(\"\\r\\n\") for l in f.readlines()]\n\n return lines\n\n\ndef loadPlainMatrix(path):\n lines = loadLines(path)\n\n gt = []\n for l in lines:\n parts = l.split(\" \")\n current = []\n for p in parts:\n if (p != \"\"):\n current.append(float(p))\n gt.append(current)\n\n return np.array(gt)\n\n\ndef dotProduct(v1, v2):\n return np.sum(np.multiply(v1, v2))\n\n\ndef calculateAngle(v1, v2):\n return 180.0 * np.arccos(dotProduct(v1, v2) / np.sqrt(dotProduct(v1, v1) * dotProduct(v2, v2))) / np.pi\n\n\ndef calculateAngularStatistics(angles1, angles2):\n n = angles1.shape[0]\n\n angles = []\n for i in range(n):\n angles.append(calculateAngle(angles1[i, :], angles2[i, :]))\n\n return (np.mean(angles), np.median(angles), np.max(angles))\n\n\ndef getBatches(data, batchSize=64):\n from math import ceil\n\n n = int(ceil(float(len(data)) / batchSize))\n\n batches = []\n\n for i in range(n):\n currentIdx = i * batchSize\n nextIdx = (i + 1) * batchSize\n batches.append(data[currentIdx:nextIdx, :])\n\n return batches\n","sub_path":"tests/helpers.py","file_name":"helpers.py","file_ext":"py","file_size_in_byte":2939,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"425418027","text":"class UserMessage:\n def __init__(self, ical_url, id):\n self.ical_url = ical_url\n self.id = id\n\n\nclass ScheduleRequestMessage:\n def __init__(self, length, date_range, time_range, ideal_days=0, ideal_times=0):\n self.length = length\n self.date_range = date_range\n self.time_range = time_range\n self.ideal_days = ideal_days\n self.ideal_times = ideal_times","sub_path":"Messages.py","file_name":"Messages.py","file_ext":"py","file_size_in_byte":406,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"154191628","text":"from flexmock import flexmock\nfrom utils import utils\nfrom infrastructure_manager import *\ntry:\n from unittest import TestCase\nexcept ImportError:\n from unittest.case import TestCase\n\nclass TestInfrastructureManager(TestCase):\n def setUp(self):\n flexmock(utils).should_receive('get_secret').and_return('secret')\n\n def tearDown(self):\n flexmock(utils).should_receive('get_secret').reset()\n\n def test_initialize(self):\n i = InfrastructureManager()\n self.assertEquals('secret', i.secret)\n\n def test_describe_instances(self):\n i = InfrastructureManager()\n params1 = {}\n result1 = i.describe_instances(params1, 'secret1')\n self.assertFalse(result1['success'])\n self.assertEquals(result1['reason'],\n InfrastructureManager.REASON_BAD_SECRET)\n\n # test the scenario where we fail to give describe_instances a\n # reservation id\n params2 = {}\n result2 = i.describe_instances(params2, 'secret')\n self.assertFalse(result2['success'])\n self.assertEquals(result2['reason'],\n 'no ' + InfrastructureManager.PARAM_RESERVATION_ID)\n\n # test what happens when a caller fails to give describe instances\n # a reservation id that's in the system\n params3 = {InfrastructureManager.PARAM_RESERVATION_ID: 'boo'}\n result3 = i.describe_instances(params3, 'secret')\n self.assertFalse(result3['success'])\n self.assertEquals(result3['reason'],\n InfrastructureManager.REASON_RESERVATION_NOT_FOUND)\n\n # test what happens when a caller gives describe_instances a reservation\n # id that is in the system\n id = '0000000000'\n params4 = {InfrastructureManager.PARAM_RESERVATION_ID: id}\n vm_info = {\n 'public_ips': ['public-ip'],\n 'private_ips': ['private-ip'],\n 'instance_ids': ['i-id']\n }\n status_info = {\n 'success': True,\n 'reason': 'received run request',\n 'state': InfrastructureManager.STATE_RUNNING,\n 'vm_info': vm_info\n }\n i.reservations.put(id, status_info)\n result4 = i.reservations.get(id)\n self.assertEquals(result4, i.describe_instances(params4, \"secret\"))\n\n params5 = json.dumps(params4)\n self.assertEquals(result4, i.describe_instances(params5, \"secret\"))\n\n try:\n i.describe_instances('foo', 'bar')\n self.fail('Must throw an exception')\n except Exception:\n pass\n\n try:\n i.describe_instances({'reservation_id': 'foo'}, {})\n self.fail('Must throw an exception')\n except TypeError:\n pass\n\n def test_run_instances(self):\n i = InfrastructureManager()\n\n params1 = {}\n result1 = i.run_instances(params1, 'secret1')\n self.assertFalse(result1['success'])\n self.assertEquals(result1['reason'],\n InfrastructureManager.REASON_BAD_SECRET)\n\n params2 = {}\n result2 = i.run_instances(params2, 'secret')\n self.assertFalse(result2['success'])\n self.assertEquals(result2['reason'], 'no infrastructure')\n\n params3 = {'infrastructure': 'ec2'}\n result3 = i.run_instances(params3, 'secret')\n self.assertFalse(result3['success'])\n self.assertEquals(result3['reason'], 'no num_vms')\n\n params4 = {'infrastructure': 'ec2', 'num_vms': 0}\n result4 = i.run_instances(params4, 'secret')\n self.assertFalse(result4['success'])\n self.assertEquals(result4['reason'],\n InfrastructureManager.REASON_BAD_VM_COUNT)\n\n try:\n result5 = i.run_instances('foo', 'bar')\n self.fail('Must throw an exception')\n except Exception:\n pass\n\n def test_terminate_instances(self):\n i = InfrastructureManager()\n\n params1 = {}\n result1 = i.terminate_instances(params1, 'secret1')\n self.assertFalse(result1['success'])\n self.assertEquals(result1['reason'],\n InfrastructureManager.REASON_BAD_SECRET)\n\n params2 = {}\n result2 = i.terminate_instances(params2, 'secret')\n self.assertFalse(result2['success'])\n self.assertEquals(result2['reason'], 'no infrastructure')\n\n try:\n result3 = i.terminate_instances('foo', 'bar')\n self.fail('Must throw an exception')\n except Exception:\n pass\n","sub_path":"InfrastructureManager/tests/test_infrastructure_manager.py","file_name":"test_infrastructure_manager.py","file_ext":"py","file_size_in_byte":4034,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"523893301","text":"import spacy\nfrom nltk.stem import SnowballStemmer\n\nfrom etl.scholarship import AcademicLevel\n\n\nnlp = spacy.load('es_core_news_md')\nstemmer = SnowballStemmer('spanish')\n\n\ndef calculate_academic_level(item: dict):\n if 'description' not in item:\n return item\n\n item['academicLevel'] = match_academic_level(item['description']).value\n\n return item\n\n\ndef match_academic_level(raw_text):\n doc = first_sentence(raw_text)\n word = subtree_matcher(doc)[1]\n\n if word is None:\n return AcademicLevel.OTHERS\n\n for child in word.children:\n if child.dep_ == 'nmod':\n word = child\n break\n\n return academic_level(word.text)\n\n\ndef first_sentence(text):\n doc = nlp(text)\n return list(doc.sents)[0]\n\n\ndef subtree_matcher(doc):\n subject_ = None\n object_ = None\n\n for token in doc:\n if token.dep_ == 'nsubj':\n subject_ = token\n\n if token.dep_ == 'obj':\n object_ = token\n\n if subject_ and object_:\n break\n\n return (subject_, object_)\n\n\ndef academic_level(token):\n token = stemmer.stem(token)\n\n if is_undergraduate(token):\n return AcademicLevel.UNDERGRADUATE\n\n if is_postgraduate(token):\n return AcademicLevel.POSTGRADUATE\n\n return AcademicLevel.OTHERS\n\n\ndef is_undergraduate(token):\n words = ['estudiantes', 'bachilleres']\n words = map(stemmer.stem, words)\n return token in set(words)\n\n\ndef is_postgraduate(token):\n words = ['profesionales', 'doctorado', 'maestria', 'postdoctorado', 'posdoctorado']\n words = map(stemmer.stem, words)\n return token in set(words)\n","sub_path":"etl/icetex/tasks/calculate_academic_level.py","file_name":"calculate_academic_level.py","file_ext":"py","file_size_in_byte":1622,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"74029445","text":"#!/usr/bin/env python\n# -*- coding:utf-8 -*-\n# Author : Lysander Tseng\n# Email : qingshan.tseng@gmail.com\n# Time : 2020/2/18 14:48\n# User : Magic\n# Product : PyCharm\n# Project : 030\n# File : 030.3.py\n# Intro : 编写一个程序,用户输入开始搜索的路径,\n# 查找该路径下(包含子文件夹内)所有的视频格式文件\n# (要求查找mp4, rmvb, avi的格式即可),并创建一\n# 个文件(videoList.txt)存放所有找到的文件的路径\nimport os # 引入os模块\n\n\ndef format_find(this_path, ext_list, target_path):\n \"\"\"定义一个格式查找函数,用于查找指定目录中指定拓展名的文件并将其路径记录在一个文本文件中\"\"\"\n\n path_list = []\n for i in os.walk(this_path):\n for j in i[2]:\n for k in ext_list:\n if k in j:\n path_list.append(i[0] + '\\\\' + j)\n else:\n path_str = '\\n\\n'.join(path_list)\n with open(target_path, 'w', encoding='utf-8') as f:\n f.write(path_str)\n print(\"保存成功!\")\n\n\ndef path_trans(raw_path):\n \"\"\"定义一个路径字符串转换函数,用于将当前操作系统中的路径字符串转换成标准路径\"\"\"\n\n path_list = raw_path.split('\\\\') # 将传入的原始路径字符串以\\为标志切割并保存在列表path_list中\n standard_path = os.sep.join(path_list) # 再将os.sep插入到每个元素之间并将其重新拼接成标准字符串\n return standard_path\n\n\nactual_this_path = path_trans(input(\"请输入待查找的初始目录:\"))\nactual_ext_list = input(\"请输入需要查找的拓展名[以','号间隔]:\").split(',')\nactual_target_path = path_trans(input(\"请输入需要保存的路径:\"))\n\nformat_find(actual_this_path, actual_ext_list, actual_target_path)\n","sub_path":"Python/Homework/030/030.3.py","file_name":"030.3.py","file_ext":"py","file_size_in_byte":1804,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"618563410","text":"import os\nimport random\nimport time\nimport json\nimport torch\nimport torchvision\nimport numpy as np\nimport pandas as pd\nimport warnings\nfrom datetime import datetime\nfrom torch import nn, optim\nfrom collections import OrderedDict\nfrom torch.autograd import Variable\nfrom torch.utils.data import DataLoader\nfrom sklearn.model_selection import train_test_split, StratifiedKFold\nfrom timeit import default_timer as timer\nfrom IPython import embed\nfrom pipelineTrain import *\nfrom MosNet import *\nfrom Xception import Xception\nfrom configTrain import configTrain\n\n# set cudnn\n# os.environ[\"CUDA_VISIBLE_DEVICES\"] = configTrain.gpus\n# torch.backends.cudnn.benchmark = True\n# warnings.filterwarnings('ignore')\nrandom.seed(configTrain.seed)\nnp.random.seed(configTrain.seed)\ntorch.manual_seed(configTrain.seed)\ntorch.cuda.manual_seed_all(configTrain.seed)\n\n\ndef main():\n fold = 0\n # mkdirs\n if not os.path.exists(configTrain.weights):\n os.mkdir(configTrain.weights)\n if not os.path.exists(configTrain.best_models):\n os.mkdir(configTrain.best_models)\n if not os.path.exists(configTrain.logs):\n os.mkdir(configTrain.logs)\n if not os.path.exists(configTrain.weights + configTrain.model_name + os.sep + str(fold) + os.sep):\n os.makedirs(configTrain.weights + configTrain.model_name + os.sep + str(fold) + os.sep)\n if not os.path.exists(configTrain.best_models + configTrain.model_name + os.sep + str(fold) + os.sep):\n os.makedirs(configTrain.best_models + configTrain.model_name + os.sep + str(fold) + os.sep)\n # get model and optimizer\n # model = MosNet()\n model = Xception()\n # model = torch.nn.DataParallel(model)\n model.cuda()\n\n # set optimizer and loss function\n optimizer = optim.Adam(model.parameters(), lr=configTrain.lr, amsgrad=False, weight_decay=configTrain.weight_decay)\n criterion = nn.CrossEntropyLoss().cuda()\n\n # some parameters\n start_epoch = 0\n best_precision1 = 0\n best_precision_save = 0\n resume = False\n\n # restart the training process\n if resume:\n checkpoint = torch.load(configTrain.best_models + str(fold) + \"/model_best.pth.tar\")\n start_epoch = checkpoint[\"epoch\"]\n fold = checkpoint[\"fold\"]\n best_precision1 = checkpoint[\"best_precision1\"]\n model.load_state_dict(checkpoint[\"state_dict\"])\n optimizer.load_state_dict(checkpoint[\"optimizer\"])\n\n # read image files\n train_data_list = get_files(configTrain.train_data, \"train\")\n val_data_list = get_files(configTrain.val_data, \"val\")\n\n # load train_data and validation_data with DataLoader\n train_dataloader = DataLoader(FaceDataset(train_data_list), batch_size=configTrain.batch_size, shuffle=True,\n pin_memory=True)\n val_dataloader = DataLoader(FaceDataset(val_data_list, train=False), batch_size=configTrain.batch_size * 2,\n shuffle=True, pin_memory=False)\n\n # Decays the learning rate by gamma every step_size epochs\n scheduler = optim.lr_scheduler.StepLR(optimizer, step_size=configTrain.stepLR_size, gamma=0.1)\n\n # define metrics for loss and accuracy\n train_losses = AverageMeter()\n train_top1 = AverageMeter()\n valid_loss = [np.inf, 0, 0]\n\n # set the module in training mode\n model.train()\n\n # train\n for epoch in range(start_epoch, configTrain.epochs):\n scheduler.step(epoch)\n train_progressor = ProgressBar(mode=\"Train\", epoch=epoch, total_epoch=configTrain.epochs,\n model_name=configTrain.model_name, total=len(train_dataloader))\n # global iter\n for iter, (input, target) in enumerate(train_dataloader):\n # switch to continue train process\n train_progressor.current = iter\n\n # get loss and accuracy\n model.train()\n input = Variable(input).cuda()\n target = Variable(torch.from_numpy(np.array(target)).long()).cuda()\n output = model(input)\n loss = criterion(output, target)\n\n precision1_train, precision2_train = accuracy(output, target, topk=(1, 2))\n\n # update batch's loss and accuracy to total\n train_losses.update(loss.item(), input.size(0))\n train_top1.update(precision1_train[0], input.size(0))\n\n # set loss and accuracy to progress bar\n train_progressor.current_loss = train_losses.avg\n train_progressor.current_top1 = train_top1.avg\n # backward\n optimizer.zero_grad()\n loss.backward()\n optimizer.step()\n\n # progress bar vision\n train_progressor()\n train_progressor.done()\n\n # evaluate every epoch\n valid_loss_accuracy = evaluate(val_dataloader, model, criterion, epoch)\n\n # determine whether to save model\n is_best = valid_loss_accuracy[1] > best_precision1\n\n # save the best precision\n best_precision1 = max(valid_loss_accuracy[1], best_precision1)\n try:\n best_precision_save = best_precision1.cpu().data.numpy()\n except:\n pass\n save_checkpoint({\n \"epoch\": epoch + 1,\n \"model_name\": configTrain.model_name,\n \"state_dict\": model.state_dict(),\n \"best_precision1\": best_precision1,\n \"optimizer\": optimizer.state_dict(),\n \"fold\": fold,\n \"valid_loss\": valid_loss_accuracy,\n }, is_best, fold)\n\n\nif __name__ ==\"__main__\":\n main()\n\n\n","sub_path":"models/train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":5534,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"288072760","text":"\n\nfrom xai.brain.wordbase.nouns._mania import _MANIA\n\n#calss header\nclass _MANIAS(_MANIA, ):\n\tdef __init__(self,): \n\t\t_MANIA.__init__(self)\n\t\tself.name = \"MANIAS\"\n\t\tself.specie = 'nouns'\n\t\tself.basic = \"mania\"\n\t\tself.jsondata = {}\n","sub_path":"xai/brain/wordbase/nouns/_manias.py","file_name":"_manias.py","file_ext":"py","file_size_in_byte":231,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"209459872","text":"import cv2\n\ndef kaze_match(im1_path, im2_path):\n # load the image and convert it to grayscale\n im1 = cv2.imread(im1_path)\n im2 = cv2.imread(im2_path)\n gray1 = cv2.cvtColor(im1, cv2.COLOR_BGR2GRAY)\n gray2 = cv2.cvtColor(im2, cv2.COLOR_BGR2GRAY)\n\n # initialize the AKAZE descriptor, then detect keypoints and extract\n # local invariant descriptors from the image\n detector = cv2.AKAZE_create()\n (kps1, descs1) = detector.detectAndCompute(gray1, None)\n (kps2, descs2) = detector.detectAndCompute(gray2, None)\n\n print(\"keypoints: {}, descriptors: {}\".format(len(kps1), descs1.shape))\n print(\"keypoints: {}, descriptors: {}\".format(len(kps2), descs2.shape))\n\n # Match the features\n bf = cv2.BFMatcher(cv2.NORM_HAMMING)\n # matches = bf.match(descs1, descs2)\n matches = bf.knnMatch(descs1, descs2, k=2) # typo fixed\n # print(matches)\n\n # Apply ratio test\n good = []\n for m, n in matches:\n\n if m.distance < 0.9*n.distance:\n good.append([m])\n\n # cv2.drawMatchesKnn expects list of lists as matches.\n im3 = cv2.drawMatchesKnn(im1, kps1, im2, kps2, good[1:20], None, flags=2)\n cv2.imshow(\"AKAZE matching\", im3)\n cv2.waitKey(0)\n\nkaze_match('img/graf1.png', 'img/graf3.png')","sub_path":"mytask1/akaze_test2.py","file_name":"akaze_test2.py","file_ext":"py","file_size_in_byte":1254,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"548414577","text":"from flask import Flask, render_template, redirect, request\nimport os\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker, scoped_session\nfrom pytz import timezone\nfrom datetime import datetime\napp=Flask(__name__)\nengine=create_engine(os.getenv(\"DATABASE_URL\"))\ndb=scoped_session(sessionmaker(bind=engine))\n\n\n@app.route(\"/\")\n\ndef index():\n\treturn render_template(\"index.html\")\n\n@app.route(\"/\")\ndef query(query):\n\ttemplate=\"Authors
\"\n\tauthors=db.execute(\"SELECT * FROM authors WHERE name LIKE :query\", {\"query\": query+\"%\"}).fetchall()\n\tbooks=db.execute(\"SELECT * FROM books WHERE name LIKE :query\", {\"query\": query+\"%\"}).fetchall()\n\tif authors:\n\t\tfor i in authors:\n\t\t\ttemplate+='\"+str(i[1])+\"\"\n\ttemplate+=\"Books!
\"\n\tif books:\n\t\tfor i in books:\n\t\t\ttemplate+='\"+str(i[1])+\", \"+str(i[3])+\"\"\n\n\treturn template\n@app.route(\"/book\")\ndef book():\n\tid=int(request.args.get(\"q\"))\n\tbook=db.execute(\"SELECT * FROM books WHERE id=:id\", {\"id\":id}).fetchall()\n\tauthor=db.execute(\"SELECT * FROM authors WHERE id=:author_id\", {\"author_id\": book[0][2]}).fetchall()\n\treturn render_template(\"book.html\", book=book[0], author=author[0])\n@app.route(\"/author\")\ndef author():\n\tid=request.args.get(\"q\")\n\tid=int(id)\n\tauthor=db.execute(\"SELECT * FROM authors WHERE id=:id\", {\"id\":id}).fetchall()\n\tbooks=db.execute(\"SELECT * FROM books WHERE author_id=:id\", {\"id\": author[0][0]}).fetchall()\n\treturn render_template(\"author.html\", author=author, books=books)\n@app.route(\"/create_book\", methods=[\"GET\", \"POST\"])\ndef create_book():\n\tif request.method==\"GET\":\n\t\treturn render_template(\"add_books.html\")\n\telse:\n\t\tname=request.form.get(\"name\")\n\t\tauthor=request.form.get(\"author\")\n\t\tyear=request.form.get(\"year\")\n\t\n\t\tif not name or not author or not year:\n\t\t\treturn render_template(\"add_books.html\")\n\t\ttry:\n\t\t\ty=int(year)\n\t\t\tif y<0:\n\t\t\t\treturn render_template(\"add_books.html\")\n\t\t\tif y>datetime.now(timezone(\"US/Eastern\")).year:\n\t\t\t\treturn render_template(\"add_books.html\")\n\n\t\texcept:\n\t\t\treturn render_template(\"add_books.html\")\n\t\tauthor_id=db.execute(\"SELECT * FROM authors WHERE name=:author\", {\"author\": author}).fetchall()\n\t\tif len(author_id)==0:\n\t\t\treturn render_template(\"add_books.html\")\n\t\tsame=db.execute(\"SELECT * FROM books WHERE name=:name AND author_id=:author_id AND year=:year\",{\"name\":name, \"author_id\": author_id[0][0], \"year\":year}).fetchall()\n\t\tif len(same)!=0:\n\t\t\treturn render_template(\"add_books.html\")\n\t\tdb.execute(\"INSERT INTO books(name, author_id, year) VALUES(:name, :author_id, :year)\", {\"name\": name, \"author_id\": author_id[0][0], \"year\": year})\n\t\tdb.commit()\n\t\treturn redirect(\"/\")\n\n@app.route(\"/create_author\", methods=[\"GET\", \"POST\"])\ndef create_author():\n\tif request.method==\"GET\":\n\t\treturn render_template(\"add_authors.html\")\n\tname=request.form.get(\"name\")\n\tif not name:\n\t\treturn render_template(\"add_authors.html\")\n\tsame= db.execute(\"SELECT * FROM authors WHERE name=:name\", {\"name\": name}).fetchall()\n\tif len(same)!=0:\n\t\treturn render_template(\"add_authors.html\")\n\tdb.execute(\"INSERT INTO authors(name) VALUES(:name)\", {\"name\": name})\n\tdb.commit()\n\treturn redirect(\"/\")\n@app.route(\"/delete_author\")\ndef delete_author():\n\tid=int(request.args.get(\"q\"))\n\tdb.execute(\"DELETE FROM books WHERE author_id=:id\", {\"id\": id})\n\tdb.execute(\"DELETE FROM authors WHERE id=:id\", {\"id\": id})\n\tdb.commit()\n\treturn redirect(\"/\")\n\n@app.route(\"/delete_book\")\ndef delete_book():\n\tid=int(request.args.get(\"q\"))\n\tdb.execute(\"DELETE FROM books WHERE id=:id\", {\"id\":id})\n\tdb.commit()\n\treturn redirect(\"/\")\n\n","sub_path":"application.py","file_name":"application.py","file_ext":"py","file_size_in_byte":3651,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"424257622","text":"\nactive = \"#45ABc5\"\ngreen = \"#00ff00\"\nerror = \"#ff0000\"\nalert = \"#fa8700\"\nwarn = \"#faf700\"\nwhite = \"#ffffff\"\n\n\nforeground = \"#858585\"\nbackground = \"#202020\"\n\nsep = \" \"\n\nws_active = \"#3e6700\"\nws_active_txt = \"#cccccc\"\nws_urgent = alert\nws_inactive = \"#303030\"\nws_inactive_txt = \"#888888\"\nws_sep_col = background\n\n\ninterval = 0.2\nwifi_name = \"wlp3s0\"\n\ndzen_args = [\"-dock\", \"-ta\", \"l\", \"-fn\", \"monospace:bold:size=10\", \\\n \"-bg\", background, \"-fg\", foreground, \"-h\", \"15\"]\n","sub_path":"i3/.pybar/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":553,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"227752097","text":"import json\nimport pprint\nimport numpy as np\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nimport matplotlib.pyplot as plt\n\nfrom sklearn.metrics.pairwise import cosine_similarity\n\nvectorizer = TfidfVectorizer()\n\n\npp = pprint.PrettyPrinter(indent=4)\nwith open('dev-v1.1.json') as json_file:\n data = json.load(json_file)\n\n # print(data['data'][0]['paragraphs'][0]['qas'][0]['answers'][0]['text'].keys())\n Database = []\n for articles in data['data']:\n article = {}\n article['Q'] = []\n for index, paragraph in enumerate(articles['paragraphs']):\n article[index] = paragraph['context']\n for question in paragraph['qas']:\n article['Q'].append([question['question'],index])\n Database.append(article)\n break\n\n resultList = {}\n for index in range(1,15):\n resultList[index] = []\n numCorrect = 0\n num = 0\n for article in Database:\n Questions = article['Q']\n del article['Q']\n paragraphs = [article[ele] for ele in article.keys()]\n paragraphs.insert(0,'')\n for question in Questions:\n paragraphs[0] = question[0]\n X = vectorizer.fit_transform(paragraphs)\n simi = cosine_similarity(X[0:1], X)\n simi[0][0] = 0\n correct = simi[0][question[1]+1]\n\n\n simi[0].sort()\n reverse = simi[0][::-1]\n mean = sum(reverse[0:10])/10\n\n # if (correct - mean)>0.05:\n # numCorrect +=1\n num+=1\n\n # print(top5)\n # resultList.append( 1 if np.argmax(simi) == question[1]+1 else 0)\n # plt.ylabel('cosine similarity')\n # # plt.xlabel('rank')\n for i in range(0,10):\n if reverse[i] == correct and i==0:\n numCorrect+=1\n # # plt.plot(i, reverse[i], 'ro',label = 'correct')\n # else:\n # # plt.plot(i, reverse[i], 'bo')\n #\n # plt.legend()\n #\n # plt.show()\n\n print(num,numCorrect,numCorrect/num)\n\n\n\n\n\n","sub_path":"explore_squad.py","file_name":"explore_squad.py","file_ext":"py","file_size_in_byte":2125,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"582776813","text":"import glob\n\nimport pytest\n\nimport archetypal as ar\n\n# configure archetypal\nar.config(log_console=True, log_file=True, use_cache=True,\n data_folder='.temp/data', logs_folder='.temp/logs',\n imgs_folder='.temp/imgs', cache_folder='.temp/cache',\n umitemplate='../data/BostonTemplateLibrary.json')\n\n# # Uncomment this block to test different file variations\n# files_to_try = ['./input_data/problematic/*.idf',\n# './input_data/regular/*.idf',\n# './input_data/umi_samples/*.idf']\n# ids = ['problematic', 'regular', 'umi_samples']\n\nfiles_to_try = ['./input_data/regular/*.idf']\nids = ['regular']\n\n\n@pytest.fixture(scope='module', params=files_to_try, ids=ids)\ndef template(fresh_start, request):\n \"\"\"Instantiate an umi template placeholder. Calls in the fresh_start\n function to clear the cache folder\"\"\"\n idf = glob.glob(request.param)\n idf = ar.copy_file(idf)\n # idf = './input_data/AdultEducationCenter.idf'\n wf = './input_data/CAN_PQ_Montreal.Intl.AP.716270_CWEC.epw'\n a = ar.Template(idf, wf)\n\n yield a\n\n\n@pytest.fixture(scope='module')\ndef test_template_withcache():\n \"\"\"Instantiate an umi template placeholder. Does note call fresh_start\n function so that caching can be used\"\"\"\n idf = glob.glob('/Users/samuelduchesne/Dropbox/Polytechnique/Doc/software'\n '/archetypal-dev/data/necb/NECB_2011_Montreal_idf/*.idf')\n idf = ar.copy_file(idf)\n # idf = './input_data/AdultEducationCenter.idf'\n wf = './input_data/CAN_PQ_Montreal.Intl.AP.716270_CWEC.epw'\n a = ar.Template(idf, wf)\n\n yield a\n\n\n@pytest.fixture(scope='module')\ndef sql(test_template_withcache):\n sql = test_template_withcache.run_eplus(silent=False, processors=-1)\n yield sql\n\n\ndef test_materials_gas(template):\n template.materials_gas = ar.materials_gas(template.idfs)\n assert not template.materials_gas.empty\n\n\ndef test_materials_glazing(template):\n template.materials_glazing = ar.materials_glazing(template.idfs)\n template.materials_glazing = ar.newrange(template.materials_gas,\n template.materials_glazing)\n return template.materials_glazing\n\n\ndef test_materials_opaque(template):\n template.materials_opaque = ar.materials_opaque(template.idfs)\n template.materials_opaque = ar.newrange(template.materials_glazing,\n template.materials_opaque)\n return template.materials_opaque\n\n\ndef test_constructions_opaque(template):\n template.constructions_opaque = ar.constructions_opaque(template.idfs,\n template.materials_opaque)\n template.constructions_opaque = ar.newrange(template.materials_opaque,\n template.constructions_opaque)\n return template.constructions_opaque\n\n\ndef test_constructions_windows(template):\n template.constructions_windows = ar.constructions_windows(template.idfs,\n template.materials_glazing)\n template.constructions_windows = ar.newrange(template.constructions_opaque,\n template.constructions_windows)\n return template.constructions_windows\n\n\ndef test_day_schedules(template):\n template.day_schedules = ar.day_schedules(template.idfs)\n return template.day_schedules\n\n\ndef test_week_schedules(template):\n template.week_schedules = ar.week_schedules(template.idfs,\n template.day_schedules)\n template.week_schedules = ar.newrange(template.day_schedules,\n template.week_schedules)\n return template.week_schedules\n\n\ndef test_year_schedules(template):\n template.year_schedules = ar.year_schedules(template.idfs,\n template.week_schedules)\n template.year_schedules = ar.newrange(template.week_schedules,\n template.year_schedules)\n return template.year_schedules\n\n\n# Zones\ndef test_zone_information(template, sql):\n template.zone_details = ar.zone_information(sql)\n\n\ndef test_zone_loads(template, sql):\n template.zone_loads = ar.zone_loads(sql)\n\n\ndef test_zone_ventilation(template, sql):\n template.zone_ventilation = ar.zone_ventilation(sql)\n\n\ndef test_zone_condition(template, sql):\n template.zone_conditioning = ar.zone_conditioning(sql)\n\n\ndef test_zone_condition_dev(test_template_withcache, sql):\n test_template_withcache.zone_conditioning = ar.zone_conditioning(sql)\n print(test_template_withcache.zone_conditioning)\n\n\ndef test_to_json(test_template_withcache):\n test_template_withcache.read()\n json = test_template_withcache.to_json(orient='records')\n print(json)\n\n\ndef test_to_json_std():\n files = glob.glob(\"./input_data/STD/*idf\")\n files = ar.copy_file(files)\n wf = './input_data/CAN_PQ_Montreal.Intl.AP.716270_CWEC.epw'\n a = ar.Template(files, wf)\n a.read()\n json = a.to_json(orient='records')\n print(json)\n","sub_path":"tests/test_core.py","file_name":"test_core.py","file_ext":"py","file_size_in_byte":5112,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"456170456","text":"#!/usr/bin/env python\r\n# coding: utf-8\r\n\r\n# In[ ]:\r\n\r\n\r\n# Import required packages\r\nimport pybullet as p\r\nimport pybullet_data\r\nimport numpy as np\r\nimport cv2\r\nimport time\r\nimport matplotlib.pyplot as plt\r\n\r\n\r\np.connect(p.GUI) \r\np.setAdditionalSearchPath(pybullet_data.getDataPath())\r\nplane = p.loadURDF(\"plane.urdf\")\r\np.setGravity(0, 0, -30)\r\n\r\nhuskypos = [0, 0, 1]\r\n\r\n# Load URDFs\r\nhusky = p.loadURDF(\"husky/husky.urdf\", huskypos[0], huskypos[1], huskypos[2])\r\ncheetah = p.loadURDF(\"mini_cheetah/mini_cheetah.urdf\",-5,+2, huskypos[2])\r\n\r\ntarget= p.loadURDF(\"block0.urdf\",3, +2, huskypos[2])\r\n\r\n\r\nmaxForce = 55#Force of car #Newton.m\r\ndist_init=0#initial distance\r\nf=50#force of cheetah #Newton.m\r\nm1=0.01\r\nd=50\r\ns=0.15# distance travelled by cheetah for calling runcheetah() once\r\nl=0\r\nl1=0 #adjust the time.sleep(l1) \r\n# Function to set cheetah's initial orientation(need was understood by experiment)\r\ndef setPos():\r\n l2=0.1\r\n n=0\r\n while(n=-np.pi:\r\n speed= -sp\r\n elif error>0.5 and error3:\r\n sp= int((mag(targetVec)-1.5)*10)\r\n m= (sp//10)-1 # To set kp and kd according to the basespeed\r\n else:\r\n sp=0\r\n sx= (mag(targetVec)/3)*h-1# Dependence on speed of enemy\r\n print(sp)\r\n speed_correction, last_error=speedCorrection(targetVec, frontVec, last_error, m) \r\n turn(speed_correction, sp, sx)\r\n\r\n for k, v in keys.items():\r\n if (k == p.B3G_UP_ARROW and (v & p.KEY_IS_DOWN) and PID_CONTROL==False):\r\n targetVel = 20\r\n for joint in range(2,6):\r\n p.setJointMotorControl2(husky,joint, p.VELOCITY_CONTROL, targetVelocity =targetVel, force = maxForce)\r\n\r\n p.stepSimulation()\r\n\r\n if (k == p.B3G_UP_ARROW and (v & p.KEY_WAS_RELEASED) and PID_CONTROL==False):\r\n targetVel = 0\r\n for joint in range(2, 6):\r\n p.setJointMotorControl2(husky, joint, p.VELOCITY_CONTROL,targetVelocity = targetVel,force = maxForce)\r\n\r\n if (k == p.B3G_DOWN_ARROW and (v & p.KEY_IS_DOWN) and PID_CONTROL==False):\r\n targetVel = -20\r\n for joint in range(2,6):\r\n p.setJointMotorControl2(husky,joint, p.VELOCITY_CONTROL, targetVelocity =targetVel, force = maxForce)\r\n\r\n p.stepSimulation()\r\n\r\n if (k == p.B3G_DOWN_ARROW and (v & p.KEY_WAS_RELEASED) and PID_CONTROL==False):\r\n targetVel = 0\r\n for joint in range(2, 6):\r\n p.setJointMotorControl2(husky, joint, p.VELOCITY_CONTROL,targetVelocity = targetVel,force = maxForce)\r\n\r\n if (k == p.B3G_RIGHT_ARROW and (v & p.KEY_IS_DOWN) and PID_CONTROL==False):\r\n targetVel = 5\r\n for joint in range(1,3):\r\n p.setJointMotorControl2(husky,2*joint, p.VELOCITY_CONTROL, targetVelocity =targetVel,force = maxForce)\r\n for joint in range(1,3):\r\n p.setJointMotorControl2(husky,2*joint+1, p.VELOCITY_CONTROL,targetVelocity =-1*targetVel,force = maxForce)\r\n\r\n p.stepSimulation()\r\n if (k == p.B3G_RIGHT_ARROW and (v & p.KEY_WAS_RELEASED)):\r\n targetVel = 0\r\n for joint in range(2, 6):\r\n p.setJointMotorControl2(husky, joint, p.VELOCITY_CONTROL,targetVelocity = targetVel,force = maxForce)\r\n\r\n if (k == p.B3G_LEFT_ARROW and (v & p.KEY_IS_DOWN)and PID_CONTROL==False):\r\n targetVel = 5\r\n for joint in range(1,3):\r\n p.setJointMotorControl2(husky,2* joint+1, p.VELOCITY_CONTROL,targetVelocity = targetVel,force = maxForce)\r\n for joint in range(1,3):\r\n p.setJointMotorControl2(husky,2* joint, p.VELOCITY_CONTROL,targetVelocity =-1* targetVel,force = maxForce)\r\n\r\n p.stepSimulation()\r\n if (k == p.B3G_LEFT_ARROW and (v & p.KEY_WAS_RELEASED)):\r\n targetVel = 0\r\n for joint in range(2, 6):\r\n p.setJointMotorControl2(husky, joint, p.VELOCITY_CONTROL, targetVelocity =targetVel,force = maxForce)\r\n\r\n p.stepSimulation()\r\n ### Press c to start auto mode\r\n if (k == ord('c') and (v & p.KEY_WAS_TRIGGERED)):\r\n print(\"\\nPID Control-on/Auto mode on\")\r\n PID_CONTROL = True\r\n if (k == ord('r') and (v & p.KEY_WAS_TRIGGERED)):\r\n print(\"\\nPID Control-off ,back to manual\")\r\n PID_CONTROL = False\r\n ### Press m to change mode \r\n if (k == ord('m') and (v & p.KEY_WAS_TRIGGERED)):\r\n b=~b\r\n print(\"\\nVisual mode\")\r\n ### Press p to take a picture from driver's view \r\n if (k == ord('p') and (v & p.KEY_IS_DOWN)):\r\n rear_bump= p.getLinkState(husky,9)\r\n front_bump= p.getLinkState(husky,8)\r\n v1= front_bump[0][0]-rear_bump[0][0]\r\n v2= front_bump[0][1]-rear_bump[0][1]\r\n v3= front_bump[0][2]-rear_bump[0][2]\r\n frontVec= np.array(list((v1, v2, v3)))\r\n t= (frontVec[0]**2 + frontVec[1]**2 + frontVec[2]**2)**(0.5)\r\n frontDrn= frontVec/t\r\n s1=20\r\n eyePos= [front_bump[0][0], front_bump[0][1], front_bump[0][2]]\r\n targetPos= list(frontDrn*s1)\r\n upVec= [0,0,1]\r\n view_matrix= p.computeViewMatrix(eyePos, targetPos, upVec)\r\n\r\n width= 512\r\n height= 512\r\n aspect_ratio= width/height\r\n near= 0.1\r\n far= 10\r\n fov= 120\r\n projection_matrix= p.computeProjectionMatrixFOV(fov, aspect_ratio, near, far)\r\n\r\n images= p.getCameraImage(width, height, view_matrix, projection_matrix, shadow=True, renderer= p.ER_BULLET_HARDWARE_OPENGL)\r\n image= images[2].reshape((height, width, 4))\r\n image= cv2.cvtColor(image, cv2.COLOR_RGBA2BGRA) \r\n cv2.imshow('Driver_View', image)\r\n cv2.waitKey(0)\r\n cv2.destroyAllWindows()\r\n ## Visual instructing mode(a little slower due to waitkey, background should be light coloured than instrucor's hand)\r\n else:\r\n #update your droid cam code for better control and visualisation\r\n cap= cv2.VideoCapture(\"http://192.168.43.53:4747/video\")\r\n \r\n while(True):\r\n _,image= cap.read()\r\n #gray= cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\r\n #mask= cv2.adaptiveThreshold(gray, 255, cv2.ADAPTIVE_THRESH_GAUSSIAN_C, cv2.THRESH_BINARY_INV, 641, 2)\r\n #mask= cv2.erode(mask, (5,5), iterations=10)\r\n image=cv2.cvtColor(image,cv2.COLOR_BGR2RGB) \r\n image=cv2.rotate(image,cv2.ROTATE_90_CLOCKWISE)\r\n lower=np.array([80,60,50])\r\n high=np.array([165,126,110])\r\n mask=cv2.inRange(image,lower,high)\r\n length= mask.shape[1]\r\n bredth= mask.shape[0]\r\n ex= length//3\r\n screen1= mask[:, 0:ex]\r\n screen2= mask[:, ex:(length-ex)]\r\n screen3= mask[:, (length-ex):]\r\n c1,_= cv2.findContours(screen1, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\r\n c2,_= cv2.findContours(screen2, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\r\n c3,_= cv2.findContours(screen3, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\r\n \r\n \r\n \r\n if len(c1)!=0:\r\n c= sorted(c1, key= cv2.contourArea, reverse= True)[0]\r\n a1= cv2.contourArea(c)\r\n else:\r\n a1=0\r\n if len(c2)!=0:\r\n c= sorted(c2, key= cv2.contourArea, reverse= True)[0]\r\n a2= cv2.contourArea(c)\r\n else:\r\n a2=0\r\n if len(c3)!=0:\r\n c= sorted(c3, key= cv2.contourArea, reverse= True)[0]\r\n a3= cv2.contourArea(c)\r\n else:\r\n a3=0\r\n a4= max(a1, a2, a3)\r\n if a4==a3:\r\n ##speed= 40##((sp+sx)/3)*2\r\n speed1= 160\r\n speed2= -160\r\n print(\"R\")\r\n\r\n elif a4==a1:\r\n #speed= -40#-((sp+sx)/3)*2\r\n speed1= -160\r\n speed2= 160\r\n print(\"L\")\r\n else:\r\n speed1= 160\r\n speed2= 160\r\n print(\"F\")\r\n targetVel_R = speed1\r\n targetVel_L = speed2\r\n for joint in range(1,3):\r\n p.setJointMotorControl2(husky,2* joint, p.VELOCITY_CONTROL, targetVelocity =targetVel_R,force = maxForce)\r\n for joint in range(1,3):\r\n p.setJointMotorControl2(husky,2* joint+1, p.VELOCITY_CONTROL,targetVelocity =targetVel_L,force = maxForce)\r\n p.stepSimulation()\r\n mask[:,ex]=255\r\n mask[:, length-ex]=255\r\n cv2.imshow(\"mask\", mask)\r\n k= cv2.waitKey(1)\r\n if k==27:\r\n b=0\r\n cap.release()\r\n cv2.destroyAllWindows()\r\n break\r\np.disconnect()\r\n\r\n\r\n","sub_path":"Robotics_Summer_camp.py","file_name":"Robotics_Summer_camp.py","file_ext":"py","file_size_in_byte":23516,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"85284448","text":"# ------------------------------------------------------------------------------------------------------------------------------------------\n# region FEATURE SELECTION\n\ndef get_ds_best_features(dfX, dfY, dataset_name, cluster_topic, get_past_years_x_y, kept_years, n_best_features):\n x_data, scaler = scale_array_between_0_1(dfX)\n\n le = LabelEncoder()\n y_data = le.fit_transform(dfY[cluster_topic]).tolist()\n\n select_k_best = SelectKBest(chi2, k=n_best_features)\n select_k_best.fit_transform(x_data, y_data)\n selected_cols = select_k_best.get_support()\n\n # List best features\n cols = list(dfX.columns[selected_cols])\n print_check(dataset_name, cols)\n\n folder_name = sanitize_string(dataset_name)\n sub_folder_name = 'BEST_FEATURES'\n cluster_by = sanitize_string(cluster_topic)\n file_path = '{}/{}/COLS_{}'.format(folder_name, sub_folder_name, cluster_by)\n save_as_pickle(cols, file_path)\n\n return cols\n\n\ndef get_best_features():\n for ds_name in DS_NAMES:\n # --\n df = read_pickle(ds_name)\n dfX, dfY = get_past_years_x_y(df, MyCols.NB_OF_PAST_YRS_KEPT)\n\n # --\n for cluster_topic in CLUSTER_TOPICS:\n for n in N_BEST_FEATURES:\n COLS = get_ds_best_features(dfX, dfY, wka.cat_type_of_program_fr, get_program_type_xy, MyClusters.NB_MODELING_YEARS, n)\n\n # Get scree charts\n best_cols = COLS + [wka.cat_type_of_program_fr, wka.yr_treatment]\n BEST_DS_COLS['{}-{}-{}'.format(sanitize_string(ds_name), sanitize_string(cluster_topic), n)] = best_cols\n\n # -- Customization\n # --\n # df_x = drop_columns(df_x, get_cols_with_prefix(df_x, 'BIN_REFERRAL_'), errors='ignore')\n # df_x = drop_columns(df_x, get_cols_with_prefix(df_x, 'NUM_PROD_'), errors='ignore')\n # df_x = drop_columns(df_x, get_cols_with_prefix(df_x, 'BIN_MAIN_PROD_'), errors='ignore')\n df_x = drop_columns(df_x, get_cols_with_prefix(df_x, 'COUNT_'), errors='ignore')\n # df_x = drop_columns(df_x, get_cols_with_prefix(df_x, 'BIN_PROD_'), errors='ignore')\n df_x = drop_columns(df_x, get_cols_with_prefix(df_x, 'BIN_NATIONALITY_'), errors='ignore')\n df_x = drop_columns(df_x, get_cols_with_prefix(df_x, 'BIN_PREV_TREAT_'), errors='ignore')\n","sub_path":"tmp.py","file_name":"tmp.py","file_ext":"py","file_size_in_byte":2267,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"312458236","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\n@Author:Fangyang\n@User:haizhi\n@File:JUDGEMENT_WENSHU_for_jilin\n@time:2018/8/8下午5:57\n@Software:PyCharm\n\n\"\"\"\nimport time\nimport random\nimport re\n\nimport json\nimport requests\nimport pymysql\nimport pymongo\nfrom bs4 import BeautifulSoup\n#config_of_mysql = pymysql.connect(\"localhost\", \"root\", \"fangyang\", \"test\")\nconfig_of_mongodb = pymongo.MongoClient('mongodb://localhost:27017')\n\n\nheader = {'User-Agent:': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_13_6) AppleWebKit/537.36 \\\n (KHTML, like Gecko) Chrome/67.0.3396.99 Safari/537.36'}\n\n\nclass Judgement_wenshu(object):\n def __init__(self, database_name, url):\n self.database_name = database_name\n self.url = url\n self.id_list = []\n self.title_list = []\n self.Pageurl_list = []\n self.source_page_list = []\n self.wenshu_content = []\n\n def link_database(self):#默认使用mongodb\n # noinspection PyBroadException\n try:\n if self.database_name == \"mysql\":\n return 0\n elif self.database_name == \"mongodb\":\n return config_of_mongodb\n except Exception:\n print(\"缺少数据库名称,默认使用mongodb\")\n return config_of_mongodb\n\n def start_request(self, ip):\n proxy = {\n 'http': 'http://{}'.format(ip),\n 'https': 'https:{}'.format(ip)\n }\n try:\n context = requests.get(self.url).text\n except Exception as e:\n print(e + \"访问失败\")\n return -1\n soup = BeautifulSoup(context, \"html.parser\")\n content_url = \"http://www.jlsfy.gov.cn:8080/susong51/cpws/loadPrintContent.htm?id=\"#详情页的获取页面\n list_context = soup.find_all(onclick=re.compile('javascript:cpwsDetail'))\n for tag in list_context:\n self.id_list.append(tag['onclick'][23:-3])\n temp = str(tag.contents[1]).split('>')\n self.title_list.append(temp[3][:-3])\n for id in self.id_list:\n url1 = content_url + id\n self.Pageurl_list.append(url1)\n if len(self.id_list) == 0:\n return 0\n page_content = requests.get(url1, header).text\n print(\"获取到\"+url1+\"的数据\")\n temp_js = json.loads(page_content)\n self.source_page_list.append(temp_js['data'])\n print(str(url1) + \" \" + \"ok\")\n # 获取到信息以后剔除HTML标签\n pat = re.compile('>(.*?)<')\n p = ''.join(pat.findall(temp_js['data']))\n self.wenshu_content.append(p)\n return 1\n\n def write_to_database(self):\n global config_of_mongodb\n if self.databasename == \"mysql\":\n for i in range(len(self.title_list)):\n sql = \"INSERT IGNORE INTO judgement_wenshu(wenshu_content, \\\n case_name, _site_record_id,topic,source,source_page,PageUrl,_in_time) \\\n VALUES ('%s', '%s', '%s','%s','%s','%s','%s','%s')\" % \\\n (self.wenshu_content[i],\n self.title_list[i],\n \"www.jlsfy.gov.cn:8080\",\n \"0\",\n \"judgement_wenshu\",\n \"诉讼无忧新版采集脚本v0_吉林高院省\",\n self.source_page_list[i],\n self.Pageurl_list[i],\n time.time())\n #cursor = config_of_mysql.cursor()\n # cursor.execute(sql)\n # config_of_mysql.commit()\n return 1\n elif self.database_name == \"mongodb\":\n for i in range(len(self.title_list)):\n mycol = config_of_mongodb['test']['judgement_wenshu']\n mycol.insert_one({\"wenshu_content\": self.wenshu_content[i], 'case_name': self.title_list[i],\n 'site_record_id': 'www.jlsfy.gov.cn:8080',\n 'topic': 'judgement_wenshu',\n 'source': '诉讼无忧新版采集脚本v0',\n 'source_page': self.source_page_list[i],\n 'PageUrl': self.Pageurl_list[i],\n '_in_time': time.strftime(\"%Y-%m-%d %H:%M:%S\", time.localtime())\n }\n )\n return 1\n\n def work(self):\n myip=config_of_mongodb['proxy']\n ip = []\n myproxy = myip['raw_proxy'].find()\n for i in myproxy:\n ip.append(i['proxy'])\n self.start_request(random.choice(ip))\n self.writetodatabase()\n\n\nif __name__ == '__main__':\n\n cjws = Judgement_wenshu('mongodb', 'http://www.jlsfy.gov.cn:8080/susong51/fymh/750/cpws.htm?page=1')\n cjws.work()\n \n\n\n\n\n\n\n\n\n\n\n","sub_path":"JUDGEMENT_WENSHU_for_jilin.py","file_name":"JUDGEMENT_WENSHU_for_jilin.py","file_ext":"py","file_size_in_byte":4936,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"15909120","text":"import torch\nimport torch.nn as nn\n\nimport speechbrain as sb\nfrom speechbrain.nnet.pooling import StatisticsPooling\nfrom speechbrain.nnet.CNN import Conv1d\nfrom speechbrain.nnet.linear import Linear\nfrom speechbrain.nnet.normalization import BatchNorm1d\nfrom speechbrain.lobes.models.dual_path import Decoder\n\n\nclass PASEEncoder(torch.nn.Module):\n def __init__(\n self,\n activation=torch.nn.PReLU,\n use_sincnet=True,\n in_channels=1,\n blocks_channels=[128,256,100],\n blocks_kernel_sizes=[11,11,1],\n blocks_strides=[2,2,1],\n ):\n super().__init__()\n self.blocks = nn.ModuleList()\n\n first_hidden_layer = Conv1d\n self.blocks.append(\n first_hidden_layer(\n in_channels=in_channels,\n out_channels=blocks_channels[0],\n kernel_size=blocks_kernel_sizes[0],\n stride=blocks_strides[0],\n )\n )\n in_channels = blocks_channels[0]\n \n for block_index in range(1, len(blocks_channels)-1):\n out_channels = blocks_channels[block_index]\n self.blocks.extend(\n [\n Conv1d(\n in_channels=in_channels,\n out_channels=out_channels,\n kernel_size=blocks_kernel_sizes[block_index],\n stride=blocks_strides[block_index],\n ),\n BatchNorm1d(input_size=out_channels),\n activation(),\n ]\n )\n in_channels = blocks_channels[block_index]\n self.blocks.extend(\n [\n Conv1d(\n in_channels=in_channels,\n out_channels=blocks_channels[block_index + 1],\n kernel_size=blocks_kernel_sizes[block_index + 1],\n stride=blocks_strides[block_index + 1],\n ),\n BatchNorm1d(input_size=blocks_channels[block_index + 1], affine=False),\n ]\n )\n\n def forward(self, x, *args, **kwargs):\n for layer in self.blocks:\n try:\n x = layer(x, *args, **kwargs)\n except TypeError:\n x = layer(x)\n return x\n","sub_path":"transfer/models/transfer_model.py","file_name":"transfer_model.py","file_ext":"py","file_size_in_byte":2283,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"14799376","text":"# Standard imports\nimport cv2\nimport sys\nimport numpy as np \n \n# Read images\nsrc = cv2.imread(sys.argv[1])\ndst = cv2.imread(sys.argv[2])\n \n \n# Create a rough mask around the airplane.\nsrc_mask = np.zeros(src.shape, src.dtype)\nsrc_mask[16:190,134:310] = 1.0\n#poly = np.array([ [16,80], [30,54], [151,63], [254,37], [298,90], [272,134], [43,122] ], np.int32)\n#cv2.fillPoly(src_mask, [poly], (255, 255, 255))\n \n# This is where the CENTER of the airplane will be placed\ncenter = (0,0)\n \n# Clone seamlessly.\noutput = cv2.seamlessClone(src, dst, src_mask, center, cv2.NORMAL_CLONE)\n \n# Save result\ncv2.imwrite(\"opencv-seamless-cloning-example.jpg\", output);\n","sub_path":"New_Driver/seamless_clone.py","file_name":"seamless_clone.py","file_ext":"py","file_size_in_byte":652,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"329145062","text":"from django.contrib import admin\n\nfrom blog.models import Entry, UserProfile\n\n\nclass EntryAdmin(admin.ModelAdmin):\n fieldsets = [\n (None, {'fields': ['title']}),\n (None, {'fields': ['author']}),\n ('Date Published', {'fields': ['pub_date']}),\n ('Dose of Happy', {'fields': ['content']}),\n ('Happy Level', {'fields': ['rating']}),\n ]\n list_display = ('title', 'pub_date', 'author',)\n\n\nclass UserProfileAdmin(admin.ModelAdmin):\n fieldsets = [\n (None, {'fields': ['user']}),\n ('About', {'fields': ['bio']}),\n ('Picture', {'fields': ['picture']}),\n ]\n list_display = ('user', 'bio', 'picture',)\n\n\nadmin.site.register(Entry, EntryAdmin)\nadmin.site.register(UserProfile, UserProfileAdmin)\n","sub_path":"blog/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":784,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"589498336","text":"\"\"\"\nAn implementation of the Probabilistic Ensembles with Trajectory Sampling (PETS) algorithm\nfrom Chua et al (2018).\n\n\"\"\"\nimport sys\nimport gym\nimport torch\n\nimport rlkit.torch.pytorch_util as ptu\nfrom rlkit.torch.PETS import Model, MPCPolicy, PETSTrainer\nfrom rlkit.envs.wrappers import NormalizedBoxEnv\nfrom rlkit.data_management.env_replay_buffer import EnvReplayBuffer\nfrom rlkit.launchers.launcher_util import setup_logger\nfrom rlkit.samplers.data_collector import MdpPathCollector\nfrom rlkit.torch.torch_rl_algorithm import TorchBatchRLAlgorithm\nfrom custom.mountain_car_continuous import mountain_car_continuous_reward\nfrom custom.cartpole_swingup import CartPoleSwingUpEnv, cartpole_swingup_reward_v1\n\nptu.set_gpu_mode(True)\n\n\ndef experiment(variant):\n # expl_env = NormalizedBoxEnv(gym.make('BipedalWalker-v3'))\n # eval_env = NormalizedBoxEnv(gym.make('BipedalWalker-v3'))\n # expl_env = NormalizedBoxEnv(CartPoleSwingUpEnv())\n # eval_env = NormalizedBoxEnv(CartPoleSwingUpEnv())\n from custom.mjcartpole import CartpoleEnv, get_cp_reward # This import will fail if you don't have Mujoco.\n expl_env = CartpoleEnv()\n eval_env = CartpoleEnv()\n # expl_env = NormalizedBoxEnv(gym.make('MountainCarContinuous-v0'))\n # eval_env = NormalizedBoxEnv(gym.make('MountainCarContinuous-v0'))\n assert variant['policy']['num_particles'] % variant['model']['num_bootstrap'] == 0, \"There must be an even number of particles per bootstrap\" # NOQA\n assert variant['algorithm_kwargs']['num_trains_per_train_loop'] % variant['model']['num_bootstrap'] == 0, \"Must be an even number of train steps per bootstrap\" # NOQA\n obs_dim = expl_env.observation_space.low.size\n action_dim = expl_env.action_space.low.size\n\n model = Model(\n hidden_sizes=variant['model']['hidden_sizes'],\n obs_dim=obs_dim,\n action_dim=action_dim,\n num_bootstrap=variant['model']['num_bootstrap'],\n rew_function=get_cp_reward,\n # env=expl_env,\n # rew_function=mountain_car_continuous_reward # for now\n )\n policy = MPCPolicy(\n model=model,\n obs_dim=obs_dim,\n action_dim=action_dim,\n num_particles=variant['policy']['num_particles'],\n cem_horizon=variant['policy']['cem_horizon'],\n cem_iters=variant['policy']['cem_iters'],\n cem_popsize=variant['policy']['cem_popsize'],\n cem_num_elites=variant['policy']['cem_num_elites'],\n sampling_strategy=variant['policy']['sampling_strategy'],\n optimizer=variant['policy']['optimizer'],\n opt_freq=variant['policy']['opt_freq'],\n cem_alpha=variant['policy']['cem_alpha'],\n )\n trainer = PETSTrainer(expl_env,\n policy,\n model,\n lr=variant['lr'])\n eval_path_collector = MdpPathCollector(\n eval_env,\n policy,\n )\n expl_path_collector = MdpPathCollector(\n expl_env,\n policy,\n )\n replay_buffer = EnvReplayBuffer(\n variant['replay_buffer_size'],\n expl_env,\n )\n algorithm = TorchBatchRLAlgorithm(\n trainer=trainer,\n exploration_env=expl_env,\n evaluation_env=eval_env,\n exploration_data_collector=expl_path_collector,\n evaluation_data_collector=eval_path_collector,\n replay_buffer=replay_buffer,\n **variant['algorithm_kwargs']\n )\n algorithm.to(ptu.device)\n algorithm.train()\n\n\nif __name__ == '__main__':\n name = sys.argv[1]\n torch.set_num_threads(8)\n variant = dict(\n policy=dict(\n num_particles=20,\n cem_horizon=25,\n cem_iters=5,\n cem_popsize=400,\n cem_num_elites=40,\n sampling_strategy='TSinf',\n optimizer='CEM',\n cem_alpha=0.1,\n opt_freq=1,\n ),\n algorithm_kwargs=dict(\n num_epochs=5,\n num_eval_steps_per_epoch=200,\n num_trains_per_train_loop=100000,\n num_expl_steps_per_train_loop=500,\n min_num_steps_before_training=10000,\n max_path_length=200,\n batch_size=256,\n ),\n # policy=dict(\n # num_particles=10,\n # cem_horizon=25,\n # cem_iters=2,\n # cem_popsize=10,\n # cem_num_elites=3,\n # sampling_strategy='TS1',\n # optimizer='CEM',\n # cem_alpha=0.1,\n # opt_freq=25,\n # ),\n # algorithm_kwargs=dict(\n # num_epochs=3000,\n # num_eval_steps_per_epoch=400,\n # num_trains_per_train_loop=100,\n # num_expl_steps_per_train_loop=200,\n # min_num_steps_before_training=200,\n # max_path_length=200,\n # batch_size=256,\n # ),\n model=dict(\n num_bootstrap=5,\n hidden_sizes=[500, 500, 500],\n ),\n replay_buffer_size=int(1e7),\n lr=0.001,\n )\n\n setup_logger(name, variant=variant)\n # import pudb; pudb.set_trace()\n experiment(variant)\n","sub_path":"examples/pets.py","file_name":"pets.py","file_ext":"py","file_size_in_byte":5410,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"79443023","text":"import copy\n\nfrom pure_mcts.keyset import KeySet\nfrom pure_mcts.mcts_dpw import MCTSStochastic, StochasticAction, StochasticState\nimport numpy as np\nfrom race_components.race_ol_uct import strategic_rollout\n\nfrom helpers import (argmax, is_atari_game, copy_atari_state, restore_atari_state, stable_normalizer)\n\n\nclass RaceStochasticState(StochasticState):\n def __init__(self, index, r, terminal, parent_action, na, signature, budget, env=None, max_depth=200, owner=None):\n assert owner is not None, \"Owner parameter must be specified for RaceStochasticState class constructor\"\n self.owner = owner\n self.end_turn = env.has_transitioned()\n\n super(RaceStochasticState, self).__init__(index, r, terminal, parent_action, na, signature, budget,\n env=env, max_depth=max_depth)\n\n if self.terminal or terminal:\n self.V = np.zeros(env.agents_number)\n\n action_list = env.get_available_actions(owner)\n self.child_actions = [RaceStochasticAction(a, parent_state=self, owner=owner) for a in action_list]\n\n def random_rollout(self, budget, env, max_depth=200):\n return strategic_rollout(env, budget, max_depth=200, terminal=self.terminal, root_owner=self.owner)\n\nclass RaceStochasticAction(StochasticAction):\n def __init__(self, index, parent_state, Q_init=0.0, owner=None):\n assert owner is not None, \"Owner must be specified for constructor of RaceStochasticAction class\"\n self.owner = owner\n super(RaceStochasticAction, self).__init__(index, parent_state, Q_init=Q_init)\n\n def add_child_state(self, s1, r, terminal, signature, budget, env=None, max_depth=200):\n child_state = RaceStochasticState(s1, r, terminal, self, self.parent_state.na, signature, budget, env=env,\n max_depth=max_depth, owner=env.get_next_agent())\n self.child_states.append(child_state)\n\n sk = KeySet(s1)\n\n # s1_hash = s1.tostring()\n # self.state_indeces[sk.__hash__()] = self.n_children\n self.state_indices[sk] = self.n_children\n self.n_children += 1\n return child_state, child_state.remaining_budget\n\nclass RaceMCTSStochastic(MCTSStochastic):\n def __init__(self, root, root_index, model, na, gamma, alpha=0.6, depth_based_bias=False, owner=None):\n super(RaceMCTSStochastic, self).__init__(root, root_index, model, na, gamma,\n alpha=alpha, depth_based_bias=depth_based_bias)\n\n assert owner is not None, \"Owner must be specified for RaceMCTSStochastic constructor\"\n self.owner = owner\n\n def search(self, n_mcts, c, Env, mcts_env, budget, max_depth=200):\n ''' Perform the MCTS search from the root '''\n is_atari = is_atari_game(Env)\n if is_atari:\n snapshot = copy_atari_state(Env) # for Atari: snapshot the root at the beginning\n else:\n mcts_env = copy.deepcopy(Env) # copy original Env to rollout from\n # else:\n # restore_atari_state(mcts_env, snapshot)\n\n # Check that the environment has been copied correctly\n try:\n sig1 = mcts_env.get_signature()\n sig2 = Env.get_signature()\n if sig1.keys() != sig2.keys():\n raise AssertionError\n if not all(np.array_equal(sig1[key], sig2[key]) for key in sig1):\n raise AssertionError\n except AssertionError:\n print(\"Something wrong while copying the environment\")\n sig1 = mcts_env.get_signature()\n sig2 = Env.get_signature()\n print(sig1.keys(), sig2.keys())\n exit()\n\n if self.root is None:\n # initialize new root\n self.root = RaceStochasticState(self.root_index, r=0.0, terminal=False, parent_action=None,\n na=self.na, signature=Env.get_signature(),\n env=mcts_env, budget=budget, owner=self.owner)\n else:\n self.root.parent_action = None # continue from current root\n if self.root.terminal:\n raise (ValueError(\"Can't do tree search from a terminal state\"))\n\n while budget > 0:\n state = self.root # reset to root for new trace\n if not is_atari:\n mcts_env = copy.deepcopy(Env) # copy original Env to rollout from\n else:\n restore_atari_state(mcts_env, snapshot)\n mcts_env.seed()\n st = 0\n while not state.terminal:\n bias = c * self.gamma ** st / (1 - self.gamma) if self.depth_based_bias else c\n action = state.select(c=bias)\n k = np.ceil(self.beta * action.n ** self.alpha)\n if k >= action.n_children:\n s1, r, t, _ = mcts_env.step(action.index)\n # if action.index == 0 and not np.array_equal(s1.flatten(), action.parent_state.index.flatten()):\n # print(\"WTF\")\n if mcts_env.has_transitioned():\n budget -= 1\n if action.get_state_ind(s1) != -1:\n state = action.child_states[action.get_state_ind(s1)] # select\n state.r = r\n else:\n state, budget = action.add_child_state(s1, r, t, mcts_env.get_signature(), budget, env=mcts_env,\n max_depth=max_depth - st) # expand\n break\n else:\n state = action.sample_state()\n mcts_env.set_signature(state.signature)\n if state.terminal:\n budget -= 1\n\n if mcts_env.has_transitioned():\n st += 1\n\n # Back-up\n R = np.zeros(mcts_env.agents_number)\n\n if not state.terminal:\n R = copy.deepcopy(state.V)\n state.update()\n agents_reward = copy.deepcopy(state.r)\n while state.parent_action is not None: # loop back-up until root is reached\n owner = state.parent_action.owner # rewards are stored in the state following the action, which has different owner\n if not state.terminal:\n if state.end_turn:\n agents_reward = copy.deepcopy(state.r)\n R[owner] = agents_reward[owner] + self.gamma * R[owner]\n else:\n R = copy.deepcopy(state.r)\n action = state.parent_action\n action.update(R[action.owner])\n state = action.parent_state\n state.update()","sub_path":"race_components/race_dpw.py","file_name":"race_dpw.py","file_ext":"py","file_size_in_byte":6803,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"319885443","text":"# Program to flatten a nested list\n\ndef flattenMat1(mat):\n # Nested List Naive\n flatten_matrix = [] \n \n for sublist in matrix: \n for val in sublist: \n flatten_matrix.append(val) \n return flatten_matrix\n \ndef flattenMat2(mat):\n # Nested List Comprehension \n return [val for sublist in matrix for val in sublist]\n \n #checking program \nmatrix = [[1, 2, 3], [4, 5], [6, 7, 8, 9]] \nprint(flattenMat2(matrix)) \n\n ","sub_path":"Shimon_Labs/flattenMatrix.py","file_name":"flattenMatrix.py","file_ext":"py","file_size_in_byte":457,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"351621295","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nfrom selenium import webdriver\nfrom selenium.common.exceptions import NoSuchElementException, StaleElementReferenceException\nimport time\nimport tweepy\nimport datetime\n\n\nclass Scraper:\n def __init__(self, query, delay_scrolling, limit):\n \"\"\"\n :param query:\n :param delay_scrolling:\n :param limit:\n \"\"\"\n self.query = query.lower()\n self.delay_scrolling = delay_scrolling\n self.limit = limit\n self.driver = webdriver.Firefox()\n self.tweet_selector = 'li.js-stream-item'\n self.id_selector = '.time a.tweet-timestamp'\n self.ids = []\n\n def format_day(self, date):\n return date.strftime('%Y-%m-%d')\n\n def form_url(self, query, since, until):\n p1 = 'https://twitter.com/search?q='\n p2 = query + '%20since%3A' + since + '%20until%3A' + until + '&src=typd'\n return p1 + p2\n\n def end_dates(self):\n date_today = datetime.date.today()\n dates_list = [datetime.date.today()]\n index = 1\n month_temp = date_today.month\n year_temp = date_today.year\n flag = False\n while (month_temp != date_today.month) or (year_temp != (date_today.year - 1)):\n index = index + 1\n month_temp = date_today.month - index\n if month_temp <= 0:\n month_temp = month_temp + 12\n if not flag:\n year_temp = year_temp - 1\n flag = True\n dates_list.append(datetime.date(year_temp, month_temp,\n self.last_day_of_month(datetime.date(year_temp, month_temp, 1)).day))\n del dates_list[-1]\n return dates_list\n\n def last_day_of_month(self, any_day):\n next_month = any_day.replace(day=28) + datetime.timedelta(days=4) # this will never fail\n return next_month - datetime.timedelta(days=next_month.day)\n\n def run_scrapping(self):\n for date in self.end_dates():\n start_date = self.format_day(datetime.date(date.year, date.month, 1))\n end_date = self.format_day(date)\n url = self.form_url(self.query, start_date, end_date)\n self.scrap_period(url)\n\n def extract_ids(self, stamps):\n ids = []\n for tweet in stamps:\n try:\n id = tweet.get_attribute('data-conversation-id')\n # id = tweet.find_element_by_css_selector(id_selector).get_attribute('href').split('/')[-1]\n ids.append(id)\n except StaleElementReferenceException as e:\n print('lost element reference', tweet)\n return ids\n\n def scrap_period(self, url):\n self.driver.get(url)\n time.sleep(self.delay_scrolling)\n try:\n found_tweets = self.driver.find_elements_by_css_selector(self.tweet_selector)\n # scroll down the browser until the results end\n increment = 10\n while len(found_tweets) <= self.limit and len(found_tweets) >= increment:\n print('scrolling down to load more tweets')\n self.driver.execute_script('window.scrollTo(0, document.body.scrollHeight);')\n time.sleep(self.delay_scrolling)\n found_tweets = self.driver.find_elements_by_css_selector(self.tweet_selector)\n increment = increment + 10\n # collect tweets id\n tweet_stamps = self.driver.find_elements_by_css_selector(self.id_selector)\n self.ids.extend(self.extract_ids(tweet_stamps))\n except NoSuchElementException:\n print('no tweets on this period')\n\n\nclass WebTweets:\n def __init__(self, query, delay, limit, class_type, api):\n self.scraper = Scraper(query, delay, limit)\n self.scraper.run_scrapping()\n self.ids = self.scraper.ids\n self.api = api\n self.class_type = class_type\n self.tweets = []\n\n def mine_scraped_ids(self, scrap_ids):\n ids_number = len(scrap_ids)\n scrapped_tweets = []\n start_i = 0\n end_i = 99\n stop = False\n while not stop:\n if end_i > ids_number:\n end_i = ids_number - 1\n stop = True\n scrapped_tweets.extend(self.api.statuses_lookup(scrap_ids[start_i:end_i], tweet_mode='extended'))\n start_i = end_i + 1\n end_i = end_i + 99\n return scrapped_tweets\n\n def filter_cls(self, tweet, cls):\n item_dict = tweet._json\n item_dict['class'] = cls\n return item_dict\n\n def get_tweets(self):\n temp_tweets = self.mine_scraped_ids(self.ids)\n for scraped_tweet in temp_tweets:\n self.tweets.append(self.filter_cls(scraped_tweet, self.class_type))\n return self.tweets\n\n\ndef set_api():\n with open('tokens', mode='r', encoding='utf-8') as myfile:\n lines = myfile.readlines()\n consumer_key = lines[0].strip()\n consumer_secret = lines[1].strip()\n access_token = lines[2].strip()\n access_secret = lines[3].strip()\n auth = tweepy.OAuthHandler(consumer_key, consumer_secret)\n auth.set_access_token(access_token, access_secret)\n api = tweepy.API(auth)\n return api\n\n\nif __name__ == '__main__':\n scraper = WebTweets('bdsfail', 5, 10, 'against', set_api())\n tweets = scraper.scrape_web()\n print(tweets)\n","sub_path":"tweets_collect_tool/scrape.py","file_name":"scrape.py","file_ext":"py","file_size_in_byte":5381,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"406193868","text":"class Settings():\n\n def __init__(self):\n self.screen_width = 1200\n self.screen_height = 600\n self.bg_color = (187,255,255)\n\n self.bullet_speed_factor = 20\n self.bullet_width = 3\n self.bullet_height = 15\n self.bullet_color = 60, 60, 60\n\n self.alien_speed_factor = 20\n self.fleet_drop_speed = 10\n # fleet_direction为1表示向右移,为-1表示向左移\n self.fleet_direction = 1\n\n self.ship_limit = 3 #ship还剩三条命\n self.alien_points = 50 #单个分数\n\n\n\n\n","sub_path":"Alien_Invasion/settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":560,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"307238331","text":"#!/usr/bin/python3\n\"\"\"\nDefine city routes.\n\"\"\"\n\n\nfrom flask import request, abort, jsonify\nfrom models import storage\nfrom models.city import City\nfrom models.state import State\nfrom api.v1.views import app_views\n\n\n@app_views.route(\"/states//cities\", methods=[\"GET\", \"POST\"])\ndef cities(state_id):\n \"\"\"Define /cities route with GET and POST methods\n\n POST - Create a new city\n GET - Get a list of all cities\n \"\"\"\n state = storage.get('State', state_id)\n if state is None:\n return abort(404)\n\n # GET\n if request.method == \"GET\":\n return jsonify([city.to_dict() for city in state.cities])\n\n # POST\n doc = request.get_json(silent=True)\n if doc is None:\n return \"Not a JSON\", 400\n if doc.get(\"name\") is None:\n return \"Missing name\", 400\n doc['state_id'] = state_id\n city = City(**doc)\n city.save()\n return jsonify(city.to_dict()), 201\n\n\n@app_views.route(\"/cities/\", methods=[\"GET\", \"DELETE\", \"PUT\"])\ndef city(city_id):\n \"\"\"Define /cities/ with GET, PUT and DELETE methodes\n\n GET - get a city with the given id\n PUT - Update the city with the given id\n DELETE - Deletes the cityy with the givem id\n \"\"\"\n city = storage.get('City', city_id)\n if city is None:\n abort(404)\n\n # GET\n if request.method == \"GET\":\n return jsonify(city.to_dict())\n\n # PUT\n elif request.method == \"PUT\":\n doc = request.get_json(silent=True)\n if doc is None:\n return \"Not a JSON\", 400\n\n for k, v in doc.items():\n if k not in (\"id\", \"created_at\", \"updated_at\"):\n setattr(city, k, v)\n city.save()\n return jsonify(city.to_dict())\n\n # DELETE\n city.delete()\n city.save()\n return jsonify({})\n","sub_path":"api/v1/views/cities.py","file_name":"cities.py","file_ext":"py","file_size_in_byte":1790,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"487752269","text":"#!/usr/bin/python3\n\"\"\"Gets data from an API\"\"\"\nimport json\nimport requests\n\n\ndef recurse(subreddit, hot_list=[], after=None):\n \"\"\"Shows hot post\"\"\"\n hot_post = []\n url = 'https://www.reddit.com/r/{}/hot.json'.format(subreddit)\n agent = {'User-Agent': 'requested'}\n parameters = {'after': after, 'limit': '100'}\n response = requests.get(url, headers=agent,\n params=parameters, allow_redirects=False)\n if response.status_code in [302, 404]:\n return None\n else:\n posts = response.json()['data']['children']\n after = response.json()['data']['after']\n for hot in posts:\n hot_list.append(hot)\n if after is not None:\n recurse(subreddit, hot_list, after)\n return hot_list\n","sub_path":"0x16-api_advanced/2-recurse.py","file_name":"2-recurse.py","file_ext":"py","file_size_in_byte":779,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"638133861","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Apr 11 21:41:48 2017\n\n@author: ram\n\"\"\"\n\nimport serial\n\nclass value:\n def __init__(self, strPort):\n self.ser = serial.Serial(strPort, 9600)\n\n def update(self,i):\n try: \n print(i)\n line = self.ser.readline()\n print(line)\n data = [float(val) for val in line.split(',')]\n print(data)\n print(rects)\n print(data)\n for rect, yi in zip(rects, data):\n print(yi)\n print(rect.set_height(yi))\n \n except KeyboardInterrupt:\n print('exiting')\n \n\n return rects\n\n # clean up\n def close(self):\n # close serial\n self.ser.flush()\n self.ser.close() ","sub_path":"graph_matplot/class/serialvalue.py","file_name":"serialvalue.py","file_ext":"py","file_size_in_byte":734,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"105059751","text":"from Panda import *\r\n# When you create a triangle or rectangle, you can choose which points in texture space each vertex goes with\r\n# In texture space, the lower left corner is P2(0,0) and the upper right is P2(1,1).\r\n\r\n\r\n\r\n# Here are four points in space. Build two triangles and texture them with realpanda.jpg\r\np1 = P3(-1, 0, -1)\r\np2 = P3(1, 0, -1)\r\np3 = P3(-1, 0, 1)\r\np4 = P3(1,0, 1)\r\n# For a triangle, the texture points are P2(0,0), P2(1,0), and P2(0, 1). You can alter\r\n# these by keyword parameters . Create a slider the more the second triangle to the side.\r\n# Change the texP3 coordinate below and see what happens.\r\nm = triangle(p1, p2, p3, texture = \"realpanda.jpg\")\r\nn = triangle(p3, p4, p2, texture = \"realpanda.jpg\", texP1 = P2(0, 1), texP2 = P2(1,1), texP3 = P2(1,0), position = P3(slider(), 0, 0))\r\nstart()","sub_path":"CompletedPandaHandouts/src/Texture and Geometry/06-texturecoords.py","file_name":"06-texturecoords.py","file_ext":"py","file_size_in_byte":825,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"353791074","text":"#! /usr/bin/env python\n# -*- coding: utf-8 -*-\n#\nfrom __future__ import print_function,division\n\nimport rospy\n\nfrom std_msgs.msg import Empty\n\nfrom larvae_behavior_classifier.msg import Behavior\n\nfrom larvae_behavior_classifier.msg import Features\n\nfrom mightex_controller.msg import CmdCurrent,CmdChannel\n\nfrom lavis_stimuli_controller.msg import StochasticStimuli\n\nimport random\n\n\nclass StochasticController(object):\n def __init__(self,*args,**kwargs):\n rospy.loginfo('Initializing stochastic_node...')\n self._initialized = False\n\n self._behavior_sub = rospy.Subscriber('behavior',Behavior,self._behavior_callback)\n self._features_sub = rospy.Subscriber('features',Features,self._features_callback)\n\n self._cmd_light_current_pub = rospy.Publisher('cmd_current',CmdCurrent,queue_size=1)\n self._cmd_light_off_pub = rospy.Publisher('cmd_off',CmdChannel,queue_size=1)\n self._cmd_all_lights_off_pub = rospy.Publisher('cmd_all_off',Empty,queue_size=1)\n\n self._start_sub = rospy.Subscriber('~start',StochasticStimuli,self._start_callback)\n self._stop_sub = rospy.Subscriber('~stop',Empty,self._stop_callback)\n\n self._feedback_period = rospy.get_param('~feedback_period')\n\n self._last_update_time = 0\n self._last_bandit = 0\n self._running = False\n self._stopped = False\n self._begin = False\n self.firstindex = 0\n\n\n self._stochastic_stimulus_channel = 0\n self._stochastic_stimulus_current = 0\n self._stochastic_left_probability = 0\n self._stochastic_right_probability = 0\n self._stochastic_background_channel = 0\n self._stochastic_background_current = 0\n self._stochastic_basal_current = 0\n\n self.impulsewidth = 10\n self.game_length = 500\n self.impulse_period = 1500\n self.s_filtered = 0\n self.bend_state = 0\n self.left_current = 0\n self.right_current = 0\n\n self.begin_run_forward_two_sec = 0\n self.run_forward_two_sec = False\n self.begin_thirteen_sec_wait = 0\n self.wait_thirteen_sec = 0\n self.begin_bandit_game = 0\n self.play_bandit_game = False\n\n rospy.loginfo('stochastic_node initialized!')\n self._initialized = True\n\n def _features_callback(self,data):\n if (self._begin == True):\n self.firstindex = data.index\n rospy.loginfo('first index: %d', self.firstindex)\n self._begin = False\n self.s_filtered = data.s_filtered\n self.index = data.index\n\n def _behavior_callback(self,data):\n\n if self._initialized and self._running:\n time_now = rospy.get_time()\n if (time_now - self._last_update_time) >= self._feedback_period:\n self._last_update_time = time_now\n\n cmd_current = CmdCurrent()\n cmd_current.channel = self._stochastic_background_channel\n cmd_current.current = 0\n self._cmd_light_current_pub.publish(cmd_current)\n\n # See if bandit is being played.\n # self._features_pub(play_bandit_game)\n if (self.play_bandit_game == False):\n # See if neither wait sequences are running\n # There is a 2 second buffer where larva must run in a forward line.\n # After completion of the 2 second run, there is an 13 second wait.\n if (self.run_forward_two_sec == False and self.wait_thirteen_sec == False):\n # If not, then attempt to initiate the 2-sec forward/run.\n if (data.b_bend == False and data.b_run):\n self.run_forward_two_sec = True\n self.begin_run_forward_two_sec = self.index\n rospy.loginfo('Initiate 2-sec run: %d', self.begin_run_forward_two_sec)\n # Beginning the 13-sec wait after finishing the 2-sec forward run\n # ensures both wait sequences never occur simultaneously\n elif (self.run_forward_two_sec == True):\n # See if it's been 2 seconds\n if (self.index - self.begin_run_forward_two_sec > 50*2):\n self.run_forward_two_sec = False\n # Initialize 13-sec wait sequence\n self.wait_thirteen_sec = True\n self.begin_thirteen_sec_wait = self.index\n rospy.loginfo('Begin 13-sec wait, Index: %d', self.begin_thirteen_sec_wait)\n # See if larvae still runs forward. Otherwise reset.\n else:\n if (data.b_bend == True or data.b_run == False):\n self.run_forward_two_sec = False\n # In general, it is good to ensure that the 13 second wait does not begin falsely.\n self.wait_thirteen_sec = False\n\n # 13 second wait.\n elif (self.wait_thirteen_sec == True):\n # See if it's been 13 seconds. If so, begin bandit game.\n if (self.index - self.begin_thirteen_sec_wait > 50*13):\n self.wait_thirteen_sec = False\n self.play_bandit_game = True\n self.begin_bandit_game = self.index\n rospy.loginfo('End 13-sec wait, Index: %d', self.index)\n # If not, then we do nothing.\n # Every time we don't play the bandit game, we make sure everything is quiescent.\n # rospy.loginfo(\"STOPPED: %d\", (self.index - self.firstindex + 1 + 500) % 1500)\n cmd_current = CmdCurrent()\n cmd_current.channel = self._stochastic_stimulus_channel\n cmd_current.current = 0\n self._cmd_light_current_pub.publish(cmd_current)\n\n # The bandit game is being played. Now we count indices:\n else:\n if ((self.index - self.begin_bandit_game) <= self.impulse_width ):\n cmd_current = CmdCurrent()\n cmd_current.channel = self._stochastic_background_channel\n # I might be able to speed up the frame-rate here if I sample from a small array, approximately gaussian.\n cmd_current.current = int(random.gauss(self._stochastic_background_current,20)) + self._stochastic_basal_current\n self._cmd_light_current_pub.publish(cmd_current)\n rospy.loginfo(\"BLUE: %d\", cmd_current.current)\n\n # Now we want to play the game for as long as the larva does not run forward for 2 seconds, after at least 10 seconds of gameplay.\n elif (self.index - self.begin_bandit_game) > self.impulse_width and (self.index - self.begin_bandit_game) <= self.game_length :\n # else:\n if data.b_bend and data.b_left == False and data.b_right == False:\n self.bend_state = 0\n elif data.b_bend and data.b_left and (self.s_filtered < 0.8):\n if self.bend_state == 0 or self.bend_state == 1:\n cmd_current = CmdCurrent()\n cmd_current.channel = self._stochastic_stimulus_channel\n self.bernoulli_left = random.random()\n self.left_current = self._stochastic_stimulus_current*(self.bernoulli_left < self._stochastic_left_probability)\n cmd_current.current = self.left_current\n self._cmd_light_current_pub.publish(cmd_current)\n self.bend_state = -1\n rospy.loginfo('Bernoulli left: %f, Current Left: %d',self.bernoulli_left, self.left_current)\n elif self.bend_state == -1:\n # Here we make sure sampling is every cast, not every frame.\n cmd_current = CmdCurrent()\n cmd_current.channel = self._stochastic_stimulus_channel\n cmd_current.current = self.left_current\n self._cmd_light_current_pub.publish(cmd_current)\n # rospy.loginfo('Current Left: %d',self.left_current)\n elif data.b_bend and data.b_right and (self.s_filtered < 0.8):\n if self.bend_state == 0 or self.bend_state == -1:\n cmd_current = CmdCurrent()\n cmd_current.channel = self._stochastic_stimulus_channel\n self.bernoulli_right = random.random()\n self.right_current = self._stochastic_stimulus_current*(self.bernoulli_right < self._stochastic_right_probability)\n cmd_current.current = self.right_current\n self.bend_state = 1\n self._cmd_light_current_pub.publish(cmd_current)\n rospy.loginfo('Bernoulli Right: %f, Current Right: %d',self.bernoulli_right, self.right_current)\n elif self.bend_state == 1:\n cmd_current = CmdCurrent()\n cmd_current.channel = self._stochastic_stimulus_channel\n cmd_current.current = self.right_current\n self._cmd_light_current_pub.publish(cmd_current)\n # rospy.loginfo('Current Right: %d',self.right_current)\n else:\n cmd_current = CmdCurrent()\n cmd_current.channel = self._stochastic_stimulus_channel\n cmd_current.current = 0\n self._cmd_light_current_pub.publish(cmd_current)\n else:\n cmd_current = CmdCurrent()\n cmd_current.channel = self._stochastic_stimulus_channel\n cmd_current.current = 0\n self._cmd_light_current_pub.publish(cmd_current)\n self.play_bandit_game = False\n self.wait_thirteen_sec = False\n self.run_forward_two_sec = False\n rospy.loginfo('Bandit game is finished. Index: %d', self.index)\n\n # Below is the protocol for extending the game past self.game_length\n # if (self.index - self.begin_bandit_game) > self.game_length:\n # if (self.run_forward_two_sec == False):\n # if (data.b_bend == False and data.b_run == True):\n # # We begin keeping track of how long larvae has been running forward, but only after 10 seconds of game play.\n # # There are reasons for and against this, but the case that wins this line of reasoning for me\n # # is the one where the larvae learns to perform a single-cast. Say a larva single-casts in ~2 seconds, begins running\n # # forward for 2 seconds. During the next 6 seconds, it might decide to cast once more. This extra sample is extremely useful\n # # because it informs the larva whether it made the right decision. Without it, the larva will learn that the best decision\n # # is the one that is made quickly. This is something I would like to avoid.\n #\n # # Mainly, this more-than-10-seconds condition is for extended multi-casts, and so we should enforce a 10-second game.\n # self.run_forward_two_sec = True\n # self.begin_run_forward_two_sec = self.index\n # else:\n # # We send protocol to thirteen second wait once larva has begun running forward for two seconds.\n # if (self.index - self.begin_run_forward_two_sec > 50*1):\n # self.play_bandit_game = False\n # self.wait_thirteen_sec = True\n # self.run_forward_two_sec = False\n # self.begin_thirteen_sec_wait = self.index\n # rospy.loginfo('Bandit game is finished. Index: %d', self.index)\n # else:\n # if (data.b_bend == True or data.b_run == False):\n # self.run_forward_two_sec = False\n\n def _start_callback(self,stimuli):\n self._stopped = False\n self._begin = True\n # rospy.loginfo('begin here')\n self._stochastic_stimulus_channel = stimuli.stochastic_stimulus_channel\n self._stochastic_stimulus_current = stimuli.stochastic_current\n self._stochastic_left_probability = stimuli.stochastic_left_probability\n self._stochastic_right_probability = stimuli.stochastic_right_probability\n self._stochastic_background_channel = stimuli.stochastic_background_channel\n self._stochastic_background_current = stimuli.stochastic_background_current\n\n cmd_current = CmdCurrent()\n cmd_current.channel = self._stochastic_background_channel\n cmd_current.current = 0\n self._cmd_light_current_pub.publish(cmd_current)\n\n rospy.loginfo('stochastic sleeping for {0} secs before starting'.format(stimuli.wait.to_sec()))\n rospy.sleep(stimuli.wait)\n # check to see if stopped while waiting\n if not self._stopped:\n rospy.loginfo('head_stochastic starting with stochastic_stimulus_current = {0}, stochastic_left_probability = {1}, stochastic_right_probability = {2}'.format(stimuli.stochastic_current,stimuli.stochastic_left_probability,stimuli.stochastic_right_probability))\n self._running = True\n\n def _stop_callback(self,data):\n self._running = False\n self._stopped = True\n rospy.loginfo('stochastic stopping')\n self._cmd_all_lights_off_pub.publish(Empty())\n\n\nif __name__ == '__main__':\n try:\n rospy.init_node('stochastic_node')\n hcc = StochasticController()\n rospy.spin()\n except rospy.ROSInterruptException:\n pass\n","sub_path":"lavis_stimuli_controller/nodes/stochastic_node.py","file_name":"stochastic_node.py","file_ext":"py","file_size_in_byte":14787,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"46808157","text":"import sqlite3\n\nclass DbDateTime:\n def __init__(self):\n self._db = None\n\n @property\n def db(self):\n if self._db == None:\n self._db = sqlite3.connect(\":memory:\")\n return self._db\n\n def datetime(self,time,modifier=None):\n if modifier != None:\n parameters=(time,modifier)\n sql = \"select julianday(datetime(?,?))\"\n else:\n parameters=(time,)\n sql = \"select julianday(datetime(?))\"\n cursor = self.db.execute(sql,parameters)\n return (list(cursor))[0][0]\n\n def format(self,julianday,format='%Y-%m-%d %H:%M:%S utc'):\n parameters=(format,julianday)\n sql = \"select strftime(?,?)\"\n cursor = self.db.execute(sql,parameters)\n return (list(cursor))[0][0]\n\n def now(self):\n return self.datetime('now')\n","sub_path":"dbdatetime.py","file_name":"dbdatetime.py","file_ext":"py","file_size_in_byte":843,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"621845190","text":"import io\n\nfrom unittest import TestCase\nfrom unittest.mock import Mock, patch\nfrom mondas.bot import Bot, Answer\n\n\nclass BotTest(TestCase):\n\n def setUp(self):\n self.bot = Bot(\"123\")\n\n def testYouCanAskAQuestion(self):\n bot = self.bot\n\n response = bot.ask(\"This is a question\")\n\n self.assertIsNotNone(response)\n\n def testItCanSayThings(self):\n bot = self.bot\n\n bot.say(\"This is a response\")\n\n self.assertEqual([\"This is a response\"], bot.output)\n\n def testItCanBeCleared(self):\n bot = self.bot\n\n bot.say(\"This is a response\")\n bot.say(\"This is another response\")\n\n self.assertEqual(\n [\"This is a response\", \"This is another response\"],\n bot.output\n )\n bot.clear()\n\n self.assertEqual([], bot.output)\n\n def testItHasASession(self):\n self.assertIsNotNone(self.bot.session_id)\n\n def testItHasAnApiAIObject(self):\n bot = self.bot\n\n self.assertIsNotNone(bot.ai)\n self.assertEqual(\"123\", bot.ai.client_access_token)\n\n def testAskFiresAnApiAIRequest(self):\n bot = self.bot\n mockApi = Mock()\n mockRequest = Mock()\n mockApi.text_request.return_value = mockRequest\n mockRequest.getresponse.return_value = io.StringIO(\"{}\")\n bot.ai = mockApi\n\n bot.ask(\"This is a test\")\n\n mockApi.text_request.assert_called()\n mockRequest.getresponse.assert_called()\n self.assertEqual(\"This is a test\", mockRequest.query)\n\n def testHandlesResponses(self):\n handler = Mock()\n handler.action = \"TestAction\"\n\n bot = self.bot\n\n bot.handlers[\"TestAction\"] = handler\n\n answer = Answer()\n answer.status = True\n answer.action = \"TestAction\"\n\n bot.handle(answer)\n \n handler.handle.assert_called_with(answer)\n","sub_path":"tests/bot/test_bot.py","file_name":"test_bot.py","file_ext":"py","file_size_in_byte":1882,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"116335127","text":"#!/usr/bin/python3\n\n\"\"\"Runs data against the cl model.\n\"\"\"\n\n__author__ = \"Martin Eigenmann\"\n__license__ = \"unlicence\"\n__version__ = \"0.0.1\"\n__email__ = \"martin.eigenmann@unisg.ch\"\n__status__ = \"Prototpye\"\n\nimport pandas as pd\nimport sys\nimport torch\nimport numpy as np\n\nclass NN(torch.nn.Module):\n def __init__(self, input_dimension,number_of_target_classes):\n super().__init__()\n # stack layers\n self.module_list = torch.nn.ModuleList()\n ## stacks contain of linear + nonlinear layers\n self.module_list.append(torch.nn.Linear(input_dimension,20))\n self.module_list.append(torch.nn.Tanh())\n #\n self.module_list.append(torch.nn.Linear(20,20))\n self.module_list.append(torch.nn.Tanh())\n #\n self.module_list.append(torch.nn.Linear(20,10))\n self.module_list.append(torch.nn.Tanh())\n # map to output layer\n self.module_list.append(torch.nn.Linear(10,number_of_target_classes))\n \n def forward(self,x):\n for module in self.module_list:\n x = module(x)\n return x\n\n\nwith torch.no_grad():\n df = pd.read_csv(sys.stdin)\n\n device = torch.device('cpu')\n model2 = NN(17,2)\n model2.load_state_dict(torch.load('/processors/torch', map_location=device))\n model2.eval()\n data_to_predict = df.drop(['b_start', 'b_duration', 'e_start', 'e_duration', 'e_distance','study','subject'], axis=1).to_numpy()[0]\n prediction = model2(torch.from_numpy(data_to_predict).float())\n predicted_class = np.argmax(prediction)\n\n b_start = int(df.iloc[0].b_start)\n e_start = int(df.iloc[0].e_start)\n study = df.iloc[0].study\n subject = df.iloc[0].subject\n\n print(f'prediction_class,b_start,e_start,study,subject')\n print(f'{predicted_class},{b_start},{e_start},{study},{subject}')\n","sub_path":"nifi/processors/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":1815,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"263842425","text":"#coding=utf-8\nfrom django.conf.urls import patterns, include, url\nfrom ansibleweb import views\n\nurlpatterns = [\n #url(r'$', views.ansible,name='ansible'),\n url(r'^cmd/$', views.ansiblecmd,name='ansiblecmd'),\n url(r'^batch/$',views.ansiblebatch, name='ansiblebatch'),\n url(r'^detail/(?P\\d+)/$',views.ansibledetail, name='ansibledetail'),\n url(r'^retry/(?P\\d+)/$', views.ansibleretry, name='ansibleretry'),\n url(r'^rerun/(?P\\d+)/$', views.ansiblererun, name='ansiblererun'),\n url(r'^group/$',views.ansiblegroup,name='ansiblegroup'),\n url(r'^log/$', views.ansiblelog,name='ansiblelog'),\n url(r'^inspection/$', views.ansibleinspection,name='ansibleinspection'),\n #url(r'^key/$', views.ansiblekey,name='ansiblekey'),\n #url(r'^playbook/$', views.ansibleplaybook,name='ansibleplaybook'),\n]","sub_path":"ansibleweb/ansible_urls.py","file_name":"ansible_urls.py","file_ext":"py","file_size_in_byte":827,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"244835067","text":"import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport streamlit as st\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.svm import SVC\nfrom sklearn.linear_model import LinearRegression \nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.metrics import r2_score, mean_absolute_error, mean_squared_error, mean_squared_log_error\n\n@st.cache()\ndef prediction(car_df, car_width, engine_size, horse_power, drive_wheel_fwd, car_comp_buick):\n\tX = car_df.iloc[:, :-1]\n\ty = car_df[\"price\"]\n\tX_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.3, random_state = 42)\n\tlin_reg = LinearRegression()\n\tlin_reg.fit(X_train, y_train)\n\tscore = lin_reg.score(X_train, y_train)\n\tprice = lin_reg.predict([[car_width, engine_size, horse_power, drive_wheel_fwd, car_comp_buick]])\n\tprice = price[0]\n\ty_test_pred = lin_reg.predict(X_test)\n\ttest_r2_score = r2_score(y_test, y_test_pred)\n\ttest_mae = mean_absolute_error(y_test, y_test_pred)\n\ttest_msle = mean_squared_log_error(y_test, y_test_pred)\n\ttest_rmse = np.sqrt(mean_squared_error(y_test, y_test_pred))\n\treturn(price, score, test_r2_score, test_mae, test_msle, test_rmse)\ndef app(car_df):\n\tst.subheader(\"Select Values\")\n\tcar_width = st.slider(\"Car Width\", float(car_df[\"carwidth\"].min()), float(car_df[\"carwidth\"].max()))\n\tengine_size = st.slider(\"Engine Size\", float(car_df[\"enginesize\"].min()), float(car_df[\"enginesize\"].max()))\n\thorse_power = st.slider(\"Horse Power\", float(car_df[\"horsepower\"].min()), float(car_df[\"horsepower\"].max()))\n\td_fwd = st.radio(\"Is it a Forward Drive Wheel Car?\", (\"Yes\", \"No\"))\n\tif d_fwd == \"No\":\n\t\td_fwd = 0\n\telse:\n\t\td_fwd = 1\n\tcom_bui = st.radio(\"Is the car manufactured by Buick\", (\"Yes\", \"No\"))\n\tif com_bui == \"No\" :\n\t\tcom_bui = 0\n\telse:\n\t\tcom_bui = 1\n\tif st.button(\"Predict\"):\n\t\tst.subheader(\"Prediction Results\")\n\t\tprice, score, car_r2, car_mae, car_msle, car_rmse = prediction(car_df, car_width, engine_size, horse_power, d_fwd, com_bui)\n\t\tst.success(f\"The predicted price of the car: {int(price)}\")\n\t\tst.info(f\"Accuracy score of the model: {score}\")\n\t\tst.info(f\"R2 Score: {car_r2}\")\n\t\tst.info(f\"Mean Absolute Error: {car_mae}\")\n\t\tst.info(f\"Root Mean Squared Error: {car_rmse}\")\n\t\tst.info(f\"Mean Squared Log Error: {car_msle}\")","sub_path":"predict.py","file_name":"predict.py","file_ext":"py","file_size_in_byte":2292,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"88422758","text":"\"\"\"Contains all events of a given day.\"\"\"\nclass EventList:\n\n errors = [] # Contains errors that occur during scrape\n\n def __init__(self, day, events):\n \"\"\"Construct an event list object.\n\n Args:\n day (str): The calendar date of the events\n events (list): A list of events\n\n Returns:\n EventList: An event list object.\n \"\"\"\n self.day = day\n self.events = events\n\n def to_json(self):\n \"\"\"Returns a JSON representation of the object.\n\n Returns: dict: A JSON representation of the object.\n \"\"\"\n j = {}\n j['day'] = self.day\n j['events'] = [e.to_json() for e in self.events]\n j['errors'] = self.errors\n return j","sub_path":"python/modules/EventList.py","file_name":"EventList.py","file_ext":"py","file_size_in_byte":744,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"182767292","text":"import boto3\n\ndynamo = boto3.resource('dynamodb',\n endpoint_url='http://localhost:8000',\n region_name='dummy',\n aws_access_key_id='dummy',\n aws_secret_access_key='dummy')\n\nt = dynamo.Table('CS411')\nt.delete()\n","sub_path":"tools/drop_local_table.py","file_name":"drop_local_table.py","file_ext":"py","file_size_in_byte":301,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"198758304","text":"# from . import cv2_transforms as transforms\n# import torchvision.transforms as transforms\n\n\ndef build_transforms(cfg_transforms):\n cfg_transforms = cfg_transforms.copy()\n transforms_list = list()\n for item in cfg_transforms:\n transforms_type = item.pop(\"type\")\n\n if transforms_type == 'transforms_type':\n if item['backend'] == 'cv2':\n from . import cv2_transforms as transforms\n else:\n # import torchvision.transforms as transforms\n from . import transforms as transforms\n else:\n transforms_params = item\n # for debug\n # print(transforms_params)\n if hasattr(transforms, transforms_type):\n # for debug\n # print('transforms: ', transforms)\n # print(getattr(transforms, transforms_type))\n transforms_list.append(getattr(transforms, transforms_type)(**transforms_params))\n else:\n raise ValueError(\"\\'type\\'{} of transforms is not defined!!!\".format(transforms_type))\n\n # print(transforms_list)\n return transforms.Compose(transforms_list)\n","sub_path":"classification/data/transforms/build.py","file_name":"build.py","file_ext":"py","file_size_in_byte":1168,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"220845901","text":"import pymysql\nimport logging\nimport datetime\nimport xmltodict\n\n\nfile = 'purchaseNotice_Brianskaya_obl_20160801_000000_20160831_235959_001.xml'\n\nfile_log = './log_tenders223/tenders223_ftp_' + str(datetime.date.today()) + '.log'\nlogging.basicConfig(level=logging.DEBUG, filename=file_log,\n format='%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n\n\ndef generator_univ(c):\n if type(c) == list:\n for i in c:\n yield i\n else:\n yield c\n\n\nclass Tender223:\n\n def __init__(self, tender223):\n self.tender223 = None\n tender_body = None\n tender_purchase = None\n tender_root = list(tender223.keys())\n tender_child = tender223[tender_root[0]]\n tender_child_2_list = list(tender_child.keys())\n for i in tender_child_2_list:\n if i.startswith('body') or i.startswith('ns2:body') or i.startswith('oos:body'):\n tender_body = tender_child[i]\n break\n if tender_body is not None:\n tender_item_list = list(tender_body.keys())\n for j in tender_item_list:\n if j.startswith('item') or j.startswith('ns2:item') or j.startswith('oos:item'):\n tender_purchase = tender_body[j]\n break\n if tender_purchase is not None:\n tender_purchase_list = list(tender_purchase.keys())\n for m in tender_purchase_list:\n if m.startswith('purchase') or m.startswith('ns2:purchase') or m.startswith('oos:purchase'):\n self.tender223 = tender_purchase[m]\n break\n\n @property\n def id_t(self):\n id_ten = 0\n if 'ns2:guid' in self.tender223:\n id_ten = self.tender223['ns2:guid']\n elif 'oos:guid' in self.tender223:\n id_ten = self.tender223['oos:guid']\n elif 'guid' in self.tender223:\n id_ten = self.tender223['guid']\n if id_ten is None:\n id_ten = 0\n return id_ten\n\n @property\n def purchaseNumber(self):\n purchaseNumber = ''\n if 'ns2:registrationNumber' in self.tender223:\n purchaseNumber = self.tender223['ns2:registrationNumber']\n elif 'oos:registrationNumber' in self.tender223:\n purchaseNumber = self.tender223['oos:registrationNumber']\n elif 'registrationNumber' in self.tender223:\n purchaseNumber = self.tender223['registrationNumber']\n if purchaseNumber is None:\n purchaseNumber = ''\n return purchaseNumber\n\n @property\n def docPublishDate(self):\n docPublishDate = ''\n if 'ns2:publicationDateTime' in self.tender223:\n docPublishDate = self.tender223['ns2:publicationDateTime']\n elif 'oos:publicationDateTime' in self.tender223:\n docPublishDate = self.tender223['oos:registrationNumber']\n elif 'publicationDateTime' in self.tender223:\n docPublishDate = self.tender223['publicationDateTime']\n if docPublishDate is None:\n docPublishDate = ''\n return docPublishDate\n\n def organizer_mainInfo(self, d):\n try:\n organizer_mainInfo = self.tender223['ns2:placer']['mainInfo'][d]\n except Exception:\n try:\n organizer_mainInfo = self.tender223['oos:placer']['mainInfo'][d]\n except Exception:\n try:\n organizer_mainInfo = self.tender223['placer']['mainInfo'][d]\n except Exception:\n organizer_mainInfo = ''\n\n if organizer_mainInfo is None:\n organizer_mainInfo = ''\n return organizer_mainInfo\n\n @property\n def placingWay_code(self):\n placingWay_code = ''\n if 'ns2:purchaseMethodCode' in self.tender223:\n placingWay_code = self.tender223['ns2:purchaseMethodCode']\n elif 'oos:purchaseMethodCode' in self.tender223:\n placingWay_code = self.tender223['oos:purchaseMethodCode']\n elif 'purchaseMethodCode' in self.tender223:\n placingWay_code = self.tender223['purchaseMethodCode']\n if placingWay_code is None:\n placingWay_code = ''\n return placingWay_code\n\n def electronicPlaceInfo(self, d):\n try:\n electronicPlaceInfo = self.tender223['ns2:electronicPlaceInfo'][d]\n except Exception:\n try:\n electronicPlaceInfo = self.tender223['oos:electronicPlaceInfo'][d]\n except Exception:\n try:\n electronicPlaceInfo = self.tender223['electronicPlaceInfo'][d]\n except Exception:\n electronicPlaceInfo = ''\n\n if electronicPlaceInfo is None:\n electronicPlaceInfo = ''\n return electronicPlaceInfo\n\n def get_attachments(self):\n if 'ns2:attachments' in self.tender223:\n if self.tender223['ns2:attachments'] is None:\n return []\n elif 'document' in self.tender223['ns2:attachments']:\n return generator_univ(self.tender223['ns2:attachments']['document'])\n else:\n return []\n elif 'oos:attachments' in self.tender223:\n if self.tender223['oos:attachments'] is None:\n return []\n elif 'document' in self.tender223['oos:attachments']:\n return generator_univ(self.tender223['oos:attachments']['document'])\n else:\n return []\n elif 'attachments' in self.tender223:\n if self.tender223['attachments'] is None:\n return []\n elif 'document' in self.tender223['attachments']:\n return generator_univ(self.tender223['attachments']['document'])\n else:\n return []\n else:\n return []\n\n def attach(self, attachment, d):\n try:\n ret_attach = attachment[d]\n except Exception:\n ret_attach = ''\n if ret_attach is None:\n ret_attach = ''\n return ret_attach\n\n def customer(self, d):\n try:\n customer_ret = self.tender223['ns2:customer']['mainInfo'][d]\n except Exception:\n try:\n customer_ret = self.tender223['oos:customer']['mainInfo'][d]\n except Exception:\n try:\n customer_ret = self.tender223['customer']['mainInfo'][d]\n except Exception:\n customer_ret = ''\n\n if customer_ret is None:\n customer_ret = ''\n return customer_ret\n\n def contact(self, d):\n try:\n contact_ret = self.tender223['ns2:contact'][d]\n except Exception:\n try:\n contact_ret = self.tender223['oos:contact'][d]\n except Exception:\n try:\n contact_ret = self.tender223['contact'][d]\n except Exception:\n contact_ret = ''\n\n if contact_ret is None:\n contact_ret = ''\n return contact_ret\n\n def get_lots(self):\n if 'ns2:lots' in self.tender223:\n if self.tender223['ns2:lots'] is None:\n return []\n elif 'lot' in self.tender223['ns2:lots']:\n return generator_univ(self.tender223['ns2:lots']['lot'])\n else:\n return []\n elif 'ns2:lot' in self.tender223:\n if self.tender223['ns2:lot'] is None:\n return []\n else:\n return generator_univ(self.tender223['ns2:lot'])\n elif 'oos:lots' in self.tender223:\n if self.tender223['oos:lots'] is None:\n return []\n elif 'oos:lot' in self.tender223['oos:lots']:\n return generator_univ(self.tender223['oos:lots']['oos:lot'])\n else:\n return []\n elif 'oos:lot' in self.tender223:\n if self.tender223['oos:lot'] is None:\n return []\n else:\n return generator_univ(self.tender223['oos:lot'])\n elif 'lots' in self.tender223:\n if self.tender223['lots'] is None:\n return []\n elif 'lot' in self.tender223['lots']:\n return generator_univ(self.tender223['lots']['lot'])\n else:\n return []\n elif 'lot' in self.tender223:\n if self.tender223['lot'] is None:\n return []\n else:\n return generator_univ(self.tender223['lot'])\n else:\n return []\n\n def get_lotitems(self, lot):\n if 'lotData' in lot:\n if 'lotItems' in lot['lotData']:\n if lot['lotData']['lotItems'] is None:\n return []\n elif 'lotItem' in lot['lotData']['lotItems']:\n return generator_univ(lot['lotData']['lotItems']['lotItem'])\n else:\n return []\n else:\n return []\n else:\n return []\n\n def lot_max_price(self, lot):\n try:\n lot_max_price = lot['lotData']['initialSum']\n except Exception:\n lot_max_price = ''\n if lot_max_price is None:\n lot_max_price = ''\n return lot_max_price\n\n def okpd2_code(self, item):\n try:\n okpd2_code = item['okpd2']['code']\n except Exception:\n okpd2_code = ''\n if okpd2_code is None:\n okpd2_code = ''\n return okpd2_code\n\n\ndef parser_type_223(doc, path_xml, filexml, reg, reg_id):\n tender = Tender223(doc)\n id_t = tender.id_t\n lots = tender.get_lots()\n for lot in lots:\n lot_max_price = tender.lot_max_price(lot)\n print(lot_max_price)\n lotitems = tender.get_lotitems(lot)\n for lotitem in lotitems:\n okpd2_code = tender.okpd2_code(lotitem)\n print(okpd2_code)\n\n\n\ndef parser(doc, path_xml, filexml, reg, reg_id):\n global file_log\n try:\n parser_type_223(doc, path_xml, filexml, reg, reg_id)\n except Exception:\n logging.exception(\"Ошибка при парсинге тендера типа 223: \")\n with open(file_log, 'a') as flog:\n flog.write('Ошибка при парсинге тендера типа 223 ' + ' ' + path_xml + ' ' + '\\n\\n\\n')\n\nwith open(file) as f:\n doc = xmltodict.parse(f.read())\nparser(doc, file, file, 32, 32)","sub_path":"test223.py","file_name":"test223.py","file_ext":"py","file_size_in_byte":10463,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"631359078","text":"from flask import Flask, redirect, render_template, request, jsonify\n\napp = Flask(__name__)\n\napp.config['DEBUG'] = True\n\n@app.route(\"/\")\ndef hello():\n return \"Hello World\"\n\n@app.route(\"/greeting\")\ndef greeting():\n return render_template('index.html', name='Akilah')\n\n@app.route(\"/pie\")\ndef pie():\n ingredients=['apples', 'butter', 'sugar', 'flour']\n return jsonify({'pie ingredient': ingredients[0]})\n\ningredients=['apples', 'butter', 'sugar', 'flour', 'baking soda']\n@app.route(\"/recipe\", methods=['GET','POST'])\ndef recipe():\n if request.method == 'POST':\n new_ingredient = request.form[\"ingredient\"]\n ingredients.append(new_ingredient)\n return redirect ('/recipe')\n else:\n return render_template('pie.html', name=\"apple pie\", ingredients=ingredients)\n\nif __name__ == \"__main__\":\n app.run()","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":841,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"295183505","text":"#!/usr/bin/env python3\nimport pathlib\nimport sys\n\nimport setuptools\nfrom setuptools.command.install import install\nimport os\nfrom genos.version import VERSION\n\n\nhere = pathlib.Path(__file__).parent.resolve()\n\n# Get the long description from the README file\nlong_description = (here / 'README.md').read_text(encoding='utf-8')\n\nclass VerifyVersionCommand(install):\n \"\"\"Custom command to verify that the git tag matches our version\"\"\"\n description = 'verify that the git tag matches our version'\n\n def run(self):\n tag = os.getenv('CIRCLE_TAG')\n\n if tag != VERSION:\n info = \"Git tag: {0} does not match the version of this app: {1}. \" \\\n \"Update version in src/genos/version.py\".format(\n tag, VERSION\n )\n sys.exit(info)\n\n\nsetuptools.setup(\n name='genos',\n version=VERSION,\n description='Instantiate objects and call functions using dictionary configs in Python using Genos.',\n long_description=long_description,\n long_description_content_type='text/markdown',\n url='https://github.com/Neural-Space/genos',\n author='Kushal Jain',\n author_email='kushal@neuralspace.ai',\n keywords='instantiation, objects, recursive instantiation, functio call, config instantiate',\n license='MIT',\n classifiers=[\n \"Development Status :: 4 - Beta\",\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: MIT License\",\n \"Topic :: Software Development :: Build Tools\",\n \"Topic :: Software Development :: Libraries :: Python Modules\",\n \"Topic :: Internet\",\n \"Programming Language :: Python :: 3.7\",\n ],\n install_requires=[\n \"omegaconf~=2.0.4\"\n ],\n python_requires='>=3.7, <4',\n package_dir={'': 'src'},\n packages=setuptools.find_packages(where='src'),\n cmdclass={\n 'verify': VerifyVersionCommand,\n }\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1894,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"42468811","text":"#-*_coding:utf8-*-\r\nfrom bs4 import BeautifulSoup\r\nfrom selenium import webdriver\r\nfrom pandas import ExcelWriter\r\nimport pandas as pd\r\nimport requests\r\nimport time\r\nimport re\r\n\r\nptt='https://www.ptt.cc'\r\nurlget=[]\r\npo=[]\r\nhot_name=[]\r\n#取得熱門子版url\r\nhot_url='https://www.ptt.cc/bbs/hotboards.html'\r\nres=requests.get(hot_url)\r\n\r\nhot_soup=BeautifulSoup(res.text,'lxml')\r\n#熱門子版總url+版名\r\nfind_hot_url=hot_soup.find_all('div',class_='b-ent')\r\nfor show_hot_url in find_hot_url:\r\n\tlist_hot_url=show_hot_url.find('a')['href']\r\n\turlget.append(list_hot_url)\r\n\tlist_hot_name=show_hot_url.find('div',class_='board-name').text\r\n\thot_name.append(list_hot_name)\r\nprint('取得熱門子版url')\r\n#熱門子版PO文總數\r\nfind_hot_push=hot_soup.find_all('div','board-nuser')\r\nfor show_hot_push in find_hot_push:\r\n\tlist_hot_push=show_hot_push.find('span').text\r\n\tpo.append(list_hot_push)\r\nfor x in urlget:\r\n\turl2=(ptt+x)\t\r\n\r\n#寫入\r\nurl = 'https://www.ptt.cc/bbs/Gossiping/index38933.html' \r\n\r\n#response = requests.get(url)\r\n#滿18進入頁面\r\n#response = requests.get(url, cookies={'over18': '1'}) \r\n#content = response.content\r\n\r\n#翻頁/這和一般取得連結方式不同\r\n#next=[]\r\n#next_page=soup.find(\"div\",class_=\"btn-group btn-group-paging\")\r\n#next2=next_page.find_all(\"a\",href=True)\r\n#for p in next2:\r\n\t#u=p['href']\r\n\t#next.append(u) \r\n#next_url=\"next :\"+ptt+next[1]\r\n\r\n#collect info 空集合\r\ntheme=[]\r\ndate=[]\r\nauthor_id=[]\r\nrecommend=[]\r\nhttp=[]\r\nabout_author=[]\r\n#---\r\n\r\narticle_date=[]\r\narticle_author=[]\r\narticle=[]\r\narticle_ip=[]\r\narticle_msg=[]\r\nartilce_old=[]\r\narticle_old_author=[]\r\nmsg_date=[]\r\nmsg_id=[]\r\nmsg_rec=[]\r\n\r\n\r\n#單筆數據測試\r\n#data_all{'title':entry.find}\r\n#a=soup.find(\"div\",class_=\"r-ent\")\r\n#print(a.find(\"div\",class_=\"title\").text)#標題\r\n#print(ptt+a.find(\"a\")[\"href\"])#網址\r\n#print(a.find(\"div\",class_=\"author\").text)#作者\r\n#print(a.find(\"span\"))#推文數!=0才能+text\r\n#print(a.find(\"div\",class_=\"date\").text)\r\n#b=a.find(\"div\",class_=\"item\")\r\n#print(ptt+b.find(\"a\")[\"href\"])#該作者其他文章\r\n\r\n#定義////刪除\r\ndel_word=\"刪除\"\r\ntargetIP = u'※ 發信站: 批踢踢實業坊'\r\n#隱藏\r\noption = webdriver.ChromeOptions()\r\noption.add_argument('headless')\r\n#自動化網頁設定\r\ndriver=webdriver.Chrome( )\r\ndriver.get(url)\r\ndriver.find_element_by_name('yes').click()\r\nprint(\"進入網址: \"+url)\r\nsoup = BeautifulSoup(driver.page_source,'lxml')\r\nprint(\"瀏覽版名: \"+soup.title.text)#該版名\r\n\r\n\r\n\r\n#find data迴圈\r\ncount=1;\r\npage=0;\r\nwhile page<1: #指定筆數\r\n\t \r\n\tinfo_all =soup.find_all(\"div\", class_=\"r-ent\")\r\n\tprint(\"第\"+str(count)+\"筆\")\r\n\tcount=count+1\r\n\t \r\n\t#作者相關文章設定空值則不顯示(測試)\r\n\tinfo_about=soup.find_all(\"div\",class_=\"item\")\r\n\tfor infos3 in info_about:\r\n\t\tabout=ptt+infos3.find(\"a\")[\"href\"]\r\n\t\tif about is None:\r\n\t\t\tabout=\"can't find\"\r\n\t\t\tabout_author.append(about)\r\n\t\telse:\r\n\t\t\tabout_author.append(about)\r\n\t#推文數\r\n\tfor infos2 in info_all:\r\n\t\trec=infos2.find(\"span\")\r\n\t\tif rec is None:\r\n\t\t\trec=\"0\"\r\n\t\t\trecommend.append(rec)\r\n\t\telse:\r\n\t\t\trecommend.append(rec.text)\r\n\t\t\t\r\n\tfor infos in info_all:\r\n\t title1=infos.find(\"div\",class_=\"title\").text \r\n\t #theme.append(title1)\r\n\t author1=infos.find(\"div\", class_=\"author\").text\r\n\t author2=author1\r\n\t author_id.append(author2)\r\n\t date1=infos.find(\"div\",class_=\"date\").text\r\n\t date2=date1\r\n\t date.append(date2) \r\n\t#被刪除文章 文章和網址顯示del \r\n\t if (del_word) in title1:\r\n\t \tweb=\"[None]文章被刪除\"\r\n\t \thttp_none=\"[None]網址被移除\"\r\n\t \ttheme.append(web)\r\n\t \thttp.append(http_none)\r\n\t \r\n\t else:\r\n\t \ttheme.append(title1)\r\n\t \tweb2=infos.find(\"a\")[\"href\"]\r\n\t \thttp.append(ptt+web2)\r\n\t#取內文\r\n\t\r\n\r\n\t\t#print()\r\n\t\r\n\t#翻頁x path位置定義 易出錯\r\n\tdriver.find_element_by_xpath('//*[@id=\"action-bar-container\"]/div/div[2]/a[2]').click()\r\n\t#driver.find_element_by_link_text('上頁').click()\r\n\tpage=page+1\r\n\ttime.sleep(0.1)\r\n\t\r\n\r\nprint(\"結束,已讀取所有頁面,等待讀取相關內文\")\r\ndriver.close()\r\nfor http_web in http:\r\n\tif http_web==\"[None]網址被移除\":\r\n\t\tarticle_date.append(\"-\")\r\n\t\tarticle_author.append(\"-\")\r\n\t\tarticle.append(\"-\")\r\n\t\tarticle_ip.append(\"-\")\r\n\t\tarticle_msg.append(\"-\")\r\n\t\tartilce_old.append(\"-\")\r\n\t\tarticle_old_author.append(\"-\")\r\n\t\tmsg_date.append(\"-\")\r\n\t\tmsg_id.append(\"-\")\r\n\t\tmsg_rec.append(\"-\")\r\n\telse:\r\n\t\thttp_get=requests.get(http_web,cookies={'over18': '1'})\r\n\t\tsoup_http=BeautifulSoup(http_get.text,'lxml')\r\n\t\tfind_push_content=soup_http.find_all('div',class_='push')\r\n\t\tptt_content=soup_http.find_all('meta', property=\"og:description\")\r\n\t\tip = soup_http.find(string=re.compile(targetIP))\r\n\t\tip = re.search(r\"[0-9]*\\.[0-9]*\\.[0-9]*\\.[0-9]*\", ip).group()\r\n\t\tarticle_ip.append(ip)\r\n\tfor p_content in find_push_content:\r\n\t\tpush_content=p_content.find('span',class_='f3 push-content').text\r\n\t\tarticle_msg.append(push_content)\r\n\t\t\r\n\tfor ptt_content2 in ptt_content:\r\n\t\tptt_content3=ptt_content2.get('content')\r\n\t\tarticle.append(ptt_content3)\r\n\t\r\n\t\t\r\n \r\n\r\nprint(\"等待輸出csv\")\t\t\r\n\r\n\r\n#輸出\r\n\r\n#for a in theme:\r\n\t#print(a)\r\n#for b in http:\r\n\t#print(b)\r\n#for c in author_id:\r\n\t#print(c)\r\n#for d in date:\r\n\t#print(d)\r\n#for e in about_author:\r\n\t#print(e)\r\n#for f in recommend:\r\n\t#print(f)\t \r\n#---\r\n#for g in article:\r\n\t#print(g)\r\n#for h in article_msg:\r\n\t#print(h)\r\n#for i in article_ip:\r\n\t#print(i)\t \r\n#dict={\"推薦\":recommend,\"主題\":theme,\"作者\":author_id,\t\"日期\":date,\t\"網址\":http,\"發文ip\":article_ip}\t\r\n#print(len(recommend), len(theme), len(author_id),len(date),len(http),len(about_author),len(article))\r\n#df=pd.DataFrame(dict)\r\n#writer=pd.ExcelWriter('pttdata.xlsx')\r\n#df.to_excel(writer,index=False,encoding='utf8')\r\n#writer.save()\r\n#print(\"已寫入csv檔\")\r\n\r\n\r\n","sub_path":"python爬蟲/ptt_data.py","file_name":"ptt_data.py","file_ext":"py","file_size_in_byte":5771,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"303001866","text":"import string\n#read in and set up \nread('results/*contree')\ninTrees = var.trees\n\nvar.trees = []\nMRP=func.readAndPop('supertrees/mrpStrictConsTree.nex')\nMajR=func.readAndPop('supertrees/mrpMajRuleConsTree.nex')\nSR2008=func.readAndPop('SR2008_cons.nex')\nSR2008.name='SR2008'\n#SPA=func.readAndPop('SPA_f.nex')\n#SPA.name='SPA'\n#SPA_support=func.readAndPop('SPA_supp_f.nex')\n#SPA_support.name='SPA_support'\n#MRP.name='MRP'\nMajR.name='MajR'\na=func.readAndPop('supertrees/mrpStrictConsTree.nex')\n#SPA.taxNames=a.taxNames\nMRP.taxNames=a.taxNames\n#inTrees.taxNames = list(string.uppercase[:10])\n#MajR.taxNames=a.taxNames\n\n\n#trees=[total_evidence, MRP,MajR, SR2008,SPA, SPA_support]\ntrees=[MRP, MajR]\ntaxNames=list(string.uppercase[:10])\n\n\n\n#tt=Trees(trees=trees, taxNames=MRP.taxNames)\ntt=Trees(trees=trees, taxNames=taxNames)\ntt.inputTreesToSuperTreeDistances(inTrees)\n\n#dm = tt.topologyDistanceMatrix('scqdist')\n#dm.writeNexus()\n#dm.writeNexus('qd_distances.nex')\n#paup execute rf_dist nj save\n\n\n\n\n\n","sub_path":"SIMULATED_DATASET/100_B_inputdistances.py","file_name":"100_B_inputdistances.py","file_ext":"py","file_size_in_byte":992,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"316503227","text":"class ListNode(object):\n\n def __init__(self, val):\n self.val = val\n self.next = None\n\n\nclass LinkedList(object):\n\n def __init__(self, first, *k):\n self.head = ListNode(first)\n temp = self.head\n for i in k:\n temp.next = ListNode(i)\n temp = temp.next\n\n def reverse(self):\n prev = None\n while self.head:\n temp = self.head.next\n self.head.next = prev\n prev = self.head\n self.head = temp\n self.head = prev\n return repr(self)\n\n def __repr__(self):\n s = '{}'.format(self.head.val)\n temp = self.head.next\n while temp:\n s += ','\n s += str(temp.val)\n temp = temp.next\n return \"LinkedList({})\".format(s)\n\n\"\"\"\n由于单向链表每个节点只有一个后继指针,因此有些操作十分复杂:\n例如反向遍历链表,获得链表中一个元素前面的元素。这些操作单链表都需要多次遍历,十分的浪费资源。\n因此我们引入了双向链表这一概念。\n双向链表较之单向链表最大的区别就是一个节点不仅仅拥有后继指针指向它的后一个节点,\n而且还拥有一个前驱指针指向它前一个节点。因此,对于双向链表来说。\n反向遍历链表和访问链表中某个元素前一个元素都可以通过它的前驱指针较为简单的完成。\n\n\"\"\"\nclass DoubleLinked():\n\n def __init__(self, val):\n self.val = val\n self.pre = self.next = None\n\n\nclass DoubleList():\n\n def __init__(self, first, *k):\n self.head = DoubleLinked(first)\n temp = self.head\n for i in k:\n item = DoubleLinked(i)\n item.pre = temp\n temp.next = item\n temp = temp.next\n self.tail = temp\n\n def reverse(self):\n pre = None\n while self.head:\n # 1.遍历每一个��点 2.pre和next互唤。\n temp = self.head\n self.head = self.head.next\n temp.next = temp.pre\n temp.pre = self.head\n self.head = temp\n\n def __repr__(self):\n s = 'head:{}'.format(self.head.val)\n temp = self.head.next\n while temp:\n s += ','\n s += str(temp.val)\n temp = temp.next\n return s","sub_path":"BasicDataStructure/LinkedList/pyLinked.py","file_name":"pyLinked.py","file_ext":"py","file_size_in_byte":2329,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"228419989","text":"\nfrom flask_sqlalchemy import SQLAlchemy\nfrom app import db\nfrom app.auditorium.models import Auditorium\nfrom app.movie.models import Movie\nfrom datetime import datetime\nfrom datetime import date\nfrom datetime import time\nfrom app.movie.models import Movie\n\n\nclass Screening (db.Model):\n __tablename__ = \"screening\"\n id = db.Column('id', db.Integer, primary_key = True)\n movie_id = db.Column('movie_id', db.Integer, db.ForeignKey('movie.id'))\n auditorium_id = db.Column('auditorium_id', db.Integer, db.ForeignKey('auditorium.id'))\n screening_start_time = db.Column('screening_start_time', db.Time)\n screening_date = db.Column('screening_date',db.Date)\n\n db.relationship('Auditorium', foreign_keys='auditorium_id')\n db.relationship('Movie', foreign_keys='movie_id')\n\n def __init__(self,movie_id,auditorium_id,screening_start_time,screening_date):\n self.movie_id = movie_id\n self.auditorium_id = auditorium_id\n self.screening_start_time = screening_start_time\n self.screening_date = screening_date\n\n def to_dict(self):\n return {\n 'id': self.id,\n 'movie_id': self.movie_id,\n 'auditorium_id': self.auditorium_id,\n 'screening_start_time': str(self.screening_start_time),\n 'screening_date': str(self.screening_date)\n }\n\n def to_dict_dates(self):\n return {\n 'date' : str(self.screening_date)\n }\n\n def to_dict_shows(self):\n return {\n 'screening_start_time' : str(self.screening_start_time)\n }\n def __repr__(self):\n return \"'Screening' { 'movie_id': %r , 'auditorium_id': %r, 'screening_start_time':%r , 'screening_date': %r}\"%(self.movie_id,self.auditorium_id,str(self.screening_start_time),str(self.screening_date))\n","sub_path":"src/app/screening/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":1802,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"641388496","text":"import numpy as np\nimport tensorflow as tf\nimport time\nfrom matplotlib.pyplot import imshow\nimport matplotlib.pyplot as plt\nimport tensorflow.contrib.keras as keras\nfrom sklearn.utils import shuffle\nfrom sklearn.metrics import mean_squared_error\nfrom tf_models import KERAS\nimport preprocessing\nimport scoring\nimport output\nfrom sklearn.model_selection import train_test_split\nimport math\n\n############################################ Configuration ###################################################\n# Decide whether self-evaluation or final submission\nfinal_submission = True\nTrain_split = 8./10\n\n## General\nBLoadData = 1\nBFinalPrediction = 1\n\n# Hyperparameters\n# epochs = 100\n# param = 30\n# layers = 2\n# batch_size = 32\n\nepochs = 90\nparam = 42\nlayers = 3\nbatch_size = 32\n\n# Gridsearch\nBGridSearch = 1\nepoch_list = [75, 90, 100, 120]\nparam_list = [24, 30, 36, 42]\nlayer_list = [1, 2, 3, 4]\nbatch_size_list = [16, 32, 64, 128]\n\n## Postprocessing\n## Score\nBRMSEScore = 0\nBAccuracy = 1\n##############################################################################################################\n\n## Load Data\nif BLoadData == 1:\n dataloader = preprocessing.loadfiles('C:\\\\Users\\\\fabri\\\\git\\\\Intro-to-ML\\\\Task2\\\\Raw_Data') # Define datafolder - HomePC\n # dataloader = preprocessing.loadfiles('C:\\\\Users\\\\fabri\\\\git\\\\Intro-to-ML\\\\Task0\\\\Raw_Data') # Surface\n X_train = dataloader.loadX_train()\n y_train = dataloader.loady_train()\n X_test = dataloader.loadX_test()\n print('X_train shape: ', X_train.shape)\n print('y_train shape: ', y_train.shape)\n \n\n## Train / Test Split \nif BFinalPrediction == 0:\n X_train, X_test, y_train, y_test = train_test_split(X_train, y_train, test_size=0.33, random_state=32) \n\nif BGridSearch == 0 or BFinalPrediction == 1: \n ## Create one hot format\n y_train_onehot = keras.utils.to_categorical(y_train)\n print('First 3 labels: ', y_train[:3])\n print('First 3 onehot labels:\\n', y_train_onehot[:3])\n\n # build model:\n model = KERAS.build(X_train, y_train_onehot, param, layers)\n # train model:\n trained_model, losses = KERAS.fit(model, X_train, y_train_onehot, epochs, batch_size)\n # predict labels:\n y_pred = KERAS.predict(trained_model, X_test)\n\n ## Score\n if BAccuracy == 1 and BFinalPrediction == 0:\n scorer = scoring.score()\n score = scorer.Accuracy(y_test, y_pred)\n print('Accuracy score is = ', repr(score))\n\nif BGridSearch == 1 and BFinalPrediction == 0: \n iters = len(epoch_list)*len(param_list)*len(layer_list)*len(batch_size_list)\n i = 0\n score_best = 0\n while i < iters:\n epochs = epoch_list[i%len(epoch_list)]\n param = param_list[math.floor((i/len(param_list))%len(epoch_list))]\n layers = layer_list[math.floor((i/(len(param_list)*len(layer_list)))%len(epoch_list))]\n batch_size = batch_size_list[math.floor((i/(len(param_list)*len(layer_list)*len(batch_size_list)))%len(epoch_list))]\n print('epoch = ', repr(epochs), '| param = ', repr(param), '| layers = ', repr(layers), '| batch_size = ', repr(batch_size))\n ## Create one hot format\n y_train_onehot = keras.utils.to_categorical(y_train)\n print('First 3 labels: ', y_train[:3])\n print('First 3 onehot labels:\\n', y_train_onehot[:3])\n\n # build model:\n model = KERAS.build(X_train, y_train_onehot, param, layers)\n # train model:\n trained_model, losses = KERAS.fit(model, X_train, y_train_onehot, epochs, batch_size)\n # predict labels:\n y_pred = KERAS.predict(trained_model, X_test)\n\n ## Score\n if BAccuracy == 1 and BFinalPrediction == 0:\n scorer = scoring.score()\n score = scorer.Accuracy(y_test, y_pred)\n print('Accuracy score is = ', repr(score))\n \n if score > score_best:\n epoch_best = epochs\n param_best = param\n layers_best = layers\n batch_size_best = batch_size\n score_best = score\n \n i += 1\n print('epoch = ', repr(epoch_best), '| param = ', repr(param_best), '| layers = ', repr(layers_best), '| batch_size = ', repr(batch_size_best), '| score = ', repr(score_best))\n\n## Output Generation\nif BFinalPrediction == 1:\n datasaver = output.savetask2('C:\\\\Users\\\\fabri\\\\git\\\\Output', 'C:\\\\Users\\\\fabri\\\\git\\\\Intro-to-ML\\\\Task2\\\\Raw_Data') # Savepath, Datapath\n datasaver.saveprediction(y_pred)","sub_path":"Task2/CodeFabrice/gridsearch.py","file_name":"gridsearch.py","file_ext":"py","file_size_in_byte":4424,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"185502568","text":"#! /usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport os\nimport subprocess\nimport time\n\n# pdftk Whetten1989.pdf output uncompressed.pdf uncompress\n# LANG=C sed -n '/^\\/Annots/!p' uncompressed.pdf > stripped.pdf\n# pdftk stripped.pdf output final.pdf compress\n\nos.chdir('remove_comments')\n\npdfs = [line for line in subprocess.check_output(\n \"find -name '*.pdf'\", shell=True).splitlines()]\n\nif os.path.isfile('uncompressed.pdf') or os.path.isfile('stripped.pdf'):\n input('remove uncompressed.pdf or stripped.pdf')\n\nfor pdf in pdfs:\n pdf = pdf.decode('UTF-8').replace('./', '')\n os.rename(pdf, '_' + pdf)\n command = \"pdftk \" + '_' + pdf + \" output uncompressed.pdf uncompress && LANG=C \" + \\\n \"sed -n '/^\\/Annots/!p' uncompressed.pdf > stripped.pdf && pdftk stripped.pdf output \" \\\n + pdf + \" compress\"\n # print(command)\n subprocess.run(command, shell=True, check=True)\n time.sleep(2)\n os.remove('uncompressed.pdf')\n os.remove('stripped.pdf')\n os.remove('_' + pdf)\n","sub_path":"remove-pdf-comments.py","file_name":"remove-pdf-comments.py","file_ext":"py","file_size_in_byte":1009,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"502912634","text":"import pandas as pd\nimport numpy as np\nfrom sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.model_selection import train_test_split\n\ndf = pd.read_csv(r\"C:\\Users\\User\\PycharmProjects\\ML_Project\\data\\Pre-processed_Sarcasm_Headlines_Dataset.csv\")\n\nheadlines = df['headline']\ny = df['is_sarcastic']\n\nheadlines_train, headlines_test, y_train, y_test = train_test_split(headlines, y, test_size=0.10, random_state=1000)\n\n# Create dictionary of words from training set\nvectorizer = CountVectorizer()\nvectorizer.fit(headlines_train)\n\nX_train = vectorizer.transform(headlines_train)\nX_test = vectorizer.transform(headlines_test)\n\n# Note: class1 is not sarcastic and class2 is sarcastic\n# y_train_array = y_train\ny_train_class1_inds = np.where(y_train==0) # Indices of all non sarcastic elements\ny_train_class2_inds = np.where(y_train==1) # Indices of all sarcastic elements\nX_train_class1 = X_train[y_train_class1_inds]\nX_train_class2 = X_train[y_train_class2_inds]\n\n# X_train_np = X_train.toarray()\n# get shape of matrix - no of rows = no of elements\nclass1_shapeR, class1_shapeC = X_train_class1.shape\nclass2_shapeR, class2_shapeC = X_train_class2.shape\n\nprob_class1 = class1_shapeR/(class1_shapeR+class2_shapeR)\nprob_class2 = class2_shapeR/(class1_shapeR+class2_shapeR)\n\n# Less efficient but easier to code... Might change later\nX_train_class1_arr = X_train_class1.toarray()\nX_train_class2_arr = X_train_class2.toarray()\n\n\n","sub_path":"Naive_Bayes/Naive_Bayes.py","file_name":"Naive_Bayes.py","file_ext":"py","file_size_in_byte":1436,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"139400829","text":"import pytest\nimport ujson\nfrom company_management.models import Company\nfrom . models import MLData\nfrom rest_framework.test import APIClient\nfrom contextlib import contextmanager\nimport os\nimport csv\n\n\n@pytest.fixture\ndef api():\n return APIClient()\n\n\n@contextmanager\ndef generate_csv():\n with open('temp.csv', 'w', newline='') as csvfile:\n spamwriter = csv.writer(csvfile, dialect='excel')\n spamwriter.writerow(['cust_ID', 'owner_first_name'])\n spamwriter.writerow(['1', 'qw'])\n spamwriter.writerow(['2', 'qw1'])\n spamwriter.writerow(['3', 'qw2'])\n try:\n yield\n finally:\n os.remove('temp.csv')\n\n\n@contextmanager\ndef generate_csv_2():\n with open('temp1.csv', 'w', newline='') as csvfile:\n spamwriter = csv.writer(csvfile, dialect='excel')\n spamwriter.writerow(['cust_ID', 'owner_first_name'])\n spamwriter.writerow(['1', 'qe'])\n spamwriter.writerow(['2', 'qe1'])\n spamwriter.writerow(['3', 'qe2'])\n try:\n yield\n finally:\n os.remove('temp1.csv')\n\n\ndef create_ml_data(api):\n com_instance = Company.objects.create(name=\"test1\")\n with generate_csv():\n with open('temp.csv') as fp:\n response = api.post('/api/ml_data/data_set/%d/' % 0, {\"company\": com_instance.id,\n \"file\": fp, \"name\": \"test1\"})\n return response\n\n\n@pytest.mark.django_db\ndef test_all_data_sets(api):\n com_instance_1 = Company.objects.create(name=\"test1\")\n MLData.objects.create(name=\"test1\", company=com_instance_1)\n MLData.objects.create(name=\"test2\", company=com_instance_1)\n com_instance_2 = Company.objects.create(name=\"test2\")\n MLData.objects.create(name=\"test1\", company=com_instance_2)\n MLData.objects.create(name=\"test2\", company=com_instance_2)\n request = api.get('/api/ml_data/all_data_sets/', )\n\n assert 200 == request.status_code\n\n body = ujson.loads(request.content)\n assert 2 == len(body)\n assert body[0] == {'label': 'test1', 'data': com_instance_1.id, 'children': [{'label': 'test1', 'data': 1},\n {'label': 'test2', 'data': 2}]\n }\n assert body[1] == {'label': 'test2', 'data': com_instance_2.id, 'children': [{'label': 'test1', 'data': 3},\n {'label': 'test2', 'data': 4}]\n }\n\n\n@pytest.mark.django_db\ndef test_data_set_get(api):\n com_instance = Company.objects.create(name=\"test1\")\n data_instance = MLData.objects.create(name=\"test1\", company=com_instance)\n request = api.get('/api/ml_data/data_set/%d/' % data_instance.id)\n assert 200 == request.status_code\n assert ujson.loads(request.content) == {'id': 5, 'name': 'test1', 'company': {'id': com_instance.id, 'name': 'test1'},\n 'key_data': {'name': ''}, 'data': None}\n\n\n@pytest.mark.django_db\ndef test_data_set_post(api):\n request = create_ml_data(api)\n assert 200 == request.status_code\n data = request.json()\n ml_instance = MLData.objects.get(name=\"test1\")\n assert ml_instance.id == data[\"id\"]\n\n\n@pytest.mark.django_db\ndef test_data_set_put(api):\n com_instance = Company.objects.create(name=\"test1\")\n data_instance = MLData.objects.create(name=\"test1\", company=com_instance)\n with generate_csv():\n with open('temp.csv') as fp:\n request = api.put('/api/ml_data/data_set/%d/' % data_instance.id, {\"company\": com_instance.id,\n \"file\": fp, \"name\": \"test1\"})\n assert 200 == request.status_code\n data = request.json()\n assert data_instance.id == data[\"id\"]\n\n\n@pytest.mark.django_db\ndef test_data_set_delete(api):\n com_instance = Company.objects.create(name=\"test1\")\n data_instance = MLData.objects.create(name=\"test1\", company=com_instance)\n request = api.delete('/api/ml_data/data_set/%d/' % data_instance.id)\n assert 200 == request.status_code\n assert 0 == len(MLData.objects.filter(name=\"test1\"))\n\n\n@pytest.mark.django_db\ndef test_data_set_row(api):\n create_ml_data(api)\n ml_instance = MLData.objects.get(name=\"test1\")\n request = api.put('/api/ml_data/data_set_row/%d/' % ml_instance.id, {\"row_index\": 1,\n \"col_name\": \"owner_first_name\",\n \"value\": 2})\n assert 200 == request.status_code\n ml_instance = MLData.objects.get(name=\"test1\")\n ml_data = ml_instance.data\n for i in ml_data:\n if i[\"row_index\"] == 1:\n assert i[\"owner_first_name\"] == '2'\n if i[\"row_index\"] == 2:\n assert i[\"owner_first_name\"] != '2'\n\n\n@pytest.mark.django_db\ndef test_map_new_data_set(api):\n create_ml_data(api)\n mapped_columns = ujson.dumps([{\"origin_column\": \"owner_first_name\", \"new_set_column\": \"owner_first_name\"}])\n primary_mapping = ujson.dumps({\"origin_column\": \"cust_ID\", \"new_set_column\": \"cust_ID\"})\n origin_column = ujson.dumps([\"cust_ID\", \"owner_first_name\"])\n new_set_column = ujson.dumps([\"cust_ID\", \"owner_first_name\"])\n context = {\n \"mapped_columns\": mapped_columns,\n \"primary_columns\": primary_mapping,\n \"origin_column\": origin_column,\n \"new_set_column\": new_set_column\n }\n ml_instance = MLData.objects.get(name=\"test1\")\n with generate_csv_2():\n with open('temp1.csv') as fp:\n context[\"file\"] = fp\n request = api.post('/api/ml_data/map_new_data_set/%d/' % ml_instance.id, context)\n\n assert 200 == request.status_code\n\n ml_instance = MLData.objects.get(name=\"test1\")\n ml_data = ml_instance.data\n for i in ml_data:\n if i[\"row_index\"] == 1:\n assert i[\"owner_first_name\"] == 'qe'\n if i[\"row_index\"] == 2:\n assert i[\"owner_first_name\"] == 'qe1'\n if i[\"row_index\"] == 3:\n assert i[\"owner_first_name\"] == 'qe2'\n","sub_path":"ml_data_care/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":6094,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"375333842","text":"import pprint\nimport time\n\nfrom luminateapi.luminate_python import LuminateV2Client\n\nLUMINATE_URL = 'https://api..luminatesec.com'\nAPI_KEY = ''\nAPI_SECRET = ''\nVERIFY_CERTIFICATE = True\n\nTEST_USER_EMAIL = \"\"\n\nif __name__ == '__main__':\n # Create a V2 Client\n luminate_client = LuminateV2Client(LUMINATE_URL,\n API_KEY,\n API_SECRET,\n VERIFY_CERTIFICATE)\n\n # Block all users with this email\n block_res = luminate_client.block_user_by_email(TEST_USER_EMAIL)\n pprint.pprint(block_res)\n\n # Unblock the same user\n unblock_res = luminate_client.unblock_user_by_email(TEST_USER_EMAIL)\n pprint.pprint(unblock_res)\n\n # Creating SSH Application\n ssh_app_res = luminate_client.create_app(\"client-ssh-test\",\n \"description\",\n \"SSH\",\n \"tcp://127.0.0.1:8000\",\n [\"some_linux_user\"])\n pprint.pprint(ssh_app_res)\n\n # Creating HTTP Application\n http_app_res = luminate_client.create_app(\"client-http-app\",\n \"description\",\n \"HTTP\",\n \"http://127.0.0.1:8080\",\n None)\n pprint.pprint(http_app_res)\n\n # Creating a Site\n site_res = luminate_client.create_site(\"site-test-client\", \"description\")\n pprint.pprint(site_res)\n\n # Binding the Application to Site\n luminate_client.bind_app_to_site(http_app_res['id'], site_res['id'])\n\n # Get user information\n user_info_res = luminate_client.get_user(TEST_USER_EMAIL)\n pprint.pprint(user_info_res)\n\n # Finding local IDP\n local_idp = None\n for idp in user_info_res:\n if idp['name'] == 'local':\n local_idp = idp\n assert (local_idp is not None)\n\n local_user = local_idp['users'][0]\n\n # Assigning the user to the created application\n assignment_res = luminate_client.assign_entity_to_app(http_app_res['id'],\n local_user['id'],\n local_user['identity_provider_id'],\n local_user['repository_type'],\n \"User\")\n print(assignment_res)\n\n # Destroy User Session\n destroy_res = luminate_client.destroy_user_sessions_by_email(TEST_USER_EMAIL)\n pprint.pprint(destroy_res)\n\n # Getting HTTP and SSH logs\n query = {\n \"free_text\": \"\",\n \"from_date\": int((time.time() - 1000) * 1000),\n \"to_date\": int(time.time() * 1000),\n }\n http_logs_res = luminate_client.get_access_logs(2, query, None)\n pprint.pprint(http_logs_res)\n\n ssh_logs_res = luminate_client.get_ssh_access_logs(2, query, None)\n pprint.pprint(ssh_logs_res)\n","sub_path":"luminateapi/examples/example.py","file_name":"example.py","file_ext":"py","file_size_in_byte":3119,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"500148387","text":"import gym\nimport numpy as np\nimport time\nfrom IPython.display import clear_output\n\nenv = gym.make(\"FrozenLake-v0\")\nactionSpace_size = env.action_space.n\nstateSpace_size = env.observation_space.n\n\nq_table = np.zeros((stateSpace_size, actionSpace_size))\n\nalpha = 0.1 # learning rate\nepsilon = 1 # e-greedy parameter\nepsilon_min = 0.0\nepsilon_max = 1\nepsilon_decay =0.001\n\ngamma = 0.99 # discount rate\n\nnum_eps = 10000\nmax_steps = 100\n\nreward_list = []\n\nfor episode in range(num_eps):\n current_state = env.reset()\n done = False\n episode_reward = 0\n for j in range(max_steps):\n if np.random.rand() > epsilon:\n action = np.argmax(q_table[current_state, :])\n else:\n action = env.action_space.sample() #random sample\n\n new_state, reward, done, info = env.step(action) #single time step\n #print(reward)\n new_action = np.argmax(q_table[new_state, :])\n q_star = reward + gamma*np.max(q_table[new_state, :]) #bellman equation\n\n q_table[current_state, action] = (1-alpha)*q_table[current_state,action] +(alpha)*q_star\n\n current_state = new_state\n episode_reward += reward\n\n if done:\n break\n\n\n epsilon = epsilon_min + (epsilon_max-epsilon_min)*np.exp(-epsilon_decay*episode)\n reward_list.append(episode_reward)\n\nreward_perThousand = np.split(np.array(reward_list), num_eps/1000)\nprint(\"Average reward per thousand episodes \\n \", np.mean(reward_perThousand, 1))\n\nfor episode in range(3):\n current_state = env.reset()\n done = False\n print(\"Episode number: \", episode+1, \"\\n\\n\")\n time.sleep(1)\n for step in range(max_steps):\n clear_output(wait= True)\n env.render()\n\n action = np.argmax(q_table[current_state, :])\n new_state, reward, done, info = env.step(action)\n time.sleep(0.3)\n if done:\n clear_output(wait=True)\n env.render()\n if reward:\n print(\"***** You found the frisbee *****\")\n time.sleep(3)\n elif not reward:\n print(\"***** You fell into a hole *****\")\n time.sleep(3)\n clear_output(wait= True)\n break\n current_state = new_state\n\n'''\n#create environment\nactions =\nstates =\n\nr= np.random.rand()\nepsilon= 1\nalpha= 01\ngamma=0.9\nq_table= zeros(statesize, actionsize)\nr_table=\n\n# start game\nfor num iters:\n reset environment\n for each episode:\n s= current state\n if r>epsilon:\n #exploitation- choose action with maximum qval\n else:\n #explore - take random action\n a= env.actionspace.sample\n q_star= \n \n loss= q_star-q[s,a]\n \n qtable[s,a]= (1-alpha)*qtable[s,a]+(alpha)*q_star\n \n epsilon *=0.9 #decay\n\n'''","sub_path":"Q-learning/first_game.py","file_name":"first_game.py","file_ext":"py","file_size_in_byte":2807,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"307143572","text":"import sys\n#Ex:1\n\n# for i in range(5):\n# for j in range(i):\n# print (\"* \", end=\"\")\n# print(\"\")\n\n# for i in range(5,0,-1):\n# for j in range(i):\n# print(\"* \", end=\"\")\n# print('')\n\n\nnum__ = input(\"Enter a number:\" )\nnum__ = num__ + 1 if num__ % 2 == 0 else num__\nmid_row = num__ // 2 + 1 \nfor i in range(1, num__ + 1):\n\tif i < mid_row:\n\t\tprint(i + \"* \")\n\telse:\n\t\tj = mid_row * 2 - i\n\t\tprint(j + \"* \")\n\n#Ex:2\nfor i in range(1,11):\n\tif i == 1:\n\t\tprevious = 1\n\telse:\n\t\tprevious = i - 1\n\tprint(f\"{previous} + {i} = {previous + i}\")\n\n#Ex:3\n\nnum = input(\"Enter a number:\\n\")\n\nif not (num and num.isdigit()):\n\tsys.exit()\nelse:\n\tnum = int(num)\n\nfor i in range(1, num // 2 + 1):\n\tif num % i == 0:\n\t\tprint(\"The divisors of \",num, \"is\",i)\n\n\n\n#Ex:4\n\nfactorial_num = input(\"Tell me number:\\n\")\nfactorial = 1\n\nif not (factorial_num and factorial_num.isdigit()):\n\tsys.exit()\nelse:\n\tfactorial_num = int(factorial_num)\n\nif factorial_num < 0:\n print(\"Factorial does not exist for negative number\")\nelif factorial_num == 0:\n print(\"The factorial of 0 is 1\")\nelse:\n for i in range(1,factorial_num + 1):\n factorial = factorial*i\n print(\"The factorial of\",factorial_num,\"is\",factorial)\n\n\n\n#Ex:5\n\nnum_1, num_2 = 0, 1\n\nwhile num_2 <= 50:\n\tprint(num_2, \" \", end=\"\")\n\tnum_1, num_2 = num_2, num_1 + num_2\n\n","sub_path":"homework_4.py","file_name":"homework_4.py","file_ext":"py","file_size_in_byte":1323,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"53384093","text":"from pyplasm import *\nfrom scipy import *\n\ngrigio_granito = [0.803,0.752,0.690];\n\n#definisco una colonna\ndef circR2(p): \n\treturn [2*COS(p[0]),0]\n\ndef circR1(p): \n\treturn [COS(p[0]),0]\ndef domaincirc(n): \n\treturn INTERVALS(2*PI)(n)\n\ncircBase=MAP(circR2)(domaincirc(25))\ncircBase=JOIN([circBase])\ncircBase=T([1,2,3])([12,12,6])(circBase)\n\ncircSopra=MAP(circR1)(domaincirc(25))\ncircSopra=JOIN([circSopra])\ncircSopra=T([1,2,3])([12,12,20])(circSopra)\ncolonna=JOIN([circBase,circSopra])\n\ncolonna=COLOR(grigio_granito)(STRUCT([colonna]))\n#VIEW(STRUCT([colonna,scalinata]))\n\n#replico le colonne e creo i colonnati\nxmov=[T(1)(5),colonna]\nymov=[T([1,2])([0,4.64]),colonna]\ncolonnato1=STRUCT(NN(7)(xmov))\ncolonnato2=STRUCT(NN(17)(ymov))\nx2mov=[T([1,2])([0,4.64]),T([1,2])([35,0])(colonna)]\ncolonnato3=STRUCT(NN(17)(x2mov))\ny2mov=[T([1])([5]),T([1,2])([0,79])(colonna)]\ncolonnato4=STRUCT(NN(7)(y2mov))\ncolonnati=STRUCT([colonna,colonnato1,colonnato2,colonnato3,colonnato4])\n\n#seconda fila di colonne\ndef circ(p): \n\treturn [0.5*COS(p[0]),0]\n\ncircBasedentro=MAP(circR1)(domaincirc(25))\ncircBasedentro=JOIN([circBasedentro])\ncircBasedentro=T([1,2,3])([12,12,6])(circBasedentro)\n\ncircSopradentro=MAP(circ)(domaincirc(25))\ncircSopradentro=JOIN([circSopradentro])\ncircSopradentro=T([1,2,3])([12,12,20])(circSopradentro)\n\ncolonnadentro=COLOR(grigio_granito)(T([1,2,3])([5,5])(JOIN([circBasedentro,circSopradentro])))\n\nxcapitello=QUOTE([2.2])\nycapitello=QUOTE([2.5])\ncapitello=COLOR(grigio_granito)(T([1,2,3])([15.9,15.3,20.2])(INSR(PROD)([xcapitello,ycapitello,QUOTE([0.5])])))\n\ncolonnadentro=STRUCT([colonnadentro,capitello])\nmovint=[T(1)(5),colonnadentro]\ncolonnatointerno=COLOR(grigio_granito)(STRUCT(NN(5)(movint)))\n\n#colonne nella stanza piu piccola\ncircBasedentro2=MAP(circR1)(domaincirc(25))\ncircBasedentro2=JOIN([circBasedentro2])\ncircBasedentro2=T([1,2,3])([12,12,6])(circBasedentro2)\ncircSopradentro2=MAP(circ)(domaincirc(25))\ncircSopradentro2=JOIN([circSopradentro2])\ncircSopradentro2=T([1,2,3])([12,12,20])(circSopradentro2)\ncolonnadentro2=JOIN([circBasedentro2,circSopradentro2])\ncolonnastanza1=COLOR(grigio_granito)(T([1,2,3])([12,25])(colonnadentro2))\ncolonnastanza1_2=COLOR(grigio_granito)(T([1,2,3])([22,25])(colonnadentro2))\ncolonnastanza1_3=COLOR(grigio_granito)(T([1,2,3])([12,30])(colonnadentro2))\ncolonnastanza1_4=COLOR(grigio_granito)(T([1,2,3])([22,30])(colonnadentro2))\ncolonnatostanza1=STRUCT([colonnastanza1,colonnastanza1_2,colonnastanza1_3,colonnastanza1_4])\n\n#colonne nella stanza piu grande, le colonne sono piu piccole\ndef circpicc(p): \n\treturn [0.25*COS(p[0]),0]\ncircsoprastanza2=MAP(circpicc)(domaincirc(25)) #raggio 0.25\ncircsoprastanza2=JOIN([circsoprastanza2])\ncircsoprastanza2=T([1,2,3])([20,53,20])(circsoprastanza2)\ncircbasestanza2=MAP(circ)(domaincirc(25)) #raggio 0.5\ncircbasestanza2=JOIN([circbasestanza2])\ncircbasestanza2=T([1,2,3])([20,53,9])(circbasestanza2)\ncolonnastanza2=JOIN([circbasestanza2,circsoprastanza2])\n\nxmov=[T([1,2])([0,3]),colonnastanza2]\ncolonnatostanza2=COLOR(grigio_granito)(STRUCT(NN(10)(xmov)))\n\nxmov2=[T([1,2])([3,0]),colonnastanza2]\ncolonnatostanza2_2=COLOR(grigio_granito)(STRUCT(NN(5)(xmov2)))\n\nxmov3=[T([1,2])([0,3]),T(1)(18)(colonnastanza2)]\ncolonnatostanza2_3=COLOR(grigio_granito)(STRUCT(NN(10)(xmov3)))\n\ncolonnatistanza2=STRUCT([colonnatostanza2,colonnatostanza2_2,colonnatostanza2_3])\n\n\nr=STRUCT([colonnatostanza1,colonnatistanza2])\nVIEW(r)\n","sub_path":"2014-03-21/python/Exercise2.py","file_name":"Exercise2.py","file_ext":"py","file_size_in_byte":3399,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"129506704","text":"import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.autograd as autograd\nimport torch.optim as optim\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.gridspec as gridspec\nimport os\nimport argparse\nfrom torchvision.datasets import MNIST\nfrom torch.autograd import Variable\nfrom torch.utils.data import Dataset, DataLoader\nfrom torchvision import transforms, datasets\nfrom torch.optim.lr_scheduler import StepLR\nimport utils\n\ncuda = True\ncnt = 0\nlr = 0.001\nout_dir = \"out_aae3\"\nbatch_size = 100\n\nnc = 1 # number of channels\nnz = 8 # size of latent vector\nngf = 128 # decoder (generator) filter factor\nndf = 128 # encoder filter factor\nh_dim = 512 # discriminator hidden size\nlam = 10 # regulization coefficient\n\ntrainset = MNIST(root='./data/',train=True,transform=transforms.ToTensor(),download=True)\ndata_loader = DataLoader(trainset, batch_size=batch_size, shuffle=True, num_workers=20, drop_last=True)\n\n\ndef load_args():\n parser = argparse.ArgumentParser(description='PyTorch MNIST WAE-GAN')\n\n parser.add_argument('--batch-size', type=int, default=100, help='')\n parser.add_argument('--epochs', type=int, default=100, help='')\n parser.add_argument('--dim', type=int, default=8, help='')\n parser.add_argument('--epsilon', type=float, default=1e-15, help='')\n parser.add_argument('--l', type=int, default=10, help='')\n parser.add_argument('--lr', type=float, default=.001, help='')\n parser.add_argument('--e_dropout', type=bool, default=False, help='')\n parser.add_argument('--resume', type=bool, default=False, help='')\n parser.add_argument('--pretrain_epochs', type=int, default=200, help='')\n parser.add_argument('--loss', type=str, default='l2sq', help='')\n parser.add_argument('--dataset', type=str, default='mnist', help='')\n\n\n args = parser.parse_args()\n return args\n\n\nclass Decoder(nn.Module):\n def __init__(self):\n super(Decoder, self).__init__()\n # input is Z, going into a convolution\n self.linear = nn.Linear(nz, ngf*8*7*7)\n\n self.deconv1 = nn.ConvTranspose2d(ngf*8 , ngf*4, 4, 1, 0, 0, bias=False)\n self.bn1 = nn.BatchNorm2d(ngf*4)\n self.deconv2 = nn.ConvTranspose2d(ngf*4, ngf*2, 4, 1, 0, 0, bias=False)\n self.bn2 = nn.BatchNorm2d(ngf*2)\n self.deconv_out = nn.ConvTranspose2d(ngf*2, nc, 4, 2, 0, 0, bias=False)\n self.sigmoid = nn.Sigmoid()\n\n def forward(self, x):\n # print ('G in: ', x.shape)\n x = x.view(-1, nz) \n x = F.relu(self.linear(x))\n x = x.view(-1, ngf*8, 7, 7)\n x = F.relu(self.bn1(self.deconv1(x)))\n x = F.relu(self.bn2(self.deconv2(x)))\n x = self.deconv_out(x)\n # print ('G out: ', x.shape)\n x_act = self.sigmoid(x)\n\n return (x_act, x)\n\n\nclass Encoder(nn.Module):\n def __init__(self, is_training=True):\n super(Encoder, self).__init__()\n self.is_training = is_training\n # input is (nc) x 64 x 64\n self.conv1 = nn.Conv2d(nc, ndf, 4, 2, 1, bias=False)\n self.bn1 = nn.BatchNorm2d(ndf)\n # state size. (ndf) x 32 x 32\n self.conv2 = nn.Conv2d(ndf, ndf * 2, 4, 2, 1, bias=False)\n self.bn2 = nn.BatchNorm2d(ndf * 2)\n # state size. (ndf*2) x 16 x 16\n self.conv3 = nn.Conv2d(ndf * 2, ndf * 4, 4, 2, 1, bias=False)\n self.bn3 = nn.BatchNorm2d(ndf * 4)\n # state size. (ndf*4) x 8 x 8\n self.conv4 = nn.Conv2d(ndf * 4, ndf * 8, 4, 2, 1, bias=False)\n self.bn4 = nn.BatchNorm2d(ndf * 8)\n # state size. (ndf*8) x 4 x 4\n self.linear = nn.Linear(ndf*8, nz)\n\n def forward(self, x):\n #print ('E in: ', x.shape)\n if self.is_training:\n z = torch.normal(torch.zeros_like(x.data), std=0.01)\n x.data += z\n x = F.leaky_relu(self.bn1(self.conv1(x)))\n x = F.leaky_relu(self.bn2(self.conv2(x)))\n x = F.leaky_relu(self.bn3(self.conv3(x)))\n x = F.leaky_relu(self.bn4(self.conv4(x)))\n # x = self.conv5(x)\n x = x.view(-1, ndf*8)\n x = self.linear(x)\n #print ('E out: ', x.shape)\n return x\n\n\nclass Discriminator(nn.Module):\n def __init__(self):\n super(Discriminator, self).__init__()\n self.linear1 = nn.Linear(nz, h_dim)\n self.linear2 = nn.Linear(h_dim, h_dim)\n self.linear3 = nn.Linear(h_dim, h_dim)\n self.linear4 = nn.Linear(h_dim, h_dim)\n self.logits = nn.Linear(h_dim, 1)\n self.sigmoid = nn.Sigmoid()\n self.a = nn.ReLU()\n\n def forward(self, x):\n x = F.relu(self.linear1(x))\n x = F.relu(self.linear2(x))\n x = F.relu(self.linear3(x))\n x = F.relu(self.linear4(x))\n x = self.logits(x)\n x = self.sigmoid(x)\n return x\n\n\ndef load_networks():\n netE = Encoder().cuda()\n netG = Decoder().cuda()\n netD = Discriminator().cuda()\n print (netE, netG, netD)\n \n return netE, netG, netD\n\n\ndef reset_grad():\n netE.zero_grad()\n netG.zero_grad()\n netD.zero_grad()\n\n\ndef param_switch(module: nn.Module, state):\n if state == \"freeze\":\n for p in module.parameters():\n p.requires_grad = False\n else:\n for p in module.parameters():\n p.requires_grad = True\n\n\nargs = load_args()\n\nnetE, netG, netD = load_networks()\n\n# optimG = optim.Adam(netG.parameters(), lr=lr)\noptimE = optim.Adam(netE.parameters(), lr=lr)\noptimD = optim.Adam(netD.parameters(), lr=lr*0.5)\noptimAE = optim.Adam(list(netE.parameters())+list(netG.parameters()), lr=lr)\n\n# LRgen = StepLR(optimG, step_size=30, gamma=0.5)\nLRenc = StepLR(optimE, step_size=30, gamma=0.5)\nLRdis = StepLR(optimD, step_size=30, gamma=0.5)\nLRae = StepLR(optimAE, step_size=30, gamma=0.5)\n\nprint (args.resume)\nif args.resume is True:\n print (\"\\n==> Loading old weights if possible\")\n netE, optimE, _ = utils.load_model(netE, optimE, \"E_latest.pth\")\n netG, optimG, _ = utils.load_model(netG, optimG, \"G_latest.pth\")\n netD, optimD, _ = utils.load_model(netD, optimD, \"D_latest.pth\")\n optimAE = utils.load_model(optimAE, \"AE_latest.pth\")\n\n\ndef pretrain_e():\n print (\"==> Pretraining Encoder\")\n for epoch in range(args.pretrain_epochs):\n for i, images in enumerate(data_loader):\n x = Variable(images[0].cuda())\n noise = Variable(torch.rand(batch_size, nz)).cuda()\n latent = netE(x).view(batch_size, nz)\n #noise = noise.view(*noise.size(), 1, 1)\n loss = utils.pretrain_loss(latent, noise)\n loss.backward()\n optimE.step()\n reset_grad()\n print (\"Pretrain Enc iter: {}, Loss: {}\".format(i, loss.data[0]))\n if loss.data[0] < 0.1:\n print (\"Finished Pretraining Encoder\")\n return\n\nif not args.resume:\n pretrain_e()\n\none = torch.Tensor([1]).cuda()\nmone = (one * -1).cuda()\n\nfor epoch in range(args.epochs):\n for batch_idx, batch_images in enumerate(data_loader):\n \"\"\" var creation \"\"\"\n X = Variable(batch_images[0]).cuda()\n z_sample = Variable(torch.randn(batch_size, nz)).cuda()\n \n z_enc = netE(X)\n X_sample, X_logits = netG(z_enc)\n reset_grad()\n\t\n \"\"\" Regularization phase \"\"\"\n param_switch(netG, \"freeze\")\n param_switch(netE, \"freeze\")\n param_switch(netD, \"free\")\n # z adversary\n D_fake = netD(z_sample)\n z_real = netE(X)\n D_real = netD(z_real)\n D_loss_fake = torch.log(D_fake).mean()\n D_loss_real = torch.log(1-(D_real+args.epsilon)).mean()\n D_loss = D_loss_fake + D_loss_real\n D_loss_fake.backward(mone)\n D_loss_real.backward(mone)\n optimD.step()\n \n \"\"\" WAE update \"\"\"\n param_switch(netG, \"free\")\n param_switch(netE, \"free\")\n param_switch(netD, \"freeze\")\n\n z_enc = netE(X)\n\n # recon_loss = utils.ae_loss(args, X+args.epsilon, X_sample+args.epsilon)\n recon_loss = F.mse_loss(X_sample, X)\n penalty = utils.gan_loss2(args, z_sample, z_enc, netD)\n recon_loss.backward(one, retain_graph=True)\n penalty.backward(mone)\n optimAE.step()\n\n \"\"\"\n z_fake = netE(X).view(batch_size,-1)\n D_fake = netD(z_fake)\n\n #G_loss = -torch.mean(torch.log(D_fake))\n G_loss = -torch.mean(D_fake)\n\n G_loss.backward()\n optimE.step()\n reset_grad()\n\t\"\"\"\n if batch_idx % 50 == 0:\n losses = [D_loss.data[0], penalty.data[0], recon_loss.data[0]]\n print('Epoch {}; iter {}; D_loss: {:.4}; G_loss: {:.4}; recon_loss: {:.4}'\n .format(epoch, batch_idx, losses[0], losses[1], losses[2]))\n utils.save_model(netE, optimE, epoch, \"E_latest.pth\")\n utils.save_model(netG, optimAE, epoch, \"G_latest.pth\")\n utils.save_model(netD, optimD, epoch, \"D_latest.pth\")\n\n # Print and plot every now and then\n if batch_idx % 100 == 0:\n utils.save_image(netG, epoch, batch_idx, orthogonal=False)\n \n","sub_path":"test_mnist_waegan.py","file_name":"test_mnist_waegan.py","file_ext":"py","file_size_in_byte":9036,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"82320134","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Mar 05 04:35:49 2015\n\n@author: DavidDobor\n\"\"\"\n\n# part 1: an implementation of a brute-force pattern-matching algorithm\n\ndef find_brute(T, P):\n \"\"\" Return the lowest index of T at which substring P begins (or else -1).\"\"\"\n n, m = len(T), len(P)\n for i in range(n-m+1): # try every potential starting index within T\n k = 0\n while k < m and T[i + k] == P[k]:\n k += 1\n if k == m:\n return i\n return -1\n \n \nT = \"abacaabaccabacabaabb\"\nP = \"abacab\"\n\nfind_brute(T,P)","sub_path":"dynamic/LCS/brute_pattern_match.py","file_name":"brute_pattern_match.py","file_ext":"py","file_size_in_byte":562,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"589130925","text":"from pytest import mark\r\nimport sqlalchemy as sa\r\nfrom sqlalchemy_utils.types import url\r\n\r\nfrom tests import TestCase\r\n\r\n\r\n@mark.skipif('url.furl is None')\r\nclass TestURLType(TestCase):\r\n def create_models(self):\r\n class User(self.Base):\r\n __tablename__ = 'user'\r\n id = sa.Column(sa.Integer, primary_key=True)\r\n website = sa.Column(url.URLType)\r\n\r\n def __repr__(self):\r\n return 'User(%r)' % self.id\r\n\r\n self.User = User\r\n\r\n def test_color_parameter_processing(self):\r\n user = self.User(\r\n website=url.furl(u'www.example.com')\r\n )\r\n\r\n self.session.add(user)\r\n self.session.commit()\r\n\r\n user = self.session.query(self.User).first()\r\n assert isinstance(user.website, url.furl)\r\n\r\n def test_scalar_attributes_get_coerced_to_objects(self):\r\n user = self.User(website=u'www.example.com')\r\n\r\n assert isinstance(user.website, url.furl)\r\n","sub_path":"python/flask-mail-labs/.venv/lib/python2.7/site-packages/tests/types/test_url.py","file_name":"test_url.py","file_ext":"py","file_size_in_byte":979,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"329102215","text":"from sqlalchemy import desc\nfrom sqlalchemy.sql import and_, select\n\nfrom tno.db import DBBase\nfrom datetime import datetime\n\nclass DesSkybotJobDao(DBBase):\n def __init__(self, pool=True):\n super(DesSkybotJobDao, self).__init__(pool)\n\n schema = self.get_base_schema()\n self.tablename = \"des_skybotjob\"\n self.tbl = self.get_table(self.tablename, schema)\n\n def get_tablename(self):\n return self.tablename\n\n def get_tbl(self):\n return self.tbl\n\n def get_by_id(self, id):\n\n stm = select(self.tbl.c).where(and_(self.tbl.c.id == int(id)))\n\n self.debug_query(stm, True)\n\n row = self.fetch_one_dict(stm)\n\n return row\n\n def get_by_status(self, status):\n \"\"\"Retorna os jobs pelo status\n ATENÇÃO: Os jobs estão ordenados pela data de criação em ordem ascendente. está ordem é importante para o pipeline.\n\n Arguments:\n status {int} -- Status do job, como está definido no Model des/SkybotJob\n\n Returns:\n [array] -- Um Array com os jobs, tem a mesma estrutura do Model des/SkybotJob\n \"\"\"\n stm = (\n select(self.tbl.c)\n .where(and_(self.tbl.c.status == int(status)))\n .order_by(self.tbl.c.start)\n )\n\n rows = self.fetch_all_dict(stm)\n\n return rows\n\n def update_by_id(self, id, job):\n stm = (\n self.tbl.update()\n .where(self.tbl.c.id == int(id))\n .values(\n status=job[\"status\"],\n start=job[\"start\"],\n exposures=job[\"exposures\"],\n nights=job[\"nights\"], \n ccds=job[\"ccds\"], \n positions=job[\"positions\"],\n asteroids=job[\"asteroids\"],\n exposures_with_asteroid=job[\"exposures_with_asteroid\"],\n ccds_with_asteroid=job[\"ccds_with_asteroid\"],\n path=job[\"path\"],\n results=job[\"results\"],\n estimated_execution_time=job[\"estimated_execution_time\"]\n\n )\n )\n\n return self.execute(stm)\n\n def complete_job(self, id, job):\n stm = (\n self.tbl.update()\n .where(self.tbl.c.id == int(id))\n .values(\n status=job[\"status\"],\n # finish=job['finish'].strftime('%Y-%m-%d %H:%M:%S'),\n finish=job[\"finish\"],\n execution_time=job[\"execution_time\"],\n error=job[\"error\"],\n )\n )\n\n return self.execute(stm)\n","sub_path":"backend/des/dao/skybot_job.py","file_name":"skybot_job.py","file_ext":"py","file_size_in_byte":2568,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"269819803","text":"import torch\nimport numpy as np\n\nfrom typing import Dict, List, Optional\n\nfrom spacy.tokens import Token as SpacyToken\n\nfrom overrides import overrides\n\nfrom allennlp.common.checks import ConfigurationError\nfrom allennlp.nn import util\nfrom allennlp.data.tokenizers.token import Token\nfrom allennlp.data.vocabulary import Vocabulary\nfrom allennlp.data.fields import SequenceField\nfrom allennlp.data.fields.text_field import TokenList\n\n\nclass BidirectionalLanguageModelField(SequenceField[Dict[str, torch.Tensor]]):\n \"\"\"\n Taken from calypso:\n https://github.com/allenai/calypso/blob/ac934b6881787387581efaa8a646531278010652/calypso/allennlp_bridge.py#L51-L131\n\n Field for adding targets for a BidirectionalLM in a multi-task setting.\n \"\"\"\n def __init__(self,\n tokens: List[Token]) -> None:\n self.tokens = tokens\n self._indexed_tokens: Optional[Dict[str, TokenList]] = None\n self._directions = ['forward_targets', 'backward_targets']\n\n if not all([isinstance(x, (Token, SpacyToken)) for x in tokens]):\n raise ConfigurationError(\"TextFields must be passed Tokens. \"\n \"Found: {} with types {}.\".format(tokens, [type(x) for x in tokens]))\n\n @overrides\n def count_vocab_items(self, counter: Dict[str, Dict[str, int]]):\n pass\n\n @overrides\n def index(self, vocab: Vocabulary):\n # self._indexed_tokens = {'forward_targets': array of next token id,\n # 'backward_targets': array of prev token id}\n\n tokens = [token.text for token in self.tokens]\n token_ids = [vocab.get_token_index(token, namespace='lm')\n for token in tokens]\n\n n_tokens = len(tokens)\n forward_targets = np.zeros(n_tokens, dtype='int32')\n forward_targets[0:-1] = token_ids[1:]\n\n backward_targets = np.zeros(n_tokens, dtype='int32')\n backward_targets[1:n_tokens] = token_ids[0:(n_tokens - 1)]\n\n self._indexed_tokens = {\n direction: target for direction, target in\n zip(self._directions, [forward_targets, backward_targets])\n }\n\n @overrides\n def get_padding_lengths(self) -> Dict[str, int]:\n # the padding length is the number of tokens in the sequence\n if self._indexed_tokens is None:\n raise ValueError(\"self._indexed_tokens is None.\")\n return {'num_lm_targets': len(self._indexed_tokens['forward_targets'])}\n\n @overrides\n def sequence_length(self) -> int:\n return len(self.tokens)\n\n @overrides\n def as_tensor(self,\n padding_lengths: Dict[str, int],\n cuda_device: int = -1) -> Dict[str, torch.Tensor]:\n\n desired_num_tokens = padding_lengths['num_lm_targets']\n if self._indexed_tokens is None:\n raise ValueError(\"self._indexed_tokens is None.\")\n unpadded_num_tokens = len(self._indexed_tokens['forward_targets'])\n\n # need to pad indexed_tokens to the specified length then return\n # as torch tensor on device\n tensors = {}\n\n for k in ['forward_targets', 'backward_targets']:\n padded_array = np.zeros(desired_num_tokens, dtype=np.int32)\n padded_array[:unpadded_num_tokens] = self._indexed_tokens[k]\n\n tensor = torch.LongTensor(padded_array)\n if cuda_device >= 0:\n tensor = tensor.cuda(cuda_device)\n tensors[k] = tensor\n\n return tensors\n\n @overrides\n def empty_field(self, vocab):\n text_field = BidirectionalLanguageModelField([])\n text_field.index(vocab)\n return text_field\n\n @overrides\n def batch_tensors(self, tensor_list: List[Dict[str, torch.Tensor]]) -> Dict[str, torch.Tensor]:\n # This is creating a dict of {token_indexer_key: batch_tensor} for each token indexer used\n # to index this field.\n return util.batch_tensor_dicts(tensor_list)\n","sub_path":"contexteval/data/fields/bidirectional_language_model_field.py","file_name":"bidirectional_language_model_field.py","file_ext":"py","file_size_in_byte":3952,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"37389030","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\nfrom django.shortcuts import render, HttpResponse, redirect\nfrom django.utils.crypto import get_random_string\n\n# Create your views here.\ndef index(request):\n if \"count\" not in request.session:\n request.session[\"count\"] = 1\n else:\n request.session[\"count\"] += 1\n \n random_word = {\n \"word\": get_random_string(length=14),\n }\n\n return render(request, \"random_word_app/index.html\", random_word)\n\ndef reset(request):\n request.session[\"count\"] = 0\n return redirect(\"/random_word\")","sub_path":"Week3/RandomWordGeneratorAssignment/apps/random_word_app/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":586,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"373669869","text":"# Tank Game:\r\n# A game with tanks and stuff.\r\n#\r\n# by Erkalys & Florimond Manca\r\n#\r\n# Level editor\r\n\r\n# Imports\r\n\r\nfrom tkinter import *\r\nimport math\r\nimport sys\r\nimport os\r\njoin = os.path.join\r\n\r\n# chemin d'accès aux fichiers\r\n\r\npath = os.getcwd()\r\n\r\n# Variables globales\r\n\r\nedit_path = join(path, \"custom_levels\")\r\nint_matrix = [[0 for j in range(32)] for i in range(21)]\r\nIA = [] # liste correspondant aux IA blue, bluePlus, purple et purplePlus\r\npaths_matrix = []\r\nrect_matrix = [[None for j in range(32)] for i in range(21)]\r\ncolor_dict = {'white': ('#fff', 0), 'black': ('#000', 1), 'brown': ('#a60', 2),\r\n 'green': ('#0f0', 3), 'yellow': ('#ff0', 4), 'orange': ('#fc0', 5),\r\n 'blue': ('#00f', 6), 'darkblue': ('#00a', 7), 'red': ('#f00', 8), 'darkred': ('#c00', 9),\r\n 'purple': ('#f0f', 10), 'darkpurple': ('#a0a', 11), 'grey': ('#aaa', 12), 'darkgrey': ('#777', 13)}\r\ntype_dict = {'white': 'None', 'black': 'Wall', 'brown': 'Pit',\r\n 'green': 'Player', 'yellow': 'IA Yellow', 'orange': 'IA YellowPlus',\r\n 'blue': 'IA Blue', 'darkblue': 'IA BluePlus', 'red': 'IA Red', 'darkred': 'IA redPlus',\r\n 'purple': 'IA purple', 'darkpurple': 'IA PurplePlus', 'grey': \"Spawner\", 'darkgrey': \"SpawnerPlus\"}\r\n\r\n# UTILITAIRES\r\n\r\n\r\ndef convert(liste, split=False): # renvoie la chaine de caractère correspondant à la liste avec des espaces\r\n string = \"\"\r\n if not split:\r\n for i in range(len(liste)):\r\n string += str(liste[i]) + \" \"\r\n else: # split = True, la liste est une liste de tupple qu'il faut séparer\r\n for i in range((len(liste))):\r\n string += str(liste[i][0]) + \" \" + str(liste[i][1]) + \" \"\r\n return string + '\\n'\r\n\r\n\r\ndef create_level(matrice, paths):\r\n n = 1\r\n while os.path.exists(join(edit_path, \"custom_level\" + str(n) + \".txt\")):\r\n n += 1\r\n fname = \"custom_level\" + str(n) + \".txt\"\r\n fullname = join(edit_path, fname)\r\n file = open(fullname, 'w')\r\n for i in range(len(matrice)):\r\n file.write(convert(matrice[i]))\r\n if paths != []:\r\n for i in range(len(paths)):\r\n file.write(convert(paths[i], split=True))\r\n file.close()\r\n\r\n\r\ndef convert_and_sort(matrix): # liste de liste de way_points -> liste de x1 y1 x2 y2 ...\r\n sortie = []\r\n for i in range(len(matrix)):\r\n sortie.append(matrix[i])\r\n for j in range(len(matrix[i])):\r\n (a, b) = matrix[i][j].pos\r\n sortie[i][j] = (b, a)\r\n sortie.sort() # on trie les listes de la matrice\r\n for i in range(len(sortie)):\r\n for j in range(len(sortie[i])):\r\n (a, b) = sortie[i][j]\r\n sortie[i][j] = (b, a)\r\n return sortie\r\n\r\n\r\ndef export(zone_dessin, int_matrix, paths_matrix, IA):\r\n warn = Tk()\r\n warn.title(\"Warning\")\r\n\r\n texte = Label(warn, text=\"You won't be able to edit this level after its exportation\")\r\n texte.pack()\r\n\r\n back = Button(warn, text='Get back to work', command=warn.destroy)\r\n back.pack(side=LEFT)\r\n\r\n def fct(zone_dessin, int_matrix, paths_matrix, IA):\r\n create_level(int_matrix, convert_and_sort(paths_matrix))\r\n clear(zone_dessin, int_matrix, paths_matrix, IA)\r\n warn.destroy()\r\n\r\n exp = Button(warn, text='Continue', command=lambda: fct(\r\n zone_dessin, int_matrix, paths_matrix, IA))\r\n exp.pack(side=RIGHT)\r\n warn.mainloop()\r\n\r\n\r\ndef clear(zone_dessin, int_matrix, paths_matrix, IA): # réinitialise l'éditeur de niveau\r\n int_matrix = [[0 for j in range(32)] for i in range(21)]\r\n paths_matrix = []\r\n IA = []\r\n for i in range(21):\r\n for j in range(32):\r\n if i == 0 or i == 20 or j == 0 or j == 31: # on est sur les bords\r\n rect = rectangle(zone_dessin, i, j, 'black', color_dict)\r\n rect_matrix[i][j] = rect\r\n int_matrix[i][j] = color_dict['black'][1]\r\n else:\r\n rect = rectangle(zone_dessin, i, j, 'white', color_dict)\r\n rect_matrix[i][j] = rect\r\n int_matrix[i][j] = color_dict['white'][1]\r\n\r\n\r\ndef clear_paths(paths_matrix, zone_dessin, IA): # réinitialise uniquement les chemins liés aux IA\r\n paths_matrix = []\r\n for i in range(len(IA)):\r\n paths_matrix.append([])\r\n zone_dessin.delete((\"path\", str(i)))\r\n for x in IA:\r\n x.points_list = [way_point(zone_dessin, x.pos[1], x.pos[0], x.id)]\r\n\r\n# GESTION D'EVENEMENTS LIES A LA SOURIS\r\n\r\n\r\nclass mouse():\r\n ''' la classe qui gère toutes les actions du curseur dans l'interface Tkinter'''\r\n\r\n def __init__(self, fenetre, color='white'):\r\n self.color = color\r\n self.mode = 'normal'\r\n self.where = fenetre\r\n self.buttons = None\r\n self.selected = None # le rectangle de l'IA bleue sélectionnée\r\n\r\n # les fonctions d'état\r\n\r\n def mouse_normal(self, zone_dessin):\r\n # passe la souris en mode normal à l'appui sur le bouton 'Normal mode'\r\n self.mode = 'normal'\r\n self.selected = None\r\n zone_dessin.delete(\"paths\")\r\n\r\n def mouse_path_creation(self, zone_dessin):\r\n # passe la souris en mode de création de path pour IA bleue à\r\n # l'appui sur le bouton 'Path creation mode'\r\n self.mode = 'path_creation'\r\n for x in IA:\r\n pts = x.points_list\r\n for i in range(len(pts) - 1):\r\n pts[i].plot(zone_dessin)\r\n pts[i].bind(pts[i + 1], zone_dessin)\r\n\r\n def In_Outils(self, event):\r\n self.where = \"zone_outils\"\r\n\r\n def Out(self, event):\r\n self.where = None\r\n\r\n def In_Dessin(self, event):\r\n self.where = \"zone_dessin\"\r\n\r\n # les fonctions d'édition\r\n\r\n def mouse_color_click(self, event):\r\n # permet de donner à la souris la couleur qui a été choisie\r\n for button in self.buttons:\r\n if button.selected:\r\n button.highlight_button() # on enlève la précédente selection\r\n if (event.x // 25, event.y // 25) == button.pos:\r\n self.color = button.color\r\n self.mode = 'normal' # on repasse en mode normal\r\n button.highlight_button()\r\n\r\n def print_color(self, event, zone_dessin, int_matrix, paths_matrix, color_dict, IA):\r\n\r\n # permet en mode normal de déposer sur le canevas un rectangle de couleur self.color\r\n # permet en mode création de chemin de dessiner le chemin de la dernière IA sélectionnée\r\n j, i = event.x // 25, event.y // 25\r\n\r\n if 1 <= i <= 19 and 1 <= j <= 30: # on ne doit pas changer les murs extérieurs\r\n rect = rect_matrix[i][j]\r\n\r\n # Cas 1: on est en mode normal d'édition\r\n if self.mode == 'normal':\r\n if rect.color != self.color:\r\n # on a affaire à une IA necessitant un path\r\n if self.color == 'blue' or self.color == 'darkblue' or self.color == 'purple' or self.color == 'darkpurple':\r\n newIA = IA_rectangle(zone_dessin, i, j, self.color, zone_dessin, color_dict)\r\n rect_matrix[i][j] = newIA\r\n int_matrix[i][j] = color_dict[self.color][1]\r\n newIA.id = len(IA)\r\n IA.append(newIA)\r\n paths_matrix.append([])\r\n else:\r\n rect_matrix[i][j] = rectangle(zone_dessin, i, j, self.color, color_dict)\r\n int_matrix[i][j] = color_dict[self.color][1]\r\n\r\n # Cas 2: on est en mode création de chemin d'IA bleue\r\n if self.mode == 'path_creation':\r\n n = int_matrix[i][j]\r\n if n == 6 or n == 7 or n == 10 or n == 11: # on a affaire à une IA qui nécessite un path\r\n for x in IA:\r\n if x.selected:\r\n x.highlight_rect(zone_dessin) # on deselectionne toutes les IA\r\n cwia = rect_matrix[i][j]\r\n self.selected = cwia.id\r\n cwia.highlight_rect(zone_dessin)\r\n elif self.selected != None:\r\n cwia = IA[self.selected]\r\n pt = way_point(zone_dessin, i, j, cwia.id)\r\n pt.bind(cwia.points_list[-1], zone_dessin)\r\n pt.plot(zone_dessin)\r\n cwia.points_list.append(pt)\r\n paths_matrix[cwia.id] = cwia.points_list\r\n\r\n def mouse_click_left(self, event, zone_dessin, int_matrix, paths_matrix, color_dict, IA):\r\n if self.where == 'zone_dessin':\r\n self.print_color(event, zone_dessin, int_matrix, paths_matrix,\r\n color_dict, IA) # on effectue l'action de la zone de dessin\r\n elif self.where == 'zone_outils':\r\n self.mouse_color_click(event) # on effecture l'action de la zone des outils\r\n\r\n def mouse_click_right(self, event, zone_dessin, int_matrix, rect_matrix, paths_matrix, color_dict, number):\r\n # N'est activée que dans la zone de dessin (contrairement au clic gauche)\r\n if self.where == \"zone_dessin\":\r\n\r\n if self.mode == 'normal': # En mode normal, permet d'effacer le rectangle sur lequel pointe la souris\r\n j, i = event.x // 25, event.y // 25\r\n if 1 <= i <= 19 and 1 <= j <= 30: # on ne doit pas changer les murs extérieurs\r\n rect = rect_matrix[i][j]\r\n if self.mode == 'normal':\r\n rect_matrix[i][j] = rectangle(zone_dessin, i, j, \"white\", color_dict)\r\n int_matrix[i][j] = 0\r\n\r\n elif self.mode == 'path_creation': # En mode path_creation, permet d'enlever un point du chemin\r\n if self.selected != None:\r\n cwia = IA[self.selected]\r\n pts = cwia.points_list\r\n if (len(pts)) > 1:\r\n pts.pop()\r\n cwia.points_list = pts\r\n paths_matrix[cwia.id] = pts\r\n zone_dessin.delete((\"paths\", str(cwia.id)))\r\n for i in range(len(pts) - 1):\r\n pts[i].plot(zone_dessin)\r\n pts[i].bind(pts[i + 1], zone_dessin)\r\n\r\n\r\n# CLASSES DIVERSES\r\n\r\nclass rectangle():\r\n '''regroupe un rectangle du canevas, sa couleur et sa position'''\r\n\r\n def __init__(self, can, i, j, color, color_dict):\r\n rect = can.create_rectangle(25 * j, 25 * i, 25 * (j + 1), 25 * (i + 1),\r\n width=1, fill=color_dict[color][0], outline=\"black\")\r\n self.rect = rect\r\n self.color = color\r\n self.pos = (j, i)\r\n self.master = can\r\n self.selected = False\r\n\r\n\r\nclass IA_rectangle:\r\n ''' un rectangle d'une IA bleue. Contient, en plus d'un rectangle classique, le parcours de l'IA'''\r\n\r\n def __init__(self, can, i, j, color, zone_dessin, color_dict):\r\n rect = can.create_rectangle(25 * j, 25 * i, 25 * (j + 1), 25 * (i + 1),\r\n width=1, fill=color_dict[color][0], outline=\"black\")\r\n self.rect = rect\r\n self.color = color\r\n self.pos = (j, i)\r\n self.master = can\r\n self.id = -1\r\n self.selected = False\r\n # liste d'objets de classe way_point formant le chemin suivi\r\n self.points_list = [way_point(zone_dessin, i, j, self.id)]\r\n\r\n def highlight_rect(self, zone_dessin):\r\n can = self.master\r\n j, i = self.pos\r\n color = self.color\r\n if not self.selected:\r\n self.rect = can.create_rectangle(\r\n 25 * j, 25 * i, 25 * (j + 1), 25 * (i + 1), width=1, fill=color_dict[color][0], outline=\"#fe2\")\r\n self.points_list[0] = way_point(zone_dessin, i, j, self.id)\r\n else:\r\n self.rect = can.create_rectangle(\r\n 25 * j, 25 * i, 25 * (j + 1), 25 * (i + 1), width=1, fill=color_dict[color][0], outline='black')\r\n self.points_list[0] = way_point(zone_dessin, i, j, self.id)\r\n self.selected = not self.selected\r\n\r\n\r\nclass way_point():\r\n\r\n def __init__(self, can, i, j, n):\r\n rect = can.create_rectangle(25 * j + 5, 25 * i + 5, 25 * (j + 1) - 5, 25 * (i + 1) - 5,\r\n width=1, fill=\"#aaa\", outline=\"#000\", tag=(\"path\", str(n)))\r\n self.id = n\r\n self.rect = rect\r\n self.color = \"#aaa\"\r\n self.pos = (j, i)\r\n\r\n def bind(self, wp, zone_dessin): # trace une ligne entre deux waypoints\r\n (j1, i1) = self.pos\r\n (j2, i2) = wp.pos\r\n x1, y1, x2, y2 = 25 * j1 + 12, 25 * i1 + 12, 25 * j2 + 12, 25 * i2 + 12\r\n if abs(x1 - x2) > abs(y1 - y2):\r\n zone_dessin.create_line(x1, y1, x2, y1, x2, y2, fill=\"#aaa\",\r\n width=1, tag=(\"path\", str(self.id)))\r\n else:\r\n zone_dessin.create_line(x1, y1, x1, y2, x2, y2, fill=\"#aaa\",\r\n width=1, tag=(\"path\", str(self.id)))\r\n\r\n def plot(self, zone_dessin):\r\n (j, i) = self.pos\r\n zone_dessin.create_rectangle(25 * j + 8, 25 * i + 8, 25 *\r\n (j + 1) - 8, 25 * (i + 1) - 8, tag=(\"path\", str(self.id)))\r\n\r\n\r\nclass color_button:\r\n # classe des boutons-outil\r\n\r\n def __init__(self, can, x1, y1, x2, y2, color, color_dict, type_dict):\r\n rect, text = self.create_button(can, x1, y1, x2, y2, color, color_dict, type_dict)\r\n self.rectangle = rect\r\n self.pos = (x1 // 25, y1 // 25)\r\n self.master = can\r\n self.coords = (x1, y1, x2, y2)\r\n self.text = text\r\n self.color = color\r\n self.selected = False\r\n\r\n def create_button(self, can, x1, y1, x2, y2, color, color_dict, type_dict):\r\n rect = can.create_rectangle(x1, y1, x2, y2, width=1, fill=color_dict[\r\n color][0], outline='black')\r\n text = can.create_text(x2 + 4, (y1 + y2) / 2, text=type_dict[color], anchor=W)\r\n return rect, text\r\n\r\n def highlight_button(self):\r\n can = self.master\r\n x1, y1, x2, y2 = self.coords\r\n color = self.color\r\n if not self.selected:\r\n self.rectangle = can.create_rectangle(\r\n x1, y1, x2, y2, width=1, fill=color_dict[color][0], outline='#ee2')\r\n else:\r\n self.rectangle = can.create_rectangle(\r\n x1, y1, x2, y2, width=1, fill=color_dict[color][0], outline='black')\r\n self.selected = not self.selected\r\n\r\n\r\n# LA FONCTION PRINCIPALE\r\n\r\ndef level_editor():\r\n\r\n fenetre = Tk()\r\n fenetre.title(\"Tank Game Level Editor\")\r\n\r\n # la souris\r\n souris = mouse(fenetre)\r\n fenetre.bind(\r\n \"\",\r\n lambda event: souris.mouse_click_left(event,\r\n zone_dessin, int_matrix, paths_matrix, color_dict, IA))\r\n fenetre.bind(\r\n \"\",\r\n lambda event: souris.mouse_click_right(event,\r\n zone_dessin, int_matrix, rect_matrix, paths_matrix, color_dict, IA))\r\n bout = Button(fenetre, text='Exit', command=lambda: fenetre.destroy())\r\n bout.pack()\r\n\r\n # la fenetre de boutons d'action\r\n fenetre_divers = LabelFrame(fenetre, text=\"Options\")\r\n fenetre_divers.pack(side=RIGHT)\r\n b_normal = Button(fenetre_divers, text=\"Normal mode\",\r\n command=lambda: souris.mouse_normal(zone_dessin))\r\n b_normal.pack()\r\n b_path = Button(fenetre_divers, text=\"Path creation mode\",\r\n command=lambda: souris.mouse_path_creation(zone_dessin))\r\n b_path.pack()\r\n exporter = Button(fenetre_divers, text=\"Export\", command=lambda: export(\r\n zone_dessin, int_matrix, paths_matrix, IA))\r\n exporter.pack()\r\n clear_all = Button(fenetre_divers, text=\"Clear All\", command=lambda: clear(\r\n zone_dessin, int_matrix, paths_matrix, IA))\r\n clear_all.pack()\r\n clear_p = Button(fenetre_divers, text=\"Clear Paths\",\r\n command=lambda: clear_paths(paths_matrix, zone_dessin, IA))\r\n clear_p.pack()\r\n\r\n # la fenetre de boutons-outil\r\n fenetre_outils = LabelFrame(fenetre, text=\"Outils\")\r\n fenetre_outils.bind(\"\", souris.mouse_color_click)\r\n fenetre_outils.bind(\"\", souris.In_Outils)\r\n fenetre_outils.bind(\"\", souris.Out)\r\n fenetre_outils.pack(side=LEFT)\r\n zone_outils = Canvas(fenetre_outils, width=120, height=525)\r\n zone_outils.pack()\r\n\r\n # la fenetre de dessin\r\n fenetre_dessin = LabelFrame(fenetre, text=\"Map\")\r\n fenetre_dessin.bind(\"\", souris.In_Dessin)\r\n fenetre_dessin.bind(\"\", souris.Out)\r\n fenetre_dessin.pack(side=RIGHT)\r\n zone_dessin = Canvas(fenetre_dessin, width=800, height=525)\r\n zone_dessin.pack()\r\n\r\n # création des boutons-outil dans le canevas zone_outils\r\n\r\n buttons = []\r\n for color in color_dict:\r\n n = color_dict[color][1]\r\n button = color_button(zone_outils, 4, 25 * (n + 1), 4 + 25,\r\n 25 * (n + 2), color, color_dict, type_dict)\r\n buttons.append(button)\r\n souris.buttons = buttons\r\n\r\n # creation des rectangles et initialisation de la matrice\r\n\r\n for i in range(21):\r\n for j in range(32):\r\n if i == 0 or i == 20 or j == 0 or j == 31: # on est sur les bords\r\n rect = rectangle(zone_dessin, i, j, 'black', color_dict)\r\n rect_matrix[i][j] = rect\r\n int_matrix[i][j] = color_dict['black'][1]\r\n else:\r\n rect = rectangle(zone_dessin, i, j, 'white', color_dict)\r\n rect_matrix[i][j] = rect\r\n int_matrix[i][j] = color_dict['white'][1]\r\n\r\n # boucle principale\r\n fenetre.mainloop()\r\n fenetre.destroy()\r\n\r\nif __name__ == \"__main__\":\r\n level_editor()\r\n","sub_path":"src/leveleditor.py","file_name":"leveleditor.py","file_ext":"py","file_size_in_byte":17983,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"260816931","text":"from json import dump as dump_json, load as load_json\nfrom pickle import dumps as pickle_dumps, loads as pickle_loads\nfrom random import choice as random_choice\n\nfrom django.conf import settings as django_settings\nfrom redis.sentinel import Sentinel\nfrom redis.exceptions import TimeoutError\nfrom redis.sentinel import MasterNotFoundError\n\nfrom .utils import executing_test_case, executing_travis\n\nimport logging\n_logger = logging.getLogger(__name__)\nd = logging.getLogger(__name__).debug\n\n\ndef RedisSentinelCache(*args, **kwargs):\n \"\"\"\n A factory for the redis client.\n\n Returns dummy cache with hardcoded dict as the storage when executing inside travis\n \"\"\"\n if executing_travis() or kwargs.get('dummy', False):\n return _RedisSentinelCacheDummy(*args, **kwargs)\n else:\n return _RedisSentinelCache(*args, **kwargs)\n\n\nclass _RedisSentinelCache():\n\n def __init__(self, db=0, master_only=False, settings=django_settings):\n \"\"\"\n db: database index to read/write to. available indexes 0-15.\n master_only: always use master for read operations, for those times when you know you are going to\n read the same key again from cache very soon.\n settings: override redis setttings in settings.py. easier to use class from outside context of django (i.e. cron)\n \"\"\"\n if not isinstance(settings, dict):\n if not hasattr(settings, 'REDIS_SENTINEL'):\n raise Exception('Missing configuration from settings.py: REDIS_SENTINEL')\n settings = settings.REDIS_SENTINEL\n\n if not settings.get('HOSTS', None):\n raise Exception('Missing configuration from settings for REDIS_SENTINEL: HOSTS')\n if not settings.get('SERVICE', None):\n raise Exception('Missing configuration from settings for REDIS_SENTINEL: SERVICE')\n if not settings.get('TEST_DB', None):\n raise Exception('Missing configuration from settings for REDIS_SENTINEL: TEST_DB')\n if len(settings['HOSTS']) < 3:\n raise Exception('Invalid configuration in settings for REDIS_SENTINEL: HOSTS minimum number of hosts is 3')\n\n if executing_test_case():\n db = settings['TEST_DB']\n elif db == settings['TEST_DB']:\n raise Exception('Invalid db: db index %d is reserved for test suite execution.' % db)\n\n self._sentinel = Sentinel(settings['HOSTS'], password=settings['PASSWORD'], socket_timeout=settings.get('SOCKET_TIMEOUT', 0.1), db=db)\n self._service_name = settings['SERVICE']\n self._DEBUG = settings.get('DEBUG', False)\n self._read_from_master_only = master_only\n self._node_count = self._node_count()\n\n def set(self, key, value, **kwargs):\n if self._DEBUG:\n d('cache: set()...')\n pickled_data = pickle_dumps(value)\n master = self._get_master()\n\n try:\n return master.set(key, pickled_data, **kwargs)\n except (TimeoutError, MasterNotFoundError):\n if self._DEBUG:\n d('cache: master timed out or not found. no write instances available')\n # no master available\n return\n\n if self._DEBUG:\n test = master.get(key)\n if test:\n d('cache: set() successful')\n else:\n d('cache: set() unsuccessful, could not get saved data?')\n\n def get_or_set(self, key, value, **kwargs):\n \"\"\"\n Atomic set of value only if key did not exist yet.\n\n Returns True if set was successful, None if set failed (= value existed)\n\n https://redis.io/commands/setnx Not recommended any longer for distributed locks...\n https://redis.io/topics/distlock However this is also just a proposal and no official\n implementation exists yet\n \"\"\"\n return self.set(key, value, nx=True, **kwargs)\n\n def get(self, key, **kwargs):\n \"\"\"\n Randomly select slave or master for reading. Fallback to master anyway in case of errors.\n\n Use of master can be forced by using the master_only flag in the constructor.\n \"\"\"\n\n # todo allow reading from cache when master is down? could possibly serve stale data.\n # see redis.conf setting: slave-serve-stale-data\n if self._DEBUG:\n d('cache: get()...')\n\n if self._read_from_master_only:\n return self._get_from_master(key, **kwargs)\n else:\n if self._slave_chosen():\n try:\n return self._get_from_slave(key, **kwargs)\n except TimeoutError:\n pass\n # lady luck chose master, or read from slave had an error\n return self._get_from_master(key, **kwargs)\n\n def delete(self, *keys):\n if self._DEBUG:\n d('cache: delete()...')\n\n master = self._get_master()\n\n try:\n master.delete(*keys)\n except (TimeoutError, MasterNotFoundError):\n if self._DEBUG:\n d('cache: master timed out or not found. no write instances available. raising error')\n # no master available\n return\n\n if self._DEBUG:\n test = master.get(keys[0])\n if not test:\n d('cache: delete() successful')\n else:\n d('cache: delete() unsuccessful, could not delete data?')\n\n def _get_from_slave(self, key, **kwargs):\n node = self._get_slave()\n try:\n res = node.get(key, **kwargs)\n except TimeoutError:\n if self._DEBUG:\n d('cache: slave.get() timed out, trying from master instead. fail-over in process?')\n # fail-over propbably happened, and the old slave is now a master\n # (in case there was only one slave). try master instead\n raise\n else:\n return pickle_loads(res) if res is not None else None\n\n def _get_from_master(self, key, **kwargs):\n master = self._get_master()\n try:\n res = master.get(key, **kwargs)\n except (TimeoutError, MasterNotFoundError):\n if self._DEBUG:\n d('cache: master timed out also. no read instances available. returning None')\n # uh oh, no master available either. either all redis instances have hit the bucket,\n # or there is a fail-over in process, and a new master will be in line in a moment\n return None\n return pickle_loads(res) if res is not None else None\n\n def get_master(self):\n \"\"\"\n Expose the master node to permit any operation in redis-py\n \"\"\"\n return self._get_master()\n\n def _slave_chosen(self):\n return random_choice(range(0, self._node_count)) != 0\n\n def _get_master(self):\n if self._DEBUG:\n d('cache: getting master')\n return self._sentinel.master_for(self._service_name, socket_timeout=0.1)\n\n def _get_slave(self):\n if self._DEBUG:\n d('cache: getting slave')\n return self._sentinel.slave_for(self._service_name, socket_timeout=0.1)\n\n def _node_count(self):\n return len(self._sentinel.discover_slaves(self._service_name)) + 1 # +1 is master\n\n\nclass _RedisSentinelCacheDummy():\n\n \"\"\"\n A dummy redis client that writes to a file on disk.\n \"\"\"\n\n _storage_path = '/tmp/redis_dummy_storage'\n\n def __init__(self, *args, **kwargs):\n # d('Note: using dummy cache')\n pass\n\n def set(self, key, value, **kwargs):\n storage = self._get_storage()\n storage[key] = value\n self._save_storage(storage)\n\n def get(self, key, **kwargs):\n return self._get_storage().get(key, None)\n\n def get_or_set(self, key, value, **kwargs):\n if self.get(key):\n return False\n else:\n self.set(key, value,)\n return True\n\n def delete(self, key, **kwargs):\n storage = self._get_storage()\n storage.pop(key, False)\n self._save_storage(storage)\n\n def get_master(self):\n return self\n\n def flushdb(self):\n self._save_storage({})\n return True\n\n def _get_storage(self):\n try:\n with open(self._storage_path, 'r') as f:\n return load_json(f)\n except Exception as e:\n _logger.error('Could not open dummy cache file for reading at %s: %s' % (self._storage_path, str(e)))\n return {}\n\n def _save_storage(self, storage):\n try:\n with open(self._storage_path, 'w') as f:\n dump_json(storage, f)\n except Exception as e:\n _logger.error('Could not open dummy cache file for writing at %s: %s' % (self._storage_path, str(e)))\n","sub_path":"src/metax_api/utils/redis.py","file_name":"redis.py","file_ext":"py","file_size_in_byte":8742,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"572082797","text":"import json\nimport random\n\nimport requests\nfrom django.db import transaction, OperationalError\nfrom rest_framework.response import Response\nfrom rest_framework import status\nfrom rest_framework.decorators import action\nfrom rest_framework import viewsets\nfrom rest_framework.permissions import AllowAny, IsAuthenticated\nimport datetime\nimport time\n# Create your views here.\nfrom rest_framework.views import APIView\nfrom rest_framework.authtoken.models import Token\nfrom accounts.models import PhoneConfirm, User\nfrom accounts.serializers import SMSSignupPhoneCheckSerializer, SMSSignupPhoneConfirmSerializer\nfrom core.sms.utils import SMSV2Manager, MMSV1Manager\nfrom core.tools import get_client_ip\nfrom django.conf import settings\nfrom logic.models import DateTimeLotteryResult\nfrom logs.models import MMSSendLog\nfrom projects.models import Project\nfrom products.models import Product, Reward\nfrom respondent.models import RespondentPhoneConfirm, Respondent\nfrom respondent.serializers import SMSRespondentPhoneCheckSerializer, RespondentCreateSerializer, SMSRespondentPhoneConfirmSerializer\n\n\nclass SMSViewSet(viewsets.GenericViewSet):\n \"\"\"\n sms 전송시 공통으로 사용하는 viewset\n \"\"\"\n permission_classes = [AllowAny]\n\n def get_serializer_class(self):\n if self.action == 'send':\n serializer = SMSSignupPhoneCheckSerializer\n elif self.action == 'confirm':\n serializer = SMSSignupPhoneConfirmSerializer\n elif self.action == 'respondent_send':\n serializer = SMSRespondentPhoneCheckSerializer\n elif self.action == 'respondent_confirm':\n serializer = SMSRespondentPhoneConfirmSerializer\n else:\n serializer = super(SMSViewSet, self).get_serializer_class()\n return serializer\n\n @action(methods=['post'], detail=False)\n def send(self, request, *args, **kwargs):\n \"\"\"\n 회원가입시 인증번호 발송하는 api입니다.\n api: api/v1/sms/send\n method: POST\n data: {'phone'}\n \"\"\"\n data = request.data\n serializer = self.get_serializer(data=data)\n if serializer.is_valid(raise_exception=True):\n phone = serializer.validated_data['phone']\n sms_manager = SMSV2Manager()\n sms_manager.set_content()\n sms_manager.create_instance(phone=phone, kinds=PhoneConfirm.SIGN_UP)\n\n if not sms_manager.send_sms(phone=phone):\n return Response(\"Failed send sms\", status=status.HTTP_410_GONE)\n\n return Response(status=status.HTTP_200_OK)\n\n return Response(status=status.HTTP_400_BAD_REQUEST)\n\n @action(methods=['post'], detail=False)\n def confirm(self, request, *args, **kwargs):\n \"\"\"\n 회원가입시 인증번호 인증 api입니다. 인증시 다음페이지(비밀번호설정)에서 사용할 phone을 리턴합니다.\n api: api/v1/sms/confirm\n method: POST\n data: {'phone', 'confirm_key'}\n \"\"\"\n data = request.data\n serializer = self.get_serializer(data=data)\n if not serializer.is_valid(raise_exception=True):\n return Response(status=status.HTTP_400_BAD_REQUEST)\n\n return Response({'phone': serializer.validated_data['phone']}, status=status.HTTP_200_OK)\n\n @action(methods=['post'], detail=False)\n def respondent_send(self, request, *args, **kwargs):\n \"\"\"\n 설문자 인증번호 발송시 사용하는 핸드폰인증입니다.\n api: api/v1/sms/respondent_send/\n method: POST\n data: {'phone'}\n \"\"\"\n data = request.data\n serializer = self.get_serializer(data=data)\n if serializer.is_valid(raise_exception=True):\n phone = serializer.validated_data['phone']\n sms_manager = SMSV2Manager()\n sms_manager.set_respondent_content()\n sms_manager.create_respondent_send_instance(phone=phone)\n\n if not sms_manager.send_sms(phone=phone):\n return Response(\"Failed send sms\", status=status.HTTP_410_GONE)\n\n return Response(status=status.HTTP_200_OK)\n\n return Response(status=status.HTTP_400_BAD_REQUEST)\n\n @transaction.atomic\n @action(methods=['post'], detail=False)\n def respondent_confirm(self, request, *args, **kwargs):\n \"\"\"\n 설문자 인증번호 인증 api입니다. 인증시 서버에서 5-10초후 reward MMS를 발송합니다.\n api: api/v1/sms/respondent_confirm\n method: POST\n 전화번호, 인증번호 와 url에서 파싱한 project_key와 validator를 담아서 보내주어야 합니다.\n data: {'phone', 'confirm_key', 'project_key', 'validator'}\n \"\"\"\n data = request.data\n serializer = self.get_serializer(data=data)\n if not serializer.is_valid(raise_exception=True):\n return Response(status=status.HTTP_400_BAD_REQUEST)\n if not Project.objects.filter(project_hash_key=data.get('project_key')).exists():\n return Response(status=status.HTTP_404_NOT_FOUND)\n self.data = serializer.validated_data\n self._set_project()\n self._create_respondent()\n\n # 여기까지가 유저 당첨확인 및 생성\n item_name = ''\n if self.is_win:\n self._set_random_reward()\n\n phone = self.data.get('phone')\n brand = self.reward.product.item.brand.name\n item_name = self.reward.product.item.name\n item_url = self.reward.reward_img.url\n due_date = self.reward.due_date\n\n if type(item_url) is tuple:\n item_url = ''.join(item_url)\n\n if type(item_name) is tuple:\n item_name = ''.join(item_name)\n\n mms_manager = MMSV1Manager()\n mms_manager.set_content(brand, item_name, due_date)\n success, code = mms_manager.send_mms(phone=phone, image_url=item_url)\n if not success: # TODO : status 로 클라에서 재전송버튼 활성화? 아니면 슬랙알림만?\n MMSSendLog.objects.create(code=code, phone=phone, item_name=item_name, item_url=item_url,\n due_date=due_date, brand=brand)\n\n # TODO: 당첨자 안나온 상품 있으면 한번에 보내기\n self.reward.winner_id = self.respondent.id\n self.reward.save()\n item_name = self.reward.product.item.short_name\n\n return Response({'id': self.project.id,\n 'is_win': self.is_win,\n 'item_name': item_name}, status=status.HTTP_200_OK)\n\n def _set_random_reward(self): # TODO: 에러날경우 패스 혹은 문의하기로\n reward_queryset = Reward.objects.filter(winner_id__isnull=True) \\\n .select_related('product', 'product__item', 'product__project', 'product__item__brand')\n remain_rewards = reward_queryset.filter(product__project=self.project)\n remain_rewards_id = list(remain_rewards.values_list('id', flat=True))\n remain_rewards_price = list(remain_rewards.values_list('product__item__price', flat=True))\n reward_weight = list(map(lambda x: round(1 / x * (sum(remain_rewards_price) / len(remain_rewards_price)))\n , remain_rewards_price))\n random_reward_id_by_weight = random.choices(remain_rewards_id, weights=reward_weight)[0]\n self.reward = reward_queryset.get(id=random_reward_id_by_weight)\n\n def _set_project(self):\n project_queryset = Project.objects.filter(project_hash_key=self.data.get('project_key'))\\\n .prefetch_related('respondents', 'respondents__phone_confirm', 'products', 'products__rewards') # 2021.07.07 [d-o-d.io 리뉴얼 ]추가 ####\n\n self.project = project_queryset.get(project_hash_key=self.data.get('project_key'))\n self.phone_confirm = RespondentPhoneConfirm.objects.filter(phone=self.data.get('phone'),\n confirm_key=self.data.get('confirm_key'),\n is_confirmed=True).first()\n\n def _create_respondent(self):\n self.is_win = self._am_i_winner()\n data = {'project': self.project.id,\n 'phone_confirm': self.phone_confirm.id,\n 'is_win': self.is_win}\n\n serializer = RespondentCreateSerializer(data=data)\n serializer.is_valid(raise_exception=True)\n self.respondent = serializer.save()\n\n def _am_i_winner(self):\n # 프로젝트 생성자는 무조건 꽝!\n if self.phone_confirm.phone == self.project.owner.phone:\n return False\n # 2021.07.07 [d-o-d.io 리뉴얼 ]추가 ####\n if self.project.products.filter(rewards__isnull=True).exists():\n return False\n # 2021.07.07 [d-o-d.io 리뉴얼 ]추가 ####\n self.lucky_times = self.project.select_logics.last().lottery_times.filter(is_used=False)\n now = datetime.datetime.now()\n self.valid_lucky_times = self.lucky_times.filter(lucky_time__lte=now)\n if not self.valid_lucky_times.exists():\n # 당첨 안된 경우\n return False\n else:\n try:\n with transaction.atomic(using='default'):\n vlt = DateTimeLotteryResult.objects.select_for_update(nowait=True)\\\n .filter(logic__project=self.project).filter(is_used=False, lucky_time__lte=now).first()\n vlt.is_used = True\n vlt.save()\n val = True\n except OperationalError:\n val = False\n except DateTimeLotteryResult.DoesNotExist:\n val = False\n return val\n\n\nclass SendMMSAPIView(APIView):\n \"\"\"\n 당첨자 3초 후 문자전송을 위해 만듬 (20210622)\n\n [DEPRECATED] 발송실패가 많음. (20210625)\n \"\"\"\n permission_classes = [AllowAny]\n\n def post(self, request, *args, **kwargs):\n\n ip = get_client_ip(request)\n if ip not in settings.ALLOWED_HOSTS:\n return Response(status=status.HTTP_400_BAD_REQUEST)\n\n data = request.data\n phone = data.get('phone')\n brand = data.get('brand')\n item_name = data.get('item_name')\n item_url = data.get('item_url')\n due_date = data.get('due_date')\n time.sleep(3) # wait 3 seconds\n mms_manager = MMSV1Manager()\n mms_manager.set_content(brand, item_name, due_date)\n success, code = mms_manager.send_mms(phone=phone, image_url=item_url)\n if not success:\n MMSSendLog.objects.create(code=code, phone=phone, item_name=item_name, item_url=item_url, due_date=due_date)\n # MMSSendLog.objects.create(code=code, phone=phone, item_name=item_name, item_url=item_url, due_date=due_date)\n return Response(status=status.HTTP_200_OK)\n\n","sub_path":"DOD-main-server/dod/core/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":10932,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"185543488","text":"\"\"\"\nPerform file organizing and metadata tasks for GeoTIFF files from MicaSense RedEdge.\nBy: Emily Sturdivant, esturdivant@usgs.gov, https://github.com/esturdivant-usgs\nRequires Python 3.2+\n\nThe user inputs the following:\n- input image directory (imgdir_in): directory where the raw files are stored and sorted by flight number\n- output image directory (imgdir_out): directory to be created and populated with the processed files\n- minimum and maximum altitude for files to process\n- mission info (survey ID, camera ID, values for metadata)\n\nNote:\n- We expect flight folder names to start with 'f' and end with a number. The number in the folder name will be used as the flight ID.\n\n\nThis script does the following:\n 1. For each flight folder copy files to a flight folder in the out dir\n 1.a. Get flight IDs (integer) from the folder names in 'imgdir_in'.\n 1.b. Copy TIF files in each flight folder to new flight folder in 'imgdir_in' using the naming scheme fXX, ignoring intermediate folders (e.g. 0000SET, 000)\n 2. Change photo metadata: rename and assign metadata values\n 2.a. Rename files with survey ID, flight ID, sensor ID, datetime, and original filename.\n 2.b. Add flight ID to UserComment tag in file header.\n 3. Add standard metadata values.\n 3.a. Add the standard WHCMSC Exif tags, with specific values input by the user.\n 3.b. Copy the CreateDate value to the GPSDate and GPSTime tags.\n 4. Copy files within the input altitude range to a separate directory tree.\n\"\"\"\n#%%\nimport pandas as pd\nimport os\nimport subprocess\nimport shutil\nimport glob\n\n#%% Set source dir\n# Only used flights 11-32, i.e. photos collected on 8/7 and 8/8. RedEdge surveying on 8/6 (flights 2–10) had many gaps.\nimgdir_in = r\"/Volumes/stor/Projects/2019009FA_PlumIsland/RAW/photos/mica_images\"\nimgdir_out = r\"/Volumes/stor/Projects/2019009FA_PlumIsland/PROCESSED/final_photos/micasense\"\n\n# If calibration files have already been separated out, they will need to be processed too.\ncal_separated = True\n\n# Altitude filtering values\nalt_min = 73\nalt_max = 88\n\n# Mission info\nfan = '2019-009-FA' # input('Field activity number (e.g. \"2017-010-FA\"): ')\ncam_id = 'm01'\n# flight_id = 'f06'\n# doi =\n\n# WHSC EXIF population\nsurveyid = fan.replace(\"-\",\"\")\nsite = \"Plum Island, MA\"\ncredit = \"U.S. Geological Survey\"\ncomment = \"\"\"One band of a multispectral image from a low-altitude aerial survey in {1}. Captured with a MicaSense RedEdge at a target altitude of 80 m. WHCMSC field activity number {0} (https://cmgds.marine.usgs.gov/fan_info.php?fa={0}).\"\"\".format(fan, site) # used for caption, ImageDescription, Caption-Abstract\nkeywords = \"{}; Massachusetts; {}; UAS; nadir; multispectral; USGS\".format(site, fan)\nartist = \"WHCMSC AIM Group\"\ncontact = \"WHSC_data_contact@usgs.gov\"\n\n#%% Functions\ndef write_WHSC_exiftags(imgdir, credit, comment, keywords, artist, contact, run=True, recursive=True):\n # Tags that will be identical for all images in the folder\n tagvalues = {}\n tagvalues['imgdir'] = imgdir\n tagvalues['credit'] = credit\n tagvalues['artist'] = artist\n tagvalues['contact'] = contact\n tagvalues['comment'] = comment\n tagvalues['keywords'] = keywords\n tagvalues['copyright'] = \"Public Domain. Please credit {credit}\".format(**tagvalues)\n # Write to EXIF\n substance = \"\"\"-Artist=\"{artist} \" -Credit=\"{credit} \" -Contact=\"{contact} \" -comment=\"{comment} \" -sep \"; \" -keywords=\"{keywords} \" -Caption=\"{comment} \" -Copyright=\"{copyright} \" -CopyrightNotice=\"{copyright} \" -Caption-Abstract=\"{comment} \" -ImageDescription=\"{comment} \" \"\"\".format(**tagvalues)\n tagvalues['substance'] = substance\n if run:\n if recursive:\n cmd = \"\"\"exiftool {substance} -overwrite_original -r {imgdir}\"\"\".format(**tagvalues)\n else:\n cmd = \"\"\"exiftool {substance} -overwrite_original {imgdir}\"\"\".format(**tagvalues)\n subprocess.check_call(cmd, shell=True)\n print(\"Updated Exif headers in directory: {}\".format(imgdir))\n return\n else:\n return(substance)\n\ndef filter_by_exif(imgdir, keep_dir, min, max, tag='GPSAltitude'):\n # Filter image files by indicated EXIF tag. Copy files that match the criteria to a keep folder.\n # Print altitudes to CSV\n exif_csv = imgdir+\"_{}.csv\".format(tag)\n cmd1 = \"\"\"exiftool -csv -{} -n -r {} > {}\"\"\".format(tag, imgdir, os.path.join(imgdir, exif_csv))\n subprocess.check_call(cmd1, shell=True)\n # Move photos with values between min and max (inclusive) to keep folder\n # Convert CSV To DF\n df = pd.read_csv(exif_csv)\n # for each entry, if altitude is between alt_min and alt_max (inclusive), move file to the keep folder\n keepct = 0 # keep counter\n rct = 0 # remove counter\n for index, row in df.iterrows():\n fpath = row.SourceFile\n fname = os.path.basename(row.SourceFile)\n alt = row.loc[tag]\n if alt < alt_min or alt > alt_max:\n rct += 1\n else:\n try:\n shutil.copy2(fpath, os.path.join(keep_dir, fname))\n keepct += 1\n except:\n pass\n # Print counts\n print(\"Files matching criteria: {} out of {}\".format(keepct, keepct+rct))\n # Check altitudes - print to CSV and DF\n exif_csv = keep_dir+\"_keepers.csv\"\n cmd1 = \"\"\"exiftool -csv -{} -n {} > {}\"\"\".format(tag, keep_dir, exif_csv)\n subprocess.check_call(cmd1, shell=True)\n df = pd.read_csv(exif_csv)\n #\n return(df)\n\n#%% For each flight folder: get flight number, copy files to out dir, rename files\n# Copy files to new flight folders, ignoring intermediate folders\nprint(\"\\n***** Creating new folder structure and copying files *****\".format(fnum, flight_id))\nfdirs = glob.glob(os.path.join(imgdir_in, 'f*[0-9]'))\nfor fltdir in fdirs:\n # Convert folder name to flight ID\n fnum = int(''.join(filter(str.isdigit, os.path.basename(fltdir))))\n flight_id = 'f{0:02d}'.format(fnum)\n # Create folder for renamed photos\n print(\"Creating folder and copying files for flight {} ({})\".format(fnum, flight_id))\n fltdir_out = os.path.join(imgdir_out, flight_id)\n try: shutil.rmtree(fltdir_out)\n except FileNotFoundError: pass\n os.makedirs(fltdir_out, exist_ok=True)\n # List TIF files and copy to processing directory, ignoring intermediate folders\n imglist = glob.glob(os.path.join(fltdir, '**/*.tif'), recursive=True)\n for img in imglist:\n shutil.copy2(img, os.path.join(fltdir_out, os.path.basename(img)))\n # Check that there are the expected number of files in the directory.\n if len(imglist) < len(os.listdir(fltdir_out)):\n print(\"WARNING: There are extra files in the flight directory.\")\n elif len(imglist) > len(os.listdir(fltdir_out)):\n print(\"WARNING: There are too few files in flight directory.\")\n\n#%% Change photo metadata: Rename, Assign Exif tag values,\nprint(\"\\n***** Running ExifTool for each flight folder to rename files and add flight number. *****\".format(fnum, flight_id))\nfor fltdir_out in glob.glob(os.path.join(imgdir_out, 'f*[0-9]')):\n flight_id = os.path.basename(fltdir_out)\n num_renamed = len(glob.glob(os.path.join(fltdir_out, '{}*'.format(surveyid))))\n if num_renamed > 0:\n print(\"ALERT: There are already {} files in the {} folder that may have been renamed. We'll skip this folder.\".format(num_renamed, flight_id))\n continue\n fnum = int(''.join(filter(str.isdigit, flight_id)))\n # Rename and save flight ID in UserComment, uses ExifTool\n cmd1 = \"\"\"exiftool -d {}_{}{}_%Y%m%dT%H%M%SZ_%%f.%%e \"-filename {}\"\"\".format(imgdir_out, exif_csv)\n# subprocess.check_call(cmd1, shell=True)\n# df = pd.read_csv(exif_csv)\n# df\n\n#%% Same process, but all in one for-loop (older version):\n# #%% For each flight folder: get flight number, copy files to out dir, rename files\n# # List folders\n# # In each folder, rename files\n# fdirs = glob.glob(os.path.join(imgdir_in, 'f*'), recursive=True)\n# for fltdir in fdirs:\n# # Convert folder name to flight ID\n# fnum = int(''.join(filter(str.isdigit, os.path.basename(fltdir))))\n# flight_id = 'f{0:02d}'.format(fnum)\n# print(\"\\n***** Processing files from flight {} ({}) *****\".format(fnum, flight_id))\n# # Create folder for renamed photos\n# print(\"...Creating new flight folder...\"\n# fltdir_out = os.path.join(imgdir_out, flight_id)\n# try: shutil.rmtree(fltdir_out)\n# except FileNotFoundError: pass\n# os.makedirs(fltdir_out, exist_ok=True)\n# # List TIF files and copy to processing directory, ignoring intermediate folders\n# imglist = glob.glob(os.path.join(fltdir, '**/*.tif'), recursive=True)\n# for img in imglist:\n# shutil.copy2(img, os.path.join(fltdir_out, os.path.basename(img)))\n# # Check that there are the expected number of files in the directory.\n# if len(imglist) < len(os.listdir(fltdir_out)):\n# print(\"WARNING: There are extra files in the flight directory.\")\n# elif len(imglist) > len(os.listdir(fltdir_out)):\n# print(\"WARNING: There are too few files in flight directory.\")\n# # Rename and save flight ID in UserComment, uses ExifTool\n# cmd1 = \"\"\"exiftool -d {}_{}{}_%Y%m%dT%H%M%SZ_%%f.%%e \"-filename len(os.listdir(fltdir_out)):\n# print(\"WARNING: There are too few files in flight directory.\")\n# #---- But not below here:\n# # Execute altitude filter\n# print(\"Copying files matching altitude criteria into keep folder...\")\n# keep_dir = os.path.join(imgdir_out, 'keep_alt{}to{}'.format(alt_min, alt_max), flight_id)\n# os.makedirs(keep_dir, exist_ok=True) # Create keep_dir tree if it doesn't already exist.\n# # Copy files within altitude bounds to keep folder\n# keeper_values = filter_by_exif(fltdir_out, keep_dir, alt_min, alt_max, tag='GPSAltitude')\n# print(\"Number of files in final directory: {}\".format(len(os.listdir(keep_dir))))\n","sub_path":"image_RenameTagFilter.py","file_name":"image_RenameTagFilter.py","file_ext":"py","file_size_in_byte":13452,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"499835514","text":"end = True\r\nwhile end:\r\n print('Do you want to type the number(yes/no):')\r\n c = input()\r\n if c == 'yes':\r\n print('Please, type the number:')\r\n a = int(input())\r\n if a < 0:\r\n print('Negative number')\r\n elif a == 0:\r\n print('number is zero')\r\n elif a % 2 == 0:\r\n print('Even number')\r\n elif a % 2 == 1:\r\n print('odd number')\r\n elif c == 'no':\r\n print('OK,see you later')\r\n end = False\r\n else:\r\n print('Please, type \"yes\" or \"no\"')","sub_path":"home_1_2.py","file_name":"home_1_2.py","file_ext":"py","file_size_in_byte":550,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"70681900","text":"import ROOT\nimport argparse\nimport ConfigParser\nimport json\n\n# Get config file name\nparser = argparse.ArgumentParser(description='Specify config and overwrite values if desired.')\nparser.add_argument(\"--config\",help = \"Specify configuration file to use\")\nparser.add_argument(\"--tag\",help=\"Tag to add to all directory names to separate this production\")\nargs = parser.parse_args()\n\n# Read configuration\ndefaults = { \"widths\": \"[0.07,0.10,0.15]\",\n \"startPoints\": 200,\n \"stopDensePoints\": 2000,\n \"stopMediumPoints\": 3000,\n \"stopPoints\" : 7500,\n \"binnings\": \"[0]\"}\n\nconfigReader = ConfigParser.RawConfigParser(defaults)\nconfigReader.read(args.config)\n \nsection = \"IO\"\ninputBkgFile = configReader.get(section, \"inputFileName\")\ninputBkgHistName = configReader.get(section,\"binTemplateHist\")\noutputFileFormat = configReader.get(section, \"outputFileFormat\")\n \nsection = \"MassesAndWidths\"\nwidths = json.loads(configReader.get(section,\"widths\"))\nnWidths = len(widths)\nstartPoints = configReader.getint(section,\"startPoints\")\nstopDensePoints = configReader.getint(section,\"stopDensePoints\")\nstopMediumPoints = configReader.getint(section,\"stopMediumPoints\")\nstopPoints = configReader.getint(section,\"stopPoints\")\n\nsection = \"Binnings\"\nbinnings = json.loads(configReader.get(section,\"binnings\"))\nnBinnings = len(binnings)\n\n# Get histogram to use as template\ninfile = ROOT.TFile(inputBkgFile,\"READ\")\nnominalBkg = infile.Get(inputBkgHistName)\nnominalBkg.SetDirectory(0)\ninfile.Close()\n\n# Generate list of masses from given info\nmasses = []\nfor mass in range(startPoints, stopDensePoints, 50) : masses.append(float(mass))\nfor mass in range(stopDensePoints, stopMediumPoints, 100) : masses.append(float(mass))\nfor mass in range(stopMediumPoints, stopPoints, 200) : masses.append(float(mass))\n\n# Generate signals\nfor mass in masses :\n for width in widths :\n \n widthForName = int(width*100)\n onesigma = mass*width\n filename = outputFileFormat.format(int(mass),widthForName)\n\n # Generate BW\n GenericBW = ROOT.TF1(\"signal\",\"TMath::BreitWigner(x,{0},{1})\".format(mass,onesigma),0,13000)\n outfile = ROOT.TFile(filename,\"RECREATE\")\n\n for binning in binnings :\n\n # Convert to histogram\n \n low = nominalBkg.GetBinLowEdge(1)\n high = nominalBkg.GetBinLowEdge(nominalBkg.GetNbinsX()) + nominalBkg.GetBinWidth(nominalBkg.GetNbinsX())\n\n if binning==0 :\n SignalHist = ROOT.TH1D(nominalBkg)\n SignalHist.SetName(\"nominalBins\")\n SignalHist.SetTitle(\"nominalBins\")\n elif binning < 0 and binning > -1.5 :\n newBins = []\n for j in range(0, nominalBkg.GetNbinsX()+1) :\n newBins.append(nominalBkg.GetBinLowEdge(j))\n newBins.append(nominalBkg.GetBinLowEdge(j)+nominalBkg.GetBinWidth(j)/2.0)\n newBins.append(nominalBkg.GetBinLowEdge(nominalBkg.GetNbinsX())+nominalBkg.GetBinWidth(nominalBkg.GetNbinsX()))\n SignalHist = ROOT.TH1D(\"halfNominalBins\",\"halfNominalBins\",len(newBins),array(newBins))\n elif binning < -1 and binning > -99 :\n SignalHist = ROOT.TH1D(nominalBkg)\n SignalHist.Rebin(int(-1*binning))\n SignalHist.SetName(\"{0}TimesNominalBins\".format(int(-1*binning)))\n SignalHist.SetTitle(\"{0}TimesNominalBins\".format(int(-1*binning)))\n elif (binning < -99) :\n newBins = []\n for j in range(0, nominalBkg.GetNbinsX()) :\n newBins.append(nominalBkg.GetBinLowEdge(j+1)+50)\n newBins.append(nominalBkg.GetBinLowEdge(nominalBkg.GetNbinsX())+nominalBkg.GetBinWidth(nominalBkg.GetNbinsX())+50)\n SignalHist = ROOT.TH1D(\"shifted50NominalBins\",\"shifted50NominalBins\",nominalBkg.GetNbinsX(),array(newBins))\n else :\n nBins = int((high - low)/ float(binning))\n SignalHist = ROOT.TH1D(\"binsCloseTo{0}\".format(binning),\"binsCloseTo{0}\".format(binning),nBins,low,high)\n\n # Make sure histograms are clear and free, regardless\n SignalHist.Reset()\n SignalHist.SetDirectory(0)\n\n for bin in range(0, SignalHist.GetNbinsX()+2) :\n a = SignalHist.GetBinLowEdge(bin)\n b = SignalHist.GetBinLowEdge(bin)+SignalHist.GetBinWidth(bin)\n content = GenericBW.Integral(a,b)\n SignalHist.SetBinContent(bin,content)\n SignalHist.SetBinError(bin,0.)\n\n intNow = SignalHist.Integral()\n\n thisPercentage = 0\n thisInterval = 0\n remember1 = 1\n remember2 = 1\n smallestInterval = 1E6\n rememberThisPercentage = 0\n for bin1 in range(0,SignalHist.GetNbinsX()+2) :\n for bin2 in range(bin1,SignalHist.GetNbinsX()+2) :\n thisPercentage = (SignalHist.Integral(bin1,bin2))/intNow;\n thisInterval = SignalHist.GetBinLowEdge(bin2) + SignalHist.GetBinWidth(bin2) - SignalHist.GetBinLowEdge(bin1)\n if (thisPercentage >= 0.95) :\n if (thisInterval < smallestInterval) :\n remember1=bin1\n remember2=bin2\n smallestInterval=thisInterval\n rememberThisPercentage=thisPercentage\n break\n \n for bin in range(0, SignalHist.GetNbinsX()+2) :\n if (bin < remember1) or (bin > remember2) :\n SignalHist.SetBinContent(bin,0.)\n SignalHist.SetBinError(bin,0.)\n intNow = SignalHist.Integral()\n SignalHist.Scale(1.0/intNow)\n\n outfile.cd()\n SignalHist.Write()\n\n outfile.Close()\n\n","sub_path":"scripts/createBWSignals.py","file_name":"createBWSignals.py","file_ext":"py","file_size_in_byte":5407,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"178672817","text":"from unittest.mock import ANY\n\nfrom freezegun import freeze_time\nfrom rest_framework import serializers, status\nfrom rest_framework.reverse import reverse\n\nfrom ..factories.analytics import AnalyticsDataFactory, AnalyticsFactory\nfrom ..models.analytics import AnalyticsData\n\n\ndef test_analytics_data_list(api_client, django_assert_max_num_queries):\n AnalyticsDataFactory.create_batch(5)\n with django_assert_max_num_queries(35):\n r = api_client.get(reverse('analyticsdata-list'))\n assert r.status_code == status.HTTP_200_OK\n assert len(r.data) == 4\n\n\ndef test_analytics_data_filter_by_analytics(api_client, django_assert_max_num_queries):\n AnalyticsDataFactory.create_batch(3)\n analytics = AnalyticsFactory()\n AnalyticsDataFactory.create_batch(2, analytics=analytics)\n with django_assert_max_num_queries(10):\n r = api_client.get(reverse('analyticsdata-list') + f'?analytics={analytics.pk}')\n assert r.status_code == status.HTTP_200_OK\n assert len(r.data) == 4\n\n\ndef test_analytics_data_filter_by_invalid_date(api_client, django_assert_max_num_queries):\n analytics = AnalyticsFactory()\n AnalyticsDataFactory.create_batch(2, analytics=analytics)\n with django_assert_max_num_queries(10):\n r = api_client.get(reverse('analyticsdata-list') + '?from=2021/07/03')\n assert r.status_code == status.HTTP_400_BAD_REQUEST\n assert 'Invalid date format, date must be in \"YYYY-MM-DD' in r.data['detail']\n\n\ndef test_analytics_data_post(api_client, staff_user):\n api_client.force_authenticate(staff_user)\n analytics = AnalyticsFactory()\n AnalyticsDataFactory.create()\n\n with freeze_time() as frozen_time:\n r = api_client.post(reverse('analyticsdata-list'), data={\n 'analytics': analytics.pk,\n 'date': '2021-07-13T20:00:10Z',\n 'value': 300\n }, format='json')\n\n assert r.status_code == status.HTTP_201_CREATED\n assert AnalyticsData.objects.get(pk=r.data['pk']).value == 300\n assert r.data == {\n 'pk': ANY,\n 'created_date': serializers.DateTimeField().to_representation(frozen_time()),\n 'modified_date': serializers.DateTimeField().to_representation(frozen_time()),\n 'date': '2021-07-13T20:00:10Z',\n 'value': 300,\n 'analytics': analytics.pk\n }\n\n\ndef test_analytics_data_patch(api_client, staff_user):\n api_client.force_authenticate(staff_user)\n analytics_data = AnalyticsDataFactory()\n\n with freeze_time():\n r = api_client.patch(\n reverse('analyticsdata-detail', (analytics_data.pk,)),\n data={\n 'value': 430,\n },\n )\n\n assert r.status_code == status.HTTP_200_OK\n assert AnalyticsData.objects.get(pk=str(analytics_data.pk)).value == 430\n\n\ndef test_analytics_data_delete(api_client, staff_user):\n api_client.force_authenticate(staff_user)\n data = AnalyticsDataFactory()\n r = api_client.delete(reverse('analyticsdata-detail', (data.pk,)))\n\n assert r.status_code == status.HTTP_204_NO_CONTENT\n assert r.data is None\n assert AnalyticsData.objects.filter(pk=str(data.pk)).first() is None\n\n\ndef test_analytics_data_anon_post(api_client):\n r = api_client.post(reverse('analyticsdata-list'), data={'value': 100}, format='json')\n\n assert r.status_code == status.HTTP_401_UNAUTHORIZED\n","sub_path":"v1/analytics/tests/analytics_data.py","file_name":"analytics_data.py","file_ext":"py","file_size_in_byte":3332,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"115913736","text":"import zipfile\n\nimport io\n\nfrom django.db.models import Q, F, Subquery, query\nfrom django.http import HttpResponse, HttpResponseRedirect, JsonResponse\nfrom django.views.decorators.csrf import csrf_exempt\nfrom django.contrib.auth.decorators import login_required\nfrom django.shortcuts import render\nfrom inventit.models import *\nimport json\nfrom decimal import *\nimport csv\nfrom django.conf import settings\nfrom django.urls import reverse\nfrom django.contrib import messages\n\nfrom django.contrib.auth import authenticate, login, logout\nfrom inventit.EmailBackEnd import EmailBackEnd\nfrom django.db import connection\nimport pandas as pd\n\n\nimport datetime as dt\nimport pandas as pd\nimport os\nfrom django.core.files.storage import FileSystemStorage\nfrom django.shortcuts import redirect\nfrom toggle.settings import allowed_users\n\nfrom django.db import transaction\nfrom django.core.serializers import serialize\nfrom django.core.serializers.json import DjangoJSONEncoder\nfrom django.db.models import F\n\n# Global Variables\nAuth_Temp = 'Auth_Templates/'\nMain_Pages = 'inventit/'\naddition = 'inventit/addition/'\n\n\ndef Signin(request):\n return render(request, Auth_Temp + \"Sign_In.html\")\n\ndef Signup(request):\n return render(request, Auth_Temp + \"sign-up.html\")\n\ndef doLogin(request):\n if request.method!=\"POST\":\n return HttpResponse(\"Method Not Allowed
\")\n else:\n user=EmailBackEnd.authenticate(request,username=request.POST.get(\"username\"),password=request.POST.get(\"password\"))\n if user!=None:\n login(request,user)\n\n if Profile.objects.filter(user=request.user).exists():\n if request.user.is_superuser:\n return HttpResponseRedirect('/summary')\n else:\n return HttpResponseRedirect('/capture1')\n elif request.user.is_superuser:\n return HttpResponseRedirect('/summary')\n else:\n return HttpResponseRedirect('/error_temp')\n else:\n messages.error(request,\"Invalid Login Details\")\n return HttpResponseRedirect(\"/\")\n\ndef signup_admin(request):\n username=request.POST.get(\"username\")\n email=request.POST.get(\"email\")\n password=request.POST.get(\"password\")\n\n try:\n if username in allowed_users:\n user=User.objects.create_user(username=username,password=password,email=email, is_staff=True,is_active=True,is_superuser=True)\n user.is_superuser=True\n user.save()\n else:\n user=User.objects.create_user(username=username,password=password,email=email)\n user.save()\n messages.success(request,\"Your Account Was Successfully Created, Please Login to proceed. Thank you!\")\n return redirect(Signin) \n except:\n messages.error(request,\"Failed to Create Admin\")\n return redirect(Signup) \n\ndef error_temp(request):\n return render(request,addition + \"error_login.html\")\n\n\n@login_required()\ndef capture1(request):\n if Profile.objects.filter(user=request.user).exists():\n\n template = Main_Pages+\"capture.html\"\n\n\n assets=Inventory.objects.all()\n # cursor = connection.cursor()\n # query = \"select * from inventit_inventory;\"\n # cursor.execute(query)\n # result = cursor.fetchall()\n # for i in result:\n # assets = i[2]\n profile = Profile.objects.get(user=request.user)\n\n # my allowed categories\n categories = Category.objects.filter(count_1=profile.team)\n bins = Category.objects.filter(count_1=profile.team)\n\n count_header = CountHeader.objects.filter(is_active=True)\n\n count_lines = CountLines.objects.filter(\n Q(count_header=count_header.first()) & Q(category__in=categories)\n )\n \n # count_lines1 = CountLines.objects.filter( Q(category=1) )\n\n # from django.db.models import Count, Max\n\n # unique_fields = ['inventory', 'category']\n\n # duplicates = (\n # CountLines.objects.values(*unique_fields)\n # .order_by()\n # .annotate(max_id=Max('id'), count_id=Count('id'))\n # .filter(count_id__gt=1)\n # )\n\n # for duplicate in duplicates:\n # (\n # CountLines.objects\n # .filter(**{x: duplicate[x] for x in unique_fields})\n # .exclude(id=duplicate['max_id'])\n # .delete()\n # )\n\n\n context = {\"categories\": categories,\"bins\": bins,\"assets\": assets, \"count_headers\": count_header, \"lines\": count_lines, \"count\": 1, \"team\": profile.team.name}\n\n return render(request, template, context)\n\n else:\n return HttpResponseRedirect('/error_counts')\n\n\n\n@login_required()\ndef capture2(request):\n if Profile.objects.filter(user=request.user).exists():\n\n template = Main_Pages+\"capture.html\"\n\n assets=Inventory.objects.all()\n profile = Profile.objects.get(user=request.user)\n\n # my allowed categories\n categories = Category.objects.filter(count_2=profile.team)\n bins = Category.objects.filter(count_2=profile.team)\n\n count_header = CountHeader.objects.filter(is_active=True)\n\n count_lines = CountLines.objects.filter(\n Q(count_header=count_header.first()) & Q(category__in=categories)\n )\n \n \n\n context = {\"bins\":bins,\"assets\":assets, \"categories\": categories,\"count_headers\": count_header, \"lines\": count_lines, \"count\": 2, \"team\": profile.team.name}\n\n return render(request, template, context)\n else:\n return HttpResponseRedirect('/error_counts')\n\n\n@login_required()\ndef capture3(request):\n if Profile.objects.filter(user=request.user).exists():\n\n\n\n template = Main_Pages+\"capture.html\"\n\n profile = Profile.objects.get(user=request.user)\n\n # my allowed categories\n categories = Category.objects.filter(count_3=profile.team)\n\n count_header = CountHeader.objects.filter(is_active=True)\n\n count_lines = CountLines.objects.filter(\n Q(count_header=count_header.first()) & Q(category__in=categories) & Q(category__count_3__exact=profile.team)\n )\n\n\n context = {\"categories\": categories,\"count_headers\": count_header, \"lines\": count_lines, \"count\": 3, \"team\": profile.team.name}\n\n return render(request, template, context)\n\n else:\n return HttpResponseRedirect('/error_counts') \n\n\n\n@login_required()\ndef summary(request):\n template = Main_Pages+\"summary.html\"\n profile = Profile.objects.all()\n\n counts_sign_off = Team.objects.all().values_list('name', 'count_1_sign_off', 'count_2_sign_off', 'count_3_sign_off')\n\n count_header = CountHeader.objects.filter(is_active=True)\n \n inventory = Inventory.objects.all() \\\n .annotate(count_variance=F('count_summary') - F('count_theoretical')) \\\n .annotate(cost_variance=F('count_variance') * F('cost'))\n \n \n\n \n cost_variance_sum = inventory.aggregate(Sum('cost_variance'))['cost_variance__sum']\n\n summary_count1 = (\n CountLines.objects.filter(count_header=count_header.first())\n .filter(count_1__gte=0)\n .count()\n )\n summary_count2 = (\n CountLines.objects.filter(count_header=count_header.first())\n .filter(count_2__gte=0)\n .count()\n )\n summary_count3 = (\n CountLines.objects.filter(count_header=count_header.first())\n .filter(count_3__gte=0)\n .count()\n )\n total = CountLines.objects.filter(count_header=count_header.first()).count()\n\n \n \n summary_count1Percentage = 0\n summary_count2Percentage = 0\n summary_count3Percentage = 0\n\n if summary_count1:\n summary_count1Percentage = round((float(summary_count1) / float(total)) * 100)\n\n if summary_count2:\n summary_count2Percentage = round((float(summary_count2) / float(total)) * 100)\n if summary_count3:\n summary_count3Percentage = round((float(summary_count3) / float(total)) * 100)\n\n counts = {\n \"summary_count1\": summary_count1,\n \"summary_count2\": summary_count2,\n \"summary_count3\": summary_count3,\n \"summary_count1Percentage\": summary_count1Percentage,\n \"summary_count2Percentage\": summary_count2Percentage,\n \"summary_count3Percentage\": summary_count3Percentage,\n \"total\": total,\n \"counts_sign_off\": counts_sign_off\n }\n\n \n #balanced item = 0\n count_lineswithzero = Inventory.objects.filter( Q(count_summary=F('count_theoretical')) & Q(count_summary=0) & Q(count_theoretical=0) ).count()\n\n totalwithzero = Inventory.objects.all().count()\n if totalwithzero:\n percwithzero = ( count_lineswithzero / totalwithzero ) * 100\n else:\n percwithzero = 0\n \n\n \n\n #balanced item != 0\n count_lineswithoutzero = Inventory.objects.filter( Q(count_summary=F('count_theoretical')) & ~Q(count_summary=0) & ~Q(count_theoretical=0) ).count()\n totalwithoutzero = Inventory.objects.all().count()\n if totalwithoutzero:\n percwithoutzero = ( count_lineswithoutzero / totalwithoutzero ) * 100\n else:\n percwithoutzero = 0\n\n #Confirmed\n query = \"select * from inventit_Inventory where Confimed=1 and (count_summary-count_theoretical) <> 0;\"\n conresults = Inventory.objects.raw(query)\n Confirmedcount = len(conresults)\n\n # Confirmedcount = Inventory.objects.filter( Q(Confimed__exact=1)).count()\n totalc = Inventory.objects.all().count()\n if totalc:\n percc = ( Confirmedcount / totalc ) * 100\n else:\n percc = 0\n\n\n #unconfirmed\n query = \"select * from inventit_Inventory where Confimed=0 and (count_summary-count_theoretical) <> 0;\"\n uncresults = Inventory.objects.raw(query)\n unconfirmedcount = len(uncresults)\n\n\n # unconfirmedcount = Inventory.objects.filter( Q(Confimed__exact=0) ).count()\n totalunc = Inventory.objects.all().count()\n if totalunc:\n percunc = ( unconfirmedcount / totalunc ) * 100\n else:\n percunc = 0\n\n\n cursor = connection.cursor()\n query = '''\\\n Select (count_summary-count_theoretical) as variance , SUM((count_summary-count_theoretical) * cost) \n FROM inventit_inventory\n WHERE (Confimed=0 and variance <> 0)\n '''\n cursor.execute(query)\n result = cursor.fetchall()\n for i in result:\n UnConfirmed_Variance = i[1]\n\n query = '''\\\n Select (count_summary-count_theoretical) as variance , SUM((count_summary-count_theoretical) * cost) \n FROM inventit_inventory\n WHERE (Confimed=1 and variance <> 0)\n '''\n cursor.execute(query)\n result = cursor.fetchall()\n for i in result:\n Confirmed_Variance = i[1]\n\n from django.core.paginator import Paginator\n\n\n\n\n ctg = Team.objects.all()\n\n context = {\n \"count_headers\": count_header, \"Teams\":ctg,\n \"counts\": counts, \"counts_sign_off\": counts_sign_off,\"profile\":profile,\n \"cost_variance_sum\": cost_variance_sum,\n \"county\":unconfirmedcount , \"percent\":percunc, \"Confirmedcount\":Confirmedcount, \"percunc\":percc,\n \"count_lineswithzero\":count_lineswithzero, \"percwithzero\":percwithzero,\n \"count_lineswithoutzero\":count_lineswithoutzero, \"percwithoutzero\":percwithoutzero,\"Confirmed_Variance\":Confirmed_Variance,\"UnConfirmed_Variance\":UnConfirmed_Variance\n }\n\n if request.method == 'POST':\n uname = request.session.get('uname')\n check = request.POST['check'] # post value either true or false\n titles = request.POST['checko']# post value i.e. name of title\n for title in Inventory.objects.filter(id=titles):\n # return HttpResponse(title.title)\n title.Confimed = check.title() # title() capitalizes first letter of string\n title.save()\n return HttpResponse('data is saved')\n\n return render(request, template, context)\n\n@login_required()\ndef save_count_summary(request):\n if request.method == \"POST\":\n pk = request.POST.get(\"pk\")\n value = request.POST.get(\"value\")\n\n inventory = Inventory.objects.get(pk=pk)\n inventory.count_summary = value\n inventory.save()\n\n response_data = {\"result\": \"Saved!!!\"}\n\n return HttpResponse(json.dumps(response_data), content_type=\"application/json\")\n else:\n return HttpResponse(\n json.dumps({\"nothing to see\": \"this isn't happening\"}),\n content_type=\"application/json\",\n )\n\n@login_required()\ndef save_data(request):\n if request.method == \"POST\":\n pk = request.POST.get(\"pk\")\n count = request.POST.get(\"count\")\n count_type = request.POST.get(\"count_type\")\n\n response_data = {}\n\n if count_type == \"1\":\n line = CountLines.objects.filter(pk=pk)\n\n if line.values()[0].get(\"count_2\") == Decimal(count):\n line.update(count_1=count, count_3=count)\n else:\n line.update(count_1=count)\n\n elif count_type == \"2\":\n line = CountLines.objects.filter(pk=pk)\n\n # import pdb; pdb.set_trace();\n if line.values()[0].get(\"count_1\") == Decimal(count):\n line.update(count_2=count, count_3=count)\n else:\n line.update(count_2=count)\n\n else:\n line = CountLines.objects.filter(pk=pk)\n line.update(count_3=count)\n\n # Update the Inventory count summary\n count_lines = CountLines.objects.filter(\n inventory__item_code=line.first().inventory.item_code\n )\n\n sum = 0\n for count_line in count_lines:\n if count_line.count_3:\n sum += count_line.count_3\n\n inventory = Inventory.objects.filter(\n item_code=line.first().inventory.item_code\n ).first()\n\n if inventory:\n inventory.count_summary = sum\n inventory.save()\n\n response_data[\"result\"] = \"Saved!!!\"\n\n return HttpResponse(json.dumps(response_data), content_type=\"application/json\")\n else:\n return HttpResponse(\n json.dumps({\"nothing to see\": \"this isn't happening\"}),\n content_type=\"application/json\",\n )\n\n\n\n@csrf_exempt\n@login_required()\ndef sign_off(request, count):\n profile = Profile.objects.get(user=request.user)\n count_header = CountHeader.objects.filter(is_active=True)\n team = profile.team\n assets=CountLines.objects.all()\n if count == 1:\n categories = Category.objects.filter(count_1=team)\n count_lines = CountLines.objects.filter(\n Q(count_header=count_header.first()) & Q(category__in=categories)\n )\n for b in count_lines:\n if b.count_1 is None:\n \n b.count_1 = 0\n b.save()\n \n elif b.count_1 == b.count_2:\n b.count_3 = b.count_1\n b.save()\n \n \n #count_lines.filter(count_1__isnull=True).update(count_1=0)\n team.count_1_sign_off = True\n elif count == 2:\n categories = Category.objects.filter(count_2=team)\n count_lines = CountLines.objects.filter(\n Q(count_header=count_header.first()) & Q(category__in=categories)\n )\n for b in count_lines:\n if b.count_2 is None:\n \n b.count_2 = 0\n b.save()\n \n elif b.count_2 == b.count_1:\n b.count_3 = b.count_1\n b.save()\n\n #count_lines.filter(count_2__isnull=True).update(count_2=0)\n team.count_2_sign_off = True\n\n else:\n categories = Category.objects.filter(count_3=team)\n count_lines = CountLines.objects.filter(\n Q(count_header=count_header.first()) & Q(category__in=categories)\n )\n count_lines.filter(count_3__isnull=True).update(count_3=0)\n team.count_3_sign_off = True\n\n team.save()\n\n return HttpResponse(status=200)\n\n@login_required()\ndef export(request):\n output = io.StringIO()\n writer = csv.writer(output, dialect=\"excel\")\n\n inventory = export_inventory()\n countlines = export_countlines()\n\n response = HttpResponse(content_type=\"application/zip\")\n response[\"Content-Disposition\"] = \"attachment; filename=csv.zip\"\n\n z = zipfile.ZipFile(response, \"w\")\n z.writestr(\"inventory.csv\", inventory.getvalue())\n z.writestr(\"countlines.csv\", countlines.getvalue())\n\n return response\n\n\ndef export_inventory():\n output = io.StringIO()\n writer = csv.writer(output)\n writer.writerow([\"Confimed\",\"Item Code\", \"Theoretical\", \"Count Summary\"])\n\n for inventory in Inventory.objects.all():\n writer.writerow(\n [inventory.Confimed,inventory.item_code, inventory.count_theoretical, inventory.count_summary]\n )\n\n return output\n\ndef export_countlines():\n output = io.StringIO()\n writer = csv.writer(output)\n writer.writerow([\"Item Code\", \"Count 1\", \"Count 2\", \"Count 3\", \"Theoretical\"])\n\n count_header = CountHeader.objects.filter(is_active=True)\n\n for countLine in CountLines.objects.filter(count_header=count_header.first()):\n writer.writerow(\n [\n countLine.inventory.item_code,\n countLine.count_1,\n countLine.count_2,\n countLine.count_3,\n countLine.inventory.count_theoretical,\n ]\n )\n\n return output\n\n\ndef add_countline_save1(request):\n is_ajax = request.headers.get('X-Requested-With') == 'XMLHttpRequest'\n if is_ajax:\n data = json.load(request)\n ctldata = data.get('payload')\n\n bin_id=ctldata[\"category\"]\n category=Category.objects.get(id=bin_id)\n inventory_id=ctldata[\"inventory\"]\n # header_id = request.POST.get(\"time_log_count_header\")\n query = \"select * from inventit_CountHeader where is_active=True;\"\n results = CountHeader.objects.raw(query)\n for result in results:\n header_id = CountHeader.objects.get(description = result).id\n headerie = CountHeader.objects.get(id=header_id)\n if not Inventory.objects.filter(item_code=inventory_id).exists():\n messages.error(request, 'Invenotry Item Does Not Exist!!')\n context= [{'messages':'Invenotry Item Does Not Exist!!'}]\n else:\n inventor=Inventory.objects.get(item_code=inventory_id).id\n inventory=Inventory.objects.get(id=inventor)\n if not CountLines.objects.filter(category=category,inventory=inventory).exists():\n count_model=CountLines(count_header=headerie,category=category,inventory=inventory)\n count_model.save()\n messages.success(request, 'CountLine was successfully added !')\n context= [{'messages':'CountLine was successfully added !'}]\n else:\n messages.error(request, 'CountLine already exist!!!')\n context= [{'messages':'CountLine already exist!!!'}]\n return JsonResponse({'context':context})\n\n\n\ndef time_save(request):\n if request.method == 'POST': \n time_log_team=request.POST.get(\"time_log_team\")\n time_log_count_num=request.POST.get(\"time_log_count_num\")\n time_log_count_header=request.POST.get(\"time_log_count_header\")\n time_log_time=request.POST.get(\"time_log_time\")\n time_log_bin=request.POST.get(\"time_log_bin\")\n nguva=Time_Log(time_log_team=time_log_team, time_log_count_num=time_log_count_num, time_log_count_header=time_log_count_header, time_log_time=time_log_time ,time_log_bin=time_log_bin)\n nguva.save()\n return redirect(Home) \n\n\ndef Home(request):\n return render(request,Main_Pages+\"capture.html\")\n\ndef upload(request):\n Category.objects.all().delete()\n CountHeader.objects.all().delete()\n CountLines.objects.all().delete()\n Inventory.objects.all().delete()\n print('Deleted all data in tables') \n try:\n if request.method == 'POST' and request.FILES['myfile']:\n \n myfile = request.FILES['myfile'] \n fs = FileSystemStorage()\n filename = fs.save(myfile.name, myfile)\n uploaded_file_url = fs.url(filename)\n excel_file = uploaded_file_url\n print(excel_file) \n \n #for Inventory\n empexceldata = pd.read_excel (\".\"+excel_file, sheet_name='Final Inventory')\n print (empexceldata)\n print(type(empexceldata))\n dbframe = empexceldata\n for dbframe in dbframe.itertuples():\n \n obj = Inventory.objects.create(Confimed=dbframe.Confimed,item_code=dbframe.item_code, count_theoretical=dbframe.count_theoretical,\n count_summary=dbframe.count_summary, cost=dbframe.cost, description=dbframe.description)\n \n print(type(obj))\n obj.save()\n\n #for bins\n mpexceldata = pd.read_excel (\".\"+excel_file, sheet_name='Final_Bins')\n print (mpexceldata)\n print(type(mpexceldata))\n dbfram = mpexceldata\n for dbfram in dbfram.itertuples():\n obje = Category.objects.create(name=dbfram.name)\n \n print(type(obje))\n obje.save()\n\n #for header\n mpexceldata = pd.read_excel (\".\"+excel_file, sheet_name='Final Count_header')\n print (mpexceldata)\n print(type(mpexceldata))\n dbfra = mpexceldata\n for dbfra in dbfra.itertuples():\n obji = CountHeader.objects.create(description=dbfra.description)\n \n print(type(obji))\n obji.save()\n \n #for countlines\n empexceldata = pd.read_excel (\".\"+excel_file, sheet_name='Final CountLines')\n print (empexceldata)\n print(type(empexceldata))\n dbfr = empexceldata\n \n for dbfr in dbfr.itertuples():\n count_heade = CountHeader.objects.get(description=dbfr.count_header).id\n count_header=CountHeader.objects.get(id=count_heade)\n categor = Category.objects.get(name=dbfr.category).id\n category=Category.objects.get(id=categor)\n inventor=Inventory.objects.get(item_code=dbfr.item_code).id\n inventory=Inventory.objects.get(id=inventor)\n print(count_header,category,inventory)\n objo=CountLines(count_header=count_header,category=category,inventory=inventory)\n objo.save()\n\n return redirect(Home) \n\n except Exception as identifier: \n print(identifier)\n \n return redirect(Home) \n\n\n\ndef error_counts(request):\n return render(request,addition + \"error_login.html\")\n\ndef finishie(request):\n return render(request,addition + \"finished.html\")\n\ndef error_network(request):\n return render(request,addition + \"error_network.html\")\n\n\n\nfrom django.core import serializers\n\ndef poptable(request):\n if request.method == 'POST':\n pk = request.POST.get(\"pk\")\n a = CountLines.objects.filter(inventory__item_code=pk)\n tmpJson = serializers.serialize(\"json\",a)\n tmpObj = json.loads(tmpJson)\n dataa = json.dumps(tmpObj)\n # return JsonResponse(dataa, safe=False)\n return render(request, Main_Pages+\"summary.html\", {\"dataa\": dataa,\"pk\":a})\n # return redirect(summary,{'Countlinees':a})\n\n\n\ndef confirm_summary(request):\n if request.method == \"POST\":\n check = request.POST['check'] # post value either true or false\n titles = request.POST['checko']# post value i.e. name of title\n for title in Inventory.objects.filter(id=titles):\n # return HttpResponse(title.title)\n title.Confimed = check.title() # title() capitalizes first letter of string\n title.save()\n response_data = {\"result\": \"Saved!!!\"}\n return HttpResponse(json.dumps(response_data), content_type=\"application/json\")\n else:\n return HttpResponse(\n json.dumps({\"nothing to see\": \"this isn't happening\"}),\n content_type=\"application/json\",\n )\n\n#==============================================================POP UP SUMMARY ==================================================================\ndef Binloc(request):\n # request.is_ajax() is deprecated since django 3.1\n is_ajax = request.headers.get('X-Requested-With') == 'XMLHttpRequest'\n\n if is_ajax:\n data = json.load(request)\n todo = data.get('payload')\n todos = list(CountLines.objects.filter(inventory__item_code = todo['task']).values())\n a = CountLines.objects.filter(inventory__item_code = todo['task'])\n # for i in a:\n context = [{\n 'category':i.category.name,\n 'count_1':i.count_1,\n 'count_2':i.count_2,\n 'count_3':i.count_3,\n } for i in a]\n # context = json.dumps(context,cls=DjangoJSONEncoder)\n return JsonResponse({'context':context})\n\n#=========================================================== SUMMARY lINES ITEMS =======================================================================\n\ndef SummaryTab(request):\n # request.is_ajax() is deprecated since django 3.1\n is_ajax = request.headers.get('X-Requested-With') == 'XMLHttpRequest'\n\n if is_ajax:\n data = json.load(request)\n tabbs = data.get('payload')\n \n list_of_p_dict2 = data_perfection(tabbs)\n\n\n return JsonResponse({'context':list_of_p_dict2})\n\n\n#=================================================PERFECT SEARCH ==============================================================================================\ndef SearchFilter(request):\n # request.is_ajax() is deprecated since django 3.1\n is_ajax = request.headers.get('X-Requested-With') == 'XMLHttpRequest'\n\n if is_ajax:\n data = json.load(request)\n tabbs = data.get('payload')\n\n\n search_data = tabbs['sdata'] \n\n invdrop = tabbs['invdropdata']\n\n if invdrop == 'Unconfirmed':\n confirm = 0\n elif invdrop == 'Confirmed':\n confirm = 1\n else:\n confirm = 1\n\n\n context = data_perfection(tabbs)\n\n if tabbs['invdropdata'] != \"Show All\":\n context = [data for data in context if data['Confimed']== confirm]\n else:\n context = context\n\n # Get all the teams fed\n a = Team.objects.all()\n # Creating empty list to add teams fed\n selected_teams = []\n for i in a:\n #Adding teams available \n selected_teams.append(i.name)\n\n # Filtered Team\n chosen_team = tabbs['teamdropdata']\n c_team = [d for d in selected_teams if d.lower() == chosen_team.lower()]\n #Finding the index of filtered team\n if len(c_team) > 0:\n list_of_alpha = Alpha_allocate(selected_teams,c_team)\n contex2 = []\n for i in range(len(list_of_alpha)):\n selected_letters = list_of_alpha[i]\n \n cont = [data for data in context if data['item_code'][0].lower() == selected_letters.lower()]\n\n for g in cont :\n contex2.append(g)\n\n context = [data for data in contex2 if search_data.lower() in data['item_code'].lower()] \n else:\n context = [data for data in context if search_data.lower() in data['item_code'].lower()] \n\n\n\n return JsonResponse({'context':context})\n\n\n\n#=============================================================== Filter By Inventory and Team================================================================================\n\ndef InvFilter(request):\n # request.is_ajax() is deprecated since django 3.1\n is_ajax = request.headers.get('X-Requested-With') == 'XMLHttpRequest'\n\n if is_ajax:\n data = json.load(request)\n tabbs = data.get('payload')\n\n\n \n\n invdrop = tabbs['invdropdata']\n\n if invdrop == 'Unconfirmed':\n confirm = 0\n elif invdrop == 'Confirmed':\n confirm = 1\n else:\n confirm = 1\n\n\n context = data_perfection(tabbs)\n\n if tabbs['invdropdata'] != \"Show All\":\n context = [data for data in context if data['Confimed']== confirm]\n\n # Get all the teams fed\n a = Team.objects.all()\n # Creating empty list to add teams fed\n selected_teams = []\n for i in a:\n #Adding teams available \n selected_teams.append(i.name)\n\n # Filtered Team\n chosen_team = tabbs['teamdata']\n c_team = [d for d in selected_teams if d.lower() == chosen_team.lower()]\n #Finding the index of filtered team\n if len(c_team) > 0:\n list_of_alpha = Alpha_allocate(selected_teams,c_team)\n contex2 = []\n for i in range(len(list_of_alpha)):\n selected_letters = list_of_alpha[i]\n \n cont = [data for data in context if data['item_code'][0].lower() == selected_letters.lower()]\n\n for g in cont :\n contex2.append(g)\n\n else:\n contex2 = context\n\n\n\n\n\n return JsonResponse({'context':contex2})\n\n\n\n#================================================================= Confirm =========================================================================================\n\ndef Confirmed(request):\n # request.is_ajax() is deprecated since django 3.1\n is_ajax = request.headers.get('X-Requested-With') == 'XMLHttpRequest'\n\n if is_ajax:\n data = json.load(request)\n tabbs = data.get('payload')\n\n context = data_perfection(tabbs)\n\n context = [data for data in context if data['Confimed'] == 0]\n\n\n # Get all the teams fed\n a = Team.objects.all()\n # Creating empty list to add teams fed\n selected_teams = []\n for i in a:\n #Adding teams available \n selected_teams.append(i.name)\n\n # Filtered Team\n chosen_team = tabbs['teamdropdata']\n c_team = [d for d in selected_teams if d.lower() == chosen_team.lower()]\n #Finding the index of filtered team\n if len(c_team) > 0:\n list_of_alpha = Alpha_allocate(selected_teams,c_team)\n\n contex2 = []\n\n for i in range(len(list_of_alpha)):\n selected_letters = list_of_alpha[i]\n cont = [data for data in context if data['item_code'][0].lower() == selected_letters.lower()]\n for i in cont: \n dd = Inventory.objects.filter(Q(item_code=i['item_code']))\n for po in dd:\n po.Confimed = 1\n po.save()\n\n\n\n\n\n return JsonResponse({'context':contex2})\n\n#================================================================= UnConfirm Page =========================================================================================\n\n\ndef UnconfirmedPage(request):\n if Profile.objects.filter(user=request.user).exists():\n\n\n\n template = addition + \"CountUnconfirmed.html\"\n\n profile = Profile.objects.get(user=request.user)\n\n # my allowed categories\n categories = Team.objects.filter(name=profile.team)\n \n\n count_header = CountHeader.objects.filter(is_active=True)\n\n sqlquery = sql_statements()[0]\n \n\n cursor = connection.cursor()\n cursor.execute(sqlquery)\n row = cursor.fetchall()\n\n \n\n contex = [{\n 'pk':i[0],\n 'Confimed':i[1],\n 'item_code':i[2],\n 'count_theoretical':i[3],\n 'count_summary':i[4] ,\n 'count_variance':i[5],\n 'cost_variance':i[6],\n } for i in row ]\n contex = [data for data in contex if data['count_summary']==None] \n\n\n # Get all the teams fed\n a = Team.objects.all()\n # Creating empty list to add teams fed\n selected_teams = []\n for i in a:\n #Adding teams available \n selected_teams.append(i.name)\n\n # Filtered Team\n chosen_team = profile.team.name\n c_team = [d for d in selected_teams if d == chosen_team ]\n #Finding the index of filtered team\n if len(c_team) > 0:\n list_of_alpha = Alpha_allocate(selected_teams,c_team)\n\n contex2 = []\n for i in range(len(list_of_alpha)):\n selected_letters = list_of_alpha[i]\n\n cont = [data for data in contex if data['item_code'][0].lower() == selected_letters.lower()]\n\n for g in cont :\n contex2.append(g)\n\n\n context = {\"inventory\":contex2,\"categories\": categories,\"count_headers\": count_header, \"count\": \"Unconfirmed\", \"team\": profile.team.name}\n\n return render(request, template, context)\n\n else:\n return HttpResponseRedirect('/error_login') \n\n#================================================================= Nullify =========================================================================================\n\n\ndef Nullify(request):\n # request.is_ajax() is deprecated since django 3.1\n is_ajax = request.headers.get('X-Requested-With') == 'XMLHttpRequest'\n\n if is_ajax:\n data = json.load(request)\n tabbs = data.get('payload')\n\n context = data_perfection(tabbs)\n\n context = [data for data in context if data['Confimed'] == 0]\n\n # Get all the teams fed\n a = Team.objects.all()\n # Creating empty list to add teams fed\n selected_teams = []\n for i in a:\n #Adding teams available \n selected_teams.append(i.name)\n\n # Filtered Team\n chosen_team = tabbs['teamfil']\n c_team = [d for d in selected_teams if d.lower() == chosen_team.lower()]\n #Finding the index of filtered team\n\n if len(c_team) > 0:\n if len(c_team) > 0:\n list_of_alpha = Alpha_allocate(selected_teams,c_team)\n\n contex2 = []\n for i in range(len(list_of_alpha)):\n selected_letters = list_of_alpha[i]\n cont = [data for data in context if data['item_code'][0].lower() == selected_letters.lower()]\n for i in cont: \n dd = Inventory.objects.filter(Q(item_code=i['item_code']))\n for po in dd:\n po.count_summary = None\n po.save()\n\n\n\n return JsonResponse({'context':context})\n\n#================================================================= Confirm =========================================================================================\n\ndef sql_statements():\n \n sqlquery1 = '''\\\n SELECT a.id as id , a.Confimed as Confimed, a.item_code as item_code,a.count_theoretical as count_theoretical, a.count_summary as count_summary, \n (a.count_summary - a.count_theoretical) as count_variance, ((a.count_summary - a.count_theoretical) * a.cost) as cost_variance\n FROM inventit_inventory a\n '''\n \n sqlquery2 = '''\\\n SELECT a.id as id , a.Confimed as Confimed, a.item_code as item_code,a.count_theoretical as count_theoretical, a.count_summary as count_summary, \n (a.count_summary - a.count_theoretical) as count_variance, ((a.count_summary - a.count_theoretical) * a.cost) as cost_variance,b.count_1,b.count_2,b.count_3\n FROM inventit_inventory a\n INNER JOIN inventit_countlines b\n ON (a.id = b.inventory_id) \n '''\n sql_statement = [sqlquery1,sqlquery2]\n return sql_statement\n\n\ndef data_perfection(tabbs):\n\n if tabbs['tabbs'] == 'ALL':\n sqlquery = sql_statements()[0]\n elif tabbs['tabbs'] == 'C_failed_NV':\n sqlquery = sql_statements()[1]\n elif tabbs['tabbs'] == 'C_failed_V': \n sqlquery = sql_statements()[1]\n elif tabbs['tabbs'] == 'C_Succ_NV':\n sqlquery = sql_statements()[1]\n elif tabbs['tabbs'] == 'C_Succ_V':\n sqlquery = sql_statements()[1]\n\n cursor = connection.cursor()\n cursor.execute(sqlquery)\n row = cursor.fetchall()\n\n\n if tabbs['tabbs'] != 'ALL':\n context = [{\n 'pk':i[0],\n 'Confimed':i[1],\n 'item_code':i[2],\n 'count_theoretical':i[3],\n 'count_summary':i[4],\n 'count_variance':i[5],\n 'cost_variance':i[6],\n 'count_1':i[7],\n 'count_2':i[8],\n 'count_3':i[9],\n } for i in row]\n else:\n context = [{\n 'pk':i[0],\n 'Confimed':i[1],\n 'item_code':i[2],\n 'count_theoretical':i[3],\n 'count_summary':i[4],\n 'count_variance':i[5],\n 'cost_variance':i[6],\n } for i in row]\n\n _name = False\n contex2 = []\n list_of_p_name = []\n list_of_p_dict = []\n list_of_p_dict2 = []\n list_of_p_dict3 = []\n infection = []\n successs = None\n if tabbs['tabbs'] == 'ALL':\n context = [data for data in context]\n successs = None\n elif tabbs['tabbs'] == 'C_failed_NV':\n context = [data for data in context if data['count_variance'] == 0]\n successs = False\n elif tabbs['tabbs'] == 'C_failed_V': \n context = [data for data in context if data['count_variance'] != 0]\n successs = False\n elif tabbs['tabbs'] == 'C_Succ_NV':\n context = [data for data in context if data['count_variance'] == 0]\n successs = True\n elif tabbs['tabbs'] == 'C_Succ_V':\n context = [data for data in context if data['count_variance'] != 0]\n successs = True\n \n\n for i in range(len(context)):\n name_o_item = context[i]['item_code']\n c_name = [data for data in context if data['item_code'] == name_o_item]\n if len(c_name) > 0:\n for s in c_name:\n if successs == True:\n decis = (s['count_1'] == s['count_2'] == s['count_3'])\n if decis == True:\n p_name = s['item_code']\n if len(list_of_p_name) > 0:\n if p_name in list_of_p_name:\n pass\n else:\n list_of_p_name.append(p_name)\n else:\n list_of_p_name.append(p_name)\n else:\n if s['item_code'] not in infection:\n infection.append(s['item_code'])\n elif successs == False:\n decis = (s['count_1'] != s['count_3'] or s['count_1'] != s['count_2'] or s['count_2'] != s['count_3'])\n if decis == True:\n p_name = s['item_code']\n if len(list_of_p_name) > 0:\n if p_name in list_of_p_name:\n pass\n else:\n list_of_p_name.append(p_name)\n else:\n list_of_p_name.append(p_name)\n else:\n if s['item_code'] not in infection:\n infection.append(s['item_code'])\n else:\n list_of_p_dict2 = context\n if successs != None:\n for i in list_of_p_name:\n list_of_p_dict = [data for data in context if data['item_code'] == i]\n list_of_p_dict2.append(list_of_p_dict[0])\n\n for p in range(len(infection)):\n s = [data for data in list_of_p_dict2 if data['item_code'] == infection[p]]\n if len(s) > 0:\n list_of_p_dict2.remove(s[0])\n\n return list_of_p_dict2\n\n\ndef Alpha_allocate(selected_teams,c_team):\n t_indx = selected_teams.index(c_team[0])\n # Array of first letters to be filtered\n arry = ['a','b','c','d','e','f','g','h','i','j','k','l','m','n','o','p','q','r','s','t','u','v','w','x','y','z','1','2','3','4','5','6','7','8','9','0']\n # Length of arry \n arry_len = len(arry)\n # Length of selected teams fed\n team_len = len(selected_teams)\n # Finding how many times teams can be allocated the arry\n team_times = arry_len/team_len\n g = int(t_indx)\n # creating empty list to store first letters per team\n list_of_alpha = []\n team_times = int(team_times)\n for i in range(int(team_times)):\n list_of_alpha.append(arry[g])\n g += team_times\n\n return list_of_alpha\n","sub_path":"inventit/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":40754,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"360302472","text":"import blurplefier\n\nwith open(\"images/input.png\", \"rb\") as test_file:\n test_file_bytes = test_file.read()\n\nextension, standard_output = blurplefier.convert_image(\n test_file_bytes, blurplefier.Methods.CLASSIC\n)\n\nwith open(f\"images/output_standard.{extension}\", \"wb\") as output_file:\n output_file.write(standard_output)\n\nextension, filtered_output = blurplefier.convert_image(\n test_file_bytes, blurplefier.Methods.FILTER\n)\n\nwith open(f\"images/output_filter.{extension}\", \"wb\") as output_file:\n output_file.write(filtered_output)\n\nextension, neo_output = blurplefier.convert_image(\n test_file_bytes,\n blurplefier.Methods.FILTER,\n modifier=blurplefier.Modifiers.NEW_LIGHT,\n)\n\nwith open(f\"images/output_neofilter.{extension}\", \"wb\") as output_file:\n output_file.write(neo_output)\n","sub_path":"tests/basic_test.py","file_name":"basic_test.py","file_ext":"py","file_size_in_byte":804,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"639578518","text":"#!/usr/bin/env python3\n\nimport numpy as np\nimport sys\nimport random as rnd\nfrom datetime import datetime\n\nOnHand = {}\nFolded = {}\nDeck = []\nBehavior = {}\ncurrent = \"\"\nactions = 0\n\nPlayer = 'Player'\nppl = (Player, 'Opponent 1', 'Opponent 2', 'Opponent 3')\nsuitdict = {\"S\":0, \"H\":1, \"D\":2, \"C\":3}\nletters = [\"A\"]+[str(i) for i in range(2,11)]+[\"J\", \"Q\", \"K\"]\nheader = (\"Spade\", \"Heart\", \"Diamond\", \"Club\")\n\nclass color:\n PURPLE = '\\033[95m'\n CYAN = '\\033[96m'\n DARKCYAN = '\\033[36m'\n BLUE = '\\033[94m'\n GREEN = '\\033[92m'\n YELLOW = '\\033[93m'\n RED = '\\033[91m'\n BOLD = '\\033[1m'\n UNDERLINE = '\\033[4m'\n END = '\\033[0m'\n\n\ndef init():\n\tglobal OnHand\n\tglobal Deck\n\tglobal Folded\n\tglobal Behavior\n\tglobal current\n\tglobal actions\n\n\tfor x in ppl:\n\t\tOnHand[x] = [False]*52\n\t\tFolded[x] = 0\n\t\tBehavior[x] = []\n\tgot = {recvr:0 for recvr in ppl}\n\tfor i in range(52):\n\t\trecvr = ppl[rnd.randint(0, 3)]\n\t\twhile got[recvr] == 13:\n\t\t\trecvr = ppl[rnd.randint(0, 3)]\n\t\tOnHand[recvr][i] = True\n\t\tgot[recvr] += 1\n\t\t# Spade 7\n\t\tif i == 6:\n\t\t\tcurrent = recvr\n\t\t\tprint(\"[Game]\", recvr, \"is to serve.\")\n\tDeck = [False]*52\n\tactions = 0\n\ndef parse_input(msg):\n\tif msg[0] not in \"SsHhDdCc\":\n\t\treturn (-1, \"[Alert] Unrecognized suid \\\"{}\\\".\".format(msg[0])) \n\tsuit = suitdict[msg[0].upper()]\n\tif len(msg)<=1 or not msg[1:].strip():\n\t\treturn (-1, \"[Alert] Unable to comprehend number.\") \n\tmsg = msg[1:].strip()\n\tif msg[0] in \"AaJjQqKk\":\n\t\tnum = letters.index(msg[0].upper())\n\telse:\n\t\ttry:\n\t\t\tnum = int(msg)-1\n\t\texcept ValueError:\n\t\t\treturn (-1, \"[Alert] Unable to comprehend number.\") \n\treturn (suit*13+num, None)\n\ndef can_put(hand, deck):\n\tif not deck[6]:\n\t\treturn \t[i==6 and hand[6] for i in range(52)]\t\n\tacc = [False]*52\n\tfor s in range(4):\n\t\tfoundL = False\n\t\tfoundR = False\n\t\tfor i in range(8):\n\t\t\tif not foundL and (i==6 or deck[s*13+i+1]):\n\t\t\t\tacc[s*13+i] = hand[s*13+i]\n\t\t\t\tfoundL = True\n\t\t\tif not foundR and deck[s*13+11-i]:\n\t\t\t\tacc[s*13+12-i] = hand[s*13+12-i]\n\t\t\t\tfoundR = True\n\treturn acc\n\ndef print_set(sset, end='\\n', highlight=''):\n\tif highlight:\n\t\tacc = can_put(sset, Deck)\n\tfor i in range(4):\n\t\tprint(header[i].ljust(10, ' '), end='')\n\t\tfor j in range(0,13):\n\t\t\thead = \"\"\n\t\t\tbody = \"\"\n\t\t\ttail = \"\"\n\t\t\tif highlight and acc[i*13+j]:\n\t\t\t\thead = highlight\n\t\t\t\ttail = color.END\n\t\t\tbody = (letters[j] if sset[i*13+j] else '-').ljust(3, ' ')\n\t\t\tprint(head+body+tail,end='')\n\t\tprint(\"\")\n\tprint(\"\", end=end)\n\n\nModel = np.zeros(1)\nModel_fold = np.zeros(1)\nBData = np.load('behavior_data.npy')\n\ndef s_AI(hand, deck):\n\tcan = can_put(hand, deck)\n\tfold = not any(can)\n\tvec = np.asarray([float(x) for x in hand]+[float(x) for x in deck])\n\n\tcan_vec = np.asarray([float(x) for x in can])\n\tif fold:\n\t\tprobf = vec@Model_fold\n\t\tres = np.argmax(probf*can_vec)\n\telse:\n\t\tprob = vec@Model\n\t\tres = np.argmax(prob*can_vec)\n\t# print(\"[DEBUG]\", res)\n\tif can[res]:\n\t\t# print(\"[DEBUG]smartboi\")\n\t\treturn res\n\treturn hand.index(True) if fold else can.index(True)\n\n\ndef processBehavior(winn, weight):\n\tglobal BData\n\tglobal Model_fold\n\tglobal Model\n\tif winn is not None:\n\t\tBData = np.concatenate((BData, [[weight]+x for x in Behavior[winn]]), axis=0)\n\t# train model\n\tdata1 = []\n\tdata2 = []\n\ty1 = []\n\ty2 = []\n\tfor x in BData:\n\t\tloop = 5\n\t\t# score fold card own[52] deck[52]\n\t\tscore, fold, card = x[0:3]\n\t\tvec = x[3:]\n\t\tif score > 0:\n\t\t\tloop -= score//10\n\t\tloop = max(0, loop)\n\t\tfor i in range(loop):\t\t\n\t\t\tif fold == 1:\n\t\t\t\tdata1.append(vec)\n\t\t\t\ty1.append([float(t==card) for t in range(52)])\n\t\t\telse:\n\t\t\t\tdata2.append(vec)\n\t\t\t\ty2.append([float(t==card) for t in range(52)])\n\n\tModel_fold, res, rnk, sing = np.linalg.lstsq(data1, y1, rcond=None)\n\tModel, res, rnk, sing = np.linalg.lstsq(data2, y2, rcond=None)\n\nrnd.seed(datetime.now())\nwhile True:\n\tinit()\n\tprocessBehavior(None, 0)\n\t# main loop\n\twhile actions < 52:\n\t\t\t# print(current)\n\t\t\tpcard = -1\n\t\t\tpfold = not any(can_put(OnHand[current], Deck))\n\t\t\tif current == ppl[0]:\n\t\t\t\tprint(\"[Game] Deck:\")\n\t\t\t\tprint_set(Deck)\n\t\t\t\tprint(\"[Game] Your cards:\")\n\t\t\t\tprint_set(OnHand[current], highlight=color.YELLOW)\n\t\t\t\twhile True:\n\t\t\t\t\tpinput = input(\"[Game] PLEASE ENTER [SHDC][A-K] or [*]: \")\n\t\t\t\t\tif '*' in pinput:\n\t\t\t\t\t\t# auto complete\n\t\t\t\t\t\tpcard = s_AI(OnHand[current], Deck)\n\t\t\t\t\t\tbreak\n\t\t\t\t\tpcard, err = parse_input(pinput.lstrip())\n\t\t\t\t\tif err:\n\t\t\t\t\t\tprint(err)\n\t\t\t\t\t\tcontinue\t\t\t\t\t\n\t\t\t\t\tif not OnHand[current][pcard]:\n\t\t\t\t\t\tprint(\"[Alert] You don't have this card!\", pcard)\n\t\t\t\t\telif not pfold and (pcard%13 > 6 and not Deck[pcard-1] or pcard%13 < 6 and not Deck[pcard+1]):\n\t\t\t\t\t\tprint(\"[Alert] You can't put this card now!\", pcard)\n\t\t\t\t\telif OnHand[current][6] and pcard != 6:\n\t\t\t\t\t\tprint(\"[Alert] Please put Spade 7 first.\")\n\t\t\t\t\telse: \n\t\t\t\t\t\tbreak\n\t\t\telse:\n\t\t\t\tpcard = s_AI(OnHand[current], Deck)\n\n\t\t\tBehavior[current].append([pfold, pcard]+OnHand[current][:]+Deck[:])\n\t\t\t\n\t\t\tOnHand[current][pcard] = False\n\t\t\tDeck[pcard] = not pfold\n\t\t\tif pfold:\n\t\t\t\tFolded[current] += (pcard%13+1)\n\t\t\tactions += 1\n\n\t\t\tprint(current, \"FOLDED\" if pfold else \"PUT {} {}.\".format(header[pcard//13], letters[pcard%13]))\n\n\t\t\tcurrent = ppl[(ppl.index(current)+1)%4]\n\tprint(\"[Game] Game Over.\")\n\tprint(\"[Game] Penalty:\")\n\tprint(ppl[0].ljust(15, ' ')+ppl[1].ljust(15, ' ')+ppl[2].ljust(15, ' ')+ppl[3].ljust(15, ' '))\n\tprint(str(Folded[ppl[0]]).ljust(15, ' ')+str(Folded[ppl[1]]).ljust(15, ' ')+\\\n\t\tstr(Folded[ppl[2]]).ljust(15, ' ')+str(Folded[ppl[3]]).ljust(15, ' '))\n\twinner = min(Folded, key=Folded.get)\n\tprocessBehavior(Player, Folded[Player]-Folded[winner])\n\tif winner != Player:\t\n\t\tprocessBehavior(winner, Folded[winner])\n\tif input(\"[Game] Another game ? [Y/N]\")[0].upper() != \"Y\":\n\t\tbreak \n\nnp.save('behavior_data.npy', BData)\nprint(\"[Data] Behavior data is saved to behavior_data.npy. N = {}\".format(len(BData)))\n\n\n# print([int(x) for x in Behavior['Player'][0]])\n","sub_path":"Sevens.py","file_name":"Sevens.py","file_ext":"py","file_size_in_byte":5774,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"295101233","text":"from django.db import models\nfrom django.contrib.auth.models import User\nfrom django.db.models.signals import post_save # very commonly used\nfrom django.dispatch import receiver # goes with post_save and other signals\nfrom rest_framework.authtoken.models import Token\n\n\nclass Location(models.Model):\n city_choice = models.CharField(max_length=20)\n\n def __str__(self):\n return self.city_choice\n\n\nclass Category(models.Model):\n new_category = models.CharField(max_length=30)\n choose_main = models.ForeignKey('self', null=True, blank=True, related_name=\"sub_choice\")\n\n def __str__(self):\n return self.new_category\n\n class Meta:\n verbose_name_plural = \"Categories\"\n\n\nclass Listing(models.Model):\n title = models.CharField(max_length=40)\n description = models.TextField()\n price = models.DecimalField(max_digits=6, decimal_places=2)\n photo = models.ImageField(upload_to='listing_photos', null=True, blank=True, verbose_name='Upload a photo')\n city = models.ForeignKey(Location)\n seller = models.ForeignKey(User)\n created = models.DateTimeField(auto_now_add=True)\n pick_category = models.ForeignKey(Category)\n\n def __str__(self):\n return self.title\n\n class Meta:\n ordering = ['-created']\n\n @property\n def photo_url(self):\n if self.photo:\n return self.photo.url\n return 'http://sport.ngmnexpo.com/cliparts/2015/02/1344144.jpg'\n\n\nclass TraderProfile(models.Model):\n user = models.OneToOneField('auth.User')\n preferred_location = models.ForeignKey(Location, null=True, blank=True)\n primary_category = models.ForeignKey(Category, null=True, blank=True)\n email_address = models.EmailField(max_length=45, null=True, blank=True)\n logo = models.ImageField(upload_to='logo_images', null=True, blank=True, verbose_name='Upload a logo')\n\n @property\n def logo_url(self):\n if self.logo:\n return self.logo.url\n return 'http://static.tumblr.com/e7snt83/tU6m7t07k/pirate_patch2.jpg'\n\n\n@receiver(post_save, sender='auth.User')\ndef create_user_profile(**kwargs):\n created = kwargs.get('created')\n instance = kwargs.get('instance')\n\n if created:\n TraderProfile.objects.create(user=instance) # hooks profile to user\n\n@receiver(post_save, sender='auth.User')\ndef create_token(**kwargs): # a shortcut pass in\n created = kwargs.get(\"created\") # boilerplate\n instance = kwargs.get(\"instance\") # boilerplate\n if created:\n Token.objects.create(user=instance) # yep. standard.\n","sub_path":"main_app/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":2542,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"502364453","text":"import http.server as SimpleHTTPServer\nimport xmlrpc.server as SimpleXMLRPCServer\nimport robot.setMotorPower as RobotMotorPower\n\ndef registerRobotXmlRpcMethods(server):\n \n # Register standard XML-RPC methods.\n server.register_introspection_functions()\n \n # Register the motor power command function.\n RobotMotorPower.init()\n server.register_function(RobotMotorPower.set,'setRobotMotorPower')\n\n\n \n \n# We define a custom server request handler, capable of both handling GET and XML-RPC requests.\nclass RequestHandler(SimpleXMLRPCServer.SimpleXMLRPCRequestHandler, SimpleHTTPServer.SimpleHTTPRequestHandler):\n rpc_paths = ('/RobotControlService',)\n\n def do_GET(self):\n SimpleHTTPServer.SimpleHTTPRequestHandler.do_GET(self) \n \n \n \n \n# Start running the server ... \nif __name__ == \"__main__\":\n \n # Create our XML-RPC server.using out custom request handler that is also able to serve web pages over GET.\n port = 8080\n server = SimpleXMLRPCServer.SimpleXMLRPCServer((\"0.0.0.0\", port), RequestHandler)\n \n # Register the XML-RPC methods ...\n registerRobotXmlRpcMethods(server)\n \n # Start to server.\n server.serve_forever()\n","sub_path":"Blockly_picar-c/server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":1206,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"182576442","text":"import os\r\nimport re\r\nfrom KO_web_parse_multi_process_v2 import get_ko_data\r\nfrom KO_web_parse_multi_process_v2 import write_file\r\nimport time\r\nimport multiprocessing\r\nimport random\r\nfrom multiprocessing import Pool\r\nimport time\r\nimport shutil\r\n\r\ndef check_and_backup():\r\n if os.path.exists(\"/home/fdong/auto_protein_analysis/ko_info/KO_INFO_END.txt\"):\r\n curent_time = time.time()\r\n ko_info_end_creat_time = os.stat(\"/home/fdong/auto_protein_analysis/ko_info/KO_INFO_END.txt\").st_ctime\r\n # print(ko_info_end_creat_time)\r\n time_gap = curent_time - ko_info_end_creat_time\r\n ModifiedTime=time.localtime(os.stat(\"/home/fdong/auto_protein_analysis/ko_info/KO_INFO_END.txt\").st_mtime) #文件访问时间 \r\n y=time.strftime('%Y', ModifiedTime) \r\n m=time.strftime('%m', ModifiedTime) \r\n d=time.strftime('%d', ModifiedTime) \r\n H=time.strftime('%H', ModifiedTime) \r\n M=time.strftime('%M', ModifiedTime)\r\n\r\n if time_gap>24*60*60*1.5:\r\n backup_name = \"/home/fdong/auto_protein_analysis/ko_info/ko_info_end_backups/KO_INFO_END_{0}{1}{2}{3}{4}.txt\".format(y,m,d,H,M)\r\n shutil.copy(\"/home/fdong/auto_protein_analysis/ko_info/KO_INFO_END.txt\",backup_name)\r\n print(\"KO_INFO_END.TXT已经备份,正在更新......\")\r\n os.system(\"python3 /home/fdong/auto_protein_analysis/ko_info/KO_num_class.py\")\r\n os.remove(\"/home/fdong/auto_protein_analysis/ko_info/KO_INFO.txt\")\r\n os.remove(\"/home/fdong/auto_protein_analysis/ko_info/KO_INFO_END.txt\")\r\n\r\n\r\n\r\n\r\ndef clean_ko_info(ko_info_end,ko_middle_temp,ko_scraped,num_temp):\r\n if not ko_middle_temp:\r\n ko_scraped= []\r\n num_temp=0\r\n return (ko_scraped, num_temp)\r\n for line in ko_middle_temp:\r\n # print(line)\r\n if re.match(r\"^K\\d{5}\",line[:6]) and len(re.findall(r\"\\t\",line)) == 3: # 剔除异常行\r\n match_temp = re.match(r\"^K\\d{5}\",line[:6]).group()\r\n if match_temp not in ko_scraped: #去除重复值\r\n ko_info_end.write(line)\r\n num_temp += 1 # 已经爬取的数量\r\n ko_scraped.append(match_temp)\r\n # ko_middle_temp.close()\r\n # ko_info_end.close()\r\n return (ko_scraped,num_temp)\r\n\r\n\r\ndef get_omit_ko_list(ko_scraped,ko_all_num,num_temp):\r\n omit_ko_list = list()\r\n if (ko_all_num - num_temp)!=0:\r\n ko_all = open(\"/home/fdong/auto_protein_analysis/ko_info/KO_terms.txt\", 'r')\r\n for ko in ko_all:\r\n if ko.strip() not in ko_scraped: #去除已经爬取的\r\n omit_ko_list.append(ko.strip())\r\n # ko_all.close()\r\n return omit_ko_list\r\n\r\n\r\n\r\ndef add_omit_main():\r\n ko_all = open(\"/home/fdong/auto_protein_analysis/ko_info/KO_terms.txt\", 'r')\r\n ko_all_num = len(ko_all.readlines())\r\n ko_all.close()\r\n\r\n num_temp = 0\r\n ko_scraped = list()\r\n kegg_ko_omit = list()\r\n #if os.path.exists(\"KO_INFO_END.txt\"):\r\n # os.remove(\"KO_INFO_END.txt\")\r\n\r\n while (ko_all_num - num_temp):\r\n ko_info_end = open(\"/home/fdong/auto_protein_analysis/ko_info/KO_INFO_END.txt\", \"a+\") # 所有ko 的所有信息__终版\r\n try:\r\n ko_middle_temp = open(\"/home/fdong/auto_protein_analysis/ko_info/KO_INFO.txt\", \"r+\") # 多线程爬取的文件,中间含有遗漏\r\n except:\r\n ko_middle_temp = None\r\n # print(ko_all_num,\"\\t\",num_temp)\r\n (ko_scraped, num_temp) = clean_ko_info(ko_info_end, ko_middle_temp, ko_scraped, num_temp)\r\n omit_ko_list = get_omit_ko_list(ko_scraped,ko_all_num,num_temp)\r\n # print(\"num_temp:\\t\",num_temp)\r\n\r\n print(\"omit ko numbers:\", len(omit_ko_list))\r\n if len(omit_ko_list) <= 50:\r\n print(omit_ko_list)\r\n # if len(omit_ko_list)==2 and \"K06060\" in omit_ko_list and \"K04225\" in omit_ko_list:\r\n # break\r\n # print(\"omit_ko_list:\\t\",omit_ko_list)\r\n # print(\"kegg_ko_omit:\\t\",kegg_ko_omit)\r\n print(\"*\"*30)\r\n if omit_ko_list == kegg_ko_omit:\r\n break\r\n else:\r\n kegg_ko_omit = omit_ko_list\r\n\r\n if len(omit_ko_list) != 0:\r\n pool = multiprocessing.Pool(processes=100)\r\n # file_data = list()\r\n for line in omit_ko_list:\r\n ko_list_signle = line.strip()\r\n # file_data.append(pool.apply_async(get_ko_data, (ko_list_signle,),callback=write_file))\r\n pool.apply_async(get_ko_data, (ko_list_signle,),callback=write_file)\r\n time.sleep(random.uniform(0,0.3))\r\n pool.close()\r\n pool.join()\r\n pool.terminate()\r\n # for line in file_data:\r\n # ko_middle_temp.write(line.get())\r\n ko_middle_temp.close()\r\n ko_info_end.close()\r\n\r\nif __name__ == '__main__':\r\n start = time.time() # 计时开始\r\n #os.remove(\"/home/fdong/auto_protein_analysis/ko_info/KO_INFO_END.txt\")\r\n#os.remove(\"/home/fdong/auto_protein_analysis/ko_info/KO_INFO.txt\")\r\n\r\n check_and_backup()\r\n add_omit_main()\r\n\r\n end = time.time()\r\n m, s = divmod(end - start, 60) # 转换时间的方法\r\n h, m = divmod(m, 60)\r\n print('KO在线爬取任务完成,共耗时%02d小时%02d分%02d秒' % (h, m, s))\r\n","sub_path":"ko_info/add_ko_omit_v2.py","file_name":"add_ko_omit_v2.py","file_ext":"py","file_size_in_byte":5286,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"74838846","text":"\"\"\"Tests for GP and SP classes\"\"\"\nimport math\nimport unittest\nimport numpy as np\nfrom gpkit import (Model, Monomial, settings, VectorVariable, Variable,\n SignomialsEnabled, ArrayVariable)\nfrom gpkit.geometric_program import GeometricProgram\nfrom gpkit.small_classes import CootMatrix\nfrom gpkit.feasibility import feasibility_model\n\nNDIGS = {\"cvxopt\": 5, \"mosek\": 7, \"mosek_cli\": 5}\n# name: decimal places of accuracy\n\n\nclass TestGP(unittest.TestCase):\n \"\"\"\n Test GeometricPrograms.\n This TestCase gets run once for each installed solver.\n \"\"\"\n name = \"TestGP_\"\n # solver and ndig get set in loop at bottom this file, a bit hacky\n solver = None\n ndig = None\n\n def test_trivial_gp(self):\n \"\"\"\n Create and solve a trivial GP:\n minimize x + 2y\n subject to xy >= 1\n\n The global optimum is (x, y) = (sqrt(2), 1/sqrt(2)).\n \"\"\"\n x = Monomial('x')\n y = Monomial('y')\n prob = Model(cost=(x + 2*y),\n constraints=[x*y >= 1])\n sol = prob.solve(solver=self.solver, verbosity=0)\n self.assertEqual(type(prob.latex()), str)\n self.assertEqual(type(prob._repr_latex_()), str)\n self.assertAlmostEqual(sol(\"x\"), math.sqrt(2.), self.ndig)\n self.assertAlmostEqual(sol(\"y\"), 1/math.sqrt(2.), self.ndig)\n self.assertAlmostEqual(sol(\"x\") + 2*sol(\"y\"),\n 2*math.sqrt(2),\n self.ndig)\n self.assertAlmostEqual(sol[\"cost\"], 2*math.sqrt(2), self.ndig)\n\n def test_simple_united_gp(self):\n R = Variable('R', units=\"nautical_miles\")\n a0 = Variable('a0', 340.29, 'm/s')\n theta = Variable(r'\\theta', 0.7598)\n t = Variable('t', 10, 'hr')\n T_loiter = Variable('T_{loiter}', 1, 'hr')\n T_reserve = Variable('T_{reserve}', 45, 'min')\n M = VectorVariable(2, 'M')\n\n if R.units:\n prob = Model(1/R,\n [t >= sum(R/a0/M/theta**0.5) + T_loiter + T_reserve,\n M <= 0.76])\n sol = prob.solve(verbosity=0)\n self.assertAlmostEqual(sol[\"cost\"], 0.0005532, self.ndig)\n\n def test_trivial_vector_gp(self):\n \"\"\"\n Create and solve a trivial GP with VectorVariables\n \"\"\"\n x = VectorVariable(2, 'x')\n y = VectorVariable(2, 'y')\n prob = Model(cost=(sum(x) + 2*sum(y)),\n constraints=[x*y >= 1])\n sol = prob.solve(solver=self.solver, verbosity=0)\n self.assertEqual(sol('x').shape, (2,))\n self.assertEqual(sol('y').shape, (2,))\n for x, y in zip(sol('x'), sol('y')):\n self.assertAlmostEqual(x, math.sqrt(2.), self.ndig)\n self.assertAlmostEqual(y, 1/math.sqrt(2.), self.ndig)\n self.assertAlmostEqual(sol[\"cost\"]/(4*math.sqrt(2)), 1., self.ndig)\n\n def test_zero_lower_unbounded(self):\n x = Variable('x', value=4)\n y = Variable('y', value=0)\n z = Variable('z')\n t1 = Variable('t1')\n t2 = Variable('t2')\n\n prob = Model(z, [z >= x + t1,\n t1 >= t2,\n t2 >= y])\n sol = prob.solve(verbosity=0)\n\n def test_mdd_example(self):\n Cl = Variable(\"Cl\", 0.5, \"-\", \"Lift Coefficient\")\n Mdd = Variable(\"Mdd\", \"-\", \"Drag Divergence Mach Number\")\n m1 = Model(1/Mdd, [1 >= 5*Mdd + 0.5, Mdd >= 0.00001])\n m2 = Model(1/Mdd, [1 >= 5*Mdd + 0.5])\n m3 = Model(1/Mdd, [1 >= 5*Mdd + Cl, Mdd >= 0.00001])\n sol1 = m1.solve(solver=self.solver, verbosity=0)\n sol2 = m2.solve(solver=self.solver, verbosity=0)\n sol3 = m3.solve(solver=self.solver, verbosity=0)\n gp1, gp2, gp3 = [m.program for m in [m1, m2, m3]]\n self.assertEqual(gp1.A, CootMatrix(row=[0, 1, 2],\n col=[0, 0, 0],\n data=[-1, 1, -1]))\n self.assertEqual(gp2.A, CootMatrix(row=[0, 1],\n col=[0, 0],\n data=[-1, 1]))\n # order of variables within a posynomial is not stable\n # (though monomial order is)\n equiv1 = gp3.A == CootMatrix(row=[0, 2, 3, 2],\n col=[0, 0, 0, 0],\n data=[-1, 1, -1, 0])\n equiv2 = gp3.A == CootMatrix(row=[0, 1, 3, 2],\n col=[0, 0, 0, 0],\n data=[-1, 1, -1, 0])\n self.assertTrue(equiv1 or equiv2)\n self.assertAlmostEqual(sol1(Mdd), sol2(Mdd))\n self.assertAlmostEqual(sol1(Mdd), sol3(Mdd))\n self.assertAlmostEqual(sol2(Mdd), sol3(Mdd))\n\n def test_additive_constants(self):\n x = Variable('x')\n m = Model(1/x, [1 >= 5*x + 0.5, 1 >= 10*x])\n m.solve(verbosity=0)\n gp = m.program\n self.assertEqual(gp.cs[1], gp.cs[2])\n self.assertEqual(gp.A.data[1], gp.A.data[2])\n\n def test_zeroing(self):\n L = Variable(\"L\")\n k = Variable(\"k\", 0)\n with SignomialsEnabled():\n constr = [L-5*k <= 10]\n sol = Model(1/L, constr).solve(verbosity=0, solver=self.solver)\n self.assertAlmostEqual(sol(L), 10, self.ndig)\n self.assertAlmostEqual(sol[\"cost\"], 0.1, self.ndig)\n\n def test_singular(self):\n \"\"\"\n Create and solve GP with a singular A matrix\n \"\"\"\n if self.solver == 'cvxopt':\n # cvxopt can't solve this problem\n # (see https://github.com/cvxopt/cvxopt/issues/36)\n return\n x = Variable('x')\n y = Variable('y')\n m = Model(y*x, [y*x >= 12])\n sol = m.solve(solver=self.solver, verbosity=0)\n self.assertAlmostEqual(sol[\"cost\"], 12, self.ndig)\n\n def test_constants_in_objective_1(self):\n '''Issue 296'''\n x1 = Variable('x1')\n x2 = Variable('x2')\n m = Model(1.+ x1 + x2, [x1 >= 1., x2 >= 1.])\n sol = m.solve(solver=self.solver, verbosity=0)\n self.assertAlmostEqual(sol[\"cost\"], 3, self.ndig)\n\n def test_constants_in_objective_2(self):\n '''Issue 296'''\n x1 = Variable('x1')\n x2 = Variable('x2')\n m = Model(x1**2 + 100 + 3*x2, [x1 >= 10., x2 >= 15.])\n sol = m.solve(solver=self.solver, verbosity=0)\n self.assertAlmostEqual(sol[\"cost\"]/245., 1, self.ndig)\n\n def test_feasibility_gp_(self):\n x = Variable('x')\n m = Model(x, [x**2 >= 1, x <= 0.5])\n self.assertRaises(RuntimeWarning, m.solve, verbosity=0)\n fm = feasibility_model(m, \"max\")\n sol1 = fm.solve(verbosity=0)\n fm = feasibility_model(m, \"product\")\n sol2 = fm.solve(verbosity=0)\n self.assertTrue(sol1[\"cost\"] >= 1)\n self.assertTrue(sol2[\"cost\"] >= 1)\n\n def test_terminating_constant_(self):\n x = Variable('x')\n y = Variable('y', value=0.5)\n prob = Model(1/x, [x + y <= 4])\n sol = prob.solve(verbosity=0)\n self.assertAlmostEqual(sol[\"cost\"], 1/3.5, self.ndig)\n\n def test_check_result(self):\n \"\"\"issue 361\"\"\"\n N = 5\n L = 5.\n dx = L/(N-1)\n EI = Variable(\"EI\",10)\n p = VectorVariable(N, \"p\")\n p = p.sub(p, 100*np.ones(N))\n V = VectorVariable(N, \"V\")\n M = VectorVariable(N, \"M\")\n th = VectorVariable(N, \"th\")\n w = VectorVariable(N, \"w\")\n eps = 1E-6\n substitutions = {var: eps for var in [V[-1], M[-1], th[0], w[0]]}\n objective = w[-1]\n constraints = [EI*V.left[1:N] >= EI*V[1:N] + 0.5*dx*p.left[1:N] + 0.5*dx*p[1:N],\n EI*M.left[1:N] >= EI*M[1:N] + 0.5*dx*V.left[1:N] + 0.5*dx*V[1:N],\n EI*th.right[0:N-1] >= EI*th[0:N-1] + 0.5*dx*M.right[0:N-1] + 0.5*dx*M[0:N-1],\n EI*w.right[0:N-1] >= EI*w[0:N-1] + 0.5*dx*th.right[0:N-1] + 0.5*dx*th[0:N-1]]\n m = Model(objective, constraints, substitutions)\n sol = m.solve(verbosity=0)\n\n def test_exps_is_tuple(self):\n \"\"\"issue 407\"\"\"\n x = Variable('x')\n m = Model(x, [x >= 1])\n m.solve(verbosity=0)\n self.assertEqual(type(m.program.cost.exps), tuple)\n\nclass TestSP(unittest.TestCase):\n \"\"\"test case for SP class -- gets run for each installed solver\"\"\"\n name = \"TestSP_\"\n solver = None\n ndig = None\n\n def test_trivial_sp(self):\n x = Variable('x')\n y = Variable('y')\n with SignomialsEnabled():\n m = Model(x, [x >= 1-y, y <= 0.1])\n sol = m.localsolve(verbosity=0, solver=self.solver)\n self.assertAlmostEqual(sol[\"variables\"][\"x\"], 0.9, self.ndig)\n with SignomialsEnabled():\n m = Model(x, [x+y >= 1, y <= 0.1])\n sol = m.localsolve(verbosity=0, solver=self.solver)\n self.assertAlmostEqual(sol[\"variables\"][\"x\"], 0.9, self.ndig)\n\n def test_relaxation(self):\n x = Variable(\"x\")\n y = Variable(\"y\")\n with SignomialsEnabled():\n constraints = [y + x >= 2, y <= x]\n objective = x\n m = Model(objective, constraints)\n m.localsolve(verbosity=0)\n\n # issue #257\n\n A = VectorVariable(2, \"A\")\n B = ArrayVariable([2, 2], \"B\")\n C = VectorVariable(2, \"C\")\n with SignomialsEnabled():\n constraints = [A <= B.dot(C),\n B <= 1,\n C <= 1]\n obj = 1/A[0] + 1/A[1]\n m = Model(obj, constraints)\n m.localsolve(verbosity=0)\n\n def test_issue180(self):\n L = Variable(\"L\")\n Lmax = Variable(\"L_{max}\", 10)\n W = Variable(\"W\")\n Wmax = Variable(\"W_{max}\", 10)\n A = Variable(\"A\", 10)\n Obj = Variable(\"Obj\")\n a_val = 0.01\n a = Variable(\"a\", a_val)\n with SignomialsEnabled():\n eqns = [L <= Lmax,\n W <= Wmax,\n L*W >= A,\n Obj >= a*(2*L + 2*W) + (1-a)*(12 * W**-1 * L**-3)]\n m = Model(Obj, eqns)\n spsol = m.solve(verbosity=0, solver=self.solver)\n # now solve as GP\n eqns[-1] = (Obj >= a_val*(2*L + 2*W) + (1-a_val)*(12 * W**-1 * L**-3))\n m = Model(Obj, eqns)\n gpsol = m.solve(verbosity=0, solver=self.solver)\n self.assertAlmostEqual(spsol['cost'], gpsol['cost'])\n\n def test_trivial_sp2(self):\n x = Variable(\"x\")\n y = Variable(\"y\")\n\n # converging from above\n with SignomialsEnabled():\n constraints = [y + x >= 2, y >= x]\n objective = y\n x0 = 1\n y0 = 2\n m = Model(objective, constraints)\n sol1 = m.localsolve(x0={x: x0, y: y0}, verbosity=0, solver=self.solver)\n\n # converging from right\n with SignomialsEnabled():\n constraints = [y + x >= 2, y <= x]\n objective = x\n x0 = 2\n y0 = 1\n m = Model(objective, constraints)\n sol2 = m.localsolve(x0={x: x0, y: y0}, verbosity=0, solver=self.solver)\n\n self.assertAlmostEqual(sol1[\"variables\"][\"x\"],\n sol2[\"variables\"][\"x\"], self.ndig)\n self.assertAlmostEqual(sol1[\"variables\"][\"y\"],\n sol2[\"variables\"][\"x\"], self.ndig)\n\n def test_sp_initial_guess_sub(self):\n x = Variable(\"x\")\n y = Variable(\"y\")\n x0 = 1\n y0 = 2\n with SignomialsEnabled():\n constraints = [y + x >= 2, y <= x]\n objective = x\n m = Model(objective, constraints)\n try:\n sol = m.localsolve(x0={x: x0, y: y0}, verbosity=0,\n solver=self.solver)\n except TypeError:\n self.fail(\"Call to local solve with only variables failed\")\n self.assertAlmostEqual(sol(x), 1, self.ndig)\n self.assertAlmostEqual(sol[\"cost\"], 1, self.ndig)\n\n try:\n sol = m.localsolve(x0={\"x\": x0, \"y\": y0}, verbosity=0,\n solver=self.solver)\n except TypeError:\n self.fail(\"Call to local solve with only variable strings failed\")\n self.assertAlmostEqual(sol(\"x\"), 1, self.ndig)\n self.assertAlmostEqual(sol[\"cost\"], 1, self.ndig)\n\n try:\n sol = m.localsolve(x0={\"x\": x0, y: y0}, verbosity=0,\n solver=self.solver)\n except TypeError:\n self.fail(\"Call to local solve with a mix of variable strings \"\n \"and variables failed\")\n self.assertAlmostEqual(sol[\"cost\"], 1, self.ndig)\n\n def test_small_signomial(self):\n x = Variable('x')\n z = Variable('z')\n local_ndig = 4\n nonzero_adder = 0.1 # TODO: support reaching zero, issue #348\n with SignomialsEnabled():\n J = 0.01*(x - 1)**2 + nonzero_adder\n m = Model(z, [z >= J])\n sol = m.localsolve(verbosity=0)\n self.assertAlmostEqual(sol['cost'], nonzero_adder, local_ndig)\n self.assertAlmostEqual(sol('x'), 0.987, 3)\n\n def test_signomials_not_allowed_in_objective(self):\n with SignomialsEnabled():\n x = Variable('x')\n y = Variable('y')\n J = 0.01*((x - 1)**2 + (y - 1)**2) + (x*y - 1)**2\n m = Model(J)\n with self.assertRaises(TypeError):\n sol = m.localsolve(verbosity=0)\n\n def test_partial_sub_signomial(self):\n \"\"\"Test SP partial x0 initialization\"\"\"\n x = Variable('x')\n y = Variable('y')\n with SignomialsEnabled():\n m = Model(x, [x + y >= 1, y <= 0.5])\n m.localsolve(x0={x: 0.5}, verbosity=0)\n self.assertEqual(m.program.gps[0].constraints[0].exp[x], -1./3)\n\n\nTEST_CASES = [TestGP, TestSP]\n\nTESTS = []\nfor testcase in TEST_CASES:\n for solver in settings[\"installed_solvers\"]:\n if solver:\n test = type(testcase.__name__+\"_\"+solver,\n (testcase,), {})\n setattr(test, \"solver\", solver)\n setattr(test, \"ndig\", NDIGS[solver])\n TESTS.append(test)\n\nif __name__ == \"__main__\":\n from gpkit.tests.helpers import run_tests\n run_tests(TESTS)\n","sub_path":"gpkit/tests/t_model.py","file_name":"t_model.py","file_ext":"py","file_size_in_byte":14190,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"287393829","text":"from globals import *\r\nfrom geant_classes import *\r\n\r\nslack = 1.0*nm\r\ndiameter = 10.6*mm # detector_geometry.hole_diameter\r\nfoil_thickness = 5000.0*nm\r\nfoil_metal_thickness = 200.0*nm\r\nfoil_material = \"ParyleneN\"\r\n\r\nclass FoilGeometry(G4LogicalVolume):\r\n def __init__(self):\r\n # A round disk of very thin parylene foil, coated with Aluminum\r\n foil_solid = G4Tubs(\"foil\",\r\n 0, diameter/2,\r\n foil_thickness/2,\r\n 0, 2*pi)\r\n G4LogicalVolume.__init__(self, foil_solid, G4Material(\"Al\"), \"foil\")\r\n\r\n # Parylene Core\r\n core_solid = G4Tubs(\"foil_core\",\r\n 0, diameter/2 - slack,\r\n (foil_thickness - foil_metal_thickness)/2,\r\n 0, 2*pi)\r\n core_lvolume = G4LogicalVolume(core_solid, G4Material(foil_material), \"foil_core\")\r\n G4PVPlacement(0, [0, 0, 0],\r\n core_lvolume,\r\n \"foil_core\",\r\n self)","sub_path":"python geometry/foil_geometry.py","file_name":"foil_geometry.py","file_ext":"py","file_size_in_byte":1109,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"260360420","text":"\"\"\"Basic compute graph elements\"\"\"\nimport abc\nfrom collections import OrderedDict\nimport dataclasses as dc\nimport itertools\nimport json\nimport logging\nimport networkx as nx\nimport os\nimport pdb\nfrom pathlib import Path\nimport typing as ty\nimport pickle as pk\nfrom copy import deepcopy, copy\nfrom time import sleep\n\nimport cloudpickle as cp\nfrom filelock import FileLock\nimport shutil\nfrom tempfile import mkdtemp\n\nfrom . import state\nfrom . import auxiliary as aux\nfrom .specs import File, BaseSpec, RuntimeSpec, Result, SpecInfo, LazyField\nfrom .helpers import (\n make_klass,\n create_checksum,\n print_help,\n load_result,\n gather_runtime_info,\n save_result,\n ensure_list,\n record_error,\n get_inputs,\n)\nfrom ..utils.messenger import send_message, make_message, gen_uuid, now, AuditFlag\n\nlogger = logging.getLogger(\"pydra\")\n\ndevelop = True\n\n\nclass TaskBase:\n _api_version: str = \"0.0.1\" # Should generally not be touched by subclasses\n _version: str # Version of tool being wrapped\n _task_version: ty.Optional[\n str\n ] = None # Task writers encouraged to define and increment when implementation changes sufficiently\n _input_sets = None # Dictionaries of predefined input settings\n\n audit_flags: AuditFlag = AuditFlag.NONE # What to audit. See audit flags for details\n\n _can_resume = False # Does the task allow resuming from previous state\n _redirect_x = False # Whether an X session should be created/directed\n\n _runtime_requirements = RuntimeSpec()\n _runtime_hints = None\n\n _cache_dir = None # Working directory in which to operate\n _references = None # List of references for a task\n\n # dj: do we need it??\n input_spec = BaseSpec\n output_spec = BaseSpec\n\n # TODO: write state should be removed\n def __init__(\n self,\n name,\n inputs: ty.Union[ty.Text, File, ty.Dict, None] = None,\n audit_flags: AuditFlag = AuditFlag.NONE,\n messengers=None,\n messenger_args=None,\n cache_dir=None,\n cache_locations=None,\n ):\n \"\"\"A base structure for nodes in the computational graph (i.e. both\n ``Node`` and ``Workflow``).\n\n Parameters\n ----------\n\n name : str\n Unique name of this node\n inputs : dictionary (input name, input value or list of values)\n States this node's input names\n \"\"\"\n self.name = name\n if not self.input_spec:\n raise Exception(\"No input_spec in class: %s\" % self.__class__.__name__)\n klass = make_klass(self.input_spec)\n self.inputs = klass(\n **{\n f.name: (None if f.default is dc.MISSING else f.default)\n for f in dc.fields(klass)\n }\n )\n self.input_names = [\n field.name\n for field in dc.fields(klass)\n if field.name not in [\"_func\", \"_graph\"]\n ]\n self.state = None\n self._output = {}\n self._result = {}\n # flag that says if node finished all jobs\n self._done = False\n if self._input_sets is None:\n self._input_sets = {}\n if inputs:\n if isinstance(inputs, dict):\n inputs = {k: v for k, v in inputs.items() if k in self.input_names}\n elif Path(inputs).is_file():\n inputs = json.loads(Path(inputs).read_text())\n elif isinstance(inputs, str):\n if self._input_sets is None or inputs not in self._input_sets:\n raise ValueError(\"Unknown input set {!r}\".format(inputs))\n inputs = self._input_sets[inputs]\n self.inputs = dc.replace(self.inputs, **inputs)\n self.state_inputs = inputs\n self.audit_flags = audit_flags\n self.messengers = ensure_list(messengers)\n self.messenger_args = messenger_args\n self.cache_dir = cache_dir\n self.cache_locations = cache_locations\n\n # dictionary of results from tasks\n self.results_dict = {}\n self.plugin = None\n\n def __getstate__(self):\n state = self.__dict__.copy()\n state[\"input_spec\"] = pk.dumps(state[\"input_spec\"])\n state[\"output_spec\"] = pk.dumps(state[\"output_spec\"])\n state[\"inputs\"] = dc.asdict(state[\"inputs\"])\n return state\n\n def __setstate__(self, state):\n state[\"input_spec\"] = pk.loads(state[\"input_spec\"])\n state[\"output_spec\"] = pk.loads(state[\"output_spec\"])\n state[\"inputs\"] = make_klass(state[\"input_spec\"])(**state[\"inputs\"])\n self.__dict__.update(state)\n\n def __getattr__(self, name):\n if name == \"lzout\": # lazy output\n return LazyField(self, \"output\")\n return self.__getattribute__(name)\n\n def help(self, returnhelp=False):\n \"\"\" Prints class help\n \"\"\"\n help_obj = print_help(self)\n if returnhelp:\n return help_obj\n\n @property\n def version(self):\n return self._version\n\n # TODO: not sure what was the idea for the method (not used)\n # def save_set(self, name, inputs, force=False):\n # if name in self._input_sets and not force:\n # raise KeyError(\"Key {} already saved. Use force=True to override.\")\n # self._input_sets[name] = inputs\n\n @property\n def checksum(self):\n return create_checksum(self.__class__.__name__, self.inputs)\n\n def set_state(self, splitter, combiner=None):\n if splitter is not None:\n self.state = state.State(\n name=self.name, splitter=splitter, combiner=combiner\n )\n else:\n self.state = None\n return self.state\n\n @property\n def output_names(self):\n return [f.name for f in dc.fields(make_klass(self.output_spec))]\n\n def audit(self, message, flags=None):\n if develop:\n with open(\n Path(os.path.dirname(__file__)) / \"..\" / \"schema/context.jsonld\", \"rt\"\n ) as fp:\n context = json.load(fp)\n else:\n context = {\n \"@context\": \"https://raw.githubusercontent.com/nipype/pydra/master/pydra/schema/context.jsonld\"\n }\n if self.audit_flags & flags:\n if self.messenger_args:\n send_message(\n make_message(message, context=context),\n messengers=self.messengers,\n **self.messenger_args,\n )\n else:\n send_message(\n make_message(message, context=context), messengers=self.messengers\n )\n\n @property\n def can_resume(self):\n \"\"\"Task can reuse partial results after interruption\n \"\"\"\n return self._can_resume\n\n @abc.abstractmethod\n def _run_task(self):\n pass\n\n @property\n def cache_dir(self):\n return self._cache_dir\n\n @cache_dir.setter\n def cache_dir(self, location):\n if location is not None:\n self._cache_dir = Path(location).resolve()\n self._cache_dir.mkdir(parents=False, exist_ok=True)\n else:\n self._cache_dir = mkdtemp()\n self._cache_dir = Path(self._cache_dir)\n\n @property\n def cache_locations(self):\n return self._cache_locations + ensure_list(self._cache_dir)\n\n @cache_locations.setter\n def cache_locations(self, locations):\n if locations is not None:\n self._cache_locations = [Path(loc) for loc in ensure_list(locations)]\n else:\n self._cache_locations = []\n\n @property\n def output_dir(self):\n return self._cache_dir / self.checksum\n\n def audit_check(self, flag):\n return self.audit_flags & flag\n\n def __call__(self, **kwargs):\n return self.run(**kwargs)\n\n def run(self, **kwargs):\n self.inputs = dc.replace(self.inputs, **kwargs)\n checksum = self.checksum\n lockfile = self.cache_dir / (checksum + \".lock\")\n \"\"\"\n Concurrent execution scenarios\n\n 1. prior cache exists -> return result\n 2. other process running -> wait\n a. finishes (with or without exception) -> return result\n b. gets killed -> restart\n 3. no cache or other process -> start\n 4. two or more concurrent new processes get to start\n \"\"\"\n # TODO add signal handler for processes killed after lock acquisition\n with FileLock(lockfile):\n # Let only one equivalent process run\n # Eagerly retrieve cached\n if self.results_dict: # should be skipped if run called without submitter\n result = self.result()\n if result is not None:\n return result\n odir = self.output_dir\n if not self.can_resume and odir.exists():\n shutil.rmtree(odir)\n cwd = os.getcwd()\n odir.mkdir(parents=False, exist_ok=True if self.can_resume else False)\n # start recording provenance, but don't send till directory is created\n # in case message directory is inside task output directory\n if self.audit_check(AuditFlag.PROV):\n aid = \"uid:{}\".format(gen_uuid())\n start_message = {\"@id\": aid, \"@type\": \"task\", \"startedAtTime\": now()}\n os.chdir(odir)\n if self.audit_check(AuditFlag.PROV):\n self.audit(start_message, AuditFlag.PROV)\n # audit inputs\n # check_runtime(self._runtime_requirements)\n # isolate inputs if files\n # cwd = os.getcwd()\n if self.audit_check(AuditFlag.RESOURCE):\n from ..utils.profiler import ResourceMonitor\n\n resource_monitor = ResourceMonitor(os.getpid(), logdir=odir)\n result = Result(output=None, runtime=None, errored=False)\n try:\n if self.audit_check(AuditFlag.RESOURCE):\n resource_monitor.start()\n if self.audit_check(AuditFlag.PROV):\n mid = \"uid:{}\".format(gen_uuid())\n self.audit(\n {\n \"@id\": mid,\n \"@type\": \"monitor\",\n \"startedAtTime\": now(),\n \"wasStartedBy\": aid,\n },\n AuditFlag.PROV,\n )\n self._run_task()\n result.output = self._collect_outputs()\n except Exception as e:\n record_error(self.output_dir, e)\n result.errored = True\n raise\n finally:\n if self.audit_check(AuditFlag.RESOURCE):\n resource_monitor.stop()\n result.runtime = gather_runtime_info(resource_monitor.fname)\n if self.audit_check(AuditFlag.PROV):\n self.audit(\n {\"@id\": mid, \"endedAtTime\": now(), \"wasEndedBy\": aid},\n AuditFlag.PROV,\n )\n # audit resources/runtime information\n eid = \"uid:{}\".format(gen_uuid())\n entity = dc.asdict(result.runtime)\n entity.update(\n **{\n \"@id\": eid,\n \"@type\": \"runtime\",\n \"prov:wasGeneratedBy\": aid,\n }\n )\n self.audit(entity, AuditFlag.PROV)\n self.audit(\n {\n \"@type\": \"prov:Generation\",\n \"entity_generated\": eid,\n \"hadActivity\": mid,\n },\n AuditFlag.PROV,\n )\n save_result(odir, result)\n with open(odir / \"_node.pklz\", \"wb\") as fp:\n cp.dump(self, fp)\n os.chdir(cwd)\n if self.audit_check(AuditFlag.PROV):\n # audit outputs\n self.audit(\n {\"@id\": aid, \"endedAtTime\": now(), \"errored\": result.errored},\n AuditFlag.PROV,\n )\n return result\n\n # TODO: Decide if the following two functions should be separated\n @abc.abstractmethod\n def _list_outputs(self):\n pass\n\n def _collect_outputs(self):\n run_output = ensure_list(self._list_outputs())\n output_klass = make_klass(self.output_spec)\n output = output_klass(**{f.name: None for f in dc.fields(output_klass)})\n return dc.replace(output, **dict(zip(self.output_names, run_output)))\n\n # TODO: should change state!\n def split(self, splitter, **kwargs):\n if kwargs:\n self.inputs = dc.replace(self.inputs, **kwargs)\n # dj:??, check if I need it\n self.state_inputs = kwargs\n splitter = aux.change_splitter(splitter, self.name)\n if self.state:\n raise Exception(\"splitter has been already set\")\n else:\n self.set_state(splitter)\n return self\n\n def combine(self, combiner):\n if not self.state:\n self.split(splitter=None)\n if not self.state:\n self.fut_combiner = combiner\n return self\n # raise Exception(\"splitter has to be set first\")\n elif self.state.combiner:\n raise Exception(\"combiner has been already set\")\n self.combiner = combiner\n self.set_state(splitter=self.state.splitter, combiner=self.combiner)\n return self\n\n # TODO: was used in submitter (if not needed should be removed)\n # def checking_input_el(self, ind):\n # \"\"\"checking if all inputs are available (for specific state element)\"\"\"\n # try:\n # self.get_input_el(ind)\n # return True\n # except: # TODO specify\n # return False\n\n def get_input_el(self, ind):\n \"\"\"collecting all inputs required to run the node (for specific state element)\"\"\"\n if ind is not None:\n # TODO: doesnt work properly for more cmplicated wf\n state_dict = self.state.states_val[ind]\n input_ind = self.state.inputs_ind[ind]\n inputs_dict = {}\n for inp in set(self.input_names):\n inputs_dict[inp] = getattr(self.inputs, inp)[\n input_ind[f\"{self.name}.{inp}\"]\n ]\n return state_dict, inputs_dict\n else:\n inputs_dict = {inp: getattr(self.inputs, inp) for inp in self.input_names}\n return None, inputs_dict\n\n def to_job(self, ind):\n \"\"\" running interface one element generated from node_state.\"\"\"\n # logger.debug(\"Run interface el, name={}, ind={}\".format(self.name, ind))\n el = deepcopy(self)\n el.state = None\n _, inputs_dict = self.get_input_el(ind)\n el.inputs = dc.replace(el.inputs, **inputs_dict)\n return el\n\n # checking if all outputs are saved\n @property\n def done(self):\n if self.results_dict:\n return all([future.done() for _, (future, _) in self.results_dict.items()])\n\n def _combined_output(self):\n combined_results = []\n for (gr, ind_l) in self.state.final_groups_mapping.items():\n combined_results.append([])\n for ind in ind_l:\n result = load_result(self.results_dict[ind][1], self.cache_locations)\n combined_results[gr].append(result)\n return combined_results\n\n def result(self, state_index=None):\n \"\"\"\n :param state_index:\n :return:\n \"\"\"\n # TODO: check if result is available in load_result and\n # return a future if not\n if self.state:\n if state_index is None:\n # if state_index=None, collecting all results\n if self.state.combiner:\n return self._combined_output()\n else:\n results = []\n for (ii, val) in enumerate(self.state.states_val):\n result = load_result(\n self.results_dict[ii][1], self.cache_locations\n )\n results.append(result)\n return results\n else: # state_index is not None\n if self.state.combiner:\n return self._combined_output()[state_index]\n result = load_result(\n self.results_dict[state_index][1], self.cache_locations\n )\n return result\n else:\n if state_index is not None:\n raise ValueError(\"Task does not have a state\")\n if self.results_dict:\n checksum = self.results_dict[None][1]\n else:\n checksum = self.checksum\n result = load_result(checksum, self.cache_locations)\n return result\n\n\nclass Workflow(TaskBase):\n def __init__(\n self,\n name,\n input_spec: ty.Union[ty.List[ty.Text], BaseSpec, None] = None,\n output_spec: ty.Optional[BaseSpec] = None,\n audit_flags: AuditFlag = AuditFlag.NONE,\n messengers=None,\n messenger_args=None,\n cache_dir=None,\n cache_locations=None,\n **kwargs,\n ):\n if input_spec:\n if isinstance(input_spec, BaseSpec):\n self.input_spec = input_spec\n else:\n self.input_spec = SpecInfo(\n name=\"Inputs\",\n fields=[(name, ty.Any) for name in input_spec]\n + [(\"_graph\", ty.Any)],\n bases=(BaseSpec,),\n )\n if output_spec is None:\n output_spec = SpecInfo(\n name=\"Output\", fields=[(\"out\", ty.Any)], bases=(BaseSpec,)\n )\n self.output_spec = output_spec\n\n super(Workflow, self).__init__(\n name=name,\n inputs=kwargs,\n cache_dir=cache_dir,\n cache_locations=cache_locations,\n audit_flags=audit_flags,\n messengers=messengers,\n messenger_args=messenger_args,\n )\n\n self.graph = nx.DiGraph()\n self.name2obj = {}\n\n # store output connections\n self._connections = None\n self.node_names = []\n\n def __getattr__(self, name):\n if name == \"lzin\":\n return LazyField(self, \"input\")\n if name == \"lzout\":\n return super().__getattr__(name)\n if name in self.name2obj:\n return self.name2obj[name]\n return self.__getattribute__(name)\n\n @property\n def nodes(self):\n return self._nodes\n\n @property\n def graph_sorted(self):\n return list(nx.topological_sort(self.graph))\n\n def add(self, task):\n if not is_task(task):\n raise ValueError(\"Unknown workflow element: {!r}\".format(task))\n self.graph.add_nodes_from([task])\n self.name2obj[task.name] = task\n self._last_added = task\n other_states = {}\n for field in dc.fields(task.inputs):\n val = getattr(task.inputs, field.name)\n if isinstance(val, LazyField):\n # adding an edge to the graph if task id expecting output from a different task\n if val.name != self.name:\n self.graph.add_edge(\n getattr(self, val.name),\n task,\n from_field=val.field,\n to_field=field.name,\n )\n if val.name in self.node_names and getattr(self, val.name).state:\n # adding a state from the previous task to other_states\n other_states[val.name] = (getattr(self, val.name).state, field.name)\n # if task has connections state has to be recalculated\n if other_states:\n if hasattr(task, \"fut_combiner\"):\n task.state = state.State(\n task.name, other_states=other_states, combiner=task.fut_combiner\n )\n else:\n task.state = state.State(task.name, other_states=other_states)\n self.node_names.append(task.name)\n self.inputs._graph = self.graph_sorted\n return self\n\n def _run_task(self):\n for task in self.graph_sorted:\n # depend on prior tasks that have state\n task.inputs.retrieve_values(self)\n if task.state and not hasattr(task.state, \"states_ind\"):\n task.state.prepare_states(inputs=task.inputs)\n if task.state and not hasattr(task.state, \"inputs_ind\"):\n task.state.prepare_inputs()\n if self.plugin is None:\n task.run()\n else:\n from .submitter import Submitter\n\n with Submitter(self.plugin) as sub:\n sub.run(task)\n while not task.done:\n sleep(1)\n\n def set_output(self, connections):\n self._connections = connections\n fields = [(name, ty.Any) for name, _ in connections]\n self.output_spec = SpecInfo(name=\"Output\", fields=fields, bases=(BaseSpec,))\n\n def _list_outputs(self):\n output = []\n for name, val in self._connections:\n if not isinstance(val, LazyField):\n raise ValueError(\"all connections must be lazy\")\n output.append(val.get_value(self))\n return output\n\n\n# TODO: task has also call\ndef is_function(obj):\n return hasattr(obj, \"__call__\")\n\n\ndef is_task(obj):\n return hasattr(obj, \"_run_task\")\n\n\ndef is_workflow(obj):\n return isinstance(obj, Workflow)\n","sub_path":"pydra/engine/core.py","file_name":"core.py","file_ext":"py","file_size_in_byte":22006,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"467423749","text":"import sys\nimport argparse\n\n\ndef echo_upper(text):\n return text.upper()\n\n\ndef echo_lower(text):\n return text.lower()\n\n\ndef echo_title(text):\n return text.title()\n\n\ndef parse_args(args):\n parser = argparse.ArgumentParser(\n description='Perform transformation on input text.')\n parser.add_argument('-u', '--upper',\n help='convert text to uppercase', action='store_true')\n parser.add_argument('-l', '--lower',\n help='convert text to lowercase', action='store_true')\n parser.add_argument('-t', '--title',\n help='convert text to titlecase', action='store_true')\n parser.add_argument('text', help='text to be manipulated')\n\n return parser.parse_args(args)\n\n\ndef main():\n text = parse_args(sys.argv[1:])\n upper = text.upper\n lower = text.lower\n title = text.title\n\n if upper:\n print(echo_upper(text))\n if lower:\n print(echo_lower(text))\n if title:\n print(echo_title(text))\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"echo.py","file_name":"echo.py","file_ext":"py","file_size_in_byte":1051,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"76964690","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Apr 10 09:19:05 2019\n\n@author: gpichard\n\"\"\"\n\nimport os\nfile_path = os.path.realpath(__file__)\ndir_path = os.path.dirname(file_path)\nif os.getcwd() != dir_path:\n print(\"Need to relocate current working directory\")\n os.chdir(dir_path)\n\nimport pandas as pd\npd.options.display.float_format = '{:.3f}'.format\nimport numpy as np\nimport datetime\nimport multiprocessing\nimport linearmodels\n\n# =============================================================================\n# # Loading the monthly panel, transforming some variables and filtering some outliers\n# =============================================================================\nMonthlyVariables_db = pd.read_csv('Monthly/MonthlyVariables_db.csv', index_col = [0, 1], header = 0, parse_dates = True)\nMonthlyVariables_db.info()\nfor col in list(['Date.' + str(i) for i in np.arange(1, 9)]):\n MonthlyVariables_db[col] = pd.to_datetime(MonthlyVariables_db[col], infer_datetime_format = True)\ndel col\nMonthlyVariables_db.reset_index(inplace = True)\nMonthlyVariables_db.set_index(['RIC','YearMonth'], inplace = True)\n\n#MonthlyAvailable_db = MonthlyVariables_db.dropna()\nMonthlyAvailable_db = MonthlyVariables_db\n# Whenever, at a given time and for a given company, the number of shares detained by ETFs is equal to or larger than the number of shares outstanding, the observation is dropped\nMonthlyAvailable_db.loc[MonthlyAvailable_db.PctSharesHeldETF >= 1, 'PctSharesHeldETF'] = None\n# Dropping observations with stock price less than USD 1\nMonthlyAvailable_db.loc[MonthlyAvailable_db.Close < 1, 'Close'] = None\n\n\n# Transformations and lags needed for regressions\nMonthlyAvailable_db = MonthlyAvailable_db.assign(InvClose = pd.Series(1/MonthlyAvailable_db['Close']))\nMonthlyAvailable_db = MonthlyAvailable_db.assign(BookToMarketRatio = pd.Series(1/MonthlyAvailable_db['PriceToBVPerShare']))\n\n# Replace infinite values with NaN in order to drop them afterwards\nMonthlyAvailable_db.replace([np.inf, -np.inf], np.nan, inplace = True)\n\n\n## Summary statistics\nKeptVariables_summary = list(['Volatility', 'PctSharesHeldETF', 'BookToMarketRatio', 'CompanyMarketCap_millions', 'InvClose', 'PctBidAskSpread', 'AmihudRatio', 'RetPast12to1M', 'RetPast12to7M', 'GrossProfitability'])\nKeptVariables_headers = list(['Volatility', 'ETF Ownership', 'Book-to-market', 'Market cap. ($ Mln.)', '1/Price', 'Rel. Bid-Ask spread', 'Amihud ratio', 'Past 12-to-1-month return', 'Past 12-to-7-month return', 'Gross profitability'])\nSumStats = MonthlyAvailable_db[KeptVariables_summary].dropna().describe().transpose()\nSumStats['count'] = SumStats['count'].astype(int)\nSumStats.set_index(pd.Series(KeptVariables_headers), drop = True, inplace = True)\n# Export\n#SumStats.to_latex('../SummaryStats/SummaryTable.tex', header = ['N (obs.)', 'Mean', 'St. dev.', 'Min.', '25%', 'Median', '75%', 'Max.'])\n#SumStats.to_latex()\n# Correlation matrix\nCorrMat = MonthlyAvailable_db[KeptVariables_summary].corr()\nmask = np.tril(np.ones_like(CorrMat, dtype=np.bool), k=0)\nCorrMat.where(mask, other = '', inplace = True)\nNumberedHeader = [\"({})\".format(x) for x in np.arange(1, len(KeptVariables_summary) + 1)]\nCorrMat.set_index([pd.Series(KeptVariables_headers), pd.Series(NumberedHeader)], inplace = True, drop = True)\n\n# Export\n#CorrMat.to_latex('../SummaryStats/CorrTable.tex', index_names = True, header = NumberedHeader[:(len(NumberedHeader))])\n\n## Dynamic panel\n# Shift variables in time : independent variables and volatility lags\nMonthlyAvailable_1lag_db = MonthlyAvailable_db.groupby(level = 0)[['PctSharesHeldETF', 'CompanyMarketCap', 'InvClose', 'AmihudRatio', 'PctBidAskSpread', 'BookToMarketRatio', 'RetPast12to1M', 'RetPast12to7M','GrossProfitability', 'Volatility', 'PctSharesHeldOtherMutual', 'PctSharesHeldPension', 'PctSharesHeldHedge']].shift(1)\nMonthlyAvailable_1lag_db.columns = [s + \"_1lag\" for s in list(MonthlyAvailable_1lag_db.columns)]\nMonthlyAvailable_2lag_db = MonthlyAvailable_db.groupby(level = 0)['Volatility'].shift(2)\nMonthlyAvailable_2lag_db.name = MonthlyAvailable_2lag_db.name + \"_2lag\"\nMonthlyAvailable_3lag_db = MonthlyAvailable_db.groupby(level = 0)['Volatility'].shift(3)\nMonthlyAvailable_3lag_db.name = MonthlyAvailable_3lag_db.name + \"_3lag\"\nMonthlyAvailable_4lag_db = MonthlyAvailable_db.groupby(level = 0)['Volatility'].shift(4)\nMonthlyAvailable_4lag_db.name = MonthlyAvailable_4lag_db.name + \"_4lag\"\nMonthlyAvailable_db = pd.concat([MonthlyAvailable_db, MonthlyAvailable_1lag_db, MonthlyAvailable_2lag_db, MonthlyAvailable_3lag_db, MonthlyAvailable_4lag_db], axis = 1)\n#MonthlyAvailable_db = MonthlyAvailable_db.dropna()\nMonthlyAvailable_db.info()\n\n## Standardizing dependent and ownership variables\n## Censoring the sample at 99.99th percentile for statistics\nMonthlyAvailable_std_db = MonthlyAvailable_db[['Volatility', 'PctSharesHeldETF', 'PctSharesHeldOtherMutual', 'PctSharesHeldPension', 'PctSharesHeldHedge']]\ncolInit = list(MonthlyAvailable_std_db.columns)\nfor col in colInit:\n mean = MonthlyAvailable_std_db[col].dropna().loc[MonthlyAvailable_std_db[col].dropna() < np.percentile(MonthlyAvailable_std_db[col].dropna(), 99.99)].mean()\n std = MonthlyAvailable_std_db[col].dropna().loc[MonthlyAvailable_std_db[col].dropna() < np.percentile(MonthlyAvailable_std_db[col].dropna(), 99.99)].std()\n # replacing the top 0.01% of values with NaN\n MonthlyAvailable_std_db[col].loc[MonthlyAvailable_std_db[col] >= np.percentile(MonthlyAvailable_std_db[col].dropna(), 99.99)] = None\n MonthlyAvailable_std_db[col] = (MonthlyAvailable_std_db[col] - mean)/std\n\ndel colInit, col, mean, std\n\n# Assign the non-normalized controls\nMonthlyAvailable_std_db = MonthlyAvailable_std_db.assign(InvClose = MonthlyAvailable_db.InvClose, AmihudRatio = MonthlyAvailable_db.AmihudRatio, RetPast12to1M = MonthlyAvailable_db.RetPast12to1M, RetPast12to7M = MonthlyAvailable_db.RetPast12to7M, CompanyMarketCap = MonthlyAvailable_db.CompanyMarketCap, BookToMarketRatio = MonthlyAvailable_db.BookToMarketRatio, PctBidAskSpread = MonthlyAvailable_db.PctBidAskSpread, GrossProfitability = MonthlyAvailable_db.GrossProfitability)\n\nMonthlyAvailable_std_1lag_db = MonthlyAvailable_std_db.groupby(level = 0)[['PctSharesHeldETF', 'CompanyMarketCap', 'InvClose', 'AmihudRatio', 'PctBidAskSpread', 'BookToMarketRatio', 'RetPast12to1M', 'RetPast12to7M','GrossProfitability', 'Volatility', 'PctSharesHeldOtherMutual', 'PctSharesHeldPension', 'PctSharesHeldHedge']].shift(1)\nMonthlyAvailable_std_1lag_db.columns = [s + \"_1lag\" for s in list(MonthlyAvailable_std_1lag_db.columns)]\nMonthlyAvailable_std_2lag_db = MonthlyAvailable_std_db.groupby(level = 0)['Volatility'].shift(2)\nMonthlyAvailable_std_2lag_db.name = MonthlyAvailable_std_2lag_db.name + \"_2lag\"\nMonthlyAvailable_std_3lag_db = MonthlyAvailable_std_db.groupby(level = 0)['Volatility'].shift(3)\nMonthlyAvailable_std_3lag_db.name = MonthlyAvailable_std_3lag_db.name + \"_3lag\"\nMonthlyAvailable_std_4lag_db = MonthlyAvailable_std_db.groupby(level = 0)['Volatility'].shift(4)\nMonthlyAvailable_std_4lag_db.name = MonthlyAvailable_std_4lag_db.name + \"_4lag\"\nMonthlyAvailable_std_db = pd.concat([MonthlyAvailable_std_db, MonthlyAvailable_std_1lag_db, MonthlyAvailable_std_2lag_db, MonthlyAvailable_std_3lag_db, MonthlyAvailable_std_4lag_db], axis = 1)\n\n## First differences\nMonthlyAvailable_diff_db = MonthlyAvailable_db.groupby(level=0).diff()\n#MonthlyAvailable_diff_db.to_csv('Monthly/MonthlyAvailable_diff.csv', header = True, index = True)\n\n## Loading quarterly variance ratios panel\nQuarterlyVarianceRatio_panel = pd.read_csv('Quarterly/Stocks_5to1dVarianceRatio.csv', header = 0, parse_dates = True)\nQuarterlyVarianceRatio_panel.Date = pd.to_datetime(QuarterlyVarianceRatio_panel.Date, infer_datetime_format = True)\nQuarterlyVarianceRatio_db = pd.melt(QuarterlyVarianceRatio_panel, id_vars = ['Date'], var_name = 'RIC', value_name = 'VR')\nQuarterlyVarianceRatio_db.set_index(['RIC','Date'], inplace = True)\nQuarterlyVarianceRatio_db = QuarterlyVarianceRatio_db.assign(absVR = abs(QuarterlyVarianceRatio_db.VR - 1))\n\n# Creating quarterly dataset with the monthly regressors, using end of quarter data or, if missing, the last available values\nlevel_values = MonthlyAvailable_db.index.get_level_values\nQuarterlyAvailable_db = MonthlyAvailable_db.groupby([level_values(0)]+[pd.Grouper(freq='Q', level = -1)]).last()\n# Lags have to be constructed again at the quarterly frequency (they are still monthly)\nQuarterlyAvailable_db.drop(columns=QuarterlyAvailable_db.filter(regex=(\"lag\")).columns, inplace = True)\nQuarterlyAvailable_db = pd.concat([QuarterlyAvailable_db, QuarterlyVarianceRatio_db], axis = 1)\n\nQuarterlyAvailable_1lag_db = QuarterlyAvailable_db.groupby(level = 0)[['PctSharesHeldETF', 'CompanyMarketCap', 'InvClose', 'AmihudRatio', 'PctBidAskSpread', 'BookToMarketRatio', 'RetPast12to1M', 'RetPast12to7M','GrossProfitability', 'Volatility', 'PctSharesHeldOtherMutual', 'PctSharesHeldPension', 'PctSharesHeldHedge', 'VR', 'absVR']].shift(1)\nQuarterlyAvailable_1lag_db.columns = [s + \"_1lag\" for s in list(QuarterlyAvailable_1lag_db.columns)]\nQuarterlyAvailable_2lag_db = QuarterlyAvailable_db.groupby(level = 0)['Volatility'].shift(2)\nQuarterlyAvailable_2lag_db.name = QuarterlyAvailable_2lag_db.name + \"_2lag\"\nQuarterlyAvailable_3lag_db = QuarterlyAvailable_db.groupby(level = 0)['Volatility'].shift(3)\nQuarterlyAvailable_3lag_db.name = QuarterlyAvailable_3lag_db.name + \"_3lag\"\nQuarterlyAvailable_4lag_db = QuarterlyAvailable_db.groupby(level = 0)['Volatility'].shift(4)\nQuarterlyAvailable_4lag_db.name = QuarterlyAvailable_4lag_db.name + \"_4lag\"\nQuarterlyAvailable_db = pd.concat([QuarterlyAvailable_db, QuarterlyAvailable_1lag_db, QuarterlyAvailable_2lag_db, QuarterlyAvailable_3lag_db, QuarterlyAvailable_4lag_db], axis = 1)\n\n\n# =============================================================================\n# ## Regression estimates\n# =============================================================================\n# Model 1 : Volatility\n# Full US sample, lagged controls\n# Column \"Baseline\"\nmod1_All_Volatility = linearmodels.PanelOLS.from_formula('Volatility ~ 1 + PctSharesHeldETF_1lag + np.log(CompanyMarketCap_1lag) + InvClose_1lag + BookToMarketRatio_1lag + RetPast12to1M_1lag + EntityEffects + TimeEffects', MonthlyAvailable_db)\nmod1_fit_All_Volatility = mod1_All_Volatility.fit(cov_type = \"kernel\")\nprint(mod1_fit_All_Volatility)\nf = open('../Regression_Results/US/mod1_fit_All_Volatility.tex', 'w')\nf.write(mod1_fit_All_Volatility.summary.as_latex())\nf.close()\n\n## Full US sample with a constant term, lagged controls and 4 volatility lags\n#mod1_All_Volatility_withLags = linearmodels.PanelOLS.from_formula('Volatility ~ 1 + PctSharesHeldETF + np.log(CompanyMarketCap_1lag) + InvClose_1lag + AmihudRatio_1lag + PctBidAskSpread_1lag + BookToMarketRatio_1lag + RetPast12to7M_1lag + GrossProfitability_1lag + EntityEffects + TimeEffects + Volatility_1lag + Volatility_2lag + Volatility_3lag + Volatility_4lag', MonthlyAvailable_db)\n#mod1_fit = mod1_All_Volatility_withLags.fit(cov_type = \"kernel\")\n#print(mod1_fit)\n#f = open('../Regression_Results/US/mod1_All_Volatility_withLags.tex', 'w')\n#f.write(mod1_fit.summary.as_latex())\n#f.close()\n\n# With additional liquidity proxies and volatility lags\n# Column \"Controls + Vol. lags\"\nmod1_All_Volatility_withLags_Controls = linearmodels.PanelOLS.from_formula('Volatility ~ 1 + PctSharesHeldETF_1lag + np.log(CompanyMarketCap_1lag) + InvClose_1lag + AmihudRatio_1lag + PctBidAskSpread_1lag + BookToMarketRatio_1lag + RetPast12to1M_1lag + GrossProfitability_1lag + EntityEffects + TimeEffects + Volatility_1lag + Volatility_2lag + Volatility_3lag + Volatility_4lag', MonthlyAvailable_db)\nmod1_fit_All_Volatility_withLags_Controls = mod1_All_Volatility_withLags_Controls.fit(cov_type = \"kernel\")\nprint(mod1_fit_All_Volatility_withLags_Controls)\nf = open('../Regression_Results/US/mod1_contemporaneousControls_fit.tex', 'w')\nf.write(mod1_fit_All_Volatility_withLags_Controls.summary.as_latex())\nf.close()\n\n### Attempt to do a dynamic panel estimation through GMM. Needs further research.\n##mod1_GMM_All_Volatility_withlags = linearmodels.LinearFactorModelGMM()\n#MonthlyAvailable_db.describe().to_csv('../SummaryStats/All_Volatility_withLags.csv', header = True, index = True)\n## Summary statistics : there are a few extreme values that may bias the sample.\n#MonthlyAvailable_db.describe().to_csv('../SummaryStats/All_Volatility_withLags.csv', header = True, index = True)\n\n\n# Including other fund ownership controls\n# Column \"Inst. o\\'ship controls\"\nmod1_All_Volatility_withLags_withFundcontrols = linearmodels.PanelOLS.from_formula('Volatility ~ 1 + PctSharesHeldETF_1lag + np.log(CompanyMarketCap_1lag) + InvClose_1lag + AmihudRatio_1lag + PctBidAskSpread_1lag + BookToMarketRatio_1lag + RetPast12to1M_1lag + GrossProfitability_1lag + PctSharesHeldOtherMutual_1lag + PctSharesHeldPension_1lag + PctSharesHeldHedge_1lag + EntityEffects + TimeEffects + Volatility_1lag + Volatility_2lag + Volatility_3lag + Volatility_4lag', MonthlyAvailable_db)\nmod1_fit_All_Volatility_withLags_withFundcontrols = mod1_All_Volatility_withLags_withFundcontrols.fit(cov_type = \"kernel\")\nprint(mod1_fit_All_Volatility_withLags_withFundcontrols)\nf = open('../Regression_Results/US/mod1_fit_All_Volatility_withLags_withFundcontrols.tex', 'w')\nf.write(mod1_fit_All_Volatility_withLags_withFundcontrols.summary.as_latex())\nf.close()\n\n# Standardization could (and should?) be performed over the relevant sample of values only\n\n# Estimation on standardized, winsorized variables\n# Column \"Standardized\"\nmod1_All_Volatility_std_withLags_withFundcontrols = linearmodels.PanelOLS.from_formula('Volatility ~ 1 + PctSharesHeldETF_1lag + np.log(CompanyMarketCap_1lag) + InvClose_1lag + AmihudRatio_1lag + PctBidAskSpread_1lag + BookToMarketRatio_1lag + RetPast12to1M_1lag + GrossProfitability_1lag + PctSharesHeldOtherMutual_1lag + PctSharesHeldPension_1lag + PctSharesHeldHedge_1lag + EntityEffects + TimeEffects + Volatility_1lag + Volatility_2lag + Volatility_3lag + Volatility_4lag', MonthlyAvailable_std_db)\nmod1_fit_All_Volatility_std_withLags_withFundcontrols = mod1_All_Volatility_std_withLags_withFundcontrols.fit(cov_type = \"kernel\")\nprint(mod1_fit_All_Volatility_std_withLags_withFundcontrols)\nf = open('../Regression_Results/US/mod1_fit_All_Volatility_std_withLags_withFundcontrols.tex', 'w')\nf.write(mod1_fit_All_Volatility_std_withLags_withFundcontrols.summary.as_latex())\nf.close()\n\n## First differences, no controls except other fund holdings and volatility lags\n## The full regressors matrix is not full rank, thus we do not include all usual controls \n#mod1_All_diffVolatility_withLags_withFundcontrols = linearmodels.PanelOLS.from_formula('Volatility_std ~ 0 + PctSharesHeldETF_std + PctSharesHeldOtherMutual_std + PctSharesHeldPension_std + PctSharesHeldHedge_std + AmihudRatio_1lag + BookToMarketRatio_1lag + RetPast12to7M_1lag + GrossProfitability_1lag + EntityEffects + TimeEffects + Volatility_1lag', MonthlyAvailable_diff_db)\n#mod1_fit_All_diffVolatility_withLags_withFundcontrols = mod1_All_diffVolatility_withLags_withFundcontrols.fit(cov_type = \"kernel\")\n#print(mod1_fit_All_diffVolatility_withLags_withFundcontrols)\n#f = open('../Regression_Results/US/mod1_fit_All_diffVolatility_std_withLags_withFundcontrols.tex', 'w')\n#f.write(mod1_fit_All_diffVolatility_withLags_withFundcontrols.summary.as_latex())\n#f.close()\n#\n## Variation : there is a FirstDifferenceOLS function in the \"linearmodels\" package\n## Using the standardized variables\n#mod1_All_FirstDiffOLS_withLags_withFundcontrols = linearmodels.FirstDifferenceOLS.from_formula('Volatility ~ 0 + PctSharesHeldETF_std + CompanyMarketCap_1lag + InvClose_1lag + AmihudRatio_1lag + PctBidAskSpread_1lag + BookToMarketRatio_1lag + RetPast12to7M_1lag + GrossProfitability_1lag + PctSharesHeldOtherMutual_std + PctSharesHeldPension_std + PctSharesHeldHedge_std + TimeEffects + Volatility_1lag + Volatility_2lag + Volatility_3lag + Volatility_4lag', MonthlyAvailable_stdMod1_db)\n#mod1_fit_All_FirstDiffOLS_withLags_withFundcontrols = mod1_All_FirstDiffOLS_withLags_withFundcontrols.fit(cov_type = \"kernel\")\n#print(mod1_fit_All_FirstDiffOLS_withLags_withFundcontrols)\n\n# Models comparison for the paper body\nmod1_comp = linearmodels.panel.compare({'Baseline':mod1_fit_All_Volatility, 'Controls + Vol. lags': mod1_fit_All_Volatility_withLags_Controls, 'Inst. o\\'ship controls':mod1_fit_All_Volatility_withLags_withFundcontrols , 'Standardized': mod1_fit_All_Volatility_std_withLags_withFundcontrols})\nprint(mod1_comp)\nf = open('../Regression_Results/US/mod1_comp.tex', 'w')\nf.write(mod1_comp.summary.as_latex())\nf.close()\n\n# Model 2 : Liquidity\n# Bid-ask spread on close prices (not equal to the High-Low spread metric from Israeli 2017) - static model\nmod2_All_Liquidity_BidAsk = linearmodels.PanelOLS.from_formula('PctBidAskSpread ~ 1 + PctSharesHeldETF_1lag + np.log(CompanyMarketCap_1lag) + BookToMarketRatio_1lag + Volatility_1lag + EntityEffects + TimeEffects', MonthlyAvailable_db)\nmod2_fit_All_Liquidity_BidAsk = mod2_All_Liquidity_BidAsk.fit(cov_type = \"kernel\")\nprint(mod2_fit_All_Liquidity_BidAsk)\nf = open('../Regression_Results/US/mod2_fit_All_Liquidity_BidAsk.tex', 'w')\nf.write(mod2_fit_All_Liquidity_BidAsk.summary.as_latex())\nf.close()\n\nmod2_All_Liquidity_BidAsk_withFundcontrols = linearmodels.PanelOLS.from_formula('PctBidAskSpread ~ 1 + PctSharesHeldETF_1lag + np.log(CompanyMarketCap_1lag) + BookToMarketRatio_1lag + Volatility_1lag + EntityEffects + TimeEffects + PctSharesHeldOtherMutual_1lag + PctSharesHeldPension_1lag + PctSharesHeldHedge_1lag', MonthlyAvailable_db)\nmod2_fit_All_Liquidity_BidAsk_withFundcontrols = mod2_All_Liquidity_BidAsk_withFundcontrols.fit(cov_type = \"kernel\")\nprint(mod2_fit_All_Liquidity_BidAsk_withFundcontrols)\nf = open('../Regression_Results/US/mod2_fit_All_Liquidity_BidAsk_withFundcontrols.tex', 'w')\nf.write(mod2_fit_All_Liquidity_BidAsk_withFundcontrols.summary.as_latex())\nf.close()\n\n# Amihud ratio decomposition\nmod2_All_Liquidity_Amihud = linearmodels.PanelOLS.from_formula('AmihudNumerator ~ 1 + PctSharesHeldETF_1lag + np.log(CompanyMarketCap_1lag) + BookToMarketRatio_1lag + AmihudDenominator + EntityEffects + TimeEffects', MonthlyAvailable_db)\nmod2_fit_All_Liquidity_Amihud = mod2_All_Liquidity_Amihud.fit(cov_type = \"kernel\")\nprint(mod2_fit_All_Liquidity_Amihud)\nf = open('../Regression_Results/US/mod2_fit_All_Liquidity_Amihud.tex', 'w')\nf.write(mod2_fit_All_Liquidity_Amihud.summary.as_latex())\nf.close()\n\nmod2_All_Liquidity_Amihud_withFundcontrols = linearmodels.PanelOLS.from_formula('AmihudNumerator ~ 1 + PctSharesHeldETF_1lag + np.log(CompanyMarketCap_1lag) + BookToMarketRatio_1lag + AmihudDenominator + EntityEffects + TimeEffects + PctSharesHeldOtherMutual_1lag ', MonthlyAvailable_db)\nmod2_fit_All_Liquidity_Amihud_withFundcontrols = mod2_All_Liquidity_Amihud_withFundcontrols.fit(cov_type = \"kernel\")\nprint(mod2_fit_All_Liquidity_Amihud_withFundcontrols)\nf = open('../Regression_Results/US/mod2_fit_All_Liquidity_Amihud_withFundcontrols.tex', 'w')\nf.write(mod2_fit_All_Liquidity_Amihud_withFundcontrols.summary.as_latex())\nf.close()\n# Only the other mutual fund holdings are included to control for institutional ownership, otherwise the matrix of regressors does not have full column rank. Other mutual funds are the most available values and economically the more sizeable of the three categories at hand.\n\n# Models comparison for the paper body\nmod2_comp = linearmodels.panel.compare({'Bid-Ask':mod2_fit_All_Liquidity_BidAsk, 'Bid-Ask w/inst. o\\'ship':mod2_fit_All_Liquidity_BidAsk_withFundcontrols, 'Amihud':mod2_fit_All_Liquidity_Amihud, 'Amihud w/inst. o\\'ship':mod2_fit_All_Liquidity_Amihud_withFundcontrols})\nprint(mod2_comp)\nf = open('../Regression_Results/US/mod2_comp.tex', 'w')\nf.write(mod2_comp.summary.as_latex())\nf.close()\n# Model 3 : Efficiency\nmod3_All_VR = linearmodels.PanelOLS.from_formula('VR ~ 1 + PctSharesHeldETF_1lag + np.log(CompanyMarketCap_1lag) + InvClose_1lag + AmihudRatio_1lag + PctBidAskSpread_1lag + BookToMarketRatio_1lag + RetPast12to7M_1lag + GrossProfitability_1lag + EntityEffects + TimeEffects', QuarterlyAvailable_db)\nmod3_fit_All_VR = mod3_All_VR.fit(cov_type = \"kernel\")\nprint(mod3_fit_All_VR)\nf = open('../Regression_Results/US/mod3_fit_All_VR.tex', 'w')\nf.write(mod3_fit_All_VR.summary.as_latex())\nf.close()\n\nmod3_All_VR_withFundcontrols = linearmodels.PanelOLS.from_formula('VR ~ 1 + PctSharesHeldETF_1lag + np.log(CompanyMarketCap_1lag) + InvClose_1lag + AmihudRatio_1lag + PctBidAskSpread_1lag + BookToMarketRatio_1lag + RetPast12to7M_1lag + GrossProfitability_1lag + EntityEffects + TimeEffects + PctSharesHeldOtherMutual_1lag + PctSharesHeldPension_1lag + PctSharesHeldHedge_1lag', QuarterlyAvailable_db)\nmod3_fit_All_VR_withFundcontrols = mod3_All_VR_withFundcontrols.fit(cov_type = \"kernel\")\nprint(mod3_fit_All_VR_withFundcontrols)\nf = open('../Regression_Results/US/mod3_fit_All_VR_withFundcontrols.tex', 'w')\nf.write(mod3_fit_All_VR_withFundcontrols.summary.as_latex())\nf.close()\n\nmod3_All_absVR = linearmodels.PanelOLS.from_formula('absVR ~ 1 + PctSharesHeldETF_1lag + np.log(CompanyMarketCap_1lag) + InvClose_1lag + AmihudRatio_1lag + PctBidAskSpread_1lag + BookToMarketRatio_1lag + RetPast12to7M_1lag + GrossProfitability_1lag + EntityEffects + TimeEffects', QuarterlyAvailable_db)\nmod3_fit_All_absVR = mod3_All_absVR.fit(cov_type = \"kernel\")\nprint(mod3_fit_All_absVR)\nf = open('../Regression_Results/US/mod3_fit_All_absVR.tex', 'w')\nf.write(mod3_fit_All_absVR.summary.as_latex())\nf.close()\n\nmod3_All_absVR_withFundcontrols = linearmodels.PanelOLS.from_formula('absVR ~ 1 + PctSharesHeldETF_1lag + np.log(CompanyMarketCap_1lag) + InvClose_1lag + AmihudRatio_1lag + PctBidAskSpread_1lag + BookToMarketRatio_1lag + RetPast12to7M_1lag + GrossProfitability_1lag + EntityEffects + TimeEffects + PctSharesHeldOtherMutual_1lag + PctSharesHeldPension_1lag + PctSharesHeldHedge_1lag', QuarterlyAvailable_db)\nmod3_fit_All_absVR_withFundcontrols = mod3_All_absVR_withFundcontrols.fit(cov_type = \"kernel\")\nprint(mod3_fit_All_absVR_withFundcontrols)\nf = open('../Regression_Results/US/mod3_fit_All_absVR_withFundcontrols.tex', 'w')\nf.write(mod3_fit_All_absVR_withFundcontrols.summary.as_latex())\nf.close()\n\n# Models comparison for the paper body\nmod3_comp = linearmodels.panel.compare({'VR':mod3_fit_All_VR, 'VR w/inst. o\\'ship':mod3_fit_All_VR_withFundcontrols, 'Abs. VR':mod3_fit_All_absVR, 'Abs. VR w/inst. o\\'ship':mod3_fit_All_absVR_withFundcontrols})\nprint(mod3_comp)\nf = open('../Regression_Results/US/mod3_comp.tex', 'w')\nf.write(mod3_comp.summary.as_latex())\nf.close()","sub_path":"Eikon/PanelAnalysis_monthly.py","file_name":"PanelAnalysis_monthly.py","file_ext":"py","file_size_in_byte":22675,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"161672192","text":"import requests\nfrom bs4 import BeautifulSoup\nimport bs4\nimport re\n\n#从网络上获取大学排名网页内容\ndef getHTMLText(url):\n try:\n r = requests.get(url, timeout = 30)\n r.raise_for_status()\n r.encoding = r.apparent_encoding\n return r.text\n except:\n return 'Fail'\n\n#提取网页内容中信息到合适的数据结构\ndef fillUnivList(ulist, html):\n soup = BeautifulSoup(html, 'html.parser')\n ulist = soup.find_all(string = re.compile('大学'))[6:21]\n d = dict()\n count = 1\n for name in ulist:\n if name not in d:\n d[name] = count\n count+=1\n lst = list()\n for name, rank in d.items():\n lst.append((rank, name))\n\n lst.sort()\n print('rank\\t univ')\n for rank, name in lst:\n print(rank, name)\n #for tr in soup.find('tbody').children: #解析tbody标签所在位置\n #if isinstance(tr, bs4.element.Tag): #tbody中找到每所大学对应的tr标签,过滤掉非标签信息\n #print(soup.find_all(string = re.compile('大学'))[:20])\n #print(tr.prettify())\n #tds = tr('td') #在tr标签中找到td标签的信息,把需要的td标签存入列表\n #ulist.append( [tds[0].string, tds[1].string, tds[2].string] )\n'''\n#利用数据结构展示并输出结果\ndef printUnivList(ulist, num):\n tplt = '{0:^10}\\t{1:{3}^10}\\t{2:^10}' #{3}format函数中的第三个变量进行填充\n print(tplt.format('排名', '学校', '所在地', chr(12288)))\n for i in range (num):\n u = ulist[i]\n print(tplt.format(u[0], u[1], u[2], chr(12288)))\n'''\nif __name__ == '__main__':\n unifo = []\n url = 'http://www.zuihaodaxue.com/ARWU2018.html'\n html = getHTMLText(url)\n fillUnivList(unifo, html)\n #printUnivList(unifo, 20) #20 univs\n","sub_path":"worldUniRank_junkV1.py","file_name":"worldUniRank_junkV1.py","file_ext":"py","file_size_in_byte":1826,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"111605690","text":"import numpy as np\nimport keras\nfrom keras.layers import Conv2D\nfrom keras.models import Sequential\nfrom keras.layers import MaxPooling2D\nfrom keras.layers import Flatten\nfrom keras.layers import Dense\nfrom keras.layers import Dropout\n\n\ndef read_labels(filename, items): # 读取图片标记,也就是要学习的数字\n file_labels = open(filename, 'rb')\n file_labels.seek(8) # 标签文件的头��8个字节,略过不读\n data = file_labels.read(items)\n y = np.zeros(items)\n for i in range(items):\n y[i] = data[i]\n file_labels.close()\n return y\n\n\ny_train = read_labels('./train-labels-idx1-ubyte', 60000) # 读取60000张训练标记\ny_test = read_labels('./t10k-labels-idx1-ubyte', 10000) # 读取10000张测试标记\n\n\ndef read_images(filename, items): # 读取图片\n file_image = open(filename, 'rb')\n file_image.seek(16)\n\n data = file_image.read(items * 28 * 28)\n\n X = np.zeros(items * 28 * 28)\n for i in range(items * 28 * 28):\n X[i] = data[i] / 255\n file_image.close()\n return X.reshape(-1, 28, 28, 1) # 请注意最后这一行,形状要整形成适合卷积网络的输入\n\n\nX_train = read_images('train-images-idx3-ubyte', 60000) # 读取60000张训练图片\nX_test = read_images('./t10k-images-idx3-ubyte', 10000) # 读取10000张测试图片\n\ny_train = keras.utils.to_categorical(y_train, 10) # one hot转码\ny_test = keras.utils.to_categorical(y_test, 10) # one hot转码\n\n# 训练与验证部分\n\nmodel = Sequential()\nmodel.add(\n Conv2D(32, kernel_size=(3, 3), activation='relu',\n input_shape=(28, 28, 1))) # 32核卷积\nmodel.add(MaxPooling2D(pool_size=(2, 2))) # 2*2池化\nmodel.add(Conv2D(64, (3, 3), activation='relu')) # 64核卷积\nmodel.add(MaxPooling2D(pool_size=(2, 2))) # 2*2池化\nmodel.add(Flatten()) # 连接层,卷积和全连接网络的中介\nmodel.add(Dense(128, activation='relu')) # 全连接层\nmodel.add(Dropout(0.5)) # Dropout层\nmodel.add(Dense(10, activation='softmax'))\n\nmodel.compile(\n loss=keras.losses.categorical_crossentropy,\n optimizer=keras.optimizers.Adadelta(),\n metrics=['accuracy']) # 编译\n\nmodel.fit(\n X_train,\n y_train,\n batch_size=128,\n epochs=10,\n verbose=1,\n validation_data=(X_test, y_test)) # 训练\nscore = model.evaluate(X_test, y_test, verbose=0) # 验证\nprint('损失数:', score[0])\nprint('准确率:', score[1])\n","sub_path":"keras_cnn.py","file_name":"keras_cnn.py","file_ext":"py","file_size_in_byte":2405,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"461998957","text":"def letterCasePermutation(self, S: str) -> List[str]:\n \"\"\"\n Given a string S,\n we can transform every letter individually\n to be lowercase or uppercase to create another string.\n\n Return a list of all possible strings we could create.\n\n Examples:\n Input: S = \"a1b2\"\n Output: [\"a1b2\", \"a1B2\", \"A1b2\", \"A1B2\"]\n\n Input: S = \"3z4\"\n Output: [\"3z4\", \"3Z4\"]\n\n Input: S = \"12345\"\n Output: [\"12345\"]\n\n Note:\n S will be a string with length between 1 and 12.\n S will consist only of letters or digits.\n \"\"\"\n perms = [\"\"]\n for c in S:\n if c.isalpha():\n perms = [p + c.lower() for p in perms] + [p + c.upper() for p in perms]\n else:\n perms = [p + c for p in perms]\n\n return perms\n","sub_path":"algo/permutation/letterCase.py","file_name":"letterCase.py","file_ext":"py","file_size_in_byte":761,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"242204207","text":"import collections\nfrom collections import Iterable\n\nimport cv2\nimport numpy\nfrom object_detection.utils.visualization_utils import draw_bounding_box_on_image_array, \\\n STANDARD_COLORS\n\n\nclass DataLabel(object):\n def __init__(self):\n self.data = None\n self.labels = None\n\n\ndef draw_boxes_on_image(\n image, boxes, classes, scores, category_index, min_scores_thresh,\n line_thickness=4):\n\n if not isinstance(min_scores_thresh, Iterable):\n min_scores_thresh_per_class = {\n class_data['name']: min_scores_thresh\n for class_data in category_index.values()\n }\n elif isinstance(min_scores_thresh, dict):\n min_scores_thresh_per_class = {\n class_data['name']: min_scores_thresh['others']\n for class_data in category_index.values()\n }\n min_scores_thresh.pop('others')\n min_scores_thresh_per_class.update(min_scores_thresh)\n else:\n min_scores_thresh_per_class = {\n class_data['name']: 0.5\n for class_data in category_index.values()\n }\n\n drawn_boxes = []\n for i in range(boxes.shape[0]):\n class_name = category_index[classes[i]]['name']\n\n if scores[i] < min_scores_thresh_per_class[class_name]:\n continue\n\n drawn_boxes.append((class_name, boxes[i]))\n\n for class_name, box in drawn_boxes:\n height, width = image.shape[:2]\n ymin, xmin, ymax, xmax = box\n pt1 = pt1_x, pt1_y = (int(xmin*width), int(ymin*height))\n pt2 = pt2_x, pt2_y = (int(xmax*width), int(ymax*height))\n color = list(numpy.random.random(size=3) * 256)\n cv2.rectangle(image, pt1, pt2, color, thickness=2)\n\n font_face = cv2.FONT_HERSHEY_PLAIN\n thickness = 1\n baseline = 0\n ((text_width, text_height), _) = cv2.getTextSize(class_name, font_face, thickness, baseline)\n\n #cv2.rectangle(image, pt1, (pt1_x + text_width, pt1_y + text_height), (255, 255, 255), -1)\n cv2.putText(\n image, class_name, (pt1_x + 2*thickness, pt1_y + text_height + 2*thickness),\n fontFace=font_face, fontScale=1, color=(0, 0, 0), thickness=thickness\n )\n\n return image\n","sub_path":"utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":2211,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"564708067","text":"#\n# Determine messages and constants used in the file-system-protection module\n#\n# Determine the size of the encrypt/decrypt data_block\nDATA_BLOCK_SIZE = 32768\n\n# Determine the value returned by the function\nERROR_CODE = -1\nSUCCESS_CODE = 0\n\nACCESS_DENIED_CODE = 5\n\n# Define operating system platform\nUNKNOWN_PLATFORM = -1\nWINDOWS_PLATFORM = 0\nLINUX_PLATFORM = 1\n\nFILE_NOT_FOUND_CODE = -2\nDIR_NOT_FOUND_CODE = -3\n\nFILE_EXIST_MSG = 'File exist.'\nDIR_EXIST_MSG = 'Directory exist.'\n\nFILE_TYPE = 0\nDIR_TYPE = 1\nREGISTRY_TYPE = 2\n\nTYPE_ENCRYPT_FILE = '.enc'\n\n# Determine the messages encrypt/decrypt file and directory\nPASSWORD_INCORRECT_MSG = 'Error. Password incorrect'\nENCRYPT_FILE_SUCCESS_MSG = 'Done encrypt file.'\nENCRYPT_FILE_ERROR_MSG = '#Error in process encrypt file.'\nDECRYPT_FILE_SUCCESS_MSG = 'Done decrypt file.'\nDECRYPT_FILE_ERROR_MSG = '#Error in process decrypt file.'\nENCRYPT_DIR_SUCCESS_MSG = 'Done encrypt directory.'\nDECRYPT_DIR_SUCCESS_MSG = 'Done decrypt directory.'\n\nCONFIRM_DEL = 0\nSKIP_CODE = 1\nOVERRIDE_CODE = 2\nPASSWORD_INCORRECT_CODE = -4\n\nCONFIRM_DEL_MSG = 'Confirm override.'\nSKIP_OVERRIDE_MSG = 'Skip override.'\n\n\nDEFAULT_PASSWORD = 'bkcs'\n\n# Determine the error value / messages returned by the database query function\nCREATE_DB_ERROR = -4\nINSERT_RECORD_ERROR = -5\nUPDATE_RECORD_ERROR = -6\nDELETE_RECORD_ERROR = -7\n\nCREATE_DB_ERROR_MSG = 'There was an error creating the database.'\nCREATE_DB_SUCCESS_MSG = 'Create success database: '\n\nQUERY_TABLE_DB_ERROR_MSG = 'The error connect to database.'\n\nADD_FILE_MSG = 'The new file add to folder.'\nCHANGE_FILE_MSG = 'File is changed.'\nDELETE_FILE_MSG = 'File is deleted.'\nNOT_CHANGE_FILE_MSG = 'File isn\\'t changed.'\n\n\nFILE_CHECK_TAG = 'file_check'\nFOLDER_CHECK_TAG = 'folder_check'\nREGISTRY_CHECK_TAG = 'windows_registry'\nCHECK_LIST_TAG = 'check_list'\n\n# New sys_check add to database is value 0\nSYS_CHECK_OBJECT_NEW = 0\nSYS_CHECK_OBJECT_OLD = 1\n\nSYS_CHECK_OBJECT_IGNORE = 1\n\nSYS_CHECK_OBJECT_XML_FILE = 'xml'\nSYS_CHECK_OBJECT_CSV_FILE = 'csv'\n\n# Registry\nVALUE_CHANGE = 'Registry value change.'\nVALUE_ADD = 'Registry value add.'\nVALUE_DEL = 'Registry value deleted.'\nKEY_ADD = 'Registry key add.'\nKEY_DEL = 'Registry key deleted.'\n\nPATH_DIR_EVENT_LOG = r\"C:\\Event_Logs\"\n\n\nADD_FILE_ACTION_MSG = 'Create File'\nCHANGE_FILE_ACTION_MSG = 'Modify File'\nDELETE_FILE_ACTION_MSG = 'Delete File'\nDELETE_DIR_ACTION_MSG = 'Delete Dir'\nCHANGE_FILE_ACL = \"Modify ACL\"\nRENAME_FILE_ACTION_MSG = 'Rename'\nMOVE_FILE_ACTION_MSG = 'Move'\nRESTORE_FILE_ACTION_MSG = 'Restore'\nRECYCLE_FILE_ACTION_MSG = 'Recycle'\n\n\n# ----------------------------------- Handle Audit Linux -----------------------------------#\nAUDIT_RULE_LINUX_PATH = \"/etc/audit/rules.d/audit.rules\"\nPATH_AUDIT_LOG = \"/var/log/audit/audit.log\"\n","sub_path":"codes/program_msg.py","file_name":"program_msg.py","file_ext":"py","file_size_in_byte":2761,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"541238840","text":"import sys,time,os,random,cPickle, math\nimport traceback\nimport pygame, thread\nfrom pygame.locals import *\n \n \n \n \n \n# Skeleton class \nclass Skeleton:\n\n\n \n # Initialise pygame, and load the fonts\n def init_pygame(self, w, h): \n pygame.init()\n default_font_name = pygame.font.match_font('bitstreamverasansmono', 'verdana', 'sans')\n if not default_font_name: \n self.default_font_name = pygame.font.get_default_font() \n self.default_font = pygame.font.Font(default_font_name, 36)\n self.screen = pygame.display.set_mode((w,h)) \n #store screen size\n self.w = self.screen.get_width()\n self.h = self.screen.get_height()\n self.centre_x = self.w/2\n self.centre_y = self.h/2\n \n \n \n #initialise any surfaces that are required\n def init_surfaces(self): \n self.draw_buf = pygame.Surface(self.screen.get_size()) \n self.draw_buf.fill((255,255,255))\n \n \n\n \n # init routine, sets up the engine,\n def __init__(self, size=(800,600), draw_fn=None, tick_fn=None, event_fn=None, quit_fn=None): \n # screen size\n self.init_pygame(size[0],size[1])\n self.init_surfaces()\n self.fps = 60 \n self.clock = pygame.time.Clock() \n self.last_frame_time = time.clock()\n self.draw_fn = draw_fn\n self.tick_fn = tick_fn\n self.quit_fn = quit_fn\n self.handle_event = event_fn\n self.looping = True\n \n \n # handles shutdown\n def quit(self):\n if self.quit_fn:\n self.quit_fn()\n self.looping = False\n \n # wrapper for drawing text on screen\n def text(self, buffer, text, color, position): \n textImage = self.default_font.render(text, True, color)\n buffer.blit(textImage, position)\n \n def transparent_text(self, buffer, text, color, position, alpha=255): \n textImage = self.default_font.render(text, False, color)\n textImage.set_alpha(alpha)\n buffer.blit(textImage, position)\n \n \n def screen_text(self, text, position, color=(0,0,0)):\n self.text(self.screen, text, color, position)\n\n \n # this is the redraw code. Add drawing code between the \"LOCK\" and \"END LOCK\" sections\n def flip(self):\n self.screen.blit(self.draw_buf, (0,0)) \n \n if self.draw_fn!=None:\n self.draw_fn(self.screen)\n pygame.display.flip()\n \n \n # Get the array containing the up or down states for each key. For example, key_state[K_UP] is true if the up arrow key is pressed, [K_q] if Q is pressed, etc.\n def check_key_state(self):\n self.key_state = pygame.key.get_pressed()\n \n \n \n #frame loop. Called on every frame. all calculation shpuld be carried out here \n def tick(self): \n self.clock.tick(self.fps) \n self.handle_events() \n delta_t = time.clock() - self.last_frame_time \n self.last_frame_time = time.clock() \n self.check_key_state() \n\n if self.tick_fn!=None:\n self.tick_fn(delta_t)\n self.flip()\n\n \n \n #returns last mouse position\n def get_mouse(self):\n return self.mouse_pos\n \n #Event handlers. These are called as events arrive\n def keydown(self,event):\n return\n \n def keyup(self,event):\n return\n \n def mouseup(self,event):\n return\n \n def mousedown(self,event):\n return\n \n def mousemove(self,event):\n (self.mouse_pos) = event.pos\n return\n \n \n \n #event handling code. Calls the relevant handlers\n def handle_events(self):\n for event in pygame.event.get(): \n if event.type==KEYDOWN:\n \n if event.key==K_ESCAPE:\n self.quit()\n else:\n self.keydown(event)\n \n if event.type==KEYUP:\n self.keyup(event)\n \n if event.type == QUIT:\n self.quit()\n \n elif event.type == MOUSEBUTTONUP:\n self.mouseup(event)\n elif event.type == MOUSEBUTTONDOWN:\n self.mousedown(event)\n elif event.type == MOUSEMOTION: \n self.mousemove(event)\n if self.handle_event:\n self.handle_event(event)\n\n #main loop. Just runs tick until the program exits \n def main_loop(self):\n while self.looping:\n self.tick()\n pygame.quit()\n \n \n\n\n \n\n\n\n ","sub_path":"skeleton.py","file_name":"skeleton.py","file_ext":"py","file_size_in_byte":5092,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"530926602","text":"\"\"\"\nOperator for scraping articles from every organisation.\n\"\"\"\n\nimport json\nimport os\n\nfrom airflow.models import BaseOperator\nfrom airflow.utils.decorators import apply_defaults\n\nfrom scrapy.crawler import CrawlerProcess\nfrom scrapy.utils.project import get_project_settings\nimport scrapy.signals\n\nfrom reach.sentry import report_exception\nfrom reach.scraper.wsf_scraping import feed_storage\nfrom reach.scraper.wsf_scraping.spiders.who_iris_spider import WhoIrisSpider\nfrom reach.scraper.wsf_scraping.spiders.nice_spider import NiceSpider\nfrom reach.scraper.wsf_scraping.spiders.gov_spider import GovSpider\nfrom reach.scraper.wsf_scraping.spiders.msf_spider import MsfSpider\nfrom reach.scraper.wsf_scraping.spiders.unicef_spider import UnicefSpider\nfrom reach.scraper.wsf_scraping.spiders.parliament_spider import ParliamentSpider\nfrom reach.scraper.wsf_scraping.spiders.acme_spider import AcmeSpider\nimport reach.scraper.wsf_scraping.settings\n\n\nSPIDERS = {\n 'who_iris': WhoIrisSpider,\n 'nice': NiceSpider,\n 'gov_uk': GovSpider,\n 'msf': MsfSpider,\n 'unicef': UnicefSpider,\n 'parliament': ParliamentSpider,\n 'acme': AcmeSpider,\n}\n\n\nclass SpiderOperator(BaseOperator):\n \"\"\"\n Pulls data from the dimensions.ai to a bucket in S3.\n\n Args:\n organisation: The organisation to pull documents from.\n \"\"\"\n\n template_fields = ('dst_s3_dir',)\n\n @apply_defaults\n def __init__(self, organisation, dst_s3_dir,\n item_years, item_max, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.organisation = organisation\n self.dst_s3_dir = dst_s3_dir\n self.dst_s3_dir = dst_s3_dir\n\n self.item_count = None\n self.scraper_errors = None\n self.item_max = item_max\n self.item_years = item_years\n\n def on_item_scraped(self, item, response):\n \"\"\" Increments our count of items for reporting/future metrics. \"\"\"\n self.item_count += 1\n\n def on_item_error(self, item, response, failure):\n \"\"\"\n Records Scrapy item_error signals; these fire automatically if\n exceptions occur while saving items in the pipeline.\n \"\"\"\n self.scraper_errors.append(\n ('item_error', item, response, failure)\n )\n\n def on_manifest_storage_error(self, exception):\n \"\"\"\n Records our feed storage's error signal, as would occur when\n exceptions occur while saving the manifest to S3.\n \"\"\"\n self.scraper_errors.append(\n ('manifest_storage_error', exception)\n )\n\n @report_exception\n def execute(self, context):\n # Initialise settings for a limited scraping\n os.environ.setdefault(\n 'SCRAPY_SETTINGS_MODULE',\n 'reach.scraper.wsf_scraping.settings'\n )\n\n if not self.dst_s3_dir.startswith('s3://'):\n raise ValueError('Invalid S3 url: %s' % self.dst_s3_dir)\n\n # This monkey-patching only works because Airflow shells out to\n # a new Python interpreter for every task it runs. It thus *must*\n # remain inside execute(), so other code paths don't touch it.\n reach.scraper.wsf_scraping.settings.MAX_ARTICLE = self.item_max\n reach.scraper.wsf_scraping.settings.WHO_IRIS_YEARS = \\\n self.item_years\n\n reach.scraper.wsf_scraping.settings.FEED_URI = \\\n 'manifest' + self.dst_s3_dir\n\n settings = get_project_settings()\n self.log.info(\n \"scrapy settings: %s\",\n json.dumps(\n {k: v for k, v in settings.items()\n if isinstance(v, (str, int, float, bool))}\n )\n )\n\n process = CrawlerProcess(settings, install_root_handler=False)\n spider = SPIDERS[self.organisation]\n crawler = process.create_crawler(spider)\n\n self.item_count = None\n self.scraper_errors = []\n crawler.signals.connect(\n self.on_item_error,\n signal=scrapy.signals.item_error)\n crawler.signals.connect(\n self.on_manifest_storage_error,\n signal=feed_storage.manifest_storage_error)\n\n process.crawl(crawler) # starts reactor\n process.start() # waits for reactor to finish\n\n if self.scraper_errors:\n scraper_errors = self.scraper_errors # put into local for sentry\n self.log.error(\n 'SpiderOperator: scrapy signaled %d errors:',\n len(scraper_errors)\n )\n for tup in self.scraper_errors:\n self.log.error('DummySpiderOperator: %r', tup)\n raise Exception(\n \"%d errors occurred during scrape\" %\n len(scraper_errors)\n )\n","sub_path":"reach/airflow/tasks/spider_operator.py","file_name":"spider_operator.py","file_ext":"py","file_size_in_byte":4727,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"16612176","text":"#!/usr/bin/python\n\n#Author: Duncan Campbell\n#Written: January 21, 2015\n#Yale University\n#Description: make a combined NYU-MPA dr7 catalogue using the MPA as the base\n\n###packages###\nfrom __future__ import print_function\nimport numpy as np\nimport h5py\nimport custom_utilities as cu\nimport matplotlib.pyplot as plt\nimport sys\n\ndef main():\n \n ###make sure to change these when running in a new enviorment!###\n #location of data directories\n filepath_2 = cu.get_output_path() + 'processed_data/mpa_dr7/'\n filepath_1 = cu.get_output_path() + 'processed_data/NYU_VAGC/'\n #################################################################\n\n catalogue_2 = 'mpa_dr7_unique'\n catalogue_1 = 'nyu_vagc_dr7'\n \n f_1 = h5py.File(filepath_1+catalogue_1+'.hdf5', 'r')\n dset_1 = f_1.get(catalogue_1)\n dset_1 = np.array(dset_1)\n print(dset_1.dtype.names)\n\n f_2 = h5py.File(filepath_2+catalogue_2+'.hdf5', 'r')\n dset_2 = f_2.get(catalogue_2)\n dset_2 = np.array(dset_2)\n print(dset_2.dtype.names)\n \n #open matching files. these are created with the 'match_to_NYU_VAGC.py' script. \n filename = 'matches_into_mpa_vagc.npy'\n match_1 = np.load(filepath_1+'matches/'+filename)\n filename = 'mpa_vagc_matched_to_nyu.npy'\n match_2 = np.load(filepath_1+'matches/'+filename)\n \n \n N_col_1 = len(dset_1.dtype.descr)\n N_col_2 = len(dset_2.dtype.descr[25:]) #number of columns used from NYU VAGC\n \n names_1 = dset_1.dtype.names\n names_2 = dset_2.dtype.names[25:]\n \n dtype = np.dtype(dset_1.dtype.descr + dset_2.dtype.descr[25:])\n \n combined_data = np.recarray((len(dset_1),),dtype=dtype)\n combined_data.fill(-99)\n \n for name in names_1:\n print(name)\n combined_data[name] = dset_1[name]\n \n for name in names_2:\n print(name)\n combined_data[name][match_1] = dset_2[name][match_2]\n \n filename = 'nyu_mpa_vagc_dr7'\n savepath = filepath_1\n f = h5py.File(savepath+filename+'.hdf5', 'w')\n dset = f.create_dataset(filename, data=combined_data)\n\n\nif __name__ == '__main__':\n main()","sub_path":"combine_w_mpa_vagc.py","file_name":"combine_w_mpa_vagc.py","file_ext":"py","file_size_in_byte":2097,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"96718307","text":"import logging\n\nfrom functools import partial\n\nfrom mozilla_version.gecko import FennecVersion\n\nfrom mozapkpublisher.common.apk.history import get_expected_combos, craft_combos_pretty_names\nfrom mozapkpublisher.common.exceptions import BadApk, BadSetOfApks, NotMultiLocaleApk\nfrom mozapkpublisher.common.utils import filter_out_identical_values\n\nlogger = logging.getLogger(__name__)\n\n\n# x86* must have the highest version code. See bug 1338477 for more context.\n_ARCHITECTURE_ORDER_REGARDING_VERSION_CODE = ('armeabi-v7a', 'arm64-v8a', 'x86', 'x86_64')\n\n\nclass ExpectedPackageNamesCheck:\n def __init__(self, expected_product_types):\n self.expected_product_types = expected_product_types\n\n def validate(self, apks_metadata_per_paths):\n types = set([metadata['package_name'] for metadata in apks_metadata_per_paths.values()])\n\n if not types == set(self.expected_product_types):\n raise BadSetOfApks('Expected product types {}, found {}'.format(self.expected_product_types, types))\n logger.info('Expected product types {} found'.format(self.expected_product_types))\n\n\nclass AnyPackageNamesCheck:\n def validate(self, _):\n pass\n\n\ndef cross_check_apks(apks_metadata_per_paths, package_names_check, skip_checks_fennec, skip_check_multiple_locales,\n skip_check_same_locales, skip_check_ordered_version_codes):\n logger.info(\"Checking APKs' metadata and content...\")\n package_names_check.validate(apks_metadata_per_paths)\n\n if not skip_checks_fennec:\n singular_apk_metadata = list(apks_metadata_per_paths.values())[0]\n _check_version_matches_package_name(\n singular_apk_metadata['firefox_version'], singular_apk_metadata['package_name']\n )\n\n _check_all_apks_have_the_same_firefox_version(apks_metadata_per_paths)\n _check_all_apks_have_the_same_build_id(apks_metadata_per_paths)\n _check_all_architectures_and_api_levels_are_present(apks_metadata_per_paths)\n\n if not skip_check_multiple_locales:\n _check_all_apks_are_multi_locales(apks_metadata_per_paths)\n\n if not skip_check_same_locales:\n _check_all_apks_have_the_same_locales(apks_metadata_per_paths)\n\n if not skip_check_ordered_version_codes:\n _check_apks_version_codes_are_correctly_ordered(apks_metadata_per_paths)\n\n logger.info('APKs are sane!')\n\n\ndef _check_piece_of_metadata_is_unique(key, pretty_key, apks_metadata_per_paths):\n all_items = [metadata[key] for metadata in apks_metadata_per_paths.values()]\n unique_items = filter_out_identical_values(all_items)\n\n if not unique_items:\n raise BadSetOfApks('No {} found'.format(key))\n if len(unique_items) > 1:\n raise BadSetOfApks(\"APKs don't have the same {}. Found: {}\".format(pretty_key, unique_items))\n\n logger.info('All APKs have the same {}: {}'.format(pretty_key, unique_items[0]))\n\n\n_check_all_apks_have_the_same_firefox_version = partial(_check_piece_of_metadata_is_unique, 'firefox_version', 'Firefox version')\n_check_all_apks_have_the_same_build_id = partial(_check_piece_of_metadata_is_unique, 'firefox_build_id', 'Firefox BuildID')\n_check_all_apks_have_the_same_locales = partial(_check_piece_of_metadata_is_unique, 'locales', 'locales')\n\n\ndef _check_version_matches_package_name(version, package_name):\n sanitized_version = FennecVersion.parse(version)\n\n if (\n (package_name == 'org.mozilla.firefox' and sanitized_version.is_release) or\n # Due to project Dawn, Nightly is now using the Aurora package name. See bug 1357351.\n (package_name == 'org.mozilla.fennec_aurora' and sanitized_version.is_nightly) or\n (\n # XXX Betas aren't following the regular XX.0bY format. Instead they follow XX.0\n # (which looks like release). Therefore, we can't use sanitized_version.is_beta\n package_name == 'org.mozilla.firefox_beta'\n and sanitized_version.is_release\n and sanitized_version.minor_number == 0\n # We ensure the patch_number is undefined. Calling sanitized_version.patch_number\n # directly raises an (expected) AttributeError\n and getattr(sanitized_version, 'patch_number', None) is None\n )\n ):\n logger.info('Firefox version \"{}\" matches package name \"{}\"'.format(version, package_name))\n\n else:\n raise BadApk('Wrong version number \"{}\" for package name \"{}\"'.format(version, package_name))\n\n\ndef _check_apks_version_codes_are_correctly_ordered(apks_metadata_per_paths):\n architectures_per_version_code = {\n metadata['version_code']: metadata['architecture']\n for metadata in apks_metadata_per_paths.values()\n }\n\n if len(architectures_per_version_code) != len(apks_metadata_per_paths):\n raise BadSetOfApks('Some APKs are sharing the same version code! APKs metadata: {}'.format(\n apks_metadata_per_paths\n ))\n\n sorted_architectures_per_version_code = tuple([\n architectures_per_version_code[version_code]\n for version_code in sorted(architectures_per_version_code.keys())\n ])\n\n previous_index = -1\n for architecture in sorted_architectures_per_version_code:\n index = _ARCHITECTURE_ORDER_REGARDING_VERSION_CODE.index(architecture)\n if index <= previous_index:\n raise BadSetOfApks(\n 'APKs version codes are not correctly ordered. Expected order: {}. Order found: {}. APKs metadata: {}'.format(\n _ARCHITECTURE_ORDER_REGARDING_VERSION_CODE, sorted_architectures_per_version_code, apks_metadata_per_paths\n )\n )\n previous_index = index\n\n logger.info('APKs version codes are correctly ordered: {}'.format(architectures_per_version_code))\n\n\ndef _check_all_apks_are_multi_locales(apks_metadata_per_paths):\n for path, metadata in apks_metadata_per_paths.items():\n locales = metadata['locales']\n\n if not isinstance(locales, tuple):\n raise BadApk('Locale list is not either a tuple. \"{}\" has: {}'.format(path, locales))\n\n number_of_locales = len(locales)\n\n if number_of_locales <= 1:\n raise NotMultiLocaleApk(path, locales)\n\n logger.info('\"{}\" is multilocale.'.format(path))\n\n\ndef _check_all_architectures_and_api_levels_are_present(apks_metadata_per_paths):\n single_metadata = list(apks_metadata_per_paths.values())[0]\n firefox_version = single_metadata['firefox_version']\n\n expected_combos = get_expected_combos(firefox_version)\n\n current_combos = set([\n (metadata['architecture'], metadata['api_level'])\n for metadata in apks_metadata_per_paths.values()\n ])\n\n missing_combos = expected_combos - current_combos\n if missing_combos:\n raise BadSetOfApks('One or several APKs are missing for Firefox {}: {}'.format(\n firefox_version, craft_combos_pretty_names(missing_combos)\n ))\n\n extra_combos = current_combos - expected_combos\n if extra_combos:\n raise BadSetOfApks('One or several APKs are not allowed for Firefox {}: {}. \\\nPlease make sure mozapkpublisher has allowed them to be uploaded.'.format(\n firefox_version, craft_combos_pretty_names(extra_combos)\n ))\n\n logger.info('Every expected APK was found!')\n","sub_path":"mozapkpublisher/common/apk/checker.py","file_name":"checker.py","file_ext":"py","file_size_in_byte":7271,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"338673481","text":"import speech_recognition as sr\nimport pyaudio\nimport os\n\nr = sr.Recognizer()\n\nprint(\"Speck Now\")\ndef speek():\n with sr.Microphone() as source:\n audio = r.listen(source)\n try:\n text = r.recognize_google(audio)\n print(text)\n\n except sr.UnknownValueError:\n print(\"sorry, Unknown Value Error\")\n return\n except sr.RequestError:\n print(\"sorry, Request Error\")\n return\n\nspeek()","sub_path":"speach_to_text.py","file_name":"speach_to_text.py","file_ext":"py","file_size_in_byte":429,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"522090686","text":"# -*- coding: UTF-8 -*-\nimport datetime\nimport logging as lg\n\n\nclass InputDataUser:\n mois = {\n 1: \"Janvier\", 2: \"Février\", 3: \"Mars\", 4: \"Avril\", 5: \"Mai\", 6: \"Juin\", 7: \"Juillet\", 8: \"Août\",\n 9: \"Septembre\", 10: \"Octobre\", 11: \"Novembre\", 12: \"Décembre\"}\n jourToday: int\n\n def __init__(self, depenses=\"0.00\", description=\"\", revenumois=\"0.00\", revenueuro=\"0.00\", categories=\"\"):\n self.moisToday = datetime.date.today().month\n self.jourToday = datetime.date.today().day\n self.anneeToday = datetime.date.today().year\n # Prénom et Nom\n self.user = {\"nom\": None, \"prenom\": None}\n self.depenseEuro = depenses\n self.depenseDescription = description\n self.depenseMois = self.mois[self.moisToday]\n self.revenuMois = revenumois\n self.revenuEuro = revenueuro\n self.categories = categories\n self.revenuesMoisUsers = ()\n self.depensesMoisUsers = ()\n self.dataForDB = {\n 'Users': \"\",\n 'Dates': (str(self.moisToday) + \"/\" + str(self.jourToday) + \"/\" + str(self.anneeToday)),\n 'Mois': \"\", 'Descriptions': \"\", 'Depenses': \"\", 'Revenus': \"\", 'Categories': \"\"\n }\n\n # Users\n @property\n def get_users(self):\n if (self.user[\"nom\"] is None) or (self.user[\"prenom\"] is None):\n lg.warning(\"il n'y a pas nom et prénom\")\n else:\n print(self.user[\"nom\"], self.user[\"prenom\"])\n\n def set_users(self, nom, prenom):\n self.user[\"nom\"] = nom\n self.user[\"prenom\"] = prenom\n\n # Catégorie\n def get_categories(self):\n return self.categories\n\n def set_categories(self, categorie):\n self.categories = categorie\n\n # Dépenses\n def get_depenses(self):\n return self.depenseMois, (self.depenseDescription, self.depenseEuro)\n\n def set_depenses(self, description, depenses):\n self.depenseDescription = description\n self.depenseEuro = depenses\n return self.depenseMois, (self.depenseDescription, self.depenseEuro)\n\n # Revenues\n def get_renenues(self):\n return self.revenuEuro, self.revenuMois\n\n def set_revenues(self, revenueuro):\n self.revenuEuro = revenueuro\n\n def get_revenues_users_mois(self):\n return self.revenuesMoisUsers\n\n def set_revenues_users_mois(self, mois, revenue):\n self.revenuesMoisUsers = (mois, revenue)\n\n def get_depenses_users_mois(self):\n return self.user, self.get_categories(), self.get_depenses()\n\n def set_depenses_users_mois(self, user, categorie, mois, description, depense):\n self.depensesMoisUsers = (user, categorie, mois, description, depense)\n\n @property\n def add_data_for_db(self):\n\n self.dataForDB['Users'] = self.user\n self.dataForDB['Descriptions'] = self.depenseDescription\n self.dataForDB['Depenses'] = self.depenseEuro\n self.dataForDB['Mois'] = self.depenseMois\n self.dataForDB['Revenus'] = self.revenuEuro\n self.dataForDB['Categories'] = self.categories\n\n return self.dataForDB\n\n\ndef main():\n teste = InputDataUser()\n teste.set_users(\"Nom Teste\", \"Prénom Teste\")\n teste.set_categories(\"Catégorie Teste\")\n teste.set_depenses(\"Description Teste\", \"22.50\")\n teste.get_users\n print(teste.get_depenses())\n print(teste.get_depenses_users_mois())\n\n print(\"\\n\", teste.add_data_for_db)\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"Package/input_data.py","file_name":"input_data.py","file_ext":"py","file_size_in_byte":3445,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"649099607","text":"#!/usr/bin/python3\n\nimport helper_functions as hf\nimport RPi.GPIO as GPIO\nimport can\nimport time\nimport os\nimport queue\nfrom threading import Thread\nfrom datetime import datetime\nfrom gps3 import gps3\n\n# initial Raspi setup for CAN Shield interfacing\nled = 22\nGPIO.setmode(GPIO.BCM)\nGPIO.setwarnings(False)\nGPIO.setup(led,GPIO.OUT)\nGPIO.output(led,True)\n\n# configure Gps\ngpsd_socket = gps3.GPSDSocket()\ndata_stream = gps3.DataStream()\ngpsd_socket.connect()\ngpsd_socket.watch()\n\n# Bring up can0 interface at 500kbps\nos.system(\"sudo /sbin/ip link set can0 up type can bitrate 500000\")\ntime.sleep(0.1)\n\ntry:\n bus = can.interface.Bus(channel='can0', bustype='socketcan_native')\nexcept OSError:\n print('Cannot find PiCAN board.')\n GPIO.output(led,False)\n exit()\n\ntemperature = 0\nrpm = 0\nspeed = 0\nthrottle = 0\ndistance = 0\ndistance_total = 0\nrpi_speed_epoch_time = 0\ntime2 = 0\ntime1 = 0\nvspeed2 = 0\nvspeed1 = 0\ncurr_lat = 0\nprev_lat = 0\ncurr_lon = 0\nprev_lon = 0\nfirst_time12 = True\nlogged_data = ''\nlogged_data_can_msg = ''\n\ncount = 0\nsp_count = 0\ntime_spent_at_stop = 0.0\ntime_start_at_stop = 0.0\n#file_count = 0\noutfile = 0\noutfile_name = 0\n############################################################\nfile_open = False\n############################################################\nfile_name = ''\n\nDEPOT_BEGIN = True\nSTARTED_FROM_DEPOT = False # what is the purpose of this flag?\nSTARTED_FROM_ROUTE = True\nRETURN_TO_DEPOT = False\nCIRCULATOR = True\nFIRST_TIME_START = True\nNEW_DATA_START_LOC = False\nNEW_DATA_STOP_LOC = False\n\n# Main control starts\ntry:\n while True:\n GPIO.output(led,True)\n # Send Engine RPM request\n msg = can.Message(arbitration_id=hf.PID_REQUEST,data=[0x02,0x01,hf.ENGINE_RPM,0x00,0x00,0x00,0x00,0x00],extended_id=False)\n bus.send(msg)\n rpm_timeStamptx = datetime.now().strftime('%H:%M:%S.%f')\n #print('sent RPM mesg')\n #time.sleep(0.01)\n # waiting for RPM\n #print('waiting for rpm')\n rpm_not_rx = True\n while rpm_not_rx:\n #print('waiting for rpm')\n message = bus.recv()\n logged_data_can_msg = str(message) + '\\n'\n #print(str(message))\n if message.arbitration_id == hf.PID_REPLY and message.data[2] == hf.ENGINE_RPM:\n rpm_timeStamp = datetime.now().strftime('%H:%M:%S.%f')\n rpm = round(((message.data[3]*256) + message.data[4])/4)\n rpm_not_rx = False\n #print('rpm recieved')\n\n # Send Vehicle speed request\n #time.sleep(0.01)\n msg = can.Message(arbitration_id=hf.PID_REQUEST,data=[0x02,0x01,hf.VEHICLE_SPEED,0x00,0x00,0x00,0x00,0x00],extended_id=False)\n bus.send(msg)\n speed_timeStamptx = datetime.now().strftime('%H:%M:%S.%f')\n #print('sent Speed mesg')\n #time.sleep(0.01)\n # waiting for Speed\n #print('waiting for speed')\n speed_not_rx = True\n while speed_not_rx:\n #print('waiting for speed')\n message = bus.recv()\n logged_data_can_msg += str(message) + '\\n'\n #print(str(message))\n if message.arbitration_id == hf.PID_REPLY and message.data[2] == hf.VEHICLE_SPEED:\n rpi_time = time.time()\n speed_timeStamp = datetime.now().strftime('%H:%M:%S.%f')\n speed = message.data[3]\n vspeed2 = speed\n time2 = message.timestamp\n speed_not_rx = False\n #print('speed recieved')\n\n # Send Throttle position request\n #time.sleep(0.01)\n msg = can.Message(arbitration_id=hf.PID_REQUEST,data=[0x02,0x01,hf.THROTTLE,0x00,0x00,0x00,0x00,0x00],extended_id=False)\n bus.send(msg)\n throttle_timeStamptx = datetime.now().strftime('%H:%M:%S.%f')\n #print('sent throttle mesg')\n #time.sleep(0.01)\n # waiting for throttle\n #print('waiting for throttle')\n throttle_not_rx = True\n while throttle_not_rx:\n #print('waiting for throttle')\n message = bus.recv()\n logged_data_can_msg += str(message) + '\\n'\n #print(str(message))\n if message.arbitration_id == hf.PID_REPLY and message.data[2] == hf.THROTTLE:\n throttle_timeStamp = datetime.now().strftime('%H:%M:%S.%f')\n throttle = round((message.data[3]*100)/255)\n throttle_not_rx = False\n time.sleep(0.01)\n #print('throttle recieved')\n\n # End transmission\n GPIO.output(led,False)\n\n #time.sleep(0.2)\n\n logged_data = rpm_timeStamptx + ', ' + rpm_timeStamp + ', {0:d}, '.format(rpm) + speed_timeStamptx + ', ' + speed_timeStamp + ', {0:f}, '.format(time2) + '{0:f}, '.format(rpi_time) + '{0:d}, '.format(speed) + throttle_timeStamptx + ', ' + throttle_timeStamp + ', {0:d}, '.format(throttle)\n\n # calculate distance\n if first_time12:\n time1 = time2\n vspeed1 = vspeed2\n first_time12 = False\n # convert speed from km/h to m/s\n vspeed1 = vspeed1 * 5 / 18\n vspeed2 = vspeed2 * 5 / 18\n distance += (vspeed2 + vspeed1)*(time2 - time1)/2\n distance_total += (vspeed2 + vspeed1)*(time2 - time1)/2\n vspeed1 = vspeed2\n time1 = time2\n\n #print('exited CAN Rx LOOP')\n\n # read GPS data\n for new_data in gpsd_socket:\n #print('received GPS')\n if new_data:\n #print('recieved valid GPS data')\n data_stream.unpack(new_data)\n curr_lat = data_stream.TPV['lat']\n if curr_lat == 'n/a':\n curr_lat = 0;\n curr_lon = data_stream.TPV['lon']\n if curr_lon == 'n/a':\n curr_lon = 0;\n curr_lat_temp = curr_lat\n curr_lon_temp = curr_lon\n if prev_lat == curr_lat and prev_lon == curr_lon:\n curr_lat_temp = 0\n curr_lon_temp = 0\n else:\n prev_lat = curr_lat\n prev_lon = curr_lon\n logged_data += str(curr_lat_temp) + ', ' + str(curr_lon_temp) + ', ' + str(data_stream.TPV['time'] + ', ') + datetime.now().strftime('%H:%M:%S.%f')\n break\n else:\n continue\n\n logged_data += ', {0:d}, {1:f}, {2:f}'.format(count,distance_total,distance)\n\n if not file_open:\n file_name = 'Documents/logs/log_DOJ_stopnwait2_' + str(datetime.now()) + '.csv'\n file_name_can_msg = 'Documents/logs/log_DOJ_stopnwait2_can_msg_' + str(datetime.now()) + '.csv'\n outfile = open(file_name,'w+')\n outfile_can_msg = open(file_name_can_msg, 'w+')\n print('Logging data timestamps...')\n file_open = True\n if file_open:\n print(logged_data,file = outfile)\n print(logged_data_can_msg,file = outfile_can_msg)\n\n count += 1\n #print(logged_data)\n\nexcept KeyboardInterrupt:\n #Catch keyboard interrupt\n GPIO.output(led,False)\n #if file_open:\n #outfile.close()\n # close CAN interface\n os.system(\"sudo /sbin/ip link set can0 down\")\n print('\\n\\rKeyboard interrtupt')\n","sub_path":"Documents/smart_car_old_scripts/python_test_scripts/main_control_py3_test_stopnwait2.py","file_name":"main_control_py3_test_stopnwait2.py","file_ext":"py","file_size_in_byte":7297,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"405911443","text":"import urllib, urllib2, json\nfrom pygeocoder import Geocoder\n\n\ndef decode_address_to_coordinates(address):\n params = {\n 'address': address,\n 'sensor': 'false',\n }\n url = 'http://maps.google.com/maps/api/geocode/json?' + urllib.urlencode(params)\n response = urllib2.urlopen(url)\n result = json.load(response)\n try:\n return result['results'][0]['geometry']['location']\n except:\n return None\n\n\ndef complete_result(address):\n params = {\n 'address': address,\n 'sensor': 'false',\n }\n url = 'http://maps.google.com/maps/api/geocode/json?' + urllib.urlencode(params)\n response = urllib2.urlopen(url)\n result = json.load(response)\n try:\n return result['results'][0]\n except:\n return None\n\n\ndef get_zipcode(address):\n try:\n result = Geocoder.geocode(address)\n isValid = result.valid_address\n print(isValid, 'if valid or not')\n if isValid:\n print(address, ' is valid address.')\n else:\n print(address, ' is invalid address')\n\n print('postal code is ', result[0].postal_code)\n return result[0].postal_code\n except:\n return None\n\n\ndef get_address_using_coordinates(latitude, longitude):\n try:\n result = Geocoder.reverse_geocode(latitude, longitude)\n return result\n except:\n return None\n\n# print decode_address_to_coordinates(\"Toronto\")\n# print decode_address_to_coordinates(\"Birmingham, AL\")\n#print decode_result('Toronto')\n#print decode_result('Birmingham, AL')\n#print get_zipcode(\"451-499 24th Street North, Birmingham, AL 35203, USA\")\n#print get_zipcode(\"Birmingham, AL\")\n#address_la = decode_address_to_coordinates(\"Birmingham, AL\")\n#print address_la\n\n#result = get_address_using_coordinates(address_la['lat'],address_la['lng'])\n#print result[0]\n#print result[1]\n#print result[2]\n#print result[3]\n#print result[4]\n#print result[5]\n#print result[6]\n#print result[7]\n#print result.country__short_name\n#print result.postal_code\n#print result.street_number\n#print result.route\n#print result.administrative_area_level_1\n#print result.city\n#print result.state\n#print result.state__short_name\n#print result.country\n#print result.formatted_address\n#print result.count\n#print result.country__short_name\n#print get_address_using_coordinates(address_co[0]['])\n","sub_path":"processWords/python_scripts/address_coordinates.py","file_name":"address_coordinates.py","file_ext":"py","file_size_in_byte":2596,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"128476428","text":"import io\nimport os\nfrom distutils.file_util import copy_file\nfrom setuptools import setup, find_packages\n\n\nbase_gh_url = 'https://github.com/sowemail/'\ndir_path = os.path.abspath(os.path.dirname(__file__))\nreadme_path = os.path.join(dir_path, 'README.rst')\nversion_path = os.path.join(dir_path, 'VERSION.txt')\n\nwith io.open(readme_path, encoding='utf-8') as readme_file:\n readme = readme_file.read()\nwith io.open(version_path, encoding='utf-8') as version_file:\n version = version_file.read().strip()\n\ncopy_file(version_path,\n os.path.join(dir_path, 'sowemail', 'VERSION.txt'),\n verbose=False)\n\n\ndef parse_requirements(filename):\n \"\"\"load requirements from a pip requirements file\"\"\"\n lineiter = (line.strip() for line in open(filename))\n return [line for line in lineiter if line and\n (not line.startswith(\"#\") and not line.startswith('-'))]\n\n\nsetup(\n name='sowemail',\n version=version,\n author='Fourat ZOUARI',\n author_email='opensource@sowemail.com',\n url='{}sowemail-python'.format(base_gh_url),\n download_url='{}sowemail-python/tarball/{}'.format(base_gh_url, version),\n packages=['sowemail'],\n include_package_data=True,\n install_requires=parse_requirements('install-requirements.pip'),\n tests_require=parse_requirements('test-requirements.pip'),\n license='MIT',\n description='SoWeMail library for Python 3',\n long_description_content_type='text/x-rst',\n long_description=readme,\n keywords=[\n 'API',\n 'SOWEMAIL'],\n python_requires='>=3.6.*',\n classifiers=[\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n 'Programming Language :: Python :: 3.8',\n ]\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1775,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"603741428","text":"from django.conf.urls import patterns, url\nfrom web import views\n\nurlpatterns = patterns(\n '',\n url(r'groups/home/', views.groups_home),\n url(r'groups/ping/', views.groups_ping),\n url(r'groups/events/', views.groups_events),\n url(r'login/', views.homepage),\n url(r'^$', views.landingpage),\n)\n","sub_path":"partyup/web/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":310,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"465701161","text":"import axelrod as axl\nimport typing\nimport sys, inspect\nimport importlib\nimport logging\nimport matplotlib.pyplot as plt\n\ndef do_axelrod_tournament(players: list, output_file_name: str,\n noise: float = 0, noise_bias: bool = False,\n prob_end: float = 0, seed: int = 0) -> None:\n \"\"\"\n Do an Axelrod tournament with the given arguments and create the boxplot, payoff matrix, and winplot of the tournament\n\n Arguments:\n players: list of Player playing in the Axelrod tournament. Necessary argument\n output_file_name: the output name for the boxplot, payoff matrix, and the winplot for the current Axelrod tournament\n noise: percentage of noise (mistake, flipping a decision) happening in each player's decision\n noise_bias: if True, only allows mistake that changes cooperation to defection\n prob_end: The probability of ending a match between a pair of players after each game\n seed: random seed for the current tournament (for reproducibility)\n \"\"\"\n axl.seed(seed)\n tournament = axl.Tournament(players, noise=noise, noise_bias=noise_bias, prob_end=prob_end) # Create a tournament\n results = tournament.play() # Play the tournament\n plot = axl.Plot(results)\n plot.boxplot()\n plt.savefig('boxplot_' + output_file_name, format='png')\n plot.winplot()\n plt.savefig('winplot_' + output_file_name, format='png')\n plot.payoff()\n plt.savefig('payoff_' + output_file_name, format='png')\n plt.close('all')\n\ndef do_case_tournament(players: list, output_file_name: str,\n turns: int = axl.DEFAULT_TURNS, prob_end: float = None,\n maximum_round: int = 20, noise: float = 0,\n noise_bias: bool = False, replace_amount: int = 1,\n game: axl.Game = None, seed: int = 0) -> None:\n \"\"\"\n Do a Case simulation with the given arguments and create the population graph of the simulation\n\n Arguments:\n players: list of Player playing in the Case simulation\n output_file_name: the output name for the population graph for the current simulation\n turns: the maximum number of turns for each Match between a pair of players\n prob_end: The probability of ending a match between a pair of players after each game\n noise: percentage of noise (mistake, flipping a decision) happening in each player's decision\n noise_bias: if True, only allows mistake that changes cooperation to defection\n replace_amount: the amount of players replaced after each simulation round\n game: iterated prisoner's dilemma game parameter for the simulation\n seed: random seed for the current tournament (for reproducibility)\n \"\"\"\n axl.seed(seed)\n tournament = axl.CaseProcess(players, turns=turns, maximum_round=maximum_round,\n replace_amount=replace_amount, noise=noise, noise_bias=noise_bias, game=game,\n prob_end=prob_end)\n tournament.play()\n tournament.populations_plot()\n plt.savefig(output_file_name, format='png', bbox_inches='tight')\n plt.close()\n\n# The list of agents playing in the Axelrod's first tournament, except Graaskamp\naxl_first_players = [\n axl.TitForTat(),\n axl.TidemanAndChieruzzi(),\n axl.Nydegger(),\n axl.Grofman(),\n axl.Shubik(),\n axl.SteinAndRapoport(),\n axl.Grudger(),\n axl.Davis(),\n axl.RevisedDowning(),\n axl.Feld(),\n axl.Joss(),\n axl.Tullock(),\n axl.UnnamedStrategy(),\n axl.Random()\n]\n\n# The list of agents playing in the Axelrod's second tournament. Incomplete\naxl_second_players = [\n axl.Black(),\n axl.Borufsen(),\n axl.Cave(),\n axl.Champion(),\n axl.Colbert(),\n axl.Eatherley(),\n axl.Getzler(),\n axl.Gladstein(),\n axl.GoByMajority(),\n axl.GraaskampKatzen(),\n axl.Harrington(),\n axl.Kluepfel(),\n axl.Leyvraz(),\n axl.Mikkelson(),\n axl.MoreGrofman(),\n axl.MoreTidemanAndChieruzzi(),\n axl.RichardHufford(),\n axl.Tester(),\n axl.Tranquilizer(),\n axl.Weiner(),\n axl.White(),\n axl.WmAdams(),\n axl.Yamachi()\n]\n\n# The list of agents playing in the Stewart-Plotkin's tournament.\nsteward_plotkin_players = [\n axl.ZDExtort2(),\n axl.HardGoByMajority(),\n axl.HardTitForTat(),\n axl.HardTitFor2Tats(),\n axl.GTFT(),\n axl.ZDGTFT2(),\n axl.Calculator(),\n axl.Prober(),\n axl.Prober2(),\n axl.Prober3(),\n axl.HardProber(),\n axl.NaiveProber()\n]\n\n# The list of agents playing in the Case's simulation\ncase_players = [\n axl.Cooperator(),\n axl.Defector(),\n axl.TitForTat(),\n axl.Grudger(),\n axl.Detective(),\n axl.TitFor2Tats(),\n axl.WinStayLoseShift(),\n axl.Random()\n]\n\n# The list of agents who won or survived the most based on the previous tournaments and simulations\nbest_players = [\n axl.SteinAndRapoport(),\n axl.Grudger(),\n axl.RevisedDowning(),\n axl.TitForTat(),\n axl.Davis(),\n axl.Nydegger(),\n axl.Cave(),\n axl.Tranquilizer(),\n axl.White(),\n axl.Eatherley(),\n axl.Champion(),\n axl.Harrington(),\n axl.Defector(),\n axl.Random(),\n axl.WinStayLoseShift(),\n axl.GTFT(),\n axl.ZDGTFT2(),\n axl.TitFor2Tats(),\n axl.NaiveProber(),\n axl.HardProber()\n]\n\n# Seed list. Contains twenty different seeds\nseeds = [1, 771923, 1143728, 291358764, 901236547, 4750670, 511161, 603276, 83281327, 34293471, 918273645, 135792468, 243165978, 9100021, 43238133, 0, 19192831, 5665363, 2231145, 123456]\n\nplayers = []\nplayer_name = ''\nfor i in range(5):\n if i == 0:\n players = axl_first_players\n player_name = 'axelrod_first'\n elif i == 1:\n players = axl_first_players + axl_second_players\n player_name = 'axelrod_second'\n elif i == 2:\n players = case_players\n player_name = 'case'\n elif i == 3:\n players = steward_plotkin_players + case_players\n player_name = 'steward_plotkin'\n else:\n players = best_players\n player_name = 'best'\n for seed in seeds:\n do_axelrod_tournament(players, 'tournament_axelrod_players_' + player_name + '_param_turns_200_prob_end_0.005_seed_' + str(seed) + '.png', seed=seed, prob_end=0.005)\n do_axelrod_tournament(players, 'tournament_axelrod_players_' + player_name + '_param_turns_200_prob_end_0.005_noise_0.05_seed_' + str(seed) + '.png', seed=seed, prob_end=0.005, noise=0.05)\n do_axelrod_tournament(players, 'tournament_axelrod_players_' + player_name + '_param_turns_200_prob_end_0.005_noise_0.05_biasnoise_seed_' + str(seed) + '.png', seed=seed, prob_end=0.005, noise=0.05, noise_bias=True)\n do_axelrod_tournament(players, 'tournament_axelrod_players_' + player_name + '_param_turns_200_prob_end_0.005_noise_0.2_seed_' + str(seed) + '.png', seed=seed, noise=0.2, prob_end=0.005)\n do_axelrod_tournament(players, 'tournament_axelrod_players_' + player_name + '_param_turns_200_prob_end_0.005_noise_0.2_biasnoise_seed_' + str(seed) + '.png', seed=seed, noise=0.2, prob_end=0.005, noise_bias=True)\n do_case_tournament(players, 'tournament_case_players_' + player_name + '_param_turns_200_seed_' + str(seed) + '.png', turns=200, maximum_round=50, seed=seed, prob_end=0.005)\n do_case_tournament(players, 'tournament_case_players_' + player_name + '_param_turns_200_round_50_noise_0.05_seed_' + str(seed) + '.png', turns=200, maximum_round=50, noise=0.05, seed=seed, prob_end=0.005)\n do_case_tournament(players, 'tournament_case_players_' + player_name + '_param_turns_200_round_50_noise_0.2_seed_' + str(seed) + '.png', turns=200, maximum_round=50, noise=0.2, seed=seed, prob_end=0.005)\n do_case_tournament(players, 'tournament_case_players_' + player_name + '_param_turns_200_round_50_noise_0.05_biasnoise_seed_' + str(seed) + '.png', turns=200, maximum_round=50, noise=0.05, noise_bias=True, seed=seed, prob_end=0.005)\n do_case_tournament(players, 'tournament_case_players_' + player_name + '_param_turns_200_round_50_noise_0.2_biasnoise_seed_' + str(seed) + '.png', turns=200, maximum_round=50, noise=0.2, noise_bias=True, seed=seed, prob_end=0.005)","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":8125,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"84552861","text":"from django.db import models\nfrom django.urls import reverse\nimport uuid\n\n# Create your models here.\n\n\nclass Genre(models.Model):\n '''\n model representing book genre\n '''\n name=models.CharField(max_length=100,help_text='Enter a book genre')\n\n def __str__(self):\n '''\n string representing model object\n '''\n return self.name\n\n\nclass Book(models.Model):\n '''\n model representing a book (not a specific copy)\n '''\n\n title=models.CharField(max_length=100)\n author=models.ForeignKey('Author',on_delete=models.SET_NULL,null=True) #assuming that each book will have only one author\n summary=models.TextField(max_length=1000,help_text='Enter a brief description of the book')\n isbn=models.CharField('ISBN',max_length=13,help_text='Enter ISBN number')\n\n genre=models.ManyToManyField(Genre,help_text='Select an appropriate genre for this book')\n\n\n def display_genre(self):\n '''\n Create a string for the genre. This is required to display genre in admin\n\n '''\n return ', '.join(genre.name for genre in self.genre.all()[:3])\n\n display_genre.short_description='Genre'\n\n\n\n def __str__(self):\n '''\n string representing model object\n '''\n return self.title\n\n def get_absolute_url(self):\n '''\n returns url to access a detail record for this book\n '''\n return reverse('book-detail',args=[str(self.id)])\n\n\nclass BookInstance(models.Model):\n '''\n model representing a specific copy of a book that can be borrowed from the library\n '''\n id=models.UUIDField(primary_key=True,default=uuid.uuid4, help_text='Unique ID for this particular book across the whole library')\n book=models.ForeignKey(Book,on_delete=models.SET_NULL,null=True)\n imprint=models.CharField(max_length=200)\n language=models.ForeignKey('Language',on_delete=models.SET_NULL,null=True)\n due_back=models.DateField(null=True,blank=True)\n\n\n LOAN_STATUS=(('m','Maintenance'),('o','On loan'),('a','Available'),('r','Reserved'),)\n\n status=models.CharField(max_length=1,choices=LOAN_STATUS,blank=True,default='m',help_text='Book availability',)\n\n class Meta:\n ordering=['due_back']\n\n def __str__(self):\n '''\n string representing model object\n '''\n return f'{self.id} ({self.book.title})'\n\n\n\nclass Author(models.Model):\n '''\n model representing an author\n '''\n\n first_name=models.CharField(max_length=100)\n last_name=models.CharField(max_length=100)\n date_of_birth=models.DateField(null=True,blank=True)\n date_of_death=models.DateField('Died',null=True,blank=True)\n\n class Meta:\n ordering=['last_name','first_name']\n\n def get_absolute_url(self):\n '''\n returns the url to access a particular author instance\n '''\n return reverse('author-detail',args=[str(self.id)])\n\n\n def __str__(self):\n '''\n string representing model object\n '''\n return f'{self.last_name}, {self.first_name}'\n\n\n\nclass Language(models.Model):\n '''\n model representing a language in which a book might be available\n '''\n name=models.CharField(max_length=100,help_text='Enter script language')\n\n def __str__(self):\n '''\n string representing model object\n '''\n return self.name\n","sub_path":"locallibrary/catalog/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":3400,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"431986040","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# In[2]:\n\n\n# import libraries\nimport numpy as np\nimport pandas as pd\nimport tensorflow as tf\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.preprocessing import LabelEncoder\nimport keras\nfrom keras import layers\nfrom keras import models\nfrom keras.layers import Dropout\nfrom keras.optimizers import SGD\nfrom keras.regularizers import l1\nfrom keras import metrics\n\n\n# In[3]:\n\n\n# load 'feat.npy'\nfeat = np.load('feat.npy',allow_pickle=True)\n\n\n# In[4]:\n\n\n# load 'path.npy'\npath = np.load('path.npy', allow_pickle=True)\n\n\n# In[5]:\n\n\n# load 'train.csv'\ntrain = pd.read_csv('train.csv')\n\n\n# In[6]:\n\n\n# load 'test.csv'\ntest = pd.read_csv('test.csv')\n\n\n# In[7]:\n\n\n# create dictionary using path values as keys and feat values as values\ndictionary = {}\nfor key,val in zip(path,feat):\n dictionary[key] = val\n\n\n# In[8]:\n\n\n# match train set with dictionary's items\ny_big_train = []\nx_big_train = []\nfor i in train.values:\n if i[0] in dictionary.keys():\n y_big_train.append(i[1])\n x_big_train.append(dictionary[i[0]])\ny_big_train = np.asarray(y_big_train)\nx_big_train = np.asarray(x_big_train)\n\n\n# In[9]:\n\n\n# match test set with dictionary's items\nx_test = []\nfor i in test.values:\n if i[0] in dictionary.keys():\n x_test.append(dictionary[i[0]])\nx_test = np.asarray(x_test)\n\n\n# In[10]:\n\n\n# applying the StandardScaler from sklearn to preprocess the train - test data sets\nscaler = StandardScaler()\n\nfor i in range(94824):\n x_big_train[i] = scaler.fit_transform(x_big_train[i])\n \nfor i in range(11005):\n x_test[i] = scaler.fit_transform(x_test[i]) \n\n\n# In[11]:\n\n\n# checking the highest length of the first dimension of the train data\ncounter=0\nfor i in range(x_big_train.shape[0]):\n if x_big_train[i].shape[0] > counter:\n counter = x_big_train[i].shape[0]\n\n#Create a zero matrix for the input of train data\ntrain_data = np.zeros((94824,99,13))\n#Concatenating the values from the train data to the zero matrix\nfor i in range(94824):\n for j in range(x_big_train[i].shape[0]):\n for z in range(x_big_train[i].shape[1]):\n train_data[i][j][z] = train_data[i][j][z] + x_big_train[i][j][z]\n\n\n# In[12]:\n\n\n# checking the highest length of the first dimension of the test data\ncounter=0\nfor i in range(x_test.shape[0]):\n if x_test[i].shape[0] > counter:\n counter = x_test[i].shape[0] \n\n#Create a zero matrix for the input of spoken data\ntest_data = np.zeros((11005,99,13))\n#Concatenating the values from the spoken data to the zero matrix\nfor i in range(11005):\n for j in range(x_test[i].shape[0]):\n for z in range(x_test[i].shape[1]):\n test_data[i][j][z] = test_data[i][j][z] + x_test[i][j][z]\n\n\n# In[13]:\n\n\n# train - validation sets split\nx_train, x_val, y_train, y_val = train_test_split(train_data, y_big_train, test_size=0.2, random_state=666, shuffle=True)\n\n\n# In[14]:\n\n\n# transform the labels for the processing of the model\nlb = LabelEncoder()\nbig_train_labels = lb.fit_transform(y_big_train)\ntrain_labels = lb.fit_transform(y_train)\nval_labels = lb.fit_transform(y_val)\n\n\n# In[15]:\n\n\n# transform the above labels to binary in order to fit them in the network\nbig_train_labels = tf.keras.utils.to_categorical(big_train_labels, num_classes=None, dtype='float32')\ntrain_labels = tf.keras.utils.to_categorical(train_labels, num_classes=None, dtype='float32')\nval_labels = tf.keras.utils.to_categorical(val_labels, num_classes=None, dtype='float32')\n\n\n# In[16]:\n\n\n# apply the 1-D Convolutional Neural Network\nn_timesteps = x_train.shape[1]\nn_features = x_train.shape[2] \nn_outputs = 35\n\nmodel = models.Sequential()\nmodel.add(layers.Conv1D(filters=100, kernel_size=10, activation='relu', input_shape=(n_timesteps,n_features)))\nmodel.add(layers.MaxPooling1D())\nmodel.add(layers.BatchNormalization())\nmodel.add(layers.Conv1D(100, 10, activation='relu'))\nmodel.add(layers.MaxPooling1D())\nmodel.add(layers.BatchNormalization())\nmodel.add(layers.Conv1D(100, 10, activation='relu'))\nmodel.add(layers.MaxPooling1D())\nmodel.add(layers.BatchNormalization())\nmodel.add(layers.Flatten())\nmodel.add(layers.Dropout(0.5))\nmodel.add(layers.Dropout(0.25))\nmodel.add(layers.Dense(n_outputs, activation='softmax'))\nmodel.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])\nprint(model.summary())\n\n# In[17]:\n\n\n# fit using train set\nmodel.fit(x_train, train_labels, epochs=20)\n\n\n# In[18]:\n\n\n# evaluate using validation set\naccuracy = model.evaluate(x_val, val_labels, batch_size=32)\nprint(model.metrics_names)\nprint(accuracy)\n\n\n# In[19]:\n\n\n# predict using test set\ntest = model.predict(test_data)\ntest = np.argmax(np.round(test),axis=1)\ntest = lb.inverse_transform(test)\n\n\n# In[20]:\n\n\n# add the words' column to test set - create csv file\nnew = pd.read_csv('test.csv', delimiter=',')\nnew['word'] = test\n\n\n# In[21]:\n\n\n# save the csv file\nnew = new.to_csv(r'C:\\Users\\Gebruiker\\Downloads\\ml_challenge\\result.csv', index=False)\n\n","sub_path":"code.py","file_name":"code.py","file_ext":"py","file_size_in_byte":5029,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"220010223","text":"# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /Users/jgorset/Code/python/fandjango/fandjango/middleware.py\n# Compiled at: 2015-12-28 07:16:58\nfrom django.conf import settings\nfrom django.http import QueryDict, HttpResponseRedirect\nfrom datetime import timedelta\nfrom urlparse import parse_qs\nfrom django.core.exceptions import ImproperlyConfigured\nfrom fandjango.views import authorize_application, authorization_denied\nfrom fandjango.models import Facebook, User, OAuthToken\nfrom fandjango.settings import FACEBOOK_APPLICATION_SECRET_KEY, FACEBOOK_APPLICATION_ID, FANDJANGO_CACHE_SIGNED_REQUEST, DISABLED_PATHS, ENABLED_PATHS\nfrom fandjango.utils import is_disabled_path, is_enabled_path, get_full_path, authorization_denied_view, get_post_authorization_redirect_url\nfrom facepy import SignedRequest, GraphAPI\ntry:\n from django.utils.timezone import now\nexcept ImportError:\n from datetime import datetime\n\n def now():\n return datetime.now()\n\n\nfrom dateutil.tz import tzlocal\n\nclass BaseMiddleware:\n\n def is_valid_path(self, request):\n if ENABLED_PATHS and DISABLED_PATHS:\n raise ImproperlyConfigured('You may configure either FANDJANGO_ENABLED_PATHS or FANDJANGO_DISABLED_PATHS, but not both.')\n if DISABLED_PATHS and is_disabled_path(request.path):\n return False\n if ENABLED_PATHS and not is_enabled_path(request.path):\n return False\n return True\n\n def is_access_denied(self, request):\n return 'error' in request.GET and request.GET['error'] == 'access_denied'\n\n\nclass FacebookMiddleware(BaseMiddleware):\n \"\"\"Middleware for Facebook canvas applications.\"\"\"\n\n def process_request(self, request):\n \"\"\"Process the signed request.\"\"\"\n if hasattr(request, 'facebook') and request.facebook:\n return\n request.facebook = False\n if not self.is_valid_path(request):\n return\n if self.is_access_denied(request):\n return authorization_denied_view(request)\n if 'signed_request' not in request.REQUEST and 'signed_request' not in request.COOKIES:\n return\n if request.method == 'POST' and 'signed_request' in request.POST:\n request.POST = QueryDict('')\n request.method = 'GET'\n request.facebook = Facebook()\n try:\n request.facebook.signed_request = SignedRequest(signed_request=request.REQUEST.get('signed_request') or request.COOKIES.get('signed_request'), application_secret_key=FACEBOOK_APPLICATION_SECRET_KEY)\n except SignedRequest.Error:\n request.facebook = False\n\n if request.facebook and request.facebook.signed_request.user.has_authorized_application and not request.facebook.signed_request.user.oauth_token.has_expired:\n try:\n user = User.objects.get(facebook_id=request.facebook.signed_request.user.id)\n except User.DoesNotExist:\n oauth_token = OAuthToken.objects.create(token=request.facebook.signed_request.user.oauth_token.token, issued_at=request.facebook.signed_request.user.oauth_token.issued_at.replace(tzinfo=tzlocal()), expires_at=request.facebook.signed_request.user.oauth_token.expires_at.replace(tzinfo=tzlocal()))\n user = User.objects.create(facebook_id=request.facebook.signed_request.user.id, oauth_token=oauth_token)\n user.synchronize()\n else:\n user.last_seen_at = now()\n if 'signed_request' in request.REQUEST:\n user.authorized = True\n if request.facebook.signed_request.user.oauth_token:\n user.oauth_token.token = request.facebook.signed_request.user.oauth_token.token\n user.oauth_token.issued_at = request.facebook.signed_request.user.oauth_token.issued_at.replace(tzinfo=tzlocal())\n user.oauth_token.expires_at = request.facebook.signed_request.user.oauth_token.expires_at.replace(tzinfo=tzlocal())\n user.oauth_token.save()\n user.save()\n\n if not user.oauth_token.extended:\n try:\n user.oauth_token.extend()\n except:\n pass\n\n request.facebook.user = user\n\n def process_response(self, request, response):\n \"\"\"\n Set compact P3P policies and save signed request to cookie.\n\n P3P is a WC3 standard (see http://www.w3.org/TR/P3P/), and although largely ignored by most\n browsers it is considered by IE before accepting third-party cookies (ie. cookies set by\n documents in iframes). If they are not set correctly, IE will not set these cookies.\n \"\"\"\n response['P3P'] = 'CP=\"IDC CURa ADMa OUR IND PHY ONL COM STA\"'\n if FANDJANGO_CACHE_SIGNED_REQUEST:\n if hasattr(request, 'facebook') and request.facebook and request.facebook.signed_request:\n response.set_cookie('signed_request', request.facebook.signed_request.generate())\n else:\n response.delete_cookie('signed_request')\n return response\n\n\nclass FacebookWebMiddleware(BaseMiddleware):\n \"\"\"Middleware for Facebook auth on websites.\"\"\"\n\n def process_request(self, request):\n \"\"\"Process the web-based auth request.\"\"\"\n if hasattr(request, 'facebook') and request.facebook:\n return\n else:\n request.facebook = False\n if not self.is_valid_path(request):\n return\n if self.is_access_denied(request):\n return authorization_denied_view(request)\n request.facebook = Facebook()\n oauth_token = False\n if 'oauth_token' in request.COOKIES:\n try:\n oauth_token = OAuthToken.objects.get(token=request.COOKIES['oauth_token'])\n except OAuthToken.DoesNotExist:\n request.facebook = False\n return\n\n else:\n if 'code' in request.GET:\n try:\n graph = GraphAPI()\n response = graph.get('oauth/access_token', client_id=FACEBOOK_APPLICATION_ID, redirect_uri=get_post_authorization_redirect_url(request, canvas=False), client_secret=FACEBOOK_APPLICATION_SECRET_KEY, code=request.GET['code'])\n components = parse_qs(response)\n oauth_token, new_oauth_token = OAuthToken.objects.get_or_create(token=components['access_token'][0], issued_at=now(), expires_at=now() + timedelta(seconds=int(components['expires'][0])))\n except GraphAPI.OAuthError:\n pass\n\n if not oauth_token or oauth_token.expired:\n request.facebook = False\n return\n try:\n user = oauth_token.user\n if not user.authorized:\n request.facebook = False\n return\n user.last_seen_at = now()\n user.save()\n except User.DoesNotExist:\n graph = GraphAPI(oauth_token.token)\n profile = graph.get('me')\n try:\n user = User.objects.get(facebook_id=profile.get('id'))\n if not user.authorized:\n if new_oauth_token:\n user.last_seen_at = now()\n user.authorized = True\n else:\n request.facebook = False\n return\n except User.DoesNotExist:\n user = User.objects.create(facebook_id=profile.get('id'), oauth_token=oauth_token)\n\n user.synchronize(profile)\n old_oauth_token = None\n if user.oauth_token != oauth_token:\n old_oauth_token = user.oauth_token\n user.oauth_token = oauth_token\n user.save()\n if old_oauth_token:\n old_oauth_token.delete()\n\n if not user.oauth_token.extended:\n try:\n user.oauth_token.extend()\n except:\n pass\n\n request.facebook.user = user\n request.facebook.oauth_token = oauth_token\n return\n\n def process_response(self, request, response):\n \"\"\"\n Set compact P3P policies and save auth token to cookie.\n\n P3P is a WC3 standard (see http://www.w3.org/TR/P3P/), and although largely ignored by most\n browsers it is considered by IE before accepting third-party cookies (ie. cookies set by\n documents in iframes). If they are not set correctly, IE will not set these cookies.\n \"\"\"\n if hasattr(request, 'facebook') and request.facebook and request.facebook.oauth_token:\n if 'code' in request.REQUEST:\n path = get_full_path(request, remove_querystrings=['code', 'web_canvas'])\n response = HttpResponseRedirect(path)\n response.set_cookie('oauth_token', request.facebook.oauth_token.token)\n else:\n response.delete_cookie('oauth_token')\n response['P3P'] = 'CP=\"IDC CURa ADMa OUR IND PHY ONL COM STA\"'\n return response","sub_path":"pycfiles/fandjango-4.2.1.tar/middleware.py","file_name":"middleware.py","file_ext":"py","file_size_in_byte":9578,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"28576869","text":"from django.conf.urls import patterns, include, url\n\nfrom . import views\n\nurlpatterns = patterns('blog',\n url(r'^$', views.index, name='index'),\n url(r'^post/(?P.+?)/$', views.post_detail, name='post_detail'),\n url(r'^tag/(?P.+?)/$', views.show_tag, name='show_tag'),\n url(r'^category/(?P.+?)/$', views.show_category, name='show_category'),\n url(r'^archive/$', views.show_archive, name='show_archive'),\n url(r'^create/$', views.create_or_update_post, name='create_post'),\n url(r'^update/(?P\\d+)/$', views.create_or_update_post, name='update_post'),\n)\n","sub_path":"blog/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":607,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"620917174","text":"# ---\n# jupyter:\n# jupytext:\n# text_representation:\n# extension: .py\n# format_name: percent\n# format_version: '1.3'\n# jupytext_version: 1.6.0\n# kernelspec:\n# display_name: Python 3\n# language: python\n# name: python3\n# ---\n\n# %% [markdown]\n# # Using numerical and categorical variables together\n#\n# In this notebook, we will present typical ways to deal with **categorical\n# variables**, namely **ordinal encoding** and **one-hot encoding**.\n#\n# We will load the entire adult census dataset.\n\n# %%\nimport pandas as pd\n\ndf = pd.read_csv(\"../datasets/adult-census.csv\")\n\ntarget_name = \"class\"\ntarget = df[target_name]\n\ndata = df.drop(columns=[target_name, \"fnlwgt\"])\n\n# %% [markdown]\n# We recall that `\"education-num\"` and `\"education\"` columns contains the same\n# information. In the previous notebook, we drop `\"education-num\"` and only\n# use the latter column. We will do the same processing here.\n\n# %%\ndata = data.drop(columns=\"education-num\")\n\n# %% [markdown]\n# ## Selection based on data types\n#\n# We separate categorical and numerical variables using the `object` data types\n# as we previously saw that it only corresponds to categorical columns. We make\n# use of the `make_column_selector` to select the corresponding columns.\n\n# %%\nfrom sklearn.compose import make_column_selector as selector\n\nnumerical_columns_selector = selector(dtype_exclude=object)\ncategorical_columns_selector = selector(dtype_include=object)\n\nnumerical_columns = numerical_columns_selector(data)\ncategorical_columns = categorical_columns_selector(data)\n\n# %% [markdown]\n# Besides, we will list beforehand the categories for each categorical column\n# to avoid issues with rare categories when evaluating the model during the\n# cross-validation.\n\n# %%\ncategories = [data[column].unique()\n for column in categorical_columns]\n\n# %% [markdown]\n# ## Dispatch some columns to a specific processor\n#\n# In the previous sections, we saw that we need to treat data differently\n# depending on their nature (i.e. numerical or categorical).\n#\n# Scikit-learn provides a `ColumnTransformer` class which will send specific\n# columns to a specific transformer, making it easy to fit a single predictive\n# model on a dataset that combines both kinds of variables together\n# (heterogeneously typed tabular data).\n#\n# We can first define the columns depending on their data type:\n#\n# * **one-hot encoding** will be applied to categorical columns. Besides, we\n# will use the option `drop=\"if_binary\"` to drop one of the column since the\n# information will be correlated;\n# * **numerical scaling** numerical features which will be standardized.\n#\n# We can now create our `ColumnTransfomer` by specifying a list of triplet\n# (preprocessor name, transformer, columns). First, let's start by creating\n# a transformer for the numerical and categorical part as we did in the\n# previous notebooks.\n\n# %%\nfrom sklearn.preprocessing import OneHotEncoder, StandardScaler\n\ncategorical_preprocessor = OneHotEncoder(categories=categories,\n drop=\"if_binary\")\nnumerical_preprocessor = StandardScaler()\n\n# %% [markdown]\n# Now, we can associate each of these preprocessors with their respective\n# columns.\n\n# %%\nfrom sklearn.compose import ColumnTransformer\n\npreprocessor = ColumnTransformer([\n ('one-hot-encoder', categorical_preprocessor, categorical_columns),\n ('standard-scaler', numerical_preprocessor, numerical_columns)])\n\n# %% [markdown]\n# Finally, we can define a pipeline to stack this \"preprocessor\" with our\n# classifier (logistic regression).\n\n# %%\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.pipeline import make_pipeline\n\nmodel = make_pipeline(preprocessor, LogisticRegression(max_iter=500))\n\n# %% [markdown]\n# Starting from `scikit-learn 0.23`, the notebooks can display an interactive\n# view of the pipelines.\n\n# %%\nfrom sklearn import set_config\nset_config(display='diagram')\nmodel\n\n# %% [markdown]\n# The final model is more complex than the previous models but still follows\n# the same API:\n#\n# - the `fit` method is called to preprocess the data then train the\n# classifier;\n# - the `predict` method can make predictions on new data;\n# - the `score` method is used to predict on the test data and compare the\n# predictions to the expected test labels to compute the accuracy.\n#\n# Let's start by splitting our data into train and test sets.\n\n# %%\nfrom sklearn.model_selection import train_test_split\n\ndata_train, data_test, target_train, target_test = train_test_split(\n data, target, random_state=42)\n\n# %% [markdown]\n# Now, we can train the model on the train set.\n\n# %%\n_ = model.fit(data_train, target_train)\n\n# %% [markdown]\n# Then, we can use the raw dataset directly to the pipeline. Indeed, we don't\n# need to make any processing. All the preprocessing will be handle when\n# calling `predict`. We will give an example by predicting on the five first\n# sample from the test set.\n\n# %%\ndata_test.head()\n\n# %%\nmodel.predict(data_test)[:5]\n\n# %%\ntarget_test[:5]\n\n# %% [markdown]\n# We can get the accuracy score by calling directly the `score` method. We will\n# compute the score on the entire test set.\n\n# %%\nmodel.score(data_test, target_test)\n\n# %% [markdown]\n# ## Evaluation of the model with cross-validation\n#\n# This model can also be cross-validated as we previously did (instead of using\n# a single train-test split):\n\n# %%\nfrom sklearn.model_selection import cross_val_score\n\nscores = cross_val_score(model, data, target, cv=5)\nscores\n\n# %%\nprint(f\"The accuracy is: {scores.mean():.3f} +- {scores.std():.3f}\")\n\n# %% [markdown]\n# The compound model has a higher predictive accuracy than the two models that\n# used numerical and categorical variables in isolation.\n\n# %% [markdown]\n# ## Fitting a more powerful model\n#\n# **Linear models** are very nice because they are usually very cheap to train,\n# **small** to deploy, **fast** to predict and give a **good baseline**.\n#\n# However, it is often useful to check whether more complex models such as an\n# ensemble of decision trees can lead to higher predictive performance.\n#\n# In the following cell we try a scalable implementation of the **Gradient\n# Boosting Machine** algorithm. For this class of models, we know that contrary\n# to linear models, it is **useless to scale the numerical features** and\n# furthermore it is both safe and significantly more computationally efficient\n# to use an arbitrary **integer encoding for the categorical variables** even\n# if the ordering is arbitrary. Therefore we adapt the preprocessing pipeline\n# as follows:\n\n# %%\nfrom sklearn.experimental import enable_hist_gradient_boosting\nfrom sklearn.ensemble import HistGradientBoostingClassifier\nfrom sklearn.preprocessing import OrdinalEncoder\n\ncategorical_preprocessor = OrdinalEncoder(categories=categories)\n\npreprocessor = ColumnTransformer([\n ('categorical', categorical_preprocessor, categorical_columns)],\n remainder=\"passthrough\")\n\nmodel = make_pipeline(preprocessor, HistGradientBoostingClassifier())\n\n# %% [markdown]\n# Now that we created our model, we can check the performance of the model.\n\n# %%\n# %%time\n_ = model.fit(data_train, target_train)\n\n# %%\nmodel.score(data_test, target_test)\n\n# %% [markdown]\n# We can observe that we get significantly higher accuracies with the Gradient\n# Boosting model. This is often what we observe whenever the dataset has a\n# large number of samples and limited number of informative features (e.g. less\n# than 1000) with a mix of numerical and categorical variables.\n#\n# This explains why Gradient Boosted Machines are very popular among\n# datascience practitioners who work with tabular data.\n\n# %% [markdown]\n# In this notebook we have:\n#\n# * used a `ColumnTransformer` to apply different preprocessing for\n# categorical and numerical variables;\n# * used a pipeline to chain the `ColumnTransformer` preprocessing and\n# logistic regresssion fitting;\n# * seen that **gradient boosting methods** can outperforms the basic linear\n# approach.\n","sub_path":"python_scripts/03_categorical_pipeline_column_transformer.py","file_name":"03_categorical_pipeline_column_transformer.py","file_ext":"py","file_size_in_byte":8043,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"217006493","text":"# -*- coding: utf-8 -*-\n\"\"\"\n\n Graphics processing module\n\n\"\"\"\nfrom colorama import init, Fore, Back\ninit(autoreset=True)\n\nMEMORY = {\n 'num': None,\n 'end': None,\n 'url': None,\n 'fps': None,\n 'res': None,\n 'size': None,\n 'percents': None,\n 'filename': None\n}\nTERM: int = 79 # terminal window width\n\n\ndef set_parameters(parameters: dict = None):\n \"\"\"\n Interface to the MEMORY dict\n \"\"\"\n global MEMORY\n\n if not parameters:\n return\n\n total_digits = max([2, len(str(parameters['end']))])\n MEMORY = {\n 'num': str(parameters['num']).rjust(total_digits, '0'),\n 'end': str(parameters['end']).rjust(total_digits, '0'),\n 'url': parameters['url'],\n 'size': str(round(parameters['size']/1024/1024, 2)).rjust(6) + 'mb ',\n 'percents': parameters['percents'],\n 'filename': parameters['filename']\n }\n\n\ndef unable_to_reach(attempt: int):\n \"\"\"\n Call status if an attempt to download has failed\n \"\"\"\n statbar = attempt * '.' + (10 - attempt) * ' '\n status(prefix='\\r' + Fore.LIGHTRED_EX, stbar=statbar, perc=0, fsize=0)\n\n\ndef unaviable():\n \"\"\"\n Call status if all attempts to download has failed\n \"\"\"\n statbar = '[ FAIL ]'\n status(prefix='\\r' + Fore.RED, stbar=statbar, name=MEMORY.get('url', ''), perc=0, postfix='\\n')\n\n\ndef call_status_for_already_exist():\n \"\"\"\n Call status when file already exist\n \"\"\"\n statbar = Fore.BLACK + Back.GREEN + ' EXISTS ' + Fore.RESET + Back.RESET\n status(stbar=statbar, perc=100, postfix='\\n')\n\n\ndef call_status_for_progres(percents: int):\n \"\"\"\n Call status when we've got new percent of file\n \"\"\"\n MEMORY['percents'] = percents\n\n ready = int(percents / 10)\n remain = 10 - ready\n\n ready_bar = Fore.YELLOW + Back.YELLOW + ready * 'X' + Back.RESET\n remain_bar = Fore.YELLOW + remain * '_'\n full_bar = ready_bar + remain_bar\n\n status(prefix='\\r' + Fore.YELLOW, stbar=full_bar, perc=percents)\n\n\ndef call_status_for_complete():\n \"\"\"\n Call status when file downloading is complete\n \"\"\"\n status(prefix='\\r' + Fore.GREEN, stbar='[COMPLETE]', perc=100, postfix='\\n')\n\n\ndef status(prefix='', num='', end='', fsize=0, perc=0, name='', stbar='', postfix=''):\n \"\"\"\n Print current status\n \"\"\"\n num = num or MEMORY['num']\n end = end or MEMORY['end']\n fsize = fsize or MEMORY['size']\n name = name or MEMORY['filename']\n perc = perc or MEMORY['percents']\n perc = ' ' + str(perc).rjust(3) + '% '\n name = name or MEMORY['filename']\n textual_status = f'[{num} of {end}]{stbar}{perc}{fsize}'\n fsize = len(end) * 2 + 30\n pos = TERM - fsize - 1\n print(prefix + textual_status + name[:pos], end=postfix)\n","sub_path":"yt_graphics.py","file_name":"yt_graphics.py","file_ext":"py","file_size_in_byte":2754,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"534465652","text":"# pylint: disable=no-self-use\nfrom unittest.mock import Mock\n\nfrom PartSeg._roi_analysis.advanced_window import MeasurementSettings, MultipleInput, Properties\nfrom PartSeg._roi_analysis.advanced_window import QInputDialog as advanced_module_input\nfrom PartSeg._roi_analysis.advanced_window import QMessageBox as advanced_message_box\nfrom PartSegCore.analysis import AnalysisAlgorithmSelection\n\n\nclass TestProperties:\n def test_synchronize_voxel_size(self, qtbot, part_settings):\n widget = Properties(part_settings)\n qtbot.addWidget(widget)\n widget.lock_spacing.setChecked(True)\n widget.update_spacing()\n value = widget.spacing[1].value()\n with qtbot.waitSignal(widget.spacing[2].valueChanged, timeout=10**4):\n widget.spacing[2].setValue(value - 20)\n assert widget.spacing[1].value() == value - 20\n\n def test_pipeline_profile_show_info(\n self, qtbot, part_settings, border_rim_profile, lower_threshold_profile, sample_pipeline\n ):\n part_settings.roi_profiles[border_rim_profile.name] = border_rim_profile\n part_settings.roi_pipelines[sample_pipeline.name] = sample_pipeline\n widget = Properties(part_settings)\n widget.show()\n qtbot.addWidget(widget)\n widget.update_profile_list()\n assert widget.profile_list.count() == 1\n part_settings.roi_profiles[lower_threshold_profile.name] = lower_threshold_profile\n widget.update_profile_list()\n assert widget.profile_list.count() == 2\n assert widget.pipeline_list.count() == 1\n assert widget.info_label.toPlainText() == \"\"\n with qtbot.waitSignal(widget.profile_list.currentItemChanged, timeout=10**4):\n widget.profile_list.setCurrentRow(1)\n profile = part_settings.roi_profiles[widget.profile_list.item(1).text()]\n assert widget.info_label.toPlainText() == profile.pretty_print(AnalysisAlgorithmSelection)\n widget.pipeline_list.setCurrentRow(0)\n assert widget.info_label.toPlainText() == sample_pipeline.pretty_print(AnalysisAlgorithmSelection)\n widget.hide()\n\n def test_delete_profile(self, qtbot, part_settings, border_rim_profile, lower_threshold_profile):\n part_settings.roi_profiles[border_rim_profile.name] = border_rim_profile\n part_settings.roi_profiles[lower_threshold_profile.name] = lower_threshold_profile\n widget = Properties(part_settings)\n widget.show()\n qtbot.addWidget(widget)\n widget.update_profile_list()\n assert widget.profile_list.count() == 2\n with qtbot.waitSignal(widget.profile_list.currentItemChanged, timeout=10**4):\n widget.profile_list.setCurrentRow(0)\n assert widget.delete_btn.isEnabled()\n with qtbot.waitSignal(widget.delete_btn.clicked):\n widget.delete_btn.click()\n assert len(part_settings.roi_profiles) == 1\n assert lower_threshold_profile.name in part_settings.roi_profiles\n widget.hide()\n\n def test_rename_profile(\n self, qtbot, part_settings, border_rim_profile, lower_threshold_profile, sample_pipeline, monkeypatch\n ):\n part_settings.roi_profiles[border_rim_profile.name] = border_rim_profile\n part_settings.roi_pipelines[sample_pipeline.name] = sample_pipeline\n part_settings.roi_profiles[lower_threshold_profile.name] = lower_threshold_profile\n widget = Properties(part_settings)\n widget.show()\n qtbot.addWidget(widget)\n widget.update_profile_list()\n assert widget.profile_list.count() == 2\n monkeypatch.setattr(advanced_module_input, \"getText\", check_text(border_rim_profile.name, \"rim\"))\n widget.profile_list.setCurrentRow(0)\n widget.rename_profile()\n assert widget.profile_list.item(1).text() == \"rim\"\n assert set(part_settings.roi_profiles.keys()) == {\"rim\", lower_threshold_profile.name}\n monkeypatch.setattr(advanced_module_input, \"getText\", check_text(sample_pipeline.name, \"rim\"))\n widget.pipeline_list.setCurrentRow(0)\n widget.rename_profile()\n assert widget.pipeline_list.item(0).text() == \"rim\"\n assert set(part_settings.roi_pipelines.keys()) == {\"rim\"}\n monkeypatch.setattr(advanced_module_input, \"getText\", check_text(\"rim\", lower_threshold_profile.name))\n\n called_mock = [0]\n\n def mock_waring(*_):\n monkeypatch.setattr(widget, \"rename_profile\", _empty)\n called_mock[0] = 1\n return advanced_message_box.No\n\n monkeypatch.setattr(advanced_message_box, \"warning\", mock_waring)\n widget.profile_list.setCurrentRow(1)\n widget.rename_profile()\n assert called_mock[0] == 1\n assert widget.profile_list.item(1).text() == \"rim\"\n assert set(part_settings.roi_profiles.keys()) == {\"rim\", lower_threshold_profile.name}\n widget.hide()\n\n def test_multiple_files_visibility(self, qtbot, part_settings):\n widget = Properties(part_settings)\n qtbot.addWidget(widget)\n assert not part_settings.get(\"multiple_files_widget\")\n assert not widget.multiple_files_chk.isChecked()\n with qtbot.waitSignal(widget.multiple_files_chk.stateChanged):\n widget.multiple_files_chk.setChecked(True)\n assert part_settings.get(\"multiple_files_widget\")\n part_settings.set(\"multiple_files_widget\", False)\n assert not widget.multiple_files_chk.isChecked()\n\n\nclass TestMeasurementSettings:\n def test_create(self, qtbot, part_settings):\n widget = MeasurementSettings(part_settings)\n qtbot.addWidget(widget)\n\n def test_base_steep(self, qtbot, part_settings):\n widget = MeasurementSettings(part_settings)\n qtbot.addWidget(widget)\n widget.show()\n widget.profile_options.setCurrentRow(0)\n assert widget.profile_options.item(0).text() == \"Volume\"\n assert widget.profile_options.item(1).text() == \"Diameter\"\n assert widget.profile_options_chosen.count() == 0\n widget.choose_option()\n assert widget.profile_options_chosen.count() == 1\n widget.choose_option()\n assert widget.profile_options_chosen.count() == 1\n widget.profile_options.setCurrentRow(1)\n assert widget.profile_options_chosen.count() == 1\n widget.choose_option()\n assert widget.profile_options_chosen.count() == 2\n widget.profile_options.setCurrentRow(0)\n widget.proportion_action()\n assert widget.profile_options_chosen.count() == 2\n widget.profile_options.setCurrentRow(1)\n widget.proportion_action()\n assert widget.profile_options_chosen.count() == 3\n assert widget.profile_options_chosen.item(2).text() == \"ROI Volume/ROI Diameter\"\n\n widget.profile_options_chosen.setCurrentRow(0)\n assert widget.profile_options_chosen.item(0).text() == \"ROI Volume\"\n widget.remove_element()\n assert widget.profile_options_chosen.count() == 2\n assert widget.profile_options_chosen.item(0).text() == \"ROI Diameter\"\n\n assert not widget.save_butt.isEnabled()\n with qtbot.waitSignal(widget.profile_name.textChanged):\n widget.profile_name.setText(\"test\")\n assert widget.save_butt.isEnabled()\n\n assert len(part_settings.measurement_profiles) == 2\n with qtbot.waitSignal(widget.save_butt.clicked):\n widget.save_butt.click()\n assert len(part_settings.measurement_profiles) == 3\n\n with qtbot.waitSignal(widget.profile_name.textChanged):\n widget.profile_name.setText(\"\")\n assert not widget.save_butt.isEnabled()\n\n widget.reset_action()\n assert widget.profile_options_chosen.count() == 0\n widget.hide()\n\n\ndef test_multiple_input(qtbot, monkeypatch):\n mock = Mock()\n monkeypatch.setattr(advanced_message_box, \"warning\", mock)\n widget = MultipleInput(\n text=\"sample text\",\n help_text=\"help\",\n objects_list=[\n (\"A\", str),\n (\"B\", int, 5),\n (\"C\", float, 5.0),\n (\"D\", int),\n (\"E\", float),\n ],\n )\n qtbot.addWidget(widget)\n assert widget.result is None\n widget.accept_response()\n assert widget.result is None\n mock.assert_called_once()\n mock.reset_mock()\n widget.object_dict[\"A\"][1].setText(\"test\")\n widget.accept_response()\n mock.assert_not_called()\n assert widget.result == {\"A\": \"test\", \"B\": 5, \"C\": 5.0, \"D\": 0, \"E\": 0.0}\n\n\ndef check_text(expected, to_return):\n def _check(*_, text=None, **_kwargs):\n assert text == expected\n return to_return, True\n\n return _check\n\n\ndef _empty():\n \"\"\"\n Empty function for monkeypatching to prevent recursive call\n in `test_rename_profile` test\n \"\"\"\n","sub_path":"package/tests/test_PartSeg/roi_analysis/test_advanced_window.py","file_name":"test_advanced_window.py","file_ext":"py","file_size_in_byte":8777,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"535459255","text":"\nimport commands\nfrom numpy import *\nfrom timings import *\nfrom strCheck import *\nfrom matplotlib import *\nfrom pylab import *\nimport sys\nfrom scipy.misc import *\n\nstatus,output= commands.getstatusoutput(\"ls Activities/*_*m.txt\")\nif len(sys.argv) > 1 and sys.argv[1] == \"50\":\n\tstatus,output= commands.getstatusoutput(\"ls Activities/*_50*m.txt\")\n\nfnames = output.split('\\n')\nfn_dates = []\nfor fn in fnames:\n\tfn_dates.append(getDate(fn))\n\t\n# sort activities based on dates\t\nfn_dates, fnames = (list(t) for t in zip(*sorted(zip(fn_dates, fnames))))\n\ndates = []\nstrks_len = []\navg_pace = []\nstrks_min = []\n\n\nfor fn in fnames:\n#\tprint fn\n\tf = open(fn)\n\tdate = getDate(fn)\t\t\n\theader = f.readline()\n\tstroke_index = header.split(',').index('Total Strokes')\n\tmeters = 0. \n\tseconds = 0.\n\tstrokes = 0.\n\tfor line in f:\n\t\t# we want the interval 3 not the single splits 3.1, 3.2, etc\n\t\tif (checkIsInterval(line) == False):\n\t\t\tcontinue\n\t\tstyle = getStyle(line)\n\t\tif not style == 0 : \n\t\t\tcontinue\n#\t\tprint line\n\t\tline = strIntCorrection(line)\n\t\ttokens = line.split(',')\n\t\tmeters = meters + float(tokens[3])\n\t\tseconds = seconds + getSeconds(tokens[4])\n\t\tstrokes = strokes + float(tokens[stroke_index])\n\n\tf.close()\n\n\tif meters > 0 :\n\t\tdates.append(date)\n\t\t# SwimSmooth plot counts both arms, so I double the strokes \n\t\tstrks_min.append(2*strokes*60/seconds)\n\t\tavg_pace.append(seconds*100/meters)\n\t\n\nn=len(strks_min)\nstrks_min = strks_min[n-10:] \navg_pace = avg_pace[n-10:] \n\nformatter = FuncFormatter(toMMSS)\n\nfig, ax1 = plt.subplots()\nax1.xaxis.set_major_formatter(formatter)\nplot(avg_pace, strks_min, \"o--\")\n\nimg = imread(\"swimsmoothpic.png\")\nplt.imshow(img,zorder=0,extent=[130, 60, 30,100], aspect='auto')\nxlim(130,60)\nylim(30,100)\nshow()\n\t\n\n\n\n\n\n\n\n","sub_path":"swimsmooth.py","file_name":"swimsmooth.py","file_ext":"py","file_size_in_byte":1734,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"334589931","text":"import glob\r\n\r\n\r\n# If run from the current folder, this example outputs:\r\n# Searching for *ule.py, I found:\r\n# path_of_that_module.py\r\n# path_of_this_module.py\r\n\r\n# If run from the parent folder, it finds no such matches,\r\n# since the glob_search does not specify an absolute path\r\n\r\n\r\nglob_search = '*ule.py'\r\n\r\nlist_of_path_names = glob.glob(glob_search)\r\n\r\nprint(f'Searching for {glob_search}, I found:')\r\n\r\nfor path in list_of_path_names:\r\n print(path)\r\n","sub_path":"paths/glob_example.py","file_name":"glob_example.py","file_ext":"py","file_size_in_byte":461,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"12158534","text":"from __future__ import unicode_literals\nfrom django import template\nfrom django.contrib.contenttypes.models import ContentType\nfrom django.contrib.gis.geos import *\nfrom django.contrib.gis.measure import D\nfrom mezzanine import template\nfrom mezzanine.generic.forms import RatingForm\nfrom directory.models import Rating, CustomAddress, Main\n\nregister = template.Library()\n\n\n@register.inclusion_tag('includes/closest_objects.html', takes_context=True)\ndef closest(context):\n \"\"\"\n Template tag for rendering closest 5 main objects\n \"\"\"\n if 'page' in context and hasattr(context['page'], 'main'):\n city = context['page'].main.address.all()[0]\n else:\n return {}\n pnt = city.location\n qs = CustomAddress.objects.filter(location__distance_lte=(pnt, D(km=2000))).\\\n distance(pnt).order_by('distance').values_list('id', flat=True)[:5]\n closest_objects = Main.objects.filter(address__id__in=list(qs)).distinct()\n return {'closest_objects': closest_objects}\n\n\n@register.inclusion_tag('includes/highest_rated.html', takes_context=True)\ndef highest_rated(context):\n \"\"\"\n Template tag for rendering highest rated 5 main objects\n \"\"\"\n if 'selected_city' in context:\n city = context['selected_city']\n else:\n return {}\n highest_rated = Main.objects.filter(address__city=city).distinct()\n highest_rated_count = highest_rated.count()\n highest_rated = highest_rated.order_by('rating')[:highest_rated_count]\n return {'highest_rated': highest_rated}\n\n\n@register.inclusion_tag(\"includes/rating.html\", takes_context=True)\ndef rating_custom(context, obj):\n \"\"\"\n Override default Mezzanine's rating tag\n \"\"\"\n ct = ContentType.objects.get_for_model(obj)\n context[\"rating_counter\"] = []\n for val in range(1, 6):\n count = Rating.objects.filter(\n content_type=ct, object_pk=obj.pk, value=val).count()\n context[\"rating_counter\"].append(count)\n context[\"rating_object\"] = context[\"rating_obj\"] = obj\n context[\"rating_form\"] = RatingForm(context[\"request\"], obj)\n comments = Rating.objects.filter(content_type=ct, object_pk=obj.pk)\\\n .values_list('comment', 'value')\n # add comments to context\n context['comments'] = [\n {'comment': item[0], 'value': item[1]} for item in comments]\n ratings = context[\"request\"].COOKIES.get(\"mezzanine-rating\", \"\")\n rating_string = \"%s.%s\" % (obj._meta, obj.pk)\n context[\"rated\"] = (rating_string in ratings)\n rating_name = obj.get_ratingfield_name()\n for f in (\"average\", \"count\", \"sum\"):\n context[\"rating_\" + f] = getattr(obj, \"%s_%s\" % (rating_name, f))\n return context\n","sub_path":"directory/templatetags/directory_tags.py","file_name":"directory_tags.py","file_ext":"py","file_size_in_byte":2659,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"654399905","text":"\"\"\"\r\nland cover zonal statistics for chicago\r\n\r\nkufre\r\n\"\"\"\r\n\r\nimport rasterio as rio\r\nimport geopandas as gpd\r\nimport pandas as pd\r\nfrom rasterstats import zonal_stats\r\n\r\nraster = 'cookcountylandcover2010\\landcover_2010_cookcounty.img'\r\n\r\n# chicago tracts from chicago open data\r\nwith rio.open(raster) as src:\r\n chicago = gpd.read_file('https://data.cityofchicago.org/api/geospatial/5jrd-6zik?method=export&format=GeoJSON').to_crs(src.crs)\r\n\r\n# reading dbf file for raster categories\r\ndbf = gpd.read_file('cookcountylandcover2010\\landcover_2010_cookcounty.img.vat.dbf')\r\n\r\n# cleaning up categories\r\nclasses = [str(x).lower().replace('/','_').replace(' ','_') for x in dbf['Class']]\r\n\r\n# creating dictionary to be used in zonal_stats()\r\nland_cover = dict(zip(dbf['Value'].tolist(), classes))\r\n\r\nzonal = zonal_stats(chicago,raster,categorical = True,nodata = 0, category_map = land_cover)\r\n\r\n# blank DataFrame to be filled in with values from zonal_stats()\r\nzeros = [0] * len(chicago)\r\n\r\nresults = pd.DataFrame(dict((v,zeros) for k,v in land_cover.items()))\r\n\r\nresults.insert(0,'geoid10', chicago['geoid10'])\r\n\r\n# inserting values\r\nfor i in range(len(zonal)):\r\n row = zonal[i]\r\n for j in list(row.keys()):\r\n results.at[i,j] = row.get(j) * 4\r\n\r\nchicago = chicago.merge(results, how = 'left', on = 'geoid10')\r\n\r\n# calculating canopy percentage for each tract\r\nchicago['pct_canopy'] = chicago['tree_canopy'] / chicago.area * 100\r\n\r\n# save results\r\nresults.to_csv('chicago_land_cover.csv')\r\n\r\nchicago.to_file('chicago.gpkg', layer ='land_cover', driver=\"GPKG\")\r\n","sub_path":"chicago_trees/chicago_zonal.py","file_name":"chicago_zonal.py","file_ext":"py","file_size_in_byte":1572,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"38731252","text":"import unittest\nimport random\n\nimport numpy as np\nimport tensorflow as tf\n\nfrom tests.test_base import MNIST_MOBINENET_V2, BlockingPredictionTester, AsyncPredictionTester\n\n\nclass MnistMobilenetV2(unittest.TestCase):\n (_, _), (x_test, _) = tf.keras.datasets.mnist.load_data()\n x_test = np.reshape(x_test, (10000, 28, 28, 1))\n\n def test_session(self):\n tester = BlockingPredictionTester(MNIST_MOBINENET_V2)\n\n for i in range(0, 10):\n idx = random.randrange(0, 9999, 1)\n tester.assert_equals(np.array(self.x_test[idx:idx + 1], dtype=np.uint8))\n\n tester.close()\n\n def test_async_session(self):\n tester = AsyncPredictionTester(MNIST_MOBINENET_V2)\n\n for i in range(0, 10):\n idx = random.randrange(0, 9999, 1)\n tester.assert_equals(np.array(self.x_test[idx:idx + 1], dtype=np.uint8))\n\n tester.close()\n\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"python/furiosa-sdk-runtime/tests/test_mnist_mobinenet_v2.py","file_name":"test_mnist_mobinenet_v2.py","file_ext":"py","file_size_in_byte":940,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"38339145","text":"import numpy as np\n\nfrom megskull.network import Network\nfrom megskull.opr.all import (\n Conv2D, Pooling2D, FullyConnected, Softmax,\n\t CrossEntropyLoss, Dropout, ElementwiseAffine, Concat \n\t\t)\nfrom megskull.opr.helper.elemwise_trans import ReLU, Identity\nfrom megskull.graph.query import GroupNode\nfrom megskull.opr.netsrc import DataProvider\nimport megskull.opr.helper.param_init as pinit\nfrom megskull.opr.helper.param_init import AutoGaussianParamInitializer as G\nfrom megskull.opr.helper.param_init import ConstantParamInitializer as C\nfrom megskull.opr.regularizer import BatchNormalization as BN\nimport megskull.opr.arith as arith\n\nglobal idx\nidx = 0\n\n\"\"\"\ndef conv_bn(inp, ker_shape, stride, padding, out_chl, isrelu):\n\tglobal idx\n\tidx += 1\n\tl1 = Conv2D(\n\t\t\"conv{}\".format(idx), inp, kernel_shape = ker_shape, stride = stride, padding = padding,\n\t\toutput_nr_channel = out_chl,\n\t\tW = G(mean = 0, std = ((1 + int(isrelu)) / (ker_shape**2 * inp.partial_shape[1]))**0.5),\n\t\tnonlinearity = {True:ReLU(), False:Identity()}[isrelu]\n\t\t)\n\tl2 = BN(\"bn{}\".format(idx), l1, eps = 1e-9)\n\tl2 = ElementwiseAffine(\"bnaff{}\".format(idx), l2, shared_in_channels = False, k = C(1), b = C(0))\n\treturn l2\n\"\"\"\n\ndef bn_relu_conv(inp, ker_shape, stride, padding, out_chl, has_relu, has_bn, has_conv = True):\n\tglobal idx\n\tidx += 1\n\tif has_bn:\n\t\tl1 = BN(\"bn{}\".format(idx), inp, eps = 1e-9)\n\t\tl1 = ElementwiseAffine(\"bnaff{}\".format(idx), l1, shared_in_channels = False, k = C(1), b = C(0))\n\telse:\n\t\tl1 = inp\n\t\n\tif has_relu:\n\t\tl2 = arith.ReLU(l1)\n\telse:\n\t\tl2 = l1\n\t\n\tif not has_conv:\n\t\treturn l2\n\n\tl3 = Conv2D(\n\t\t\"conv{}\".format(idx), l2, kernel_shape = ker_shape, stride = stride, padding = padding,\n\t\toutput_nr_channel = out_chl,\n\t\tnonlinearity = Identity()\n\t\t)\n\t\n\treturn l3\n\ndef dense_block(inp, k, l):\n\tlay = inp\n\tfor i in range(l):\n\t\tcur_lay = bn_relu_conv(lay, 3, 1, 1, k, True, True)\n\t\tlay = Concat([lay, cur_lay], axis = 1)\n\treturn lay\n\ndef transition(inp, i):\n\tl1 = bn_relu_conv(inp, 1, 1, 0, inp.partial_shape[1], True, True, i != 2)\n\tglobal idx\n\tidx += 1\n\tif i != 2:\n\t\tl2 = Pooling2D(\n\t\t\t\"Pooling{}\".format(idx), l1, window = 2, mode = \"AVERAGE\"\n\t\t\t)\n\telse:\n\t\tl2 = Pooling2D(\n\t\t\t\"Pooling{}\".format(idx), l1, window = 8, stride = 8, mode = \"AVERAGE\"\n\t\t\t)\n\treturn l2\n\n\n\"\"\"\ndef res_layer(inp, chl):\n\tpre = inp\n\tinp = conv_bn(inp, 3, 1, 1, chl, True)\n\tinp = conv_bn(inp, 3, 1, 1, chl, False)\n\tinp = arith.ReLU(inp + pre)\n\treturn inp\n\ndef res_block(inp, chl, n):\n\tstride = 2\n\tif chl == 16:\n\t\tstride = 1\n\tpre = inp\n\tinp = conv_bn(inp, 3, stride, 1, chl, True)\n\tinp = conv_bn(inp, 3, 1, 1, chl, False)\n\tinp = inp + conv_bn(pre, 1, stride, 0, chl, False)\n\tinp = arith.ReLU(inp)\n\t\n\tfor i in range(n - 1):\n\t\tinp = res_layer(inp, chl)\n\t\n\treturn inp\n\"\"\"\n\ndef make_network(minibatch_size = 64):\n\tpatch_size = 32\n\tinp = DataProvider(\"data\", shape = (minibatch_size, 3, patch_size, patch_size))\n\tlabel = DataProvider(\"label\", shape = (minibatch_size, ))\n\n\tlay = bn_relu_conv(inp, 3, 1, 1, 16, False, False)\n\n\tk, l = 12, (40 - 4) // 3\n\tfor i in range(3):\n\t\tlay = transition(dense_block(lay, k, l), i)\n\t\n\t#global average pooling\n\tprint(lay.partial_shape)\n\tfeature = lay.mean(axis = 2).mean(axis = 2)\n\t#feature = Pooling2D(\"glbpoling\", lay, window = 8, stride = 8, mode = \"AVERAGE\")\n\tpred = Softmax(\"pred\", FullyConnected(\n\t\t\"fc0\", feature, output_dim = 10,\n\t\tnonlinearity = Identity()\n\t\t))\n\t\n\tnetwork = Network(outputs = [pred])\n\tnetwork.loss_var = CrossEntropyLoss(pred, label)\n\treturn network\n\nif __name__ == '__main__':\n\tmake_network()\n","sub_path":"SE/d40/network.py","file_name":"network.py","file_ext":"py","file_size_in_byte":3519,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"483822749","text":"what = input(\"Please enter the name of a txt file and directory if not in Work:\")\nthefile = open(what, \"r\")\nstring = thefile.read()\nthefile.close()\nletters = \"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\"\nwcount = 0\nlcount = 1\naword = False\nfor i in range(len(string)):\n if string[i] in (letters) and aword == False:\n aword = True\n if aword:\n if string[i] not in (letters):\n aword = False\n wcount +=1\n if i == len(string)-1 and string[i] in (letters):\n wcount +=1\n if string[i] == \"\\n\":\n lcount+=1\nprint (\"In the file \" +what+\", there are\", wcount,\"words and\",lcount,\"lines.\")\n","sub_path":"counter.py","file_name":"counter.py","file_ext":"py","file_size_in_byte":657,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"28954184","text":"numbers = [10,22,33,44,55,66,77,88,90]\n\ndef cacl_avg(values):\n sum = 0\n for count in values:\n sum += count\n avg = sum / len(values)\n\n return avg\n\ndef main():\n res = cacl_avg(numbers)\n print('{:.3f}'.format(res))\n\n\nif __name__ == '__main__':\n main()","sub_path":"datetime/numbers.py","file_name":"numbers.py","file_ext":"py","file_size_in_byte":276,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"88071310","text":"from django.http import HttpResponse\nfrom django.template import loader\nfrom bs4 import BeautifulSoup as bs4\nfrom playground import urls\n\n# Basically the render function but prettifies the html file\n# Because the programmer has severe OCD when it comes to indents.\n# Also append a few things (like the urls) into the context.\ndef prender(request, template_name, context={}, content_type='text/html', status=200, using=None):\n context['urls'] = []\n for url in urls.urlpatterns:\n if url.name[:2] != '__':\n context['urls'].append({'name':url.name.replace(\"_\", \" \").title(), 'pattern':url.pattern})\n content = loader.render_to_string(template_name, context, request, using) # Get HTML code from file\n no_new_line = content.split('\\n')\n no_new_list = ''\n for line in no_new_line:\n no_new_list = no_new_list+line\n #document = html.fromstring(no_new_list) # Get HTML string\n soup = bs4(no_new_list, features=\"lxml\")\n pretty_doc = soup.prettify()\n nowhite_doc = '' # Empty string\n\n # Remove lines that just have whitespace\n for line in pretty_doc.split('\\n'):\n if line.rstrip(): # If the line is just whitespace\n nowhite_doc = nowhite_doc+line+'\\n'\n\n # Append to the top of the document also\n return HttpResponse(nowhite_doc, content_type, status) # Spit out the string\n","sub_path":"internals/rendering.py","file_name":"rendering.py","file_ext":"py","file_size_in_byte":1363,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"88283892","text":"#!/usr/bin/env python3\n# The FLASK_APP environment variable is used to specify how to load the application.\n# flask is aware of static and templates folder\n\nfrom flask import Flask, render_template, request, url_for, flash, redirect\nimport sqlite3\nfrom werkzeug.exceptions import abort\n\napp = Flask(__name__)\napp.config['SECRET_KEY'] = 'your secret key'\n\ndef get_db_connection():\n conn = sqlite3.connect('database.db')\n conn.row_factory = sqlite3.Row\n return conn\n\ndef get_post(post_id):\n conn = get_db_connection()\n post = conn.execute('SELECT * FROM posts WHERE id = ?',(post_id,)).fetchone()\n conn.close()\n if post is None:\n abort(404)\n return post\n\n\n# flask creare URL map for this\n@app.route('/')\ndef indexroot():\n conn = get_db_connection()\n posts = conn.execute('SELECT * FROM posts').fetchall()\n conn.close()\n return render_template('index.html',mess=\"Kubernetes\", posts=posts)\n \n@app.route('/index.html')\ndef index():\n conn = get_db_connection()\n posts = conn.execute('SELECT * FROM posts').fetchall()\n conn.close()\n return render_template('index.html',mess=\"Kubernetes\", posts=posts)\n\n@app.route('/')\ndef post(post_id):\n post = get_post(post_id)\n return render_template('post.html', post=post)\n\n\n@app.route('/create', methods=('GET', 'POST'))\ndef create():\n if request.method == 'POST':\n title = request.form['title']\n content = request.form['content']\n\n if not title:\n flash('Title is required!')\n else:\n conn = get_db_connection()\n conn.execute('INSERT INTO posts (title, content) VALUES (?, ?)',\n (title, content))\n conn.commit()\n conn.close()\n return redirect(url_for('index'))\n\n return render_template('create.html')\n\n\n@app.route('//edit', methods=('GET', 'POST'))\ndef edit(id):\n post = get_post(id)\n\n if request.method == 'POST':\n title = request.form['title']\n content = request.form['content']\n\n if not title:\n flash('Title is required!')\n else:\n conn = get_db_connection()\n conn.execute('UPDATE posts SET title = ?, content = ?'\n ' WHERE id = ?',\n (title, content, id))\n conn.commit()\n conn.close()\n return redirect(url_for('index'))\n\n return render_template('edit.html', post=post)\n\n\n\n@app.route('//delete', methods=('POST',))\ndef delete(id):\n post = get_post(id)\n conn = get_db_connection()\n conn.execute('DELETE FROM posts WHERE id = ?', (id,))\n conn.commit()\n conn.close()\n flash('\"{}\" was successfully deleted!'.format(post['title']))\n return redirect(url_for('index'))\n\n@app.route('/about.html')\ndef aboutus():\n return render_template('about.html')\n\n@app.route('/dblinks.html')\ndef dblinks():\n return render_template('dblinks.html')\n\n@app.route('/monitor.html')\ndef monitor():\n return render_template('monitor.html')\n\n@app.route('/work.html')\ndef work():\n return render_template('work.html')","sub_path":"flask-blog/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":3117,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"552910687","text":"class Skok:\n def __init__(self, linija):\n dijelovi = [dio.split(',') for dio in linija.strip().split('->')]\n self.iz, self.za = tuple(dijelovi[0])\n self.u, self.s, self.pa = tuple(dijelovi[1])\n\n def odgovara(self, stanje, znak):\n return self.iz == stanje and self.za == znak\n\nclass Stroj:\n def __init__(self, linije):\n self.prazan = linije[2]\n self.stanje = linije[3]\n self.ciljna = linije[4].split(',')\n self.skokovi = [Skok(l) for l in linije[5:]]\n self.traka = []\n self.mjesto = 0\n\n def postavi(self, traka):\n self.traka = list(traka) + [self.prazan]\n\n def pokreni(self):\n koraci = 0\n\n while True:\n znak = self.traka[self.mjesto]\n skokovi = filter(lambda s: s.odgovara(self.stanje, znak), self.skokovi)\n if len(skokovi) < 1: break\n\n skok = skokovi[0]\n self.stanje = skok.u\n self.traka[self.mjesto] = skok.s\n koraci += 1\n\n if skok.pa == 'R':\n self.mjesto = self.mjesto + 1\n self.traka = self.traka + [self.prazan]\n else:\n self.mjesto = self.mjesto - 1\n\n return (koraci, 1 if self.stanje in self.ciljna else 0)\n\ndef utrlab3(definicijaTS, sadrzajTrake):\n stroj = Stroj(definicijaTS.strip('%').split('%'))\n stroj.postavi(sadrzajTrake)\n rezultat = stroj.pokreni()\n return \"%d%%%d\" % rezultat\n","sub_path":"labosi/lab-3/2010-11/by_unknown/UUTR_2010-11_lab-3.py","file_name":"UUTR_2010-11_lab-3.py","file_ext":"py","file_size_in_byte":1324,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"93185107","text":"#!/usr/bin/python3\n\nMINTIME = 100000000000000\nMINTIME = 1\n\nimport fileinput\nfrom math import gcd\nfrom collections import defaultdict\n\ndef kgv(a, b):\n\n return a * b // gcd(a, b)\n\ndef solve_next_bus(start, step, pos, id):\n\n i = 0\n\n while (start +i*step + pos) % id != 0:\n \n i += 1\n\n new_step = kgv(step, id)\n\n return start +i*step, new_step, i+1\n\ndef main():\n \n # part 1\n\n input_ = fileinput.input()\n\n arrival = int(input_.readline())\n\n buses = { pos: int(id) for pos,id in enumerate(input_.readline().strip().split(\",\")) if id.isnumeric() }\n\n departures = { id: id * (arrival // id + (1 if arrival % id else 0)) for id in buses.values() }\n\n busid, departure = min(departures.items(), key = lambda x: x[1])\n\n print(busid * (departure - arrival))\n\n # part 2\n\n timestamp = 0\n step = 1\n iterations = 0\n\n for pos, id in buses.items():\n print(timestamp, step)\n timestamp, step, i = solve_next_bus(timestamp, step, pos, id)\n iterations += i\n\n print(timestamp, iterations)\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"13/13.py","file_name":"13.py","file_ext":"py","file_size_in_byte":1035,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"459615920","text":"# -*- coding: utf-8 -*-\n# 实例23:one_hot实验\n\nimport tensorflow as tf\n\n# 标准的one-hot编码\nlabels0 = [\n [0, 0, 1],\n [0, 1, 0]\n]\n\n# 非标注的one-hot编码\nlabels1 = [\n [0.4, 0.1, 0.5],\n [0.3, 0.6, 0.1]\n]\n\n# 输出值\nlogits = [\n [2, 0.5, 6],\n [0.1, 0, 3]\n]\nresult0 = tf.nn.softmax_cross_entropy_with_logits(labels=labels0, logits=logits)\nresult1 = tf.nn.softmax_cross_entropy_with_logits(labels=labels1, logits=logits)\n\nwith tf.Session() as sess:\n print(\"result0: \", sess.run(result0))\n print(\"result1: \", sess.run(result1))\n","sub_path":"tf06/example23.py","file_name":"example23.py","file_ext":"py","file_size_in_byte":560,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"456584028","text":"from functools import reduce\n\n\nclass MrComplex(object):\n @staticmethod\n def update(game_info):\n defense_tasks = []\n reserves = []\n\n # Check for threats and available ships\n for planet in game_info.my_planets.values():\n incoming_fleets = [fleet for fleet in game_info.enemy_fleets.values() if fleet.dest.id == planet.id]\n incoming_threats = []\n turn_processed = 0\n avail_ships = planet.num_ships\n excess_ships = 0\n\n for fleet in incoming_fleets:\n new_threat = True\n for threat in incoming_threats:\n if threat.turns_remaining == fleet.turns_remaining:\n new_threat = False\n threat.num_ships += fleet.num_ships\n break\n if new_threat:\n incoming_threats.append(PlanetThreat(round(fleet.turns_remaining), fleet.num_ships))\n\n incoming_threats = sorted(incoming_threats, key=lambda incoming: incoming.turns_remaining)\n\n for threat in incoming_threats:\n def check_for_reinforcements(accum, fleet):\n if fleet.dest.id == planet.id and turn_processed < fleet.turns_remaining <= threat.turns_remaining:\n return accum + fleet.num_ships\n else:\n return accum\n\n reinforcements = reduce(check_for_reinforcements, game_info.my_fleets.values(), 0)\n\n if turn_processed == 0:\n reinforcements += round(threat.turns_remaining * planet.growth_rate)\n if reinforcements >= threat.num_ships:\n excess_ships = reinforcements - threat.num_ships\n elif reinforcements + planet.num_ships >= threat.num_ships:\n avail_ships = planet.num_ships - (threat.num_ships - reinforcements)\n excess_ships = avail_ships\n else:\n shortfall = threat.num_ships - (planet.num_ships + reinforcements)\n defense_tasks.append(DefenseTask(planet, threat.turns_remaining, shortfall))\n else:\n produced_ships = round((threat.turns_remaining - turn_processed) * planet.growth_rate)\n total_defense = produced_ships + excess_ships + reinforcements\n shortfall = threat.num_ships - total_defense\n\n if shortfall > 0:\n defense_tasks.append(DefenseTask(planet, threat.turns_remaining, shortfall))\n avail_ships = 0\n excess_ships = 0\n else:\n excess_ships = total_defense - threat.num_ships\n if avail_ships > excess_ships:\n avail_ships = excess_ships\n\n turn_processed = threat.turns_remaining\n\n if avail_ships > 0:\n reserves.append(ReserveForce(planet, avail_ships))\n\n # Send available ships to defend planets under attack\n defense_tasks = sorted(defense_tasks, key=lambda def_task: def_task.turns_remaining)\n for task in defense_tasks:\n defense_forces = []\n ships_required = task.num_ships\n\n for reserve in reserves:\n if reserve.planet.id == task.planet.id:\n continue\n distance = round(reserve.planet.distance_to(task.planet))\n if distance <= task.turns_remaining and reserve.num_ships > 0 and ships_required > 0:\n if ships_required >= reserve.num_ships:\n ships_required -= reserve.num_ships\n defense_forces.append(reserve)\n else:\n defense_forces.append(ReserveForce(reserve.planet, ships_required))\n ships_required = 0\n\n if ships_required == 0:\n for force in defense_forces:\n game_info.planet_order(force.planet, task.planet, force.num_ships)\n used_reserve = next(reserve for reserve in reserves if reserve.planet.id == force.planet.id)\n used_reserve.num_ships -= task.num_ships\n task.num_ships = 0\n\n # Attack with any remaining available ships\n reserves = [reserve for reserve in reserves if reserve.num_ships > 0]\n reserves = sorted(reserves, key=lambda res: res.num_ships, reverse=True)\n for reserve in reserves:\n for planet in game_info.not_my_planets.values():\n def get_inbound_ships(accum, fleet):\n if fleet.dest.id == planet.id:\n return accum + fleet.num_ships\n else:\n return accum\n my_ships = reduce(get_inbound_ships, game_info.my_fleets.values(), 0)\n enemy_ships = reduce(get_inbound_ships, game_info.enemy_fleets.values(), 0)\n if planet.owner_id == 0:\n neutral_ships = planet.num_ships\n ship_balance = my_ships - (neutral_ships + enemy_ships)\n if ship_balance > 0: # Already have enough ships en route to conquer\n continue\n ship_balance += reserve.num_ships\n if ship_balance > 0:\n game_info.planet_order(reserve.planet, planet, ship_balance)\n reserve.num_ships -= ship_balance\n elif planet.owner_id == 2:\n enemy_ships += round(reserve.planet.distance_to(planet) * planet.growth_rate)\n ship_balance = my_ships - enemy_ships\n if ship_balance > 0:\n continue\n ship_balance += reserve.num_ships\n if ship_balance > 0:\n game_info.planet_order(reserve.planet, planet, ship_balance)\n reserve.num_ships -= ship_balance\n\n\nclass DefenseTask(object):\n def __init__(self, planet, turns_remaining, num_ships):\n self.planet = planet\n self.turns_remaining = turns_remaining\n self.num_ships = num_ships\n\n\nclass ReserveForce(object):\n def __init__(self, planet, num_ships):\n self.planet = planet\n self.num_ships = num_ships\n\n\nclass PlanetThreat(object):\n def __init__(self, turns_remaining, num_ships):\n self.turns_remaining = turns_remaining\n self.num_ships = num_ships\n\n","sub_path":"11_spike/bots/MrComplex.py","file_name":"MrComplex.py","file_ext":"py","file_size_in_byte":6612,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"426096794","text":"from chatterbot import ChatBot\nfrom chatterbot.trainers import ListTrainer\n\nchatbot = ChatBot(\n 'করোনা বট',\n storage_adapter='chatterbot.storage.SQLStorageAdapter',\n logic_adapters=[\n 'chatterbot.logic.MathematicalEvaluation',\n 'chatterbot.logic.TimeLogicAdapter',\n 'chatterbot.logic.BestMatch',\n {\n 'import_path': 'chatterbot.logic.BestMatch',\n 'default_response': 'আমি দুঃখিত, কিন্তু আমি বুঝতে পারছিনা। আমি এখনো শিখছি।',\n 'maximum_similarity_threshold': 0.70\n }\n ],\n database_uri='sqlite:///database.sqlite3'\n)\n\n\ntrainer = ListTrainer(chatbot)\n\ntraining_data_quesans = open('ques_ans.txt').read().splitlines()\ntraining_data_personal = open('personal_ques.txt').read().splitlines()\n\ntraining_data = training_data_quesans + training_data_personal\n\ntrainer.train(training_data)\n\n","sub_path":"chatbot.py","file_name":"chatbot.py","file_ext":"py","file_size_in_byte":963,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"326993173","text":"from pyasn1.type import univ, namedtype, tag, constraint\r\n\r\nfrom .common import ConditionalTime, Credentials, InvokeId, IntUnsignedLong,\\\r\n SpaceLinkDataUnit, Time\r\nfrom .pdu import SleScheduleStatusReportInvocation, SleStopInvocation,\\\r\n SleAcknowledgement, SleScheduleStatusReportReturn\r\nfrom .raf_structure import RafParameterName, RequestedFrameQuality, AntennaId,\\\r\n CarrierLockStatus, DiagnosticRafGet, DiagnosticRafStart, FrameQuality,\\\r\n FrameSyncLockStatus, LockStatus, Notification, RafGetParameter,\\\r\n RafProductionStatus, SymbolLockStatus\r\nfrom .bind import SleBindInvocation, SleBindReturn, SlePeerAbort,\\\r\n SleUnbindInvocation, SleUnbindReturn\r\n\r\n\r\n# Incoming PDUs\r\n\r\nclass RafStartInvocation(univ.Sequence):\r\n componentType = namedtype.NamedTypes(\r\n namedtype.NamedType('invokerCredentials', Credentials()),\r\n namedtype.NamedType('invokeId', InvokeId()),\r\n namedtype.NamedType('startTime', ConditionalTime()),\r\n namedtype.NamedType('stopTime', ConditionalTime()),\r\n namedtype.NamedType('requestedFrameQuality', RequestedFrameQuality())\r\n )\r\n\r\n\r\nclass RafGetParameterInvocation(univ.Sequence):\r\n componentType = namedtype.NamedTypes(\r\n namedtype.NamedType('invokerCredentials', Credentials()),\r\n namedtype.NamedType('invokeId', InvokeId()),\r\n namedtype.NamedType('rafParameter', RafParameterName())\r\n )\r\n\r\n\r\nclass RafUserToProviderPdu(univ.Choice):\r\n componentType = namedtype.NamedTypes(\r\n namedtype.NamedType('rafBindInvocation', SleBindInvocation().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 100))),\r\n namedtype.NamedType('rafBindReturn', SleBindReturn().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 101))),\r\n namedtype.NamedType(\r\n 'rafUnbindInvocation', SleUnbindInvocation().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 102))),\r\n namedtype.NamedType('rafUnbindReturn', SleUnbindReturn().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 103))),\r\n namedtype.NamedType('rafStartInvocation', RafStartInvocation().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 0))),\r\n namedtype.NamedType('rafStopInvocation', SleStopInvocation().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 2))),\r\n namedtype.NamedType(\r\n 'rafScheduleStatusReportInvocation',\r\n SleScheduleStatusReportInvocation().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 4))),\r\n namedtype.NamedType(\r\n 'rafGetParameterInvocation', RafGetParameterInvocation().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 6))),\r\n namedtype.NamedType('rafPeerAbortInvocation', SlePeerAbort().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatSimple, 104)))\r\n )\r\n\r\n\r\n# Outgoing PDUs\r\n\r\nclass RafStartReturn(univ.Sequence):\r\n componentType = namedtype.NamedTypes(\r\n namedtype.NamedType('performerCredentials', Credentials()),\r\n namedtype.NamedType('invokeId', InvokeId()),\r\n namedtype.NamedType('result', univ.Choice(\r\n componentType=namedtype.NamedTypes(\r\n namedtype.NamedType(\r\n 'positiveResult', univ.Null().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatSimple, 0))),\r\n namedtype.NamedType(\r\n 'negativeResult', DiagnosticRafStart().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 1)))\r\n )))\r\n )\r\n\r\n\r\nclass RafTransferDataInvocation(univ.Sequence):\r\n componentType = namedtype.NamedTypes(\r\n namedtype.NamedType('invokerCredentials', Credentials()),\r\n namedtype.NamedType('earthReceiveTime', Time()),\r\n namedtype.NamedType('antennaId', AntennaId()),\r\n namedtype.NamedType('dataLinkContinuity', univ.Integer().subtype(\r\n subtypeSpec=constraint.ValueRangeConstraint(-1, 16777215))),\r\n namedtype.NamedType('deliveredFrameQuality', FrameQuality()),\r\n namedtype.NamedType('privateAnnotation', univ.Choice(\r\n componentType=namedtype.NamedTypes(\r\n namedtype.NamedType('null', univ.Null().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatSimple, 0))),\r\n namedtype.NamedType('notNull', univ.OctetString().subtype(\r\n subtypeSpec=constraint.ValueSizeConstraint(\r\n 1, 128)).subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatSimple, 1)))\r\n ))),\r\n namedtype.NamedType('data', SpaceLinkDataUnit())\r\n )\r\n\r\n\r\nclass RafSyncNotifyInvocation(univ.Sequence):\r\n componentType = namedtype.NamedTypes(\r\n namedtype.NamedType('invokerCredentials', Credentials()),\r\n namedtype.NamedType('notification', Notification())\r\n )\r\n\r\n\r\nclass FrameOrNotification(univ.Choice):\r\n componentType = namedtype.NamedTypes(\r\n namedtype.NamedType(\r\n 'annotatedFrame', RafTransferDataInvocation().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 0))),\r\n namedtype.NamedType(\r\n 'syncNotification', RafSyncNotifyInvocation().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 1)))\r\n )\r\n\r\n\r\nclass RafTransferBuffer(univ.SequenceOf):\r\n componentType = FrameOrNotification()\r\n\r\n\r\nclass RafStatusReportInvocation(univ.Sequence):\r\n componentType = namedtype.NamedTypes(\r\n namedtype.NamedType('invokerCredentials', Credentials()),\r\n namedtype.NamedType('errorFreeFrameNumber', IntUnsignedLong()),\r\n namedtype.NamedType('deliveredFrameNumber', IntUnsignedLong()),\r\n namedtype.NamedType('frameSyncLockStatus', FrameSyncLockStatus()),\r\n namedtype.NamedType('symbolSyncLockStatus', SymbolLockStatus()),\r\n namedtype.NamedType('subcarrierLockStatus', LockStatus()),\r\n namedtype.NamedType('carrierLockStatus', CarrierLockStatus()),\r\n namedtype.NamedType('productionStatus', RafProductionStatus())\r\n )\r\n\r\n\r\nclass RafGetParameterReturn(univ.Sequence):\r\n componentType = namedtype.NamedTypes(\r\n namedtype.NamedType('performerCredentials', Credentials()),\r\n namedtype.NamedType('invokeId', InvokeId()),\r\n namedtype.NamedType('result', univ.Choice(\r\n componentType=namedtype.NamedTypes(\r\n namedtype.NamedType(\r\n 'positiveResult', RafGetParameter().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext,\r\n tag.tagFormatConstructed, 0))),\r\n namedtype.NamedType(\r\n 'negativeResult', DiagnosticRafGet().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 1)))\r\n )))\r\n )\r\n\r\n\r\nclass RafProviderToUserPdu(univ.Choice):\r\n componentType = namedtype.NamedTypes(\r\n namedtype.NamedType('rafBindInvocation', SleBindInvocation().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 100))),\r\n namedtype.NamedType('rafBindReturn', SleBindReturn().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 101))),\r\n namedtype.NamedType(\r\n 'rafUnbindInvocation', SleUnbindInvocation().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 102))),\r\n namedtype.NamedType('rafUnbindReturn', SleUnbindReturn().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 103))),\r\n namedtype.NamedType('rafStartReturn', RafStartReturn().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 1))),\r\n namedtype.NamedType('rafStopReturn', SleAcknowledgement().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 3))),\r\n namedtype.NamedType('rafTransferBuffer', RafTransferBuffer().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 8))),\r\n namedtype.NamedType(\r\n 'rafScheduleStatusReportReturn',\r\n SleScheduleStatusReportReturn().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 5))),\r\n namedtype.NamedType(\r\n 'rafStatusReportInvocation', RafStatusReportInvocation().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 9))),\r\n namedtype.NamedType(\r\n 'rafGetParameterReturn', RafGetParameterReturn().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatConstructed, 7))),\r\n namedtype.NamedType('rafPeerAbortInvocation', SlePeerAbort().subtype(\r\n implicitTag=tag.Tag(\r\n tag.tagClassContext, tag.tagFormatSimple, 104)))\r\n )\r\n","sub_path":"slecommon/datatypes/raf_pdu.py","file_name":"raf_pdu.py","file_ext":"py","file_size_in_byte":9855,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"4633909","text":"#!/usr/bin/env python\n# Licensed to Cloudera, Inc. under one\n# or more contributor license agreements. See the NOTICE file\n# distributed with this work for additional information\n# regarding copyright ownership. Cloudera, Inc. licenses this file\n# to you under the Apache License, Version 2.0 (the\n# \"License\"); you may not use this file except in compliance\n# with the License. You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom crequest.middleware import CrequestMiddleware\n\nfrom desktop.conf import ENABLE_ORGANIZATIONS\n\n\ndef default_organization():\n from useradmin.models import Organization\n default_organization, created = Organization.objects.get_or_create(name='default', domain='default')\n return default_organization\n\n\ndef get_user_request_organization():\n request = CrequestMiddleware.get_request()\n return request.user.organization if request and hasattr(request, 'user') and request.user.is_authenticated() else default_organization()\n\n\ndef _fitered_queryset(queryset, by_owner=False):\n request = CrequestMiddleware.get_request()\n\n # Avoid infinite recursion on very first retrieval of the user\n if ENABLE_ORGANIZATIONS.get() and \\\n request and hasattr(request, 'user') and hasattr(request.user, '_wrapped') and type(request.user._wrapped) is not object and \\\n request.user.is_authenticated():\n if by_owner:\n filters = {'owner__organization': request.user.organization}\n else:\n filters = {'organization': request.user.organization}\n\n queryset = queryset.filter(**filters)\n\n return queryset\n","sub_path":"apps/useradmin/src/useradmin/organization.py","file_name":"organization.py","file_ext":"py","file_size_in_byte":1908,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"409710148","text":"from unittest import TestCase\n\nfrom libmt94x.transfer_failed_codes import InvalidTransferFailedCodeError\nfrom libmt94x.transfer_failed_codes import TransferFailed\nfrom libmt94x.transfer_failed_codes import TransferFailedMisc\nfrom libmt94x.transfer_failed_codes import TransferFailedSEPA\n\n\nclass TransferFailedTests(TestCase):\n def setUp(self):\n self.sepa = TransferFailedSEPA.get_instance()\n self.misc = TransferFailedMisc.get_instance()\n self.any = TransferFailed.get_instance()\n\n # SEPA codes tests\n\n def test_sepa_code_ok(self):\n reason = self.sepa.resolve_code('AC01')\n self.assertEquals(reason, 'Rekeningnummer incorrect')\n\n rv = self.sepa.code_is_valid('AC01')\n self.assertEquals(rv, True)\n\n def test_sepa_code_invalid(self):\n # This code belongs to the Misc category\n with self.assertRaises(InvalidTransferFailedCodeError):\n self.sepa.resolve_code('AC03')\n\n rv = self.sepa.code_is_valid('AC03')\n self.assertEquals(rv, False)\n\n # Misc codes tests\n\n def test_misc_code_ok(self):\n reason = self.misc.resolve_code('AC03')\n self.assertEquals(reason, 'Ongeldig rekeningnummer crediteur')\n\n rv = self.misc.code_is_valid('AC03')\n self.assertEquals(rv, True)\n\n def test_misc_code_invalid(self):\n # This code belongs to the SEPA category\n with self.assertRaises(InvalidTransferFailedCodeError):\n self.misc.resolve_code('AC01')\n\n rv = self.misc.code_is_valid('AC01')\n self.assertEquals(rv, False)\n\n # Any/All codes tests\n\n def test_any_sepa_code_ok(self):\n # Can resolve a SEPA code\n reason = self.any.resolve_code('AC01')\n self.assertEquals(reason, 'Rekeningnummer incorrect')\n\n rv = self.any.code_is_valid('AC01')\n self.assertEquals(rv, True)\n\n def test_any_misc_code_ok(self):\n # Can resolve a Misc code\n reason = self.any.resolve_code('AC03')\n self.assertEquals(reason, 'Ongeldig rekeningnummer crediteur')\n\n rv = self.any.code_is_valid('AC03')\n self.assertEquals(rv, True)\n\n def test_any_code_invalid(self):\n # This code is not defined for either SEPA nor Misc\n with self.assertRaises(InvalidTransferFailedCodeError):\n self.any.resolve_code('AC00')\n\n rv = self.any.code_is_valid('AC00')\n self.assertEquals(rv, False)\n","sub_path":"tests/test_transfer_failed_codes.py","file_name":"test_transfer_failed_codes.py","file_ext":"py","file_size_in_byte":2415,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"581448322","text":"import numpy as np\nimport pandas as pd\n\ntrain = pd.read_csv(\"../input/train.csv\", dtype={\"Age\": np.float64}, )\ntest = pd.read_csv(\"../input/test.csv\", dtype={\"Age\": np.float64}, )\n\npredictors = [\"Pclass\", \"Sex\", \"Age\", \"SibSp\", \"Parch\", \"Fare\", \"Embarked\"]\nmark=\"Survived\"\n\n\ndef harmonize_data(titanic):\n \n titanic[\"Age\"] = titanic[\"Age\"].fillna(titanic[\"Age\"].median())\n titanic[\"Age\"].median()\n \n \n \n titanic.loc[titanic[\"Sex\"] == \"male\", \"Sex\"] = 0\n titanic.loc[titanic[\"Sex\"] == \"female\", \"Sex\"] = 1\n \n titanic[\"Embarked\"] = titanic[\"Embarked\"].fillna(\"S\")\n\n titanic.loc[titanic[\"Embarked\"] == \"S\", \"Embarked\"] = 0\n titanic.loc[titanic[\"Embarked\"] == \"C\", \"Embarked\"] = 1\n titanic.loc[titanic[\"Embarked\"] == \"Q\", \"Embarked\"] = 2\n\n titanic[\"Fare\"] = titanic[\"Fare\"].fillna(titanic[\"Fare\"].median())\n\n return titanic\n\n\ndef create_submission(clf, train, test, predictors, filename):\n\n clf.fit(train[predictors], train[\"Survived\"])\n predictions = clf.predict(test[predictors])\n\n submission = pd.DataFrame({\n \"PassengerId\": test[\"PassengerId\"],\n \"Survived\": predictions\n })\n \n submission.to_csv(filename, index=False)\n\n\nfrom sklearn.model_selection import cross_val_score\n\ndef validation_scores(clf, train_data):\n scores = cross_val_score(\n clf,\n train_data[predictors],\n train_data[mark],\n cv=3\n )\n return scores.mean()\n\n\ntrain_data = harmonize_data(train)\ntest_data = harmonize_data(test)\n\n\ndef compare_metods(classifiers, train_data):\n names, scores = [], []\n for name, clf in classifiers:\n names.append(name)\n scores.append(validation_scores(clf, train_data))\n return pd.DataFrame(scores, index=names, columns=['Scores'])\n\n\n\nfrom sklearn.ensemble import GradientBoostingClassifier\n\nclassifiers = [\n (\"Gradient Boosting\", GradientBoostingClassifier(max_depth=4)),\n]\n\nres = compare_metods(classifiers, train_data)\nres\n\nprint(res)\n","sub_path":"script/gradient.py","file_name":"gradient.py","file_ext":"py","file_size_in_byte":1973,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"76572322","text":"import json\nimport base64\nimport numpy as np\nimport sys\nimport time\nimport os\nimport cv2\nfrom PIL import Image\nimport io\nimport boto3\nimport urllib\n\n\n# construct the argument parse and parse the arguments\nconfthres = 0.5\nnmsthres = 0.1\ns3_resource = boto3.resource('s3')\ndynamodb = boto3.client('dynamodb')\nTABLE_NAME = 'IMAGE_URL'\nbasic_url = 'https://s3-image-storing-bucket-fit5225.s3.amazonaws.com/'\n\n\n\ndef load_model(configpath, weightspath):\n # load our YOLO object detector trained on COCO dataset (80 classes)\n print(\"[INFO] loading YOLO from disk...\")\n net = cv2.dnn.readNetFromDarknet(configpath, weightspath)\n return net\n \ndef do_prediction(image, net, LABELS):\n (H, W) = image.shape[:2]\n # determine only the *output* layer names that we need from YOLO\n ln = net.getLayerNames()\n ln = [ln[i[0] - 1] for i in net.getUnconnectedOutLayers()]\n\n # construct a blob from the input image and then perform a forward\n # pass of the YOLO object detector, giving us our bounding boxes and\n # associated probabilities\n blob = cv2.dnn.blobFromImage(image, 1 / 255.0, (416, 416),\n swapRB=True, crop=False)\n net.setInput(blob)\n start = time.time()\n layerOutputs = net.forward(ln)\n # print(layerOutputs)\n end = time.time()\n\n # show timing information on YOLO\n print(\"[INFO] YOLO took {:.6f} seconds\".format(end - start))\n\n # initialize our lists of detected bounding boxes, confidences, and\n # class IDs, respectively\n boxes = []\n confidences = []\n classIDs = []\n\n # loop over each of the layer outputs\n for output in layerOutputs:\n # loop over each of the detections\n for detection in output:\n # extract the class ID and confidence (i.e., probability) of\n # the current object detection\n scores = detection[5:]\n # print(scores)\n classID = np.argmax(scores)\n # print(classID)\n confidence = scores[classID]\n\n # filter out weak predictions by ensuring the detected\n # probability is greater than the minimum probability\n if confidence > confthres:\n # scale the bounding box coordinates back relative to the\n # size of the image, keeping in mind that YOLO actually\n # returns the center (x, y)-coordinates of the bounding\n # box followed by the boxes' width and height\n box = detection[0:4] * np.array([W, H, W, H])\n (centerX, centerY, width, height) = box.astype(\"int\")\n\n # use the center (x, y)-coordinates to derive the top and\n # and left corner of the bounding box\n x = int(centerX - (width / 2))\n y = int(centerY - (height / 2))\n\n # update our list of bounding box coordinates, confidences,\n # and class IDs\n boxes.append([x, y, int(width), int(height)])\n\n confidences.append(float(confidence))\n classIDs.append(classID)\n\n # apply non-maxima suppression to suppress weak, overlapping bounding boxes\n idxs = cv2.dnn.NMSBoxes(boxes, confidences, confthres,\n nmsthres)\n\n # TODO Prepare the output as required to the assignment specification(DONE)\n # ensure at least one detection exists\n if len(idxs) > 0:\n # loop over the indexes we are keeping\n # create an array of the results\n result = []\n #loop over the objects and put them in an array and return the result\n for i in idxs.flatten():\n result.append(LABELS[classIDs[i]])\n return result\n \n\n \n\n\n\nbucket = 'library-bucket-fit5225'\nkey1 = 'coco.names'\nkey2 = 'yolov3-tiny.cfg'\nkey3 = 'yolov3-tiny.weights'\n\n\n\ns3_bucket = s3_resource.Bucket(bucket)\ns3_obj1 = s3_bucket.Object(key=key1)\ns3_obj2 = s3_bucket.Object(key=key2)\ns3_obj3 = s3_bucket.Object(key=key3)\n\n\n\n\n\n\ndef lambda_handler(event, context):\n \n print('Lamda handler started')\n bucket2 = event['Records'][0]['s3']['bucket']['name']\n print(bucket2)\n key = event['Records'][0]['s3']['object']['key']\n keyImage = urllib.parse.unquote_plus(key,encoding = 'utf-8')\n print(keyImage)\n s3_bucket_2 = s3_resource.Bucket(bucket2)\n s3_obj5 = s3_bucket_2.Object(key=key)\n \n # read image which is uplaoded to bucker\n \n \n # read yolo files from bucket \n #boto to get file from bucket\n #make this into an array\n labels = s3_obj1.get().get('Body').read()\n labels = str(labels)\n labels = labels[2:]\n labels.replace('\"','')\n #print(labels)\n labels = labels.split(\"\\\\n\")\n #print(labels)\n cfg = s3_obj2.get().get('Body').read()\n weights = s3_obj3.get().get('Body').read()\n\n\n img_data = s3_obj5.get().get('Body').read()\n\n #image from the s3 bucket\n img = Image.open(io.BytesIO(img_data))\n npimg = np.array(img)\n image = npimg.copy()\n image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)\n nets = load_model(cfg, weights)\n result = do_prediction(image, nets, labels)\n tags = 'abc'\n if result == None:\n tags = list(result)\n else: \n \n if len(result) > 1 :\n tags = []\n for i in result:\n if i not in tags:\n tags.append(i)\n tags = list(tags) \n else:\n tags = list(result)\n \n url = basic_url+keyImage\n print(url)\n print(tags)\n data = {}\n data['url_list'] = {'S' : url}\n data['tags'] = {'SS' : tags}\n dynamodb.put_item(TableName=TABLE_NAME, Item=data)\n print('Inserted into db')\n\n \n","sub_path":"object_detection_lambda_final.py","file_name":"object_detection_lambda_final.py","file_ext":"py","file_size_in_byte":5673,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"304916741","text":"from collections import namedtuple\nfrom itertools import cycle\nfrom typing import List\n\nimport numpy as np\nimport plotly.colors\nfrom astropy import units as u\nfrom astropy.coordinates import CartesianRepresentation\nfrom plotly.graph_objects import Figure\n\nfrom poliastro.plotting.util import BODY_COLORS, generate_label\nfrom poliastro.util import norm\n\n\nclass Trajectory(namedtuple(\"Trajectory\", [\"trajectory\", \"state\", \"label\", \"color\"])):\n pass\n\n\nclass BaseOrbitPlotter:\n \"\"\"\n Parent Class for the 2D and 3D OrbitPlotter Classes based on Plotly.\n \"\"\"\n\n def __init__(self, figure=None):\n self._figure = figure or Figure()\n self._layout = None\n\n self._trajectories = [] # type: List[Trajectory]\n\n self._attractor = None\n self._attractor_radius = np.inf * u.km\n\n self._color_cycle = cycle(plotly.colors.DEFAULT_PLOTLY_COLORS)\n\n @property\n def trajectories(self):\n return self._trajectories\n\n def _set_attractor(self, attractor):\n if self._attractor is None:\n self._attractor = attractor\n elif attractor is not self._attractor:\n raise NotImplementedError(\n f\"Attractor has already been set to {self._attractor.name}.\"\n )\n\n def set_attractor(self, attractor):\n \"\"\"Sets plotting attractor.\n\n Parameters\n ----------\n attractor : ~poliastro.bodies.Body\n Central body.\n\n \"\"\"\n self._set_attractor(attractor)\n\n def _redraw_attractor(self):\n # Select a sensible value for the radius: realistic for low orbits,\n # visible for high and very high orbits\n min_radius = min(\n [\n trajectory.represent_as(CartesianRepresentation).norm().min() * 0.15\n for trajectory, _, _, _ in self._trajectories\n ]\n or [0 * u.m]\n )\n radius = max(self._attractor.R.to(u.km), min_radius.to(u.km))\n # TODO: Remove previously plotted sphere?\n self._plot_sphere(\n radius,\n BODY_COLORS.get(self._attractor.name, \"#999999\"),\n self._attractor.name,\n )\n\n self._attractor_radius = radius\n\n def _plot_point(self, radius, color, name, center=None):\n raise NotImplementedError\n\n def _plot_sphere(self, radius, color, name, center=None):\n raise NotImplementedError\n\n def plot_trajectory(self, trajectory, *, label=None, color=None):\n \"\"\"Plots a precomputed trajectory.\n\n An attractor must be set first.\n\n Parameters\n ----------\n trajectory : ~astropy.coordinates.CartesianRepresentation\n Trajectory to plot.\n label : string, optional\n color : string, optional\n\n \"\"\"\n if self._attractor is None:\n raise ValueError(\n \"An attractor must be set up first, please use \"\n \"set_attractor(Major_Body) or plot(orbit).\"\n )\n else:\n if color is None:\n color = next(self._color_cycle)\n\n trace = self._plot_trajectory(trajectory, str(label), color, False)\n\n self._trajectories.append(\n Trajectory(trajectory, None, label, trace.line.color)\n )\n\n if not self._figure._in_batch_mode:\n return self.show()\n\n def _plot_trajectory(self, trajectory, label, color, dashed):\n raise NotImplementedError\n\n def plot(self, orbit, *, label=None, color=None):\n \"\"\"Plots state and osculating orbit in their plane.\n\n Parameters\n ----------\n orbit : ~poliastro.twobody.orbit.Orbit\n Orbit to plot.\n label : string, optional\n Label of the orbit.\n color : string, optional\n Color of the line and the position.\n\n \"\"\"\n if color is None:\n color = next(self._color_cycle)\n\n self._set_attractor(orbit.attractor)\n\n label = generate_label(orbit, label)\n trajectory = orbit.sample()\n\n trace = self._plot_trajectory(trajectory, label, color, True)\n\n self._trajectories.append(\n Trajectory(trajectory, orbit.r, label, trace.line.color)\n )\n\n # Redraw the attractor now to compute the attractor radius\n self._redraw_attractor()\n\n # Plot required 2D/3D shape in the position of the body\n radius = min(\n self._attractor_radius * 0.5, (norm(orbit.r) - orbit.attractor.R) * 0.5\n ) # Arbitrary thresholds\n self._plot_point(radius, color, label, center=orbit.r)\n\n if not self._figure._in_batch_mode:\n return self.show()\n\n def _prepare_plot(self):\n if self._attractor is not None:\n self._redraw_attractor()\n\n self._figure.layout.update(self._layout)\n\n def show(self):\n \"\"\"Shows the plot in the Notebook.\n\n Updates the layout and returns the underlying figure.\n\n \"\"\"\n self._prepare_plot()\n return self._figure\n","sub_path":"src/poliastro/plotting/_base.py","file_name":"_base.py","file_ext":"py","file_size_in_byte":5013,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"163026122","text":"#!/usr/bin/env python3\n# coding=utf-8\n\"\"\"Implement Hyman's (flawed) concurrent lock acquisition algorithm.\n\nThis module implements Hyman's concurrent lock acquisition algorithm, as laid\nout in his 1966 letter \"Comments on a Problem in Concurrent Programming\nControl.\"\n\nThe original problem is as follows: given a group of N computers working\ntogether on some task, how can we ensure that only one works on a critical\nportion fo the task at a time? Some requirements are as follows:\n\n#. If two computers attempt to enter the critical section at the same time, the\n solution can *not* be to choose a winner based on some static property of\n the computers, such as their IDs\n#. Assumptions may not be made about the relative speeds of the computers. The\n computers may change speed over time.\n#. If a computer leaves the cluster while not in the critical section, the\n remaining computers must be able to continue working on the task.\n#. Theoretically infinite waits are permissible, so long as they're\n statistically unlikely.\n\nIn addition, Hyman notes that his algorithm works when there are only two\ncomputers. The original code is poorly formatted and, to my eyes, illegible.\nFurthermore, Knuth noted that \"there are 15 syntactic ALGOL errors in 12 lines\nof program!\" The California State University, Stanislaus, has provided a\n`translation`_. Here is a slightly cleaned-up rendition of that translation,\nwhich I will use as the basis of my program:\n\n // Each of two processes, P0 and P1 executes the code below, P0 executes\n // Protocol(0,1), and P1 executes Protocol(1,0).\n\n int turn = 0;\n int flag[2] = false;\n void Protocol (int me, int you) {\n do {\n flag[me] = true;\n while (turn != me) {\n while (flag[you]) {\n /* do nothing */;\n }\n turn = me;\n }\n CriticalSection(me);\n flag[me] = false;\n RemainderSection(me);\n } while (true);\n }\n\nFor an example of how this procedure can fail, consider the following\nscenario:\n\n#. Worker 1 raises flag 1.\n# Worker 1 notes that it is not their turn to do work.\n#. Worker 1 notes that flag 0 is down.\n#. Worker 0 raises his flag.\n#. Worker 0 notes that it is his turn to do work, and proceeds into the\n critical section.\n#. Worker 1 declares that it is his turn to do work, and proceeds into the\n critical section.\n\n.. _translation:\n https://www.cs.csustan.edu/~john/Classes/Previous_Semesters/CS3750_OperatingSys_I/2009_04_Fall/hymanProb.html\n\"\"\"\nfrom threading import Thread\nfrom typing import Callable, List\n\nWORKERS = 2\n\nturn: int = 0\nflags: List[bool] = [False, False]\n\n\ndef main() -> None:\n \"\"\"Spawn several threads.\"\"\"\n threads = [Thread(target=do_work, args=(i,)) for i in range(WORKERS)]\n for thread in threads:\n thread.start()\n for thread in threads:\n thread.join()\n\n\ndef do_work(worker_id: int) -> None:\n \"\"\"Do bogus work.\n\n Do work, acquire the lock, do more work, release the lock, and do more\n work. (No work is actually done.)\n\n :param worker_id: A unique identifier for this thread. Used in lock\n management logic.\n \"\"\"\n lock_and_call(worker_id, print, f'Worker {worker_id} doing critical work.')\n\n\ndef lock_and_call(worker_id: int, function: Callable, *args, **kwargs) -> None:\n \"\"\"Acquire the global lock, call the given function, and release the lock.\n\n :param worker_id: A unique identifier for this thread. Used in lock\n management logic.\n :param function: A function to call once the lock is acquired.\n :param args: Passed to ``function``.\n :param kwargs: Passed to ``function``.\n \"\"\"\n global turn # pylint:disable=global-statement,invalid-name\n global flags # pylint:disable=global-statement,invalid-name\n\n other_worker_id = 1 - worker_id\n flags[worker_id] = True\n while turn != worker_id:\n while flags[other_worker_id]:\n pass\n turn = worker_id\n print(f'Worker {worker_id} has acquired the lock.')\n function(*args, **kwargs)\n print(f'Worker {worker_id} is releasing the lock.')\n flags[worker_id] = False\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"python/hyman-concurrent/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4218,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"284820681","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nfrom django.conf import settings\nimport django.utils.timezone\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('common', '0001_initial'),\n migrations.swappable_dependency(settings.AUTH_USER_MODEL),\n ]\n\n operations = [\n migrations.CreateModel(\n name='Contact',\n fields=[\n ('id', models.AutoField(serialize=False, verbose_name='ID', auto_created=True, primary_key=True)),\n ('created_on', models.DateTimeField(auto_now_add=True, default=django.utils.timezone.now)),\n ('modified_on', models.DateTimeField(auto_now=True, default=django.utils.timezone.now)),\n ('first_name', models.CharField(max_length=255)),\n ('last_name', models.CharField(max_length=255)),\n ('email', models.EmailField(unique=True, max_length=255)),\n ('phone', models.CharField(max_length=255)),\n ('address', models.ForeignKey(blank=True, null=True, to='common.Address', default=None)),\n ('comments', models.ManyToManyField(blank=True, to='common.Comment')),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='Vendor',\n fields=[\n ('id', models.AutoField(serialize=False, verbose_name='ID', auto_created=True, primary_key=True)),\n ('created_on', models.DateTimeField(auto_now_add=True, default=django.utils.timezone.now)),\n ('modified_on', models.DateTimeField(auto_now=True, default=django.utils.timezone.now)),\n ('name', models.CharField(unique=True, max_length=255)),\n ('comments', models.ManyToManyField(blank=True, to='common.Comment')),\n ],\n options={\n 'abstract': False,\n },\n bases=(models.Model,),\n ),\n migrations.AddField(\n model_name='contact',\n name='represents',\n field=models.ForeignKey(to='vendors.Vendor'),\n preserve_default=True,\n ),\n migrations.AddField(\n model_name='contact',\n name='user',\n field=models.OneToOneField(blank=True, null=True, to=settings.AUTH_USER_MODEL, default=None),\n preserve_default=True,\n ),\n migrations.AlterUniqueTogether(\n name='contact',\n unique_together=set([('represents', 'first_name', 'last_name')]),\n ),\n ]\n","sub_path":"apps/vendors/migrations/0001_initial.py","file_name":"0001_initial.py","file_ext":"py","file_size_in_byte":2629,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"397517503","text":"import threading\nimport socket\n\nclass Tello:\n ANY_ADDR = '0.0.0.0'\n\n DEVICE_COMMAND_ADDR = '192.168.10.1'\n DEVICE_COMMAND_SEND_PORT = 8889\n DEVICE_COMMAND_RECV_PORT = 9000\n\n DEVICE_STATE_RECV_PORT = 8890\n\n DEVICE_VIDEO_RECV_PORT = 9000\n\n def __init__(self):\n self.deviceCommandSocket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\n self.deviceCommandSocket.bind((Tello.ANY_ADDR, Tello.DEVICE_COMMAND_RECV_PORT))\n\n self.deviceStateSocket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\n self.deviceStateSocket.bind((Tello.ANY_ADDR, Tello.DEVICE_STATE_RECV_PORT))\n\n self.__recvThreadQuit = False\n self.__recvThread = threading.Thread(target=self.__recv)\n self.__recvThread.start()\n\n self.__stateThreadQuit = False\n self.__stateThread = threading.Thread(target=self.__state)\n self.__stateThread.start()\n\n def __del__(self):\n self.deviceCommandSocket.close()\n\n def __recv(self):\n while not self.__recvThreadQuit:\n data, addr = self.deviceCommandSocket.recvfrom(128)\n if len(data) > 0:\n print(\"Last command received '%s' from %s\" % (data, addr))\n else:\n print(\"Status received from %s\" % addr)\n print(\"Done receiving\")\n\n def __state(self):\n while not self.__stateThreadQuit:\n data, addr = self.deviceStateSocket.recvfrom(512)\n if len(data) > 0:\n print(\"From %s:\" % addr[0])\n print(\" %s\" % data)\n else:\n print(\"Status received from %s\" % addr)\n print(\"Done state\")\n\n def testCmd(self):\n self.deviceCommandSocket.sendto('command', (Tello.DEVICE_COMMAND_ADDR, Tello.DEVICE_COMMAND_SEND_PORT))\n\n def release(self):\n print(\"Thread joining...\")\n\n self.__recvThreadQuit = True\n self.__recvThread.join()\n\n self.__stateThreadQuit = True\n self.__stateThread.join()\n\n print(\"Thread joined\")\n\nif __name__ == \"__main__\":\n tello = Tello()\n tello.testCmd()\n tello.release()\n","sub_path":"tello.py","file_name":"tello.py","file_ext":"py","file_size_in_byte":2104,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"198007305","text":"#!/usr/bin/python\n\nimport os, codecs, re, argparse, sys\nfrom io import StringIO\nimport unicodecsv as csv\nfrom lxml import etree\n\nfrom test_anystyle_parser import *\n\n# THE FOLLOWING THREE \"for\" STATMENTS SKETCH OUT HOW THE SCRIPT MIGHT RUN: \n#\n# 1) PROVIDED WITH A SET OF EXPLICIT FOLDER/FILE COMBINATIONS; \n# 2) PROVIDED WITH A FOLDER NAME, SO AS TO PROCESS ALL THE FILES \n# IN THE FOLDER; \n# 3) TO PROCESS JUST ONE FILE, EXPLICITLY NAMED AS A TEST.\n#\n\n#for file_name in ['csv/1984-1993_11_scopus.csv', 'csv/1984-1993_5_scopus.csv', 'csv/1984-1993_6_scopus.csv', 'csv/1994-1996 scopus.csv', 'csv/1997-1998.csv']:\n\n#for temp in os.listdir('some_folder/'):\n# file_name = 'some_folder/' + temp\n\n#R/S: DEFINE ARGUMENT\n\ncsv.field_size_limit(sys.maxsize)\nparser = argparse.ArgumentParser()\nparser.add_argument(\"filename\", help=\"The Name of the File to be Processed\", nargs='+')\nargs = parser.parse_args()\n\n\nfor file_name in args.filename:\n \n # OPEN THE OUTPUT FILE, TO WHICH WE WILL RIGHT OUR RESULTS\n\n file_handle = open('for_open_refine_' + file_name.split('/')[-1], 'w')\n w = csv.writer(file_handle, encoding='utf-8')\n\n # READ THE INPUT CSV INTO data, A LIST CONTAINING ALL THE CSV ROWS\n\n data = []\n for d in csv.reader(open(file_name, 'r'), encoding='utf-8'):\n data.append(d)\n\n # WRITE THE CSV HEADER ROW; NOTE THAT WE'RE ADDING HEADINGS FOR\n # OUT NEW COLUMNS.\n\n header_row = data[0]\n w.writerow(['row_type', 'source_file', 'scopus_EID', 'reference_n'] + header_row)\n\n # PROCESS THE REMAINING ROWS IN THE CSV, ONE-BY-BY, STARTING AT ROW 2 (\"[1:]\").\n\n for row in data[1:]:\n\n # PICK THE IMPORTANT BITS FROM THE ROW\n\n authors_sequence = row[0]\n #R/S HANDLE THE JR/SR PROBLEM\n authors_sequence = re.sub(r'(\\,)(\\s[JS]r\\.?)', r'\\2', authors_sequence)\n #R/S HANDLE THE \"et al\" PROBLEM\n authors_sequence = re.sub(r'\\s?[Ee][Tt]\\.?\\s?[Aa][Ll]\\.?','', authors_sequence)\n row[0]=authors_sequence\n title = row[1]\n year = row[2]\n journal = row[3]\n #R/S CLEAN UP JOURNAL TITLES\n journal = re.sub(r'[\\,\\\\\\/\\.]','',journal)\n journal = re.sub(r'^[iI]n\\s','',journal)\n row[3]=journal\n references_sequence = row[24]\n scopus_EID = row[41]\n\n # FIX THE TRUNCATION PROBLEM\n\n if len(row[24]) > 32767:\n row[24] = row[24][:32765]\n\n # WRITE THE ORIGINAL ROW TO THE OUTPUT, ADDING OUR NEW COLUMNS.\n\n new_row = ['scopus result', file_name.replace('csv/', ''), scopus_EID, ''] + row\n w.writerow(new_row)\n \n # PICK APART THE REFERENCES\n\n if references_sequence.strip() > '':\n\n # REFERENCES ARE SEPARATED BY SEMICOLON. SPLIT ON SEMICOLON,\n # THEN PROCESS EACH OF THE INDIVIDUAL REFERENCES. \n # reference_n IS THE SEQUENCE NUMBER OF THE REFERENCE\n # (STARTING WITH 0), r IS THE ACTUAL TEXT OF THE REFERENCE.\n \n for reference_n, r in enumerate(references_sequence.split(';')):\n\n if r.strip() > '':\n\n # CALL THE RUBY ANYSTYLE PARSER, THEN PARSE THE \n # RESULTS INTO SEPARATE FIELDS.\n #\n # THIS IS QUITE SLOW. THE ANYSTYLE, WHICH IS, AS\n # FAR AS WE'RE CONCERNED, A BLACKBOX, SEEMS TO \n # EXPEND QUITE A BIT OF PROCESSING PARSING REFERENCES.\n # R/S FIX THE JR/SR PROBLEM\n r = re.sub(r'(\\,)(\\s[JS]r\\.?)', r'\\2', r)\n # R/S FIX THE \"et al\" PROBLEM\n r = re.sub(r'\\,\\s?[Ee][Tt]\\.?\\s?[Aa][Ll]\\.?','', r)\n author, title, year, journal, volume, from_page, to_page = call_anystyle_parser(r)\n \n # ANYSTYLE PARSER OFTEN (10% OF THE TIME?) LEAVES\n # BITS OF THE ARTICLE OR JOURNAL IN WITH THE YEAR.\n # THIS CODE CLEANS THAT UP.\n\n if re.sub('[0-9]', '', year.strip()) == '':\n # IT'S OKAY, SO WE DON'T HAVE TO DO ANYTHING\n pass\n else:\n\n # SPLIT THE YEAR BY GROUPS OF NUMBERS\n\n year_parts = re.split('([0-9]+)', year.strip())\n\n # WHICH ONE OF THE PARTS IS REALLY THE YEAR?\n\n year_pos = -1\n for a in range(0, len(year_parts)):\n if len(year_parts[a]) == 4 and re.sub('[0-9]', '', year_parts[a]) == '':\n year_pos = a\n break\n\n # IF WE FOUND THE ACTUAL YEAR (\"year_pos != -1\"),\n # THE PUT THE PARTS WHERE THEY REALLY BELONG\n\n if year_pos != -1:\n if year_pos > 0:\n title = title + ' ' + ''.join(year_parts[:year_pos])\n year = year_parts[year_pos]\n for a in range(year_pos + 1, len(year_parts)):\n if re.sub('[0-9]', '', year_parts[a]) == '':\n volume = volume + year_parts[a]\n else:\n journal = journal + year_parts[a]\n\n # FINALLY, WE'RE DONE, AND CAN WRITE A \"reference\"\n # ROW TO THE OUTPUT FILE.\n \n #R/S CLEAN UP JOURNAL TITLES\n journal = re.sub(r'^[iI]n\\s','',journal)\n new_row = ['reference', file_name.replace('csv/', ''), scopus_EID, reference_n] + [author, title, year, journal, volume, '', '', from_page, to_page, '', '', '', '', '', '', '', '', '', '', '', '', '', '', r]\n w.writerow(new_row)\n\n file_handle.close()\n\n\n \n","sub_path":"make_openrefine_input_2.py","file_name":"make_openrefine_input_2.py","file_ext":"py","file_size_in_byte":5991,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"89911502","text":"import pandas as pd\r\nimport requests\r\nimport openpyxl\r\n\r\n# Read the data\r\nDataPath = './Hotel_reviews_NLP/Tripadvisor Review Part3.xlsx'\r\nLocations = pd.read_excel(DataPath, Headname=None, usecols=[2, 3])\r\nData = pd.read_excel(DataPath, Headname=None)\r\nLocList = []\r\nCityList = []\r\n\r\n# Save different Addresses of hotels\r\nfor i in range(0, Locations.shape[0]):\r\n if Locations.ix[i, 0] not in LocList:\r\n LocList.append(Locations.ix[i, 0])\r\n CityList.append(Locations.ix[i, 1])\r\nKey = ',+CA&key=AIzaSyBin-hGUWXjPnlT8q_ETK17zRYFbLsuw-4'\r\nGeocoding = 'https://maps.googleapis.com/maps/api/geocode/xml?address='\r\nLatitude = []\r\nlongitude = []\r\n\r\nprint(LocList.index(Locations.ix[10000, 0]))\r\n\r\n# Get Geocoding list\r\nfor i in range(0, len(LocList)):\r\n # for i in range(0, 1):\r\n url = Geocoding + LocList[i] + ' ' + CityList[i] + Key\r\n data = requests.get(url).text\r\n print(i)\r\n Latitude.append(data[data.find('') + 5:data.find('')])\r\n longitude.append(data[data.find('') + 5:data.find('')])\r\nprint(Latitude, longitude)\r\n\r\n# Transform all addresses to Geocoding\r\nData['Latitude'] = None\r\nData['Longitude'] = None\r\nfor i in range(0, Locations.shape[0]):\r\n index = LocList.index(Locations.ix[i, 0])\r\n Data.ix[i, 15] = Latitude[index]\r\n Data.ix[i, 16] = longitude[index]\r\nData.to_excel('./Hotel_reviews_NLP/Tripadvisor Review Part3-transformed.xlsx')\r\n","sub_path":"Step2.py","file_name":"Step2.py","file_ext":"py","file_size_in_byte":1408,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"451566282","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Jul 22 15:53:40 2016\n\"\"\"\nimport shutil\nimport os, datetime\n\ninPath = \"C:/1save/taipei_parking\" #這個是來源 folder path\n\noutPath = \"C:/1save/test\" #這個是目的地 folder path\n\nb4List = os.listdir(inPath)\n\naftList = []\n\ndef insertColon(string = str, index = int):\n \"\"\"\n 在原始字串的某個位置,插入冒號.\n 第一變數放目標字串\n 第二變數放要插入的位置\n \"\"\"\n return string[:index] + ':' + string[index:]\n\n\nfor name in b4List:\n timeIn = name[:-4].replace(\"_\", \"T\")\n timeIn = insertColon(timeIn, -4)\n timeIn = insertColon(timeIn, -2)\n obj = datetime.datetime.strptime(timeIn,'%Y-%m-%dT%H:%M:%S')\n obj = obj + datetime.timedelta(hours = -1)\n outName = outPath + \"/\" + obj.strftime(\"%Y-%m-%d_%H%M%S\") + \".csv\"\n shutil.copy(inPath + \"/\" + name, outName)\n #d0 = d0.strftime(\"%Y-%m-%d_%H%M%S\")\n aftList.append(obj) #obj 是轉換成 datetime後的東西\n","sub_path":"minus1hour.py","file_name":"minus1hour.py","file_ext":"py","file_size_in_byte":968,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"160054008","text":"\"\"\"\n-*- coding: utf-8 -*-\n@Name : pos_tagging-processor.py\n@Time : 2021/3/16 0:46\n@Author : 软工1701 李澳 U201716958\n@Desc : 预处理如 PTB,CTB,人民日报语料库 等数据集的函数工具包,目前先以处理 PTB 数据集的函数集合作为代表\n\"\"\"\n\n\n# 生成词频降序排列的词汇表文件\ndef generate_ptb_vocab(src_train_data, vocab_file):\n \"\"\"\n\n Args:\n src_train_data: ptb训练集数据作为输入\n vocab_file: 输出的词汇表文件\n\n Returns:\n null\n\n References:\n https://blog.csdn.net/qq_23031939/article/details/79759344\n https://chehongshu.blog.csdn.net/article/details/85288590?utm_medium=distribute.pc_relevant.none-task-blog-BlogCommendFromMachineLearnPai2-4.control&dist_request_id=&depth_1-utm_source=distribute.pc_relevant.none-task-blog-BlogCommendFromMachineLearnPai2-4.control\n\n Build Time:\n 2021/3/16 1:30\n\n \"\"\"\n\n import codecs\n import collections\n from operator import itemgetter\n\n # 统计单词出现的频率\n counter = collections.Counter()\n with codecs.open(src_train_data, \"r\", \"utf-8\") as f:\n for line in f:\n for word in line.strip().split():\n counter[word] += 1\n\n # 按照词频顺序对单词进行排序\n # Counter 集成于 dict 类,因此也可以使用字典的方法,此类返回一个以元素为 key 、元素个数为 value 的 Counter 对象集合\n # 依据key排序 itermgetter(1)为降序\n sorted_word_to_cnt = sorted(counter.items(), key=itemgetter(1), reverse=True)\n\n # 转换成单词string的list\n sorted_words_list = [x[0] for x in sorted_word_to_cnt]\n\n # 因为稍后需要在文本换行处加入句子结束符\"\"(或者说行结束符),这里预先将其加入词汇表\n # 而因为在 ptb 数据集中,输入数据已经将低频词汇替换为了,可以作为一个“单词”在以上步骤中被一并处理,\n # 所以不需要另外在单词的list中加入低频词汇/的项\n sorted_words_list = [\"\"] + sorted_words_list\n\n # 将降序后的单词序列写入到 vocab_file的词汇表文件中\n with codecs.open(vocab_file, 'w', 'utf-8') as file_output:\n for word in sorted_words_list:\n file_output.write(word + '\\n')\n\n\n# 将数据集中的训练、测试、验证的数据文件编号化\ndef vocab_transform_index(vocab, src_train_data, src_test_data, src_valid_data, train_index_data, test_index_data,\n valid_index_data):\n \"\"\"\n\n Args:\n vocab: 创建好的词汇降序表\n src_train_data: ptb 数据集中的训练数据\n src_test_data: ptb 数据集中的测试数据\n src_valid_data: ptb 数据集中的验证数据\n train_index_data: 经过编号化后的训练数据\n test_index_data: 经过编号化后的测试数据\n valid_index_data: 经过编号化后的验证数据\n\n Returns:\n null\n\n References:\n https://blog.csdn.net/qq_23031939/article/details/79759344\n https://chehongshu.blog.csdn.net/article/details/85288590?utm_medium=distribute.pc_relevant.none-task-blog-BlogCommendFromMachineLearnPai2-4.control&dist_request_id=&depth_1-utm_source=distribute.pc_relevant.none-task-blog-BlogCommendFromMachineLearnPai2-4.control\n\n Build Time:\n 2021/3/18 0:08\n\n \"\"\"\n import codecs\n with codecs.open(vocab, 'r', 'utf-8') as vocab_f:\n # 获取降序后的单词列表 vocab_list\n vocab_list = [single_vocab.strip() for single_vocab in vocab_f.readlines()]\n # 生成单词和行数的 dict\n word_index_dict = {k: v for (k, v) in zip(vocab_list, range(len(vocab_list)))}\n\n # 根据单词获取其在字典中的编号 index 的函数 get_word_index_from_dict\n def get_word_index_from_dict(word_content):\n return word_index_dict[word_content] if word_content in word_index_dict else word_index_dict['']\n\n # 编号化单个数据集文件的函数 src_to_index\n def src_to_index(src_data, index_data):\n # 指向原始格式数据集文件和编号化后的数据集文件指针 f_in, f_out\n f_in = codecs.open(src_data, 'r', 'utf-8')\n f_out = codecs.open(index_data, 'w', 'utf-8')\n for line in f_in:\n # 每一行/每一个句子中出现的单词集合(外加句子结束符\"\")\n words = line.strip().split() + [\"\"]\n # 单词集合转换成对应的 index 集合\n index_line = ' '.join([str(get_word_index_from_dict(word)) for word in words]) + '\\n'\n # 按行转换后写道对应的文件中\n f_out.write(index_line)\n f_in.close()\n f_out.close()\n\n # 将训练数据,测试数据,验证数据分别编号化\n src_to_index(src_train_data, train_index_data)\n src_to_index(src_test_data, test_index_data)\n src_to_index(src_valid_data, valid_index_data)\n","sub_path":"postag/pos_tagging/utils/dataset/processor.py","file_name":"processor.py","file_ext":"py","file_size_in_byte":4949,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"315553644","text":"import sys\nN, M = map(int, sys.stdin.readline().split())\ntimes = []\nfor i in range(N):\n times.extend(list(map(int, sys.stdin.readline().split('/n'))))\n\ns = 1\ne = max(times)*M\nwhile s <= e:\n num = (s+e)//2\n total = 0\n for time in list(times):\n total += num // time\n if total >= M:\n result = num\n e = num-1\n else:\n s = num+1\nprint(result)","sub_path":"김민채/0808/BOJ_3079.py","file_name":"BOJ_3079.py","file_ext":"py","file_size_in_byte":382,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"50414519","text":"'''\nProblem 1.4\nWrite a function for sync(x) = (sin x) / x\nMake sure that your function handles x = 0 correctly.\n'''\n\nimport numpy as np\nimport matplotlib.pyplot as mpl\n\ndef plot_sinc(x, y):\n mpl.figure()\n mpl.plot(x, y, 'r-')\n mpl.title(\"Sinc Function\")\n mpl.xlabel(\"x\")\n mpl.ylabel(\"sync(x)\")\n mpl.show()\n\ndef my_sinc(x):\n if x == 0.0:\n return 1\n else:\n return np.sin(x)/x\n\n# Lets calculate a few values and plot it\nsamples = 200\nx = np.linspace(-30.0, 30.0, samples)\ny = np.empty_like(x)\nfor i in range(samples):\n y[i] = my_sinc(x[i])\nplot_sinc(x,y)\n\n","sub_path":"Lecture3/CompPhysPy/ch01/sinc.py","file_name":"sinc.py","file_ext":"py","file_size_in_byte":595,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"431811376","text":"# Prompt: https://leetcode.com/problems/matrix-diagonal-sum/\n# Runtime: 88 ms, faster than 55.68% of Python online submissions for Matrix Diagonal Sum.\n# Memory Usage: 13.6 MB, less than 44.32% of Python online submissions for Matrix Diagonal Sum.\n\n\nclass Solution(object):\n def diagonalSum(self, mat):\n \"\"\"\n :type mat: List[List[int]]\n :rtype: int\n \"\"\"\n sm = 0\n length = len(mat)\n for i in range(length):\n sm += mat[i][i]\n # avoid adding middle twice\n if i != length-i-1:\n sm += mat[length-i-1][i]\n \n return sm\n","sub_path":"0. Easy/1572. Matrix Diagonal Sum/matrix_diagonal_sum.py","file_name":"matrix_diagonal_sum.py","file_ext":"py","file_size_in_byte":634,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"189834718","text":"### Give a target amount n and a list of distinct coin values, return a list of the least amount of coins needed ###\nimport unittest\n\n\ndef change_coin(target_amount, coins):\n \"\"\"input a target amount 11 and coin_list [1, 5], return [5, 5, 1]\"\"\"\n coin_list = []\n if target_amount == 0:\n return coin_list\n else:\n coins.sort(reverse=True)\n if target_amount >= coins[0]:\n target_amount -= coins[0]\n coin_list.append(coins[0])\n coin_list.extend(change_coin(target_amount, coins))\n elif target_amount < coins[0]:\n coin_list.extend(change_coin(target_amount, coins[1:]))\n return coin_list\n\n\nclass Test(unittest.TestCase):\n \"\"\"Test cases\"\"\"\n\n '''test change coin'''\n def test_change_coin(self):\n self.assertEqual(change_coin(15, [1, 2, 5]), [5, 5, 5])\n self.assertEqual(change_coin(19, [1, 3, 5]), [5, 5, 5, 3, 1])\n self.assertEqual(change_coin(5, [1, 2]), [2, 2, 1])\n\n\n'''run the script'''\nif __name__ == '__main__':\n unittest.main()\n\n","sub_path":"Recursion/CoinChange.py","file_name":"CoinChange.py","file_ext":"py","file_size_in_byte":1048,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"138960923","text":"# Cheetah-Shell/modules/dir.py\n# Copyright (C) 2018 Marco Tomas Rodriguez.\n# Licensed under https://opensource.org/licenses/MIT \nimport os\nimport modules.operativesystem as eos\n\n\nclass Dir:\n def __init__(self):\n self.directory = os.getcwd()\n\n def setDir(self, directory):\n i = 0\n for letters in directory:\n if letters == '/' or letters == '\\\\':\n i = i + 1\n if directory == '..':\n if eos.getPlatform() == 'Windows':\n split = str(self.directory).split('\\\\')\n dir = \"\"\n for routes in split[:-1]:\n dir = dir + routes + '\\\\'\n self.directory = dir\n\n else:\n # Finish using some Unix based OS\n pass\n elif i < 1:\n if eos.getPlatform() == 'Windows':\n if os.path.isdir(self.directory + '\\\\' + directory):\n self.directory = self.directory + '\\\\' + directory\n else:\n print(\"Path doesn't exist\")\n elif os.path.isdir(directory):\n self.directory = directory\n else:\n return print(\"The directory not exist\")\n\n def getDir(self):\n return self.directory","sub_path":"modules/dir.py","file_name":"dir.py","file_ext":"py","file_size_in_byte":1273,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"369659098","text":"'''Optimized retrieval and in-memory storage of a small amount of information across many contacts.'''\n\nfrom process.logging import Logger as log\nfrom process.globals import config\nfrom database import db\n\nclass ContactCache(object):\n def __init__(self, require_email=False, **kw):\n self.columns = []\n self.contacts = []\n self.require_email = require_email\n\n def isEmpty(self):\n return not self.contacts.empty()\n\n def fetch(self):\n '''Load a batch of contacts into the cache'''\n query = self.buildQuery()\n\n self.contacts = []\n result = db.get_db().execute(query)\n for row in result:\n name_components = []\n keys = ['first_name', 'middle_name', 'last_name', 'organization_name']\n\n for key in keys:\n if key in row and row[key]:\n name_components.append(row[key])\n\n #TODO: consider some flatter structure:\n #self.contacts.append([\n #\trow['id'],\n #\t\" \".join(name_components),\n #\trow['email'],\n #])\n self.contacts.append({\n 'id': row['id'],\n 'name': \" \".join(name_components),\n 'email': row['email'],\n })\n\n def buildQuery(self):\n query = db.Query()\n query.columns.extend([\n \"contact.id\",\n \"contact.first_name\",\n \"contact.middle_name\",\n \"contact.last_name\",\n \"email.email\",\n \"address.street_address\",\n \"address.city\",\n \"address.postal_code\",\n \"state.abbreviation\",\n \"country.iso_code\",\n ])\n email_clause = \"civicrm_email email ON contact.id = email.contact_id\"\n if self.require_email:\n email_clause += \" AND email.email IS NOT NULL\"\n query.tables = [\n \"civicrm_contact contact\",\n email_clause,\n \"civicrm_address address ON contact.id = address.contact_id\",\n \"civicrm_country country ON address.country_id = country.id\",\n \"civicrm_state_province state ON address.state_province_id = state.id\",\n ]\n query.group_by = [\n \"contact.id\",\n ]\n query.order_by = [\n \"contact.id\",\n ]\n return query\n\nclass PagedGroup(ContactCache):\n pagesize = config.contact_cache_size\n\n def __init__(self, **kw):\n super(PagedGroup, self).__init__(**kw)\n self.offset = 0\n\n def buildQuery(self):\n query = super(PagedGroup, self).buildQuery()\n log.info(\"Limiting batch contact retrieval to {num} records.\".format(num=self.pagesize))\n query.limit = self.pagesize\n query.offset = self.offset\n return query\n\n def next(self):\n #TODO:\n #query.offset += self.pagesize\n #self.fetch()\n raise Exception(\"unimplemented\")\n\nclass TaggedGroup(PagedGroup):\n \"\"\"Select contacts based on included and excluded tags.\"\"\"\n\n def __init__(self, tag, excludetag=None, **kw):\n super(TaggedGroup, self).__init__(**kw)\n self.tag = tag\n self.excludetag = excludetag\n\n def buildQuery(self):\n query = super(TaggedGroup, self).buildQuery()\n query.tables.extend([\n \"civicrm_entity_tag entity_tag ON entity_tag.entity_id = contact.id AND entity_tag.tag_id = %(tag_id)s AND entity_tag.entity_table = 'civicrm_contact'\",\n ])\n query.params.update({\n 'tag_id': self.tag.id\n })\n\n if self.excludetag:\n query.tables.extend([\n \"civicrm_entity_tag entity_tag_not ON entity_tag_not.entity_id = contact.id AND entity_tag_not.tag_id = %(excludetag_id)s AND entity_tag_not.entity_table = 'civicrm_contact'\",\n ])\n query.where.extend([\n \"entity_tag_not.id IS NULL\"\n ])\n query.params.update({\n 'excludetag_id': self.excludetag.id\n })\n\n return query\n","sub_path":"dedupe/contact_cache.py","file_name":"contact_cache.py","file_ext":"py","file_size_in_byte":4021,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"153665581","text":"\"\"\"\n建议88:使用multiprocessing克服GIL的缺陷\n\n进程之间的通信优先考虑Pipe和Queue,而不是Lock,Event,Condition,Semaphore等 同步原语。\n使用Queue进行进程之间的通信的时候,传输的对象必须是可以序列化的,否则put操作会导致PicklingError。\nmultiprocessing.Pipe([duplex]),其中dumplex默认为True,表示为双向管道,否则为单向。它返回\n一个Connection对象的组(conn1, conn2),分别代表管道两端。它不支持进程安全,只适用在两个进程间通信,\n性能更好。\n\"\"\"\n\nfrom multiprocessing import Process, Pipe, Queue\nimport time\n\n\ndef reader_pipe(pipe):\n # 这里用output_p端\n output_p, input_p = pipe\n input_p.close() # 关闭一端,另一端接收数据\n while True:\n try:\n msg = output_p.recv() # noqa\n except EOFError:\n break\n\n\ndef writer_pipe(count, input_p):\n # 这里用input_p端\n for i in range(0, count):\n input_p.send(i)\n\n\ndef reader_queue(queue):\n while True:\n msg = queue.get()\n if (msg == 'DONE'):\n break\n\n\ndef writer_queue(count, queue):\n for ii in range(0, count):\n queue.put(ii)\n queue.put('DONE')\n\n\n# if __name__ == '__main__':\n# print('testing for pipe:')\n# for count in [10**3, 10**4, 10**5]:\n# output_p, input_p = Pipe()\n# reader_p = Process(target=reader_pipe, args=((output_p, input_p), )) # 拷贝一份pipe\n# reader_p.start()\n\n# _start = time.time()\n# output_p.close() # 当前进程的管道,关闭一端,往另一端发送数据\n# writer_pipe(count, input_p)\n# input_p.close()\n# reader_p.join()\n# print('Sending %s numbers to Pipe() took %s seconds' % (count, (time.time() - _start)))\n\n# print('testing for queue:')\n# for count in [10**3, 10**4, 10**5]:\n# queue = Queue()\n# reader_p = Process(target=reader_queue, args=((queue), ))\n# reader_p.daemon = True\n# reader_p.start()\n\n# _start = time.time()\n# writer_queue(count, queue)\n# reader_p.join()\n# print('Sending %s numbers to Queue() took %s seconds' % (count, time.time() - _start))\n\"\"\"\n尽量避免资源共享。如果不可避免,可以通过multiprocessing.Value 和 multiprocessing.Array\n或者 multiprocessing.sharedctypes来实现内存共享,也可以通过服务器进程管理器Manager()来\n实现数据和状态的共享。\n共享内存方式更快,效率更高,但Manager()使用起来更加方便,并且支持本地和远程内存共享。\n\"\"\"\nimport time\nfrom multiprocessing import Process, Value\n\n\ndef func(val):\n for i in range(10):\n time.sleep(0.1)\n with val.get_lock(): # 必须加锁同步\n val.value += 1\n\n\nif __name__ == '__main__':\n v = Value('i', 0)\n processList = [Process(target=func, args=(v, )) for i in range(10)]\n for p in processList:\n p.start()\n for p in processList:\n p.join()\n print(v.value)\n","sub_path":"suggestion88.py","file_name":"suggestion88.py","file_ext":"py","file_size_in_byte":3043,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"572912394","text":"def main():\n n = int(input(\"Insira o valor de n: \"))\n fatorial = 1\n i = 2\n while i <= n:\n fatorial = fatorial*i\n i = i + 1\n\n print(\"O fatorial de \" + str(n) + \" eh : \" + str(fatorial))\n\n\nmain()","sub_path":"fatorial-simples/python/syngon/fatorial-simples.py","file_name":"fatorial-simples.py","file_ext":"py","file_size_in_byte":222,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"562729474","text":"# -*- coding: utf-8 -*-\n\nimport dash\nfrom dash.dependencies import Input, Output\nimport dash_core_components as dcc\nimport dash_html_components as html\nimport flask\nimport pandas as pd\nimport plotly.graph_objs as go\nimport utils\nimport files_config\nimport plotly.express as px\nimport gensim\nimport nltk\nfrom nltk.tokenize import word_tokenize\nfrom dash.exceptions import PreventUpdate\nimport utils\n\nnltk.data.path.append(files_config.nltk_data_path_local)\nnltk.data.path.append(files_config.nltk_data_path_container)\n\nserver = flask.Flask(__name__)\napp = dash.Dash(__name__, server=server)\n\n# vocab with dimensionality reduction and most similar already calculated\ndf = pd.read_pickle(files_config.viz_df_filepath)\n# need to index on the vocab but also need to pass plotly a column\n# name to label the scatterplot, so will duplicate the vocab field\n# until can find a better way\ndf['index'] = df['vocab']\ndf = df.set_index('index')\n\n# Word2vec model\nmodel_word = gensim.models.Word2Vec.load(files_config.word_model_filepath)\n\ntfidf = gensim.models.TfidfModel.load(files_config.tfidf_filepath)\ndictionary = gensim.corpora.Dictionary([list(model_word.wv.vocab.keys())])\n\ndf_subjects = pd.read_pickle(files_config.search_lookup_filepath)\n\nlayout_children = [\n html.H2('FOI search/similarity'),\n html.Div(\n \"Based on a word2vec model, corpus is request bodies and subject lines from Hackney's disclosure log\"\n ),\n dcc.Graph(id=\"graph\", style={\"width\": \"75%\", \"display\": \"inline-block\"}),\n html.H5('Most similar to:'),\n dcc.Dropdown(\n id='word-dropdown',\n placeholder='Search words in model vocabulary...',\n options=[{'value': i, 'label': i} for i in df.index],\n multi=False,\n ),\n html.Div(id='most-similar'),\n html.Br(),\n html.Hr(),\n html.H5('What do you want to know?'),\n html.Div(\n 'Returns suggestions from the disclosure log. Based on cosine similarity of vectors of submitted text vs requests. These sentence/document vectors are a TF-IDF weighted average of the vectors of the constituent words.'\n ),\n html.Br(),\n dcc.Textarea(\n id='search-textarea',\n placeholder='Your request...',\n rows=50,\n style={'width': '50%'},\n ),\n html.Br(),\n html.Button('Submit', id='search_log_button'),\n html.Br(),\n html.Div(id='results-list'),\n html.Br(),\n html.Br(),\n html.Hr(),\n]\n\napp.layout = html.Div(children=layout_children)\n\n\n@app.callback(\n Output('results-list', 'children'),\n [Input('search-textarea', 'value'), Input('search_log_button', 'n_clicks')],\n)\ndef update_search_results(query, n_clicks):\n if not n_clicks:\n return html.Div()\n ctx = dash.callback_context\n input_id = ctx.triggered[0]['prop_id'].split('.')[0]\n if not (ctx.triggered and input_id == 'search_log_button'):\n raise PreventUpdate\n else:\n df_results = utils.search_log(\n query, 5, model_word, df_subjects, dictionary, tfidf\n )\n rows = []\n for i in range(len(df_results)):\n row = []\n for col in df_results.columns:\n value = df_results.iloc[i][col]\n if col == 'url':\n cell = html.Td(html.A(href=value, target=\"_blank\", children=value))\n else:\n cell = html.Td(children=value)\n row.append(cell)\n rows.append(html.Tr(row))\n return html.Table(\n # Header\n [html.Tr([html.Th(col) for col in df_results.columns])]\n + rows\n )\n\n\n@app.callback(Output('most-similar', 'children'), [Input('word-dropdown', 'value')])\ndef update_most_similar(chosen_word):\n if chosen_word:\n similar = df.loc[chosen_word]['most_similar']\n return html.Table([html.Tr(html.Td(' '.join(map(str, i)))) for i in similar])\n\n\n@app.callback(Output(\"graph\", \"figure\"), [Input('word-dropdown', 'value')])\ndef make_figure(chosen_word):\n fig = px.scatter(data_frame=df, x='x', y='y', text='vocab')\n return fig\n\n\nif __name__ == '__main__':\n # app.run_server(debug=True)\n app.run_server(debug=False, port=8080)\n","sub_path":"foi_semantic_search/dashboard.py","file_name":"dashboard.py","file_ext":"py","file_size_in_byte":4146,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"526159735","text":"import numpy as np\nimport re\n\ninputfile = \"input11\"\nexample = '''Monkey 0:\n Starting items: 79, 98\n Operation: new = old * 19\n Test: divisible by 23\n If true: throw to monkey 2\n If false: throw to monkey 3\n\nMonkey 1:\n Starting items: 54, 65, 75, 74\n Operation: new = old + 6\n Test: divisible by 19\n If true: throw to monkey 2\n If false: throw to monkey 0\n\nMonkey 2:\n Starting items: 79, 60, 97\n Operation: new = old * old\n Test: divisible by 13\n If true: throw to monkey 1\n If false: throw to monkey 3\n\nMonkey 3:\n Starting items: 74\n Operation: new = old + 3\n Test: divisible by 17\n If true: throw to monkey 0\n If false: throw to monkey 1'''\n\nalphabet = \"abcdefghijklmnopqrstuvwxyz\"\nALPHABET = alphabet.capitalize()\n\nneighbours = [(1, 0), (0, 1), (-1, 0), (0, -1)]\n\ndef textgrid_to_numbers(lines):\n return np.array([[int(x) for x in line] for line in lines], dtype=int)\n\ndef traverse(root):\n sum = 0\n for entry in root.keys():\n if entry == '..':\n continue\n if isinstance(root[entry], dict):\n sum += traverse(root[entry])\n else:\n sum += root[entry]\n global totsum\n totsum.append(sum)\n return sum\n\ndef append_if_exists(dictionary: dict, key, val):\n if key not in dictionary.keys():\n dictionary[key] = []\n\n dictionary[key].append(val)\n\ndef test(level, divisor):\n return level % divisor == 0\n\ndef main():\n file_contents = ''\n with open(inputfile) as infile:\n file_contents = infile.read()\n \n lines = file_contents.splitlines() \n # lines = example.splitlines()\n\n monkeys = []\n\n print(f\"Read {len(lines)} lines\")\n\n common_divisor = 1\n\n current_monkey = {}\n monkey_number = -1\n i = 0\n while i < len(lines):\n line = lines[i]\n assert line[:6] == \"Monkey\"\n monkey_number = int(line.split(' ')[-1][:-1]) # remove \":\"\n assert len(monkeys) == monkey_number\n current_monkey['items'] = [int(item) for item in (lines[i+1].split(': ')[1]).split(', ')]\n op_text = lines[i+2].split('new = ')[1]\n if 'old +' in op_text:\n plus = int(op_text.split(' ')[-1])\n def plus_fun(plus_):\n return lambda level: level + plus_\n current_monkey['operation'] = plus_fun(plus)\n elif 'old * old' in op_text:\n current_monkey['operation'] = lambda level: level * level\n elif 'old * ' in op_text:\n mult = int(op_text.split(' ')[-1])\n def mult_fun(mult_):\n return lambda level: level * mult_\n current_monkey['operation'] = mult_fun(mult)\n else:\n raise Exception('parsing error in op')\n\n current_monkey['test_div'] = int(lines[i + 3].split(' ')[-1])\n common_divisor *= current_monkey['test_div']\n current_monkey['test_res_true'] = int(lines[i + 4].split(' ')[-1])\n current_monkey['test_res_false'] = int(lines[i + 5].split(' ')[-1])\n current_monkey['activity'] = 0\n\n monkeys.append(current_monkey)\n current_monkey = {}\n\n i += 7\n\n # play game\n for i in range(10000):\n print(f'round {i}')\n for j, monkey in enumerate(monkeys):\n # print(monkey)\n monkey['activity'] += len(monkey['items'])\n for item in monkey['items']:\n op_level = monkey['operation'](item)\n new_level = op_level % common_divisor\n # print(f'op_level: {op_level}, new: {new_level}')\n if test(new_level, monkey['test_div']):\n monkeys[monkey['test_res_true']]['items'].append(new_level)\n else:\n # print(f'monkey {j} throws {new_level}')\n monkeys[monkey['test_res_false']]['items'].append(new_level)\n monkey['items'] = []\n\n print(sorted([monkey['activity'] for monkey in monkeys]))\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"day11.py","file_name":"day11.py","file_ext":"py","file_size_in_byte":3948,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"407541677","text":"# -*- coding: utf-8 -*-\nimport scrapy\nimport re\nfrom scrapy.http import Request\nimport urlparse\nimport datetime\n\nfrom ArticleSpider.items import JobboleArticleItem\nfrom ArticleSpider.utils.common import get_md5\n\nclass JobboleSpider(scrapy.Spider):\n name = 'jobbole'\n allowed_domains = ['blog.jobbole.com']\n start_urls = ['http://blog.jobbole.com/all-posts/']\n\n # Scrapy will make request to start_urls with parse() as call back.\n\n # extract detailed page urls from the overview page, yield a RequestForUrl + parse_detail for each page;\n # after visit the urls in current overview page, yield a RequestForNextPage + parse;\n def parse(self, response):\n # Use CSS selector instead of Xpath selector (they have the same effect) to select detailed urls list\n # select all nodes: id=archive / class=post && floated-thumb / class=post-thumb / tag=a\n detailNodes = response.css(\"#archive .post.floated-thumb .post-thumb a\")\n for detailNode in detailNodes:\n # print(detailedUrl)\n # Asynchronous request (Scrapy help to handle) for detailed page, and then call \"parse_detail\"\n # \"Request\" is in scrapy.http; \"parse.urljoin\" is in urllib, to generate a complete url (sometimes detailed url is a relative url).\n # (Assume the coverPicture for each article can only be seen on the overview page (not the detailed page), then we need get the picture and pass into parse_detail, using \"meta\")\n imgSrc = detailNode.css(\"img::attr(src)\").extract_first()\n detailUrl = detailNode.css(\"::attr(href)\").extract_first()\n yield Request(urlparse.urljoin(response.url, detailUrl), meta={\"coverImg\": imgSrc}, callback=self.parse_detail)\n\n # Find next overview page, make request and call self (parse()) again.\n # .extract_first(defaultValue) return defaultValue if no ele in extracted list; or return the first ele in list\n nextUrl = response.css(\".navigation .next.page-numbers::attr(href)\").extract_first(\"\")\n if nextUrl:\n yield Request(nextUrl, callback=self.parse)\n pass\n\n # parse information for a detailed page\n def parse_detail(self, response):\n # DOM元素id是全局唯一的;\n # scrapy 获取的源码是未执行js动态生成时的html源码(相当于browser中查看源码);\n # 而直接从browser通过开发者工具xpath Copy拿到的xpath是基于js执行动态生成后的html;\n # 因此通过从browser复制得到类似/div[1]/div[2]...这样的xpath有时是不靠谱的;\n # 而类似\"*[@id=\"...\"]\"包含id 或者 \"*[@class=\"header\"]/...\"包含实际内容class的 这样的比较靠谱\n # re_selector = response.xpath('//*[@id=\"post-114442\"]/div[1]/h1')\n\n # response.xpath(...) 返回的是SelectorList;\n # 此处Selector元素[0]:\n # (Selector or SelectorList) .extract() 返回( data or dataList)\n\n coverImg = response.meta.get(\"coverImg\", \"\")\n title = response.css(\".entry-header h1::text\").extract_first()\n date = response.xpath('//*[@class=\"entry-meta\"]/p[1]/text()')[0].extract().replace('·', '').strip()\n\n # < span data - post - id = \"114442\"class =\" btn-bluet-bigger href-style vote-post-up register-user-only \" > < i class =\"fa fa-thumbs-o-up\" > < / i > < h10 id=\"114442votetotal\" > 1 < / h10 > 赞 < / span >\n thumbUp = int(response.xpath(\"//span[contains(@class, 'vote-post-up')]/h10/text()\").extract()[0])\n\n favSpan = response.xpath(\"//span[contains(@class, 'bookmark-btn')]/text()\").extract()[0];\n favMatch = re.match(\".*?(\\d+).*\", favSpan)\n if favMatch:\n favNum = int(favMatch.group(1))\n else:\n favNum = 0\n\n comment = response.xpath(\"//a[@href='#article-comment']/span/text()\").extract()[0]\n commentMatch = re.match('.*?(\\d+).*', comment)\n if commentMatch:\n commentNum = int(commentMatch.group(1))\n else:\n commentNum = 0\n\n tagList = response.xpath(\"//p[@class='entry-meta-hide-on-mobile']/a/text()\").extract()\n # Delete \"k 评论\" tag\n tagList = [element for element in tagList if not (element.strip().endswith(\"评论\"))]\n tags = \",\".join(tagList)\n\n contentList = response.css(\"div.entry *::text\").extract()\n content = \" \".join(contentList)\n content = content.replace(\"\\t\", \" \").replace(\"\\n\", \" \").replace(\"\\r\", \" \")\n content = \" \".join(content.split())\n\n item = JobboleArticleItem()\n item[\"title\"] = title\n try:\n create_date = datetime.datetime.strptime(date, \"%Y/%m/%d\").date()\n except Exception as e:\n create_date = datetime.datetime.now().date()\n item[\"date\"] = create_date\n item[\"url\"] = response.url\n # item[\"urlObjId\"]\n # item[url] would be processed as a list in \"pipelines\"\n item[\"coverImgUrl\"] = [coverImg]\n item[\"thumbUp\"] = thumbUp\n item[\"favNum\"] = favNum\n item[\"commentNum\"] = commentNum\n item[\"tags\"] = tags\n item[\"content\"] = content\n item[\"urlObjId\"] = get_md5(response.url)\n\n #yield send item to pipeline if \"settings\" enable pipleline\n yield item\n\n pass\n","sub_path":"ArticleSpider/spiders/jobbole.py","file_name":"jobbole.py","file_ext":"py","file_size_in_byte":5379,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"312272337","text":"from data_importers.management.commands import BaseXpressDemocracyClubCsvImporter\n\n\nclass Command(BaseXpressDemocracyClubCsvImporter):\n council_id = \"NWL\"\n addresses_name = \"2021-04-16T11:45:50.143865/North West Leicestershire Democracy_Club__06May2021 (2).tsv\"\n stations_name = \"2021-04-16T11:45:50.143865/North West Leicestershire Democracy_Club__06May2021 (2).tsv\"\n elections = [\"2021-05-06\"]\n csv_delimiter = \"\\t\"\n\n def address_record_to_dict(self, record):\n uprn = record.property_urn.strip().lstrip(\"0\")\n\n if uprn in [\n \"200003503741\", # OLD FARMHOUSE, NOTTINGHAM ROAD, STAUNTON HAROLD, ASHBY-DE-LA-ZOUCH\n \"100030573013\", # THE OLD SLAUGHTER HOUSE, PARK LANE FARM, PARK LANE, CASTLE DONINGTON, DERBY\n ]:\n return None\n\n if record.addressline6 in [\n \"LE65 2RF\",\n \"DE74 2DE\",\n \"DE11 8HB\",\n \"LE67 6HP\",\n \"LE12 9TB\",\n \"LE67 5DJ\",\n \"DE12 6DW\",\n ]:\n return None\n\n return super().address_record_to_dict(record)\n\n def station_record_to_dict(self, record):\n\n # Hall Lane Methodist Church Hall Lane Whitwick LE67 5FD\n if record.polling_place_id == \"4013\":\n record = record._replace(polling_place_postcode=\"\")\n\n return super().station_record_to_dict(record)\n","sub_path":"polling_stations/apps/data_importers/management/commands/import_north_west_leicestershire.py","file_name":"import_north_west_leicestershire.py","file_ext":"py","file_size_in_byte":1370,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"415850249","text":"import numpy as np\nimport pickle\nimport sys\nfrom WGS84toENU import *\n\n__all__ = [ \"getCameraParams\" ]\n\n\ndef getCameraParams():\n cam = [{}, {}]\n for i in [0, 1]:\n cam[i]['R_to_c_from_i'] = np.array([[-1, 0, 0],\n [0, 0, -1],\n [0, -1, 0]])\n\n if i == 0:\n cam[i]['rot_x'] = deg2rad(-0.8) # better cam 1\n cam[i]['rot_y'] = deg2rad(-0.5)\n cam[i]['rot_z'] = deg2rad(-0.005)\n cam[i]['t_x'] = -0.5\n cam[i]['t_y'] = 1.1\n cam[i]['t_z'] = 0.0\n elif i == 1:\n cam[i]['rot_x'] = deg2rad(-0.61) # better cam 2\n cam[i]['rot_y'] = deg2rad(0.2)\n cam[i]['rot_z'] = deg2rad(0.0)\n cam[i]['t_x'] = 0.5\n cam[i]['t_y'] = 1.1\n cam[i]['t_z'] = 0.0\n\n cam[i]['fx'] = 2221.8\n cam[i]['fy'] = 2233.7\n cam[i]['cu'] = 623.7\n cam[i]['cv'] = 445.7\n cam[i]['KK'] = np.array([[cam[i]['fx'], 0.0, cam[i]['cu']],\n [0.0, cam[i]['fy'], cam[i]['cv']],\n [0.0, 0.0, 1.0]])\n cam[i]['f'] = (cam[i]['fx'] + cam[i]['fy']) / 2\n\n return cam\n\nif __name__ == '__main__':\n cam= getCameraParams()\n filename = 'cam_params.pickle'\n if len(sys.argv) > 1:\n filename = sys.argv[1]\n pickle.dump(cam, open(filename, 'wb'))\n\n","sub_path":"src/caffe/py_lane_label_reader/CameraParams.py","file_name":"CameraParams.py","file_ext":"py","file_size_in_byte":1424,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"334599463","text":"from magma import *\nfrom mantle import *\n\n__all__ = ['Input', 'Output']\n\ndef Input(N, inputs):\n\n print('Building input ports')\n\n A = In(Bits(N))()\n n = len(inputs)\n if n == 1:\n O = inputs[0]\n elif n == 2:\n mux = Mux(2, N)\n mux(inputs[0], inputs[1], A[0])\n O = mux.O\n elif n == 4:\n mux = Mux(4, N)\n mux(inputs[0], inputs[1], inputs[2], inputs[3], A[0:2])\n O = mux.O\n elif n == 8:\n mux = Mux(8, N)\n mux(inputs[0], inputs[1], inputs[2], inputs[3],\n inputs[4], inputs[5], inputs[6], inputs[7], A[0:3])\n O = mux.O\n\n return AnonymousCircuit(\"A\", A, \"O\", O)\n\n\n#\n# Generate a register with the same number of bits as the output\n#\ndef OutputOne(input, output, we):\n reg = Register(8, has_ce=True)\n reg(input, ce=we)\n wire(reg, output)\n\ndef Output(N, outputs):\n\n print('Building output ports')\n I = In(Bits(N))()\n A = In(Bits(N))()\n we = In(Bit)()\n\n if len(outputs) == 1:\n OutputOne(I, outputs[0], we)\n else:\n for output in outputs:\n en = Decode(i, N)(A)\n en = And2()(en, we)\n OutputOne(I, output, en)\n\n return AnonymousCircuit(\"A\", A, \"I\", I, \"WE\", we )\n\n","sub_path":"pico40/io.py","file_name":"io.py","file_ext":"py","file_size_in_byte":1230,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"378139414","text":"def compare(s1, s2):\n if not s1 and not s2:\n return True\n if not s1 or not s2:\n return False\n\n r1 = []\n i = 0\n while i < len(s1):\n if s1[i] == '-':\n r1.pop()\n i += 2\n else:\n r1.append(s1[i])\n i += 1\n\n r1 = ''.join(r1)\n print(r1)\n\n r2 = []\n i = 0\n while i < len(s2):\n if s2[i] == '-':\n r2.pop()\n i += 2\n else:\n r2.append(s2[i])\n i += 1\n\n r2 = ''.join(r2)\n print(r2)\n\n if r1 == r2:\n return True\n else:\n return False\n\n\ndef compare_2(s1, s2):\n if not s1 and not s2:\n return True\n if not s1 or not s2:\n return False\n\n i = j = 0\n l1 = []\n while i < len(s1):\n if s1[i] == '-':\n l1.append('-1')\n i += 2\n else:\n l1.append(s1[i])\n i += 1\n l2 = []\n while j < len(s2):\n if s2[j] == '-':\n l2.append('-1')\n j += 2\n else:\n l2.append(s2[j])\n j += 1\n\n print(l1, l2)\n\n minus1_count1 = 0\n i = len(l1) - 1\n minus1_count2 = 0\n j = len(l2) - 1\n while i >= 0 and j >= 0:\n while l1[i] == '-1':\n minus1_count1 += 1\n i -= 1\n while l2[j] == '-1':\n minus1_count2 += 1\n j -= 1\n\n i -= minus1_count1\n minus1_count1 = 0\n j -= minus1_count2\n minus1_count2 = 0\n\n print(i, j)\n print(l1[i], l2[j])\n if l1[i] == l2[j]:\n i -= 1\n j -= 1\n continue\n else:\n return False\n\n if minus1_count1 != minus1_count2:\n return False\n\n return True\n\n\nstring1 = '1-12-13-14-156789'\nstring2 = '567891234-1-1-1-1'\n\nprint(compare_2(string1, string2))\nprint(compare(string1, string2))\n","sub_path":"Google/输入数字比较相同.py","file_name":"输入数字比较相同.py","file_ext":"py","file_size_in_byte":1851,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"518897031","text":"#!/usr/bin/env python\nimport random \nimport math\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport os\nimport argparse\n\ndcoef = {'v1': \"0.174323\", 'v2': \"0.1918\", 'v3': \"0.1928\", 'v4': \"0.2155\", 'v5': \"0.2893\", 'v6': \"0.3308\", 'v7': \"0.347\", 'v8': \"0.4115\", 'v9':\"0.5688\", 'v10': \"1.0\"}\nD_list = [\"v1\", \"v2\", \"v3\", \"v4\", \"v5\", \"v6\", \"v7\", \"v8\", \"v9\", \"v10\"]\n\nd_uc = 26.6226e-8\n\ndef buildMOF(x, y, z):\n mof, clist, coord = [], [], []\n if len(z) == 0:\n for i in x:\n for j in y:\n clist.append([i,j])\n elif len(z) != 0:\n for i in x:\n for j in y:\n for k in z:\n clist.append([i, j, k])\n for i in range(len(clist)):\n slist = ','.join([str(n) for n in clist[i]])\n coord.append(slist)\n# random.seed(10) \n for i in range(len(coord)):\n mof.append(random.choice(D_list))\n \n dcoord = dict(zip(coord,mof))\n coord_split = [coord[i].split(',') for i in range(len(coord))]\n coef_list =[]\n for i in mof:\n coef_list.append(dcoef[i])\n \n return dcoord, mof, coord, coef_list\n\n\ndef find2Dneigh(smof):\n neigh_list = []\n smof_x = int(smof[0])\n smof_y = int(smof[1])\n mol1 = [smof_x, smof_y-1]\n mol2 = [smof_x+1, smof_y-1]\n mol3 = [smof_x+1, smof_y]\n mol4 = [smof_x+1, smof_y+1]\n mol5 = [smof_x, smof_y+1]\n list1 = [mol1, mol2, mol3, mol4, mol5]\n \n for i in range(len(list1)):\n if (list1[i][1] != B) and (list1[i][0] != A) and (list1[i][1] != -1):\n neigh_list.append(list1[i]) \n \n return neigh_list\n\n\ndef forwardN(smof):\n n = 3\n f = int((1/3)*100 + n)\n v = int((1/3)*100 + n/2)\n b = 100 - f - v\n sub_list = []\n smof_x = int(smof[0])\n smof_y = int(smof[1])\n pk = random.randint(1,100)\n if smof_x == 0 and smof_y != 0:\n if pk <= 30:\n x = 0\n y = [1] \n else:\n x = 1\n y = [0, 1]\n elif smof_x != 0 and smof_y == 0:\n if pk <= 30:\n x = 0\n y = [1]\n else: \n x = 1\n y = [0, 1]\n elif smof_x == 0 and smof_y == 0:\n if pk <= 30:\n x = 0\n y = [1]\n else: \n x = 1\n y = [0, 1]\n else: \n if pk <= b:\n x = -1\n y = [-1, 0, 1]\n elif pk > b and pk <= b + v:\n x = 0\n y = [-1, 1]\n else:\n x = 1\n y = [-1, 0, 1]\n \n for i in y: \n nei_mof1 = (smof_x + x, smof_y + i)\n if nei_mof1[0] >= 0 and nei_mof1[1] >=0 and nei_mof1[0] < A and nei_mof1[1] < A:\n sub_list.append(nei_mof1)\n \n return sub_list\n\n\ndef compNeigh(sub_list, dcoord):\n neigh_D, str_list, mof_f = [], [], []\n D_list = []\n local_d = {}\n\n for i in range(len(sub_list)):\n str_list_i = ','.join([str(n) for n in sub_list[i]])\n str_list.append(str_list_i) # '1,1'\n a = dcoord[str_list_i] # 6.0\n\n neigh_D.append(str(dcoord[str_list_i])) #'4'\n\n D_list.append(str(a))\n\n # find the next mof to walk to:\n d_t = dict(zip(str_list, D_list))\n dict2 = {}\n names = set(d_t.values())\n d = {}\n for n in names:\n d[n] = [k for k in d_t.keys() if d_t[k] == n ]\n pool = []\n \n if 'v10' in d:\n a = random.randint(1,100)\n if a <= 90:\n temp_f = random.choice(d.get('v10'))\n mof_f = list(map(int, temp_f.split(',')))\n else:\n new_d = {key:val for key, val in d.items() if key != 'v10'}\n for i in new_d:\n f = int(len(new_d[i]))*int(i[1])\n for j in range(f):\n pool.append(str(new_d[i])[1:-1].replace(\"'\", \"\"))\n temp = random.choice(random.choice(pool).split(', '))\n mof_f = list(map(int, temp.split(',')))\n\n else:\n for i in d:\n f = int(len(d[i]))*int(i[1])\n for j in range(f):\n pool.append(str(d[i])[1:-1].replace(\"'\", \"\"))\n \n tp, tp2 = [], []\n for i in pool:\n tp.append(list(i.split(', ')))\n for i in tp:\n for j in i:\n tp2.append(j)\n temp = random.choice(tp2)\n mof_f = list(map(int, temp.split(',')))\n \n \n return neigh_D, str_list, mof_f\n\n\ndef caldist(mof1, mof2):\n dist = math.sqrt(((mof2[0]-mof1[0])**2) + ((mof2[1]-mof1[1])**2)) \n #for 3D walk\n# dist = math.sqrt(((mof2[0]-mof1[0])**2) + ((mof2[1]-mof1[1])**2)) + ((mof2[2]-mof1[2])**2))\n return dist\n\n\ndef randwalk(smof, N, dcoord):\n path = [smof]\n dist = 0\n temp = 0\n time1 = 0\n time1_list, time2_list = [], []\n dt_list, dx_list = [], []\n time2 = 0\n time = 0\n sub_list = forwardN(smof)\n for i in range(N):\n neigh_D, str_list, mof_f = compNeigh(sub_list, dcoord)\n cod_smof = ','.join(map(str,smof))\n cod_mof_f = ','.join(map(str,mof_f))\n D_smof = dcoef[dcoord[cod_smof]]\n D_mof_f = dcoef[dcoord[cod_mof_f]]\n dx = caldist(mof_f, smof)\n if dcoord[cod_smof] != 'v10':\n d_time = (dx**2)/(4*float(D_smof))\n time1_list.append((dx**2)/(4*float(D_smof)))\n time1 += (dx**2)/(4*float(D_smof))\n elif dcoord[cod_mof_f] == 'v10':\n d_time = (dx)/(2*float(D_mof_f))\n time2_list.append((dx)/(2*float(D_mof_f)))\n time2 += (dx)/(2*float(D_mof_f))\n elif dcoord[cod_smof] == 'v10':\n d_time = (dx)/(2*float(D_smof))\n time2_list.append((dx)/(2*float(D_smof)))\n time2 += (dx)/(2*float(D_smof))\n\n temp += d_time\n dt_list.append(round(temp, 3))\n\n \n dist += caldist(mof_f, smof)\n dx_list.append(round(dist, 3))\n \n sub_list = forwardN(mof_f)\n \n if smof in sub_list:\n sub_list.remove(smof)\n else:\n pass\n smof = mof_f\n path.append(smof)\n time = time1 + time2\n if len(sub_list) == 0 or (mof_f[0] == A-1):\n break\n \n return path, dist, time, dt_list, dx_list\n\n\ndef cmap(path, coef_list):\n w = A \n h = B \n d = 100\n relist = np.array([float(i) for i in coef_list]).reshape(A,B)\n \n color_map = plt.imshow(relist)\n color_map.set_cmap(\"Blues_r\")\n \n plt.savefig('temp1.png')\n \n for i in path:\n x = i[0]\n y = i[1]\n relist[y][x] = 0 \n \n color_map = plt.imshow(relist)\n color_map.set_cmap(\"Blues_r\")\n\n plt.savefig('temp2.png')\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser(description='Generate 2D walk plot and data')\n parser.add_argument('-a', nargs=1, help='Input lattice dimension', required=True)\n args = parser.parse_args()\n\n if vars(args)['a'][0] != 'NULL':\n A = int(vars(args)['a'][0])\n B = A\n smof = [0, int(A/2)]\n N = A*B*10\n tot_dist = A * d_uc\n C = 0\n x = [ i for i in range(A)]\n y = [ i for i in range(B)]\n z = [ i for i in range(C)]\n dcoord, mof, coord, coef_list = buildMOF(x, y, z)\n sub_list = forwardN(smof)\n neigh_D, str_list, mof_f = compNeigh(sub_list, dcoord)\n path, dist, time, dt_list, dx_list = randwalk(smof, N, dcoord)\n cmap(path, coef_list) \n\n","sub_path":"RandomWalk/2D_random.py","file_name":"2D_random.py","file_ext":"py","file_size_in_byte":7363,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"123598838","text":"import argparse\nimport shutil\nimport sys\nimport os\n\nfrom interface import *\nfrom error import *\nfrom utils import *\n\nparser = argparse.ArgumentParser(description='Main Interface')\nparser.add_argument('--skip', help='number of images skipped', default=1, required=False)\nparser.add_argument('--noErr', help='skip error metric computation', default=False, required=False, action='store_true')\nparser.add_argument('--noDiff', help='skip difference image computation', default=False, required=False, action='store_true')\nargs = parser.parse_args()\n\nDATA_ROOT_PATH \t\t= \"/home/Workspace/data/\"\nMETA_DATA_FILENAME \t= \"cyclic_meta.csv\"\nERROR_FILENAME \t\t= \"cyclic_error.csv\"\nINSTALL_PATH \t\t= \"/home/Workspace/install\"\nSKIP_NUM\t\t\t= 1\n\ndef skipFrames(path, skipNum):\n\t\n\tfileList = os.listdir(path)\n\timgList = filterImages(fileList)\n\timgList = sorted(imgList)\n\n\tkeepNum = skipNum + 1\n\n\tfor counter, fileName in enumerate(imgList):\n\t\t\n\t\tif((counter % keepNum) != 0):\n\t\t\tos.remove(os.path.join(path, fileName))\n\ndef parseArgs():\n\t\n\tglobal SKIP_NUM\n\tSKIP_NUM = int(args.skip)\n\t\n\tif(SKIP_NUM != 0 and SKIP_NUM != 1 and SKIP_NUM != 3 and SKIP_NUM != 7):\n\t\tprint(\"Invalid number of images to be skipped (\" + str(SKIP_NUM) + \")\")\n\t\tsys.exit(1)\n\t\t\n\nif __name__ == \"__main__\":\n\t\n\tparseArgs()\n\t\n\t# Assume all images are in root directory\n\t\n\torigPath \t= os.path.join(DATA_ROOT_PATH, \"orig\")\n\ttempPath \t= os.path.join(DATA_ROOT_PATH, \"temp\")\n\toutPath \t= os.path.join(DATA_ROOT_PATH, \"out\")\n\tdiffPath \t= os.path.join(DATA_ROOT_PATH, \"diff\")\n\t\n\t# Reset directory if anything was done\n\t\n\tif(os.path.exists(os.path.join(DATA_ROOT_PATH, META_DATA_FILENAME))):\n\t\tos.remove(os.path.join(DATA_ROOT_PATH, META_DATA_FILENAME))\n\t\t\n\tif(os.path.exists(os.path.join(DATA_ROOT_PATH, ERROR_FILENAME))):\n\t\tos.remove(os.path.join(DATA_ROOT_PATH, ERROR_FILENAME))\n\t\n\tif(os.path.exists(origPath)):\n\t\tmoveAllFiles(origPath, DATA_ROOT_PATH)\n\t\tshutil.rmtree(origPath)\n\t\n\tif(os.path.exists(tempPath)):\n\t\tshutil.rmtree(tempPath)\n\t\t\n\tif(os.path.exists(outPath)):\n\t\tshutil.rmtree(outPath)\n\t\t\n\tif(os.path.exists(diffPath)):\n\t\tshutil.rmtree(diffPath)\n\t\n\t# Create directory structure\n\t\n\tos.makedirs(origPath)\n\tos.makedirs(tempPath)\n\tos.makedirs(outPath)\n\t\n\tmoveAllFiles(DATA_ROOT_PATH, origPath)\n\tcopyAllFiles(origPath, tempPath)\n\tskipFrames(tempPath, SKIP_NUM)\n\t\n\t# Do the interpolation\n\t\n\tprint(\"Interpolating images ...\")\n\t\n\tmeta_data = []\n\t\n\tif(SKIP_NUM == 0 or SKIP_NUM == 1):\n\t\tmeta_data = interpolate(tempPath, outPath, meta_data, INSTALL_PATH)\n\telif(SKIP_NUM == 3):\n\t\tmeta_data = interpolate(tempPath, outPath, meta_data, INSTALL_PATH)\n\t\tdeleteAllFiles(tempPath)\n\t\tmoveAllFiles(outPath, tempPath)\n\t\tmeta_data = interpolate(tempPath, outPath, meta_data, INSTALL_PATH)\n\telif(SKIP_NUM == 7):\n\t\tmeta_data = interpolate(tempPath, outPath, meta_data, INSTALL_PATH)\n\t\tdeleteAllFiles(tempPath)\n\t\tmoveAllFiles(outPath, tempPath)\n\t\tmeta_data = interpolate(tempPath, outPath, meta_data, INSTALL_PATH)\n\t\tdeleteAllFiles(tempPath)\n\t\tmoveAllFiles(outPath, tempPath)\n\t\tmeta_data = interpolate(tempPath, outPath, meta_data, INSTALL_PATH)\n\t\n\tshutil.rmtree(tempPath)\n\t\n\t# Write out metadata\n\t\n\tmeta_data_file = open(os.path.join(DATA_ROOT_PATH, META_DATA_FILENAME), \"w\")\n\tmeta_data_file.write(\"Avg time per img, Total time, width, height, interp img num\\n\")\n\tmeta_data_file.write(str(meta_data[0]) + \",\" + str(meta_data[1]) + \",\" + str(meta_data[2]) + \",\" + str(meta_data[3]) + \",\" + str(meta_data[4]) + \"\\n\")\n\tmeta_data_file.close()\n\t\n\t# Compute Error\n\t\n\tif(SKIP_NUM != 0):\n\t\t\n\t\tif(not args.noDiff):\n\t\t\tprint(\"Calculating Diff Imgs ...\")\n\t\t\tcomputeDiffImgs(origPath, outPath, diffPath)\n\t\t\t\n\t\tif(not args.noErr):\n\t\t\tprint(\"Calculating Errors ...\")\n\t\t\t\n\t\t\terror_data = []\n\t\t\terror_data = computeErrorMetrics(origPath, outPath, error_data)\n\t\n\t\t\t# Write err to file\n\t\n\t\t\terror_data_file = open(os.path.join(DATA_ROOT_PATH, ERROR_FILENAME), \"w\")\n\t\t\terror_data_file.write(\"Index, MSE, SSIM, MS_SSIM, PSNR, VIFP, UQI\\n\")\n\t\t\n\t\t\tcounter = 0\n\t\t\twhile(counter < len(error_data)):\n\t\t\t\terror_data_file.write(str(error_data[counter+0]) + \",\" + str(error_data[counter+1]) + \",\" + str(error_data[counter+2]) + \",\" + str(error_data[counter+3]) + \",\" + str(error_data[counter+4]) + \",\" + str(error_data[counter+5]) + \",\" + str(error_data[counter+6]) + \"\\n\")\n\t\t\t\n\t\t\t\tcounter = counter + 7;\n\t\t\n\t\t\terror_data_file.close()\n\t\t\n\tprint(\"Done.\")\n\t\n\n","sub_path":"docker/cyclic/Workspace/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4359,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"386382406","text":"newString = ''\nuserString = input(\"Please enter a string: \")\ncount = 1\nuserString = userString.lower()\n\nfor char in userString:\n if count % 2 == 0:\n newString += char.upper()\n count += 1\n else:\n newString += char\n count += 1\nprint(newString)\n\ndef splitStr(string):\n return [ch for ch in string]\n\nnewStringSplit = splitStr(newString)\nnewStringLen = len(newStringSplit)\nnewStringFinalChar = newStringLen - 1\n\nif newStringSplit[0] == \"(\" and newStringSplit[newStringFinalChar] == \")\":\n print(\"Balanced!\")\nelse:\n print(\"Not Balanced!\")\n","sub_path":"00_parentheses.py","file_name":"00_parentheses.py","file_ext":"py","file_size_in_byte":575,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"180189768","text":"from qubo_helper import Qubo\nfrom tsp_problem import TSPProblem \nfrom vrp_problem import VRPProblem\nfrom itertools import product\nimport DWaveSolvers\nimport networkx as nx\nimport numpy as np\nfrom vrp_solvers import *\n\nif __name__ == '__main__':\n\n # Some graph\n n = 20\n \"\"\"paths = np.random.random_integers(1, 10, (n, n))\n\n for u in range(n):\n paths[u][u] = 0\n\n for u in range(n):\n for (i, j) in product(range(n), range(n)):\n paths[i][j] = min(paths[i][j], paths[i][u] + paths[u][j])\"\"\"\n\n paths = np.zeros((n, n))\n for (i, j) in product(range(n), range(n)):\n paths[i][j] = abs(i - j)\n\n # Problem parameters\n sources = [5, 15]\n #sources = [0, 3, 15, 50, 77, 38, 89]\n costs = paths\n time_costs = costs\n #capacities = [n, n, n, n, n, n, n, n, n, n]\n capacities = [n, n]\n #dests = [1, 2, 16, 19, 8, 25, 55, 33, 31, 88, 97, 24, 10, 61, 48, 11, 92, 54, 38, 65]\n dests = [2, 4, 10, 19]\n weights = [1 for _ in range(0, n)]\n\n time_windows = dict()\n time_windows[2] = 10\n time_windows[4] = 15\n time_windows[19] = 10\n time_windows[10] = 20\n\n only_one_const = 100.\n order_const = 1.\n capacity_const = 0.\n time_const = 0.\n\n problem = VRPProblem(sources, costs, time_costs, capacities, dests, weights, time_windows)\n\n solver = FullQuboSolver(problem)\n #solver = AveragePartitionSolver(problem)\n\n result = solver.solve(only_one_const, order_const, capacity_const, time_const,\n solver_type = 'qbsolv', num_reads = 100)\n print(result.solution)\n #result.description()\n","sub_path":"vrp/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1588,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"109113142","text":"import os\r\nimport re\r\n\r\ndef fileslist():\r\n kartoteka = os.listdir('.')\r\n return kartoteka\r\n\r\ndef readfiles(kartoteka):\r\n texts = []\r\n for file in kartoteka:\r\n with open(file, encoding=\"utf-8\") as f:\r\n text = f.read()\r\n texts.append(text)\r\n return texts\r\n\r\ndef maketable(text_list):\r\n for text in text_list:\r\n data1 = re.search('meta content=\"(.*)\" name=\"docid\"', text)\r\n if data1:\r\n docid = data1.group(1)\r\n data2 = re.search('(.*)', text)\r\n if data2:\r\n title = data2.group(1)\r\n data3 = re.search('meta content=\"(.*)\" name=\"author\"', text)\r\n if data3:\r\n author = data3.group(1)\r\n data4 = re.search('meta content=\"(.*)\" name=\"created\"', text)\r\n if data4:\r\n created = data4.group(1)\r\n data5 = re.search('meta content=\"(.*)\" name=\"topic\"', text)\r\n if data5:\r\n topic = data5.group(1)\r\n data6 = re.search('meta content=\"(.*)\" name=\"tagging\"', text)\r\n if data6:\r\n tagging = data6.group(1)\r\n print(docid + \",\" + title + \",\" + author + \",\" + created + \",\" + topic + \",\" + tagging)\r\n \r\ndef searchabb(text_list):\r\n freqdict = {}\r\n fullist = []\r\n for text in text_list:\r\n answer = re.findall(\"[/s]([А-ЯЁA-Z]).+^[а-яёa-z][/s]\", text)\r\n for item in answer:\r\n fullist.append(item)\r\n for element in fullist:\r\n if element in freqdict:\r\n freqdict[element] = freqdict[element] + 1\r\n else:\r\n freqdict[element] = 1\r\n for key, value in freqdict.items():\r\n print(key)\r\n print(\"\\t\")\r\n print(value)\r\n \r\n\r\n \r\ndef main():\r\n maketable(readfiles(fileslist()))\r\n searchabb(readfiles(fileslist()))\r\n\r\nmain()\r\n\r\n","sub_path":"prog-examen.py","file_name":"prog-examen.py","file_ext":"py","file_size_in_byte":1831,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"229786083","text":"number = \"9,223,372,036,854,775,807\"\ncleanedNumber = ''\n\n# for loop assigns each variable char to the next number in the string\nfor char in number:\n if char in '0123456789':\n cleanedNumber = cleanedNumber + char\n\nnewNumber = int(cleanedNumber)\nprint(\"The number is {}\".format(newNumber))\n\n# for loop goes through the sequence for each string in the list\nfor state in [\"not pinin'\", \"no more\", \"a stiff\", \"breft of life\"]:\n print(\"This parrot is \" + state)\n # print(\"This parrot is {}\".format(state))\n\n# the range is between 0 and 100 and it steps by every 5 numbers\nfor i in range(0, 100, 5):\n print(\"i is {}\".format(i))\n\n# shows the times table that children learn in elementary school\n# uses nested for loops to show this\nfor i in range(1, 13):\n for j in range(1, 13):\n print(\"{1} times {0} is {2}\".format(i, j, i * j), end='\\t')\n print(' ')\n","sub_path":"loopPractice/forloops2.py","file_name":"forloops2.py","file_ext":"py","file_size_in_byte":876,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"375783613","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Apr 30 15:18:29 2018\nALDS1-4c\n@author: maezawa\n\"\"\"\ndef can_load(w, k, p):\n n = len(w)\n m = 0\n tk = 0\n i = 0\n while tk < k:\n if m + w[i] <= p:\n m += w[i]\n i += 1\n if i >= n:\n return n+1\n else:\n m = 0\n tk += 1\n return i\n\n\nn, k = list(map(int, input().split()))\nw = []\ntr = [0 for _ in range(k)]\n\nfor i in range(n):\n w.append(int(input()))\n\nmaxw = max(w)\n# =============================================================================\n# for p in range(maxw, maxw*n):\n# if can_load(w, k, p) == n:\n# print(p)\n# break\n# =============================================================================\nright = maxw*n\nleft = maxw\nwhile left 6 (= 2 * 3)\n\t\"\"\"\n\tto_return = 1\n\tfor nb in list_numbers:\n\t\tto_return *= nb\n\treturn to_return\n\t\n\n\n\n\n","sub_path":"exh/model/vars.py","file_name":"vars.py","file_ext":"py","file_size_in_byte":3318,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"376451393","text":"#! /usr/bin/env python\n\nimport tensorflow as tf\nimport numpy as np\nimport psycopg2\nfrom flask import Flask, jsonify, request, abort\nfrom preprocess import MyVocabularyProcessor\n\n# Parameters\n# ==================================================\n\n# Eval Parameters\ntf.flags.DEFINE_integer(\"batch_size\", 64, \"Batch Size (default: 64)\")\ntf.flags.DEFINE_string(\"vocab_filepath\", \"runs/1541748108/checkpoints/vocab\",\n \"Load training time vocabulary (Default: None)\")\ntf.flags.DEFINE_string(\"model\", \"runs/1541748108/checkpoints/model-33000\",\n \"Load trained model checkpoint (Default: None)\")\n\n# Misc Parameters\ntf.flags.DEFINE_boolean(\"allow_soft_placement\", True, \"Allow device soft device placement\")\ntf.flags.DEFINE_boolean(\"log_device_placement\", False, \"Log placement of ops on devices\")\n\nFLAGS = tf.flags.FLAGS\n\n# Build vocabulary\nvocab_processor = MyVocabularyProcessor(30, min_frequency=0)\nvocab_processor = vocab_processor.restore(FLAGS.vocab_filepath)\n\n\ndef char2vec(arr):\n return np.asarray(list(vocab_processor.transform(arr)))\n\n\ndef get_test_data_set(text):\n return char2vec(np.full([len(abbr_vec_arr)], text))\n\n\ndef batch_iter(data, batch_size):\n \"\"\"\n Generates a batch iterator for a dataset.\n \"\"\"\n data = np.asarray(data)\n # print(data)\n # print(data.shape)\n data_size = len(data)\n num_batches_per_epoch = int(len(data) / batch_size) + 1\n shuffled_data = data\n for batch_num in range(num_batches_per_epoch):\n start_index = batch_num * batch_size\n end_index = min((batch_num + 1) * batch_size, data_size)\n yield shuffled_data[start_index:end_index]\n\n\ndef prepare_abbr_arr():\n cur = conn.cursor()\n cur.execute(\"SELECT trim(machine_name), trim(formal_name) FROM game_alias WHERE status = 'confirmed'\")\n conn.commit()\n abbr = []\n formal = []\n for t in cur:\n abbr.append(t[0])\n formal.append(t[1])\n return np.asarray(abbr), np.asarray(formal), char2vec(abbr)\n\n\ndef predict(x1_arr):\n batches = batch_iter(list(zip(abbr_vec_arr, x1_arr, np.zeros([len(abbr_vec_arr)]))), 2 * FLAGS.batch_size)\n all_predictions = []\n all_d = []\n for db in batches:\n x1_dev_b, x2_dev_b, y_dev_b = zip(*db)\n batch_predictions, batch_acc, batch_sim = sess.run([predictions, accuracy, sim],\n {input_x1: x1_dev_b, input_x2: x2_dev_b,\n input_y: y_dev_b, dropout_keep_prob: 1.0})\n all_predictions = np.concatenate([all_predictions, batch_predictions])\n all_d = np.concatenate([all_d, batch_sim])\n\n keys = np.nonzero(all_d)\n print(keys)\n print(list(zip(abbr_arr[keys], formal_arr[keys], all_predictions[keys])))\n # print(all_predictions[keys])\n mink = np.argmin(all_predictions[keys])\n print(abbr_arr[keys][mink])\n print(all_predictions[keys][mink])\n # return 'ok'\n return formal_arr[keys][mink]\n\n\n# Initial global variables\n\napp = Flask(__name__)\nconn = psycopg2.connect(\"dbname=feedback_nlp user=postgres password=postgres host=data-platform-sh-01.nvidia.com\")\nabbr_arr, formal_arr, abbr_vec_arr = prepare_abbr_arr()\n\ngraph = tf.Graph()\nwith graph.as_default():\n session_conf = tf.ConfigProto(\n allow_soft_placement=FLAGS.allow_soft_placement,\n log_device_placement=FLAGS.log_device_placement)\n sess = tf.Session(config=session_conf)\n with sess.as_default():\n # Load the saved meta graph and restore variables\n saver = tf.train.import_meta_graph(\"{}.meta\".format(FLAGS.model))\n sess.run(tf.initialize_all_variables())\n saver.restore(sess, FLAGS.model)\n\n # Get the placeholders from the graph by name\n input_x1 = graph.get_operation_by_name(\"input_x1\").outputs[0]\n input_x2 = graph.get_operation_by_name(\"input_x2\").outputs[0]\n input_y = graph.get_operation_by_name(\"input_y\").outputs[0]\n\n dropout_keep_prob = graph.get_operation_by_name(\"dropout_keep_prob\").outputs[0]\n # Tensors we want to evaluate\n predictions = graph.get_operation_by_name(\"output/distance\").outputs[0]\n\n accuracy = graph.get_operation_by_name(\"accuracy/accuracy\").outputs[0]\n\n sim = graph.get_operation_by_name(\"accuracy/temp_sim\").outputs[0]\n\n\n@app.route('/sim/debug', methods=['POST'])\ndef sim_debug():\n data = request.get_json(force=True)\n if 'text' in data:\n _text = data['text']\n return jsonify(predict(get_test_data_set(_text)))\n else:\n return abort(406)\n\n\ndef main():\n app.run(host=\"0.0.0.0\", debug=True)\n\n\nif __name__ == '__main__':\n main()\n conn.close()\n","sub_path":"service.py","file_name":"service.py","file_ext":"py","file_size_in_byte":4685,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"478171663","text":"import OpenGL.GL as gl, colours\nfrom math import cos, sin, pi\n\ndef points(*points, colour = colours.white):\n gl.glBegin(gl.GL_POINTS)\n for point in points:\n gl.glVertex2f(*point)\n gl.glEnd()\n\ndef line(point1, point2, colour = colours.white):\n \"Draw a line from point1 of screen coordinates to point2\"\n gl.glBegin(gl.GL_LINE_STRIP)\n for point in (point1, point2):\n gl.glVertex2f(*point)\n gl.glEnd()\n \ndef lines(point_list, loop = False, fill = False, colour = colours.white):\n \"Draw a line from point to end\"\n if fill:\n gl.glBegin(gl.GL_POLYGON)\n elif loop:\n gl.glBegin(gl.GL_LINE_LOOP)\n else:\n gl.glBegin(gl.GL_LINE_STRIP)\n\n for point in point_list:\n gl.glVertex2f(*point)\n gl.glEnd()\n\ndef circle(center, radius, fill = False, num_segments = False, colour = colours.white, ):\n if not num_segments:\n num_segments = radius * 10\n \n # walk around a circle \n point_list = [ ( int(radius*cos(theta/num_segments)+center[0]), int(radius*sin(theta/num_segments)+center[1]) )\n for theta in range(int(2*pi * (num_segments))) ]\n\n lines(point_list, fill = fill, loop = True, colour = colour)","sub_path":"draw.py","file_name":"draw.py","file_ext":"py","file_size_in_byte":1211,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"200217000","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Apr 10 09:28:41 2018\n\n@author: Dean\n\"\"\"\n\nclass Node:\n def __init__(self,value=None):\n self.value = value\n self.left = None\n self.right = None\n \ndef preOrderRecur(tree):\n if not tree:\n return\n print(tree.value,end=\" \")\n preOrderRecur(tree.left)\n preOrderRecur(tree.right)\n return\n\ndef inOrderRecur(tree):\n if not tree:\n return\n inOrderRecur(tree.left)\n print(tree.value,end=\" \")\n inOrderRecur(tree.right)\n return\n\ndef posOrderRecur(tree):\n if not tree:\n return\n posOrderRecur(tree.left)\n posOrderRecur(tree.right)\n print(tree.value,end=\" \")\n return\n\n\ndef createByPreAndIn(preArr,InArr):\n if not preArr or not InArr:\n return None\n rootValue = preArr[0]\n mid = 0\n for i in InArr:\n if i == rootValue:\n break\n mid += 1\n Llen = len(InArr[:mid])\n Rlen = len(InArr[mid+1:])\n root = Node(rootValue)\n root.left = createByPreAndIn(preArr[1:Llen+1],InArr[:mid])\n root.right = createByPreAndIn(preArr[Llen+1:],InArr[mid+1:])\n return root\n\ndef createByInAndPos(InArr,posArr):\n if not InArr or not posArr:\n return None\n mid = 0\n for i in InArr:\n if i == posArr[-1]:\n break\n mid += 1\n root = Node(posArr[-1])\n Llen = mid\n root.left = createByInAndPos(InArr[:mid],posArr[:Llen])\n root.right = createByInAndPos(InArr[mid+1:],posArr[Llen:-1])\n return root\n\n#注意只有不存在度为1的节点的二叉树才能被先序和后序重构\ndef createByPreAndPos(preArr,posArr):\n if not preArr and not posArr:\n return None\n root = Node(preArr[0])\n k = 0\n leftRoot = None #左子树的根节点\n if len(preArr) > 1: #这个地方要注意,需要特殊处理,负责就报错了。\n leftRoot = preArr[1]\n for i in posArr:\n if i == leftRoot:\n break\n k += 1\n Llen = 0\n if leftRoot is not None: #同理,这个地方要特殊处理。\n Llen = k + 1 #求出左子树包含的节点数\n root.left = createByPreAndPos(preArr[1:Llen+1],posArr[:Llen])\n root.right = createByPreAndPos(preArr[Llen+1:],posArr[Llen:-1])\n return root\n \n \n\nif __name__ == \"__main__\":\n preOrder = [1,2,4,5,3,6,7]\n InOrder = [4,2,5,1,6,3,7]\n posOrder = [4,5,2,6,7,3,1]\n tree = createByPreAndIn(preOrder,InOrder)\n tree2 = createByInAndPos(InOrder,posOrder)\n tree3 = createByPreAndPos(preOrder,posOrder)\n preOrderRecur(tree)\n print()\n inOrderRecur(tree)\n print()\n posOrderRecur(tree)\n print()\n preOrderRecur(tree2)\n print()\n inOrderRecur(tree2)\n print()\n posOrderRecur(tree2)\n print()\n preOrderRecur(tree3)\n print()\n inOrderRecur(tree3)\n print()\n posOrderRecur(tree3)\n print() \n \n \n","sub_path":"算法题/程序员面试指南/python/树/先中后序数组两两组合重构二叉树.py","file_name":"先中后序数组两两组合重构二叉树.py","file_ext":"py","file_size_in_byte":2850,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"586573634","text":"# Copyright (c) Ye Liu. All rights reserved.\n\nfrom collections import OrderedDict\n\nimport torch.distributed as dist\nfrom torch._utils import (_flatten_dense_tensors, _take_tensors,\n _unflatten_dense_tensors)\nfrom torch.nn.utils import clip_grad\n\nfrom .base import HOOKS, Hook\n\n\ndef _allreduce_coalesced(tensors, world_size, bucket_size_mb):\n if bucket_size_mb > 0:\n bucket_size_bytes = bucket_size_mb * 1024 * 1024\n buckets = _take_tensors(tensors, bucket_size_bytes)\n else:\n buckets = OrderedDict()\n for tensor in tensors:\n tp = tensor.type()\n if tp not in buckets:\n buckets[tp] = []\n buckets[tp].append(tensor)\n buckets = buckets.values()\n\n for bucket in buckets:\n flat_tensors = _flatten_dense_tensors(bucket)\n dist.all_reduce(flat_tensors)\n flat_tensors.div_(world_size)\n for tensor, synced in zip(\n bucket, _unflatten_dense_tensors(flat_tensors, bucket)):\n tensor.copy_(synced)\n\n\ndef _allreduce_grads(params, coalesce, bucket_size_mb):\n grads = [\n param.grad.data for param in params\n if param.requires_grad and param.grad is not None\n ]\n world_size = dist.get_world_size()\n if coalesce:\n _allreduce_coalesced(grads, world_size, bucket_size_mb)\n else:\n for tensor in grads:\n dist.all_reduce(tensor.div_(world_size))\n\n\n@HOOKS.register()\nclass OptimizerHook(Hook):\n\n def __init__(self, interval=1):\n super(OptimizerHook, self).__init__()\n self._interval = interval\n\n def _avg_grads(self, engine):\n step_size = engine.iter_in_epoch - self._last_update_iter + 1\n for param in engine.model.parameters():\n if param.requires_grad and param.grad is not None:\n param.grad.data.div_(step_size)\n self._last_update_iter = engine.iter_in_epoch + 1\n\n def _clip_grads(self, params, cfg):\n params_with_grad = filter(\n lambda p: p.requires_grad and p.grad is not None, params)\n if len(params_with_grad) > 0:\n clip_grad.clip_grad_norm_(params_with_grad, **cfg)\n\n def before_train_epoch(self, engine):\n self._last_update_iter = 0\n engine.optimizer.zero_grad()\n\n def after_train_iter(self, engine):\n loss_type = engine.cur_stage.get('loss', 'loss')\n engine.losses[loss_type].backward()\n\n if self.every_n_iters_in_epoch(\n engine, self._interval) or self.last_iter_in_epoch(engine):\n self._avg_grads(engine)\n grad_clip = engine.cur_stage.get('grad_clip', None)\n if grad_clip is not None:\n self._clip_grads(engine.model.parameters(), grad_clip)\n engine.optimizer.step()\n engine.optimizer.zero_grad()\n\n def after_train_epoch(self, engine):\n engine.optimizer.zero_grad()\n\n\n@HOOKS.register()\nclass DistOptimizerHook(OptimizerHook):\n\n def __init__(self, interval=1, coalesce=True, bucket_size_mb=-1):\n super(DistOptimizerHook, self).__init__()\n self._interval = interval\n self._coalesce = coalesce\n self._bucket_size_mb = bucket_size_mb\n\n def _avg_grads(self, engine):\n super(DistOptimizerHook, self)._avg_grads(engine)\n _allreduce_grads(engine.model.parameters(), self._coalesce,\n self._bucket_size_mb)\n","sub_path":"nncore/engine/hooks/optimizer.py","file_name":"optimizer.py","file_ext":"py","file_size_in_byte":3421,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"529442855","text":"#!/usr/bin/env python3\n\nimport sys\nimport os\nimport shutil\nfrom jinja2 import Environment, FileSystemLoader\n\nTEMPLATE_FILETYPES = [\n \"md\"\n]\n\ndef write_file(dest, content):\n \"\"\"Write given content to the specified destination file\"\"\"\n with open(dest, 'w') as file:\n file.write(content)\n\nif len(sys.argv) < 3:\n print(\"Not enough arguments\")\n sys.exit(1)\n\nsearch_path = os.path.abspath(sys.argv[1])\noutput_path = os.path.abspath(sys.argv[2])\n\nenv = Environment(\n loader=FileSystemLoader(search_path)\n)\n\nfor template in env.list_templates():\n # extract file name\n filename = template.split(\"/\")[-1]\n # remove file from path to get relative path to specified file\n base_dir = os.path.abspath(os.path.join(search_path, template.replace(filename, \"\")))\n # create target path\n target_dir = base_dir.replace(search_path, output_path)\n target = os.path.abspath(os.path.join(target_dir, filename))\n\n # create target folder if not existent\n if not os.path.isdir(target_dir):\n os.makedirs(target_dir)\n\n # skip if file is not a plain text file\n if template.split(\".\")[-1] not in TEMPLATE_FILETYPES:\n shutil.copyfile(os.path.join(base_dir, filename), target)\n else:\n template_obj = env.get_template(template)\n \n def get_uml(path):\n \"\"\"Get content of specified file\"\"\"\n with open(os.path.join(base_dir, path), \"r\") as file:\n return file.read()\n\n # render template based on given file path\n rendered_result = template_obj.render(get_uml=get_uml) \n\n write_file(target, rendered_result)\n","sub_path":"templating.py","file_name":"templating.py","file_ext":"py","file_size_in_byte":1628,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"350758906","text":"import sqlalchemy\nfrom dnascissors.config import cfg\nfrom dnascissors.model import Base\nfrom dnascissors.model import Target\nimport requests\nimport json\n\n\ndef get_gene_id(gene_symbol, species='homo_sapiens'):\n species = species.lower().replace(' ', '_')\n url = (\"https://rest.ensembl.org/xrefs/symbol/{}/{}?\").format(species, gene_symbol)\n r = requests.get(url, headers={\"Content-Type\": \"application/json\"})\n gene = json.loads(r.text)\n if gene:\n return gene[0]['id']\n\n\ndef main():\n engine = sqlalchemy.create_engine(cfg['DATABASE_URI'])\n Base.metadata.bind = engine\n DBSession = sqlalchemy.orm.sessionmaker(bind=engine)\n dbsession = DBSession()\n targets = dbsession.query(Target).all()\n print('Update gene ids on target in {} to be Ensembl ones starting with ENSG...'.format(cfg['DATABASE_URI']))\n for target in targets:\n old_gene_id = target.gene_id\n new_gene_id = get_gene_id(target.gene_id, target.genome.species)\n if not old_gene_id == new_gene_id:\n new_gene_id = get_gene_id(target.name, target.genome.species)\n if not new_gene_id and target.name.startswith('LEPR'):\n new_gene_id = get_gene_id('LEPR', target.genome.species)\n\n if new_gene_id:\n if target.description:\n if len(target.description) >= 984:\n target.description = \"{}... [submitted_gene_id: {}]\".format(target.description[:984], old_gene_id)\n else:\n target.description = \"{} [submitted_gene_id: {}]\".format(target.description, old_gene_id)\n else:\n target.description = \"[submitted_gene_id: {}]\".format(old_gene_id)\n target.gene_id = new_gene_id\n dbsession.commit()\n print(\">>>UPDATED: {}\\t{}\\t{}\\t{}\".format(target.name, old_gene_id, new_gene_id, target.description))\n else:\n print(\"!!! NO ID: {}\\t{}\\t{}\".format(target.name, target.gene_id, target.description))\n else:\n print(\"UP-TO-DATE: {}\\t{}\\t{}\".format(target.name, target.gene_id, target.description))\n print('done')\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"python/scripts/update_target_gene_ids.py","file_name":"update_target_gene_ids.py","file_ext":"py","file_size_in_byte":2218,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"531182429","text":"# -*- coding: utf-8 -*-\n\n# Define your item pipelines here\n#\n# Don't forget to add your pipeline to the ITEM_PIPELINES setting\n# See: http://doc.scrapy.org/en/latest/topics/item-pipeline.html\nimport pymysql\n\nfrom sqlalchemy.sql import func\n\nfrom models import Company, WikipediaDb, XingCompanyDb, session\n\n\nclass GoogleSpiderPipeline(object):\n \"\"\"\n Pipeline that processes info from google spider.\n For more info: https://doc.scrapy.org/en/latest/topics/item-pipeline.html\n \"\"\"\n collection_name = 'google_items'\n\n def process_item(self, item, spider):\n if not item['update']:\n new_company = Company(name=item['company_name'], website=item['url'], website_long=item['url_long'])\n session.add(new_company)\n else:\n session.query(Company).\\\n filter((Company.name == item['company_name']) & ((Company.website == None) | (Company.website == 'NA'))).\\\n update({'website': item['url'], 'website_long': item['url_long']})\n\n def close_spider(self, spider):\n session.commit()\n\n\nclass WikipediaSpiderPipeline(object):\n \"\"\"\n Pipeline that processes info from wikipedia spider.\n For more info: https://doc.scrapy.org/en/latest/topics/item-pipeline.html\n \"\"\"\n collection_name = 'wikipedia_items'\n\n def process_item(self, item, spider):\n update = item['update']\n company_name = item['company_name']\n company_website = item['company_website']\n item = dict(summary_wikipedia_w=item['summary'], categories_wikipedia_w=item['categories'],\n revenue_wikipedia_w=item['revenue'], revenue_currency_wiki_w=item['currency'],\n branch_wikipedia_w=item['branche'], wiki_url_w=item['url'], headquarters_wiki_w=item['sitz'],\n employees_wikipedia_w=item['mitarbeiter'], company_website_w=item['wiki_company_website'],\n last_update_w=func.now())\n\n company = session.query(Company).filter_by(name=company_name, website=company_website).first()\n if not update:\n new_entry = WikipediaDb(company_name_w=company_name, timestamp_w=func.now(), wc_id=company.id, **item)\n session.add(new_entry)\n else:\n session.query(WikipediaDb).filter(WikipediaDb.company_name_w == company_name).\\\n update(item, synchronize_session=False)\n\n company.last_update = func.now()\n company.wiki_evaluation = func.now()\n company.wikipedia_url = item['wiki_url_w']\n\n def close_spider(self, spider):\n session.commit()\n\n\nclass XingSpiderPipeline(object):\n \"\"\"\n Pipeline that processes info from xing spider.\n For more info: https://doc.scrapy.org/en/latest/topics/item-pipeline.html\n \"\"\"\n collection_name = 'xing_items'\n\n def process_item(self, item, spider):\n update = item['update']\n company_name = item['company_name']\n\n item = dict(street_xing=item.get('street'), city_xing=item.get('city'), description_xing=item.get('about_us'),\n zipcode_xing=item.get('postal_code'), country_xing=item.get('country'), tel_xing=item.get('phone'),\n fax_xing=item.get('fax'), company_email_xing=item.get('email'), industry_xing=item.get('industry'),\n established_in_xing=item.get('established'), products_xing=item.get('products'),\n employees_size_xing=item.get('employees_number'), company_website_x=item.get('url'),\n last_update_x=func.now(), employees_group_xing_x=item.get('registered_employees_number'))\n\n company = session.query(Company).filter_by(name=company_name)\n if not update:\n company = company.first()\n new_entry = XingCompanyDb(company_name_x=company_name, timestamp_x=func.now(), xc_id=company.id, **item)\n session.add(new_entry)\n else:\n company = company.filter(Company.xing_page != 'NA', Company.xing_page is not None).first()\n session.query(XingCompanyDb).filter(XingCompanyDb.xc_id == company.id).\\\n update(item, synchronize_session=False)\n\n company.last_update = func.now()\n company.xing_page_update = func.now()\n company.xing_page = item['company_website_x']\n\n def close_spider(self, spider):\n session.commit()\n","sub_path":"mx_crm/pipelines.py","file_name":"pipelines.py","file_ext":"py","file_size_in_byte":4328,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"381090783","text":"__author__ = 'shenli'\nfrom Logic.ChangLunLogic import *\nimport DataFeeder.DBFeeder as dbFeeder\nfrom Data.DataModel.TradePoints import trade_points\nimport UtilTools.DateUtils as DateUtils\n\ndef runModel(stockId,startDate,endDate,freq,tradeDate):\n trade_points.delete().where(trade_points.stockid == stockId).execute()\n sds = dbFeeder.getHistDataArray(stockId,startDate,endDate,freq)\n startIndex = getTradeStartIndex(sds,tradeDate)\n sds = sds[startIndex:]\n for i in range(50,len(sds)):\n changLun = ChangLunLogic()\n changLun.initData(stockId,startDate,endDate,freq,sds[0:i])\n changLun.run()\n changLun.isReadyTrade()\n changLun = ChangLunLogic()\n changLun.initData(stockId,startDate,endDate,freq,sds)\n changLun.run()\n changLun.loadhistBuyPoints()\n changLun.saveResult()\n #changLun.showResult()\n\ndef getStartIndex(histData):\n maxData = max(histData,key=lambda x:x.getHigh())\n minData = min(histData,key=lambda x:x.getLow())\n maxIndex = histData.index(maxData)\n minIndex = histData.index(minData)\n return max(0, min(maxIndex,minIndex) - 44)\n\ndef getTradeStartIndex(histData,tradeDate):\n tradeDate = DateUtils.convertDateStrToDatetime(tradeDate)\n for i in range(0,len(histData)):\n if histData[i].getDateTime() < tradeDate:\n continue\n return i\n\nif __name__ == '__main__':\n stockId = '000001'\n startDate = '2013-09-13'\n endDate = '2014-09-13'\n freq = '30'\n tradeDate = '2013-09-13'\n runModel(stockId,startDate,endDate,freq,tradeDate)\n","sub_path":"TradeModel/ChangLunModel.py","file_name":"ChangLunModel.py","file_ext":"py","file_size_in_byte":1548,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"321086340","text":"\n# coding=utf-8\nfrom __future__ import absolute_import\n\nimport octoprint.plugin\n\nclass UICustomizerPlugin(octoprint.plugin.StartupPlugin,\n octoprint.plugin.SettingsPlugin,\n octoprint.plugin.AssetPlugin,\n octoprint.plugin.TemplatePlugin):\n\n def on_after_startup(self):\n self._logger.info(\"UI Customizer is initialized.\")\n\n def get_assets(self):\n return dict(\n js=[\"js/uicustomizer.js\",\"js/Sortable.min.js\"],\n css=[\"css/uicustomizer.css\",\"css/bootstrap-responsive.css\"],\n less=[\"less/uicustomizer.less\"]\n\n )\n\n # default settings\n def get_settings_defaults(self):\n return {\n \"rows\" : [\n {\n \"#sidebar_plugin_firmware_check_wrapper\": True,\n \"#files_wrapper\": True,\n \"#connection_wrapper\": True\n },\n {\n \"div.UICmainTabs\": True\n },\n {\n \"#UICWebCamWidget\": True,\n \"#state_wrapper\": True,\n \"#sidebar_plugin_action_command_notification_wrapper\": True\n }\n ],\n \"widths\" : [3,6,3],\n \"fluidLayout\" : True,\n \"fixedHeader\" : True,\n \"fixedFooter\" : True,\n \"hideGraphBackground\" : True,\n \"responsiveMode\": True,\n \"navbarplugintempfix\": False,\n \"addWebCamZoom\" : False\n }\n\n def get_template_configs(self):\n return [\n dict(type=\"settings\", custom_bindings=False)\n ]\n\n def get_update_information(self):\n # Define the configuration for your plugin to use with the Software Update\n # Plugin here. See https://docs.octoprint.org/en/master/bundledplugins/softwareupdate.html\n # for details.\n return dict(\n uicustomizer=dict(\n displayName=self._plugin_name,\n displayVersion=self._plugin_version,\n\n # version check: github repository\n type=\"github_release\",\n user=\"LazeMSS\",\n repo=\"OctoPrint-UICustomizer\",\n current=self._plugin_version,\n\n # update method: pip\n pip=\"https://github.com/LazeMSS/OctoPrint-UICustomizer/archive/{target_version}.zip\"\n )\n )\n\n\n__plugin_name__ = \"UI Customizer\"\n__plugin_pythoncompat__ = \">=2.7,<4\"\n\ndef __plugin_load__():\n global __plugin_implementation__\n __plugin_implementation__ = UICustomizerPlugin()\n\n global __plugin_hooks__\n __plugin_hooks__ = {\n \"octoprint.plugin.softwareupdate.check_config\": __plugin_implementation__.get_update_information\n }","sub_path":"octoprint_uicustomizer/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":2779,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"256841361","text":"import wx\n\n\ndata = [(\"Jessica Alba\", \"Pomona\", \"1981\"), \n (\"Sigourney Weaver\", \"New York\", \"1949\"),\n (\"Angelina Jolie\", \"Los Angeles\", \"1975\"),\n (\"Natalie Portman\", \"Jerusalem\", \"1981\"),\n (\"Rachel Weiss\", \"London\", \"1971\"), \n (\"Scarlett Johansson\", \"New York\", \"1984\")]\n\nclass Example(wx.Frame):\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.selected_list = []\n\n self.initUI()\n\n def initUI(self):\n hbox = wx.BoxSizer(orient=wx.HORIZONTAL)\n panel = wx.Panel(parent=self)\n\n self.list = wx.ListCtrl(parent=panel, id=wx.ID_ANY, style=wx.LC_REPORT)\n self.list.InsertColumn(col=0, heading=\"name\", width=140)\n self.list.InsertColumn(col=1, heading=\"place\", width=130)\n self.list.InsertColumn(col=2, heading=\"year\", format=wx.LIST_FORMAT_LEFT, width=90)\n\n idx = 0\n\n for i in data:\n index = self.list.InsertItem(index=idx, label=i[0])\n self.list.SetItem(index=index, column=1, label=i[1])\n self.list.SetItem(index=index, column=2, label=i[2])\n idx += 1\n\n self.list.Bind(event=wx.EVT_LIST_ITEM_SELECTED, handler=self.selectedItem)\n self.list.Bind(event=wx.EVT_LIST_ITEM_DESELECTED, handler=self.deselectedItem)\n\n\n hbox.Add(window=self.list, proportion=1, flag=wx.EXPAND)\n panel.SetSizer(sizer=hbox)\n\n self.SetTitle(\"Actresses\")\n self.Center()\n\n def selectedItem(self, event):\n selected_item = event.GetIndex()\n print(self.list.GetItem(selected_item, col=1).GetText())\n self.selected_list.append(selected_item)\n\n # print(self.selected_list)\n\n def deselectedItem(self, event):\n deselected_item = event.GetIndex()\n self.selected_list.remove(deselected_item)\n\n # print(self.selected_list)\n\n\n # def deselectedItem(self, event):\n # selected_item = event.GetText()\n # self.selected_list.remove(selected_item)\n # print(self.selected_list)\n\n\ndef main():\n app = wx.App()\n ex = Example(None)\n ex.Show()\n app.MainLoop()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"Python3/ZetCode/wxPython/7 - Advanced widgets/7.4 - actresses TEST.py","file_name":"7.4 - actresses TEST.py","file_ext":"py","file_size_in_byte":2166,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"126732262","text":"\nfrom network import MotionNet\n\nTEST=True\n\n#The following parameters must have the same value in 'training' and 'test' modes.\nnum_layer=1\nhidden_unit = 1000\ntime_step = 90\nbatch_Frame = 1\nsave_period = 300\nuse_gpu=True\nuse_cudnn=True\nModel=2 # # Only 1, 2, and 3 are possible and only works when TEST = True.\n'''Execution'''\nif TEST:\n MotionNet(TEST=TEST , Model=1 ,save_period=save_period, num_layer=num_layer , hidden_unit=hidden_unit , time_step = time_step , batch_Frame= batch_Frame , use_gpu=use_gpu , use_cudnn=use_cudnn , graphviz=False)\nelse:\n #batch learning\n MotionNet(epoch=300 , batch_size=68 , save_period=save_period, optimizer='adam', learning_rate=0.01 ,Dropout=0.2 , use_gpu=use_gpu , use_cudnn=use_cudnn ,\n TEST=TEST , num_layer=num_layer , hidden_unit=hidden_unit , time_step = time_step , batch_Frame = batch_Frame , graphviz=False )\n\n","sub_path":"DeepHumanPrediction/Code/Master_Thesis/Human Motion Analysis/Human Motion Pattern Classification/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":869,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"230774988","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('students', '0003_auto_20150324_2059'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='Exam',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('name', models.CharField(max_length=256, verbose_name='\\u041d\\u0430\\u0437\\u0432\\u0430')),\n ('date', models.DateTimeField(null=True, verbose_name='\\u0414\\u0430\\u0442\\u0430 \\u0456 \\u0447\\u0430\\u0441 \\u043f\\u0440\\u043e\\u0432\\u0435\\u0434\\u0435\\u043d\\u043d\\u044f')),\n ('teacher', models.CharField(max_length=256, null=True, blank=True)),\n ('groups', models.ForeignKey(verbose_name=b'\\xd0\\x93\\xd1\\x80\\xd1\\x83\\xd0\\xbf\\xd0\\xb0', blank=True, to='students.Group', null=True)),\n ],\n options={\n 'verbose_name': '\\u0415\\u043a\\u0437\\u0430\\u043c\\u0435\\u043d',\n 'verbose_name_plural': '\\u0415\\u043a\\u0437\\u0430\\u043c\\u0435\\u043d\\u0438',\n },\n bases=(models.Model,),\n ),\n ]\n","sub_path":"students/migrations/0004_exam.py","file_name":"0004_exam.py","file_ext":"py","file_size_in_byte":1239,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"204198287","text":"# -*- coding: cp936 -*-\n\"\"\"\n题目:已知2016年1月1日是周五,请编写打印某个月月历的函数\nprintCalendar(month),该函数根据传递的month月份,按示\n例格式打印该月月历。请编写代码测试你所编写的函数正确性。\n比如:二月份的月历输出结果为:\n 日 一 二 三 四 五 六\n 1 2 3 4 5 6\n 7 8 9 10 11 12 13\n 14 15 16 17 18 19 20\n 21 22 23 24 25 26 27\n 28 29\n\"\"\"\n\nfirst_weekday = 5 # 该年份第一天是周5\ndays = [31, 29, 31, 30, 31, 30, 31, 31, 30, 31, 30, 31] # 每个月的天数\ndef printCalendar(month):\n # 计算month月份第一天是星期几\n t=days[month-1]\n s=0\n for i in range(0,month-1,1) :\n s=s+days[i]\n k=(s-2)%7\n p=[[],[],[],[],[],[]]\n j=(t-(7-k))//7\n b=t-(7-k)-7*j\n\n # 计算该月的每一天是星期几\n d=1\n for i in range(k,7,1):\n p[0].append(d)\n d=d+1 \n for q in range(1,j+1,1):\n for i in range(0,7,1):\n p[q].append(d)\n d=d+1 \n for i in range(0,b,1):\n p[j+1].append(d)\n d=d+1\n for i in range(b,7,1):\n p[j+1].append('\\t')\n \n\n \n\n\n # 打印标题(日一二三四五六)\n print('日\\t一\\t二\\t三\\t四\\t五\\t六')\n\n # 输出月历\n print('\\t'*k,end='')\n for i in range(0,7-k,1):\n print('%d\\t' %p[0][i],end=\"\")\n print('\\t') \n for j in range(1,j+1,1):\n for i in range(0,7,1):\n print('%d\\t' %p[j][i],end='')\n print('\\t') \n for i in range(0,b,1):\n print('%d\\t' %p[j+1][i],end='')\n print('\\t'*(6-b))\n \n# 函数是否正确第一个测试用例:输出2月份月历\nmonth = 2\nprintCalendar(month)\n\n\"\"\"\n# 函数是否正确第二个测试用例:输出全年各月份月历\nmonthNames = ['一月', '二月', '三月', '四月', '五月', '六月', '七月', '八月', '九月', '十月', '十一月', '十二月']\nfor month in range(1, 12+1):\n print (' ', monthNames[month-1])\n printCalendar(month)\n\"\"\"\n","sub_path":"work_frame.py","file_name":"work_frame.py","file_ext":"py","file_size_in_byte":2093,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"615570279","text":"from django import forms\nfrom django.contrib.auth.forms import UserCreationForm, UserChangeForm\nfrom django.db import transaction\nfrom django.dispatch import receiver\nfrom django.contrib.auth.models import Group\nfrom django.views.generic import CreateView, ListView, View\nfrom django.db import transaction\nfrom django.dispatch import receiver\n# from django.forms import CheckboxSelectMultiple\n# from multiselectfield import MultiSelectField\nfrom .models import User, Vacancies\nfrom summary.models import Summary, Workexperience\nfrom category.models import Uzbekiston_provinces, Uzbekiston_region\nfrom category.models import Citizenship, Specialization\nfrom jobs.models import Vacanciess\n\n\n\n\n\n# Change OneToOne form \nclass UserForm(forms.ModelForm):\n last_name = forms.CharField(required=False, label=\"Familya\", widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'autofocus': 'autofocus', 'placeholder': 'Фамилия'}\n ))\n first_name = forms.CharField(required=False, label=\"Ism\", widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': 'Имя'}\n )) \n email = forms.CharField(required=False, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': 'Электронная почта'}\n ))\n username = forms.CharField(required=False, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': 'Логин'}\n ))\n # password = forms.CharField(required=False, widget=forms.TextInput(\n # attrs = {'class':'form-control', 'placeholder': 'Parol'}\n # ))\n\n\n class Meta:\n model = User\n fields = ('last_name', 'first_name', 'email', 'username')\n\n\n\nclass VacanciesForm(forms.ModelForm):\n phone_numer = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': 'Номер телефона'}\n ))\n name_company = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': 'Названия компании'}\n ))\n\n class Meta:\n model = Vacancies\n fields = ('phone_numer', 'name_company', 'provinces', 'region')\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.fields['provinces'].widget.attrs.update({'class': 'form-control rounded-1', 'placeholder': 'Login'}) \n self.fields['region'].widget.attrs.update({'class': 'form-control rounded-1', 'placeholder': 'Login'}) \n\n\n\n\n\n\n# ↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓\n# ↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓\nclass UserSignUpForm(UserCreationForm):\n\n first_name = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'autofocus': 'autofocus', 'placeholder': 'Имя'}\n )) \n email = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'autocomplete':'username', 'placeholder': 'Электронная почта'}\n ))\n username = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'autocomplete':'username', 'placeholder': 'Логин'}\n ))\n password1 = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'type': 'password', 'class':'form-control rounded-1', 'autocomplete':'current-password', 'placeholder': 'Пароль'}\n ))\n password2 = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'type': 'password','class':'form-control rounded-1', 'autocomplete':'current-password', 'placeholder': 'Повтор пароля'}\n ))\n \n \n class Meta(UserCreationForm.Meta):\n model = User \n\n @transaction.atomic\n def save(self, commit=True):\n user = super().save(commit=False)\n user.is_active = True\n user.first_name = self.cleaned_data.get('first_name')\n user.email = self.cleaned_data.get('email')\n user.save()\n return user\n\n # def __init__(self, *args, **kwargs):\n # super().__init__(*args, **kwargs)\n # self.fields['username'].widget.attrs.update({'class': 'form-control', 'placeholder': 'Login'}) \n # self.fields['password1'].widget.attrs.update({'class': 'form-control', 'placeholder': 'Parol'}) \n # self.fields['password2'].widget.attrs.update({'class': 'form-control', 'placeholder': 'Parolni qaytadan kiriting'}) \n\n# ↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓\n# ↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓\n\n\n\n\nYJS = (\n ('юридического лица', 'юридического лица'),\n ('физическое лицо', 'физическое лицо'),\n ('другие', 'другие'),\n )\n\n# ↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓\n# ↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓\n\nclass UservacanciesSignUpForm(UserCreationForm):\n\n last_name = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'autofocus': 'autofocus', 'placeholder': 'Фамилия'}\n ))\n first_name = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': 'Имя'}\n )) \n email = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': 'Электронная почта'}\n ))\n name_company = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': 'Названия компании'}\n ))\n choose = forms.ChoiceField(choices=YJS)\n phone_numer = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': '+9'}\n ))\n username = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': 'Логин'}\n ))\n password1 = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'type': 'password','class':'form-control rounded-1', 'placeholder': 'Пароль'}\n ))\n password2 = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'type': 'password','class':'form-control rounded-1', 'placeholder': 'Повтор пароля'}\n ))\n provinces = forms.ModelChoiceField(queryset=Uzbekiston_provinces.objects.all(), label=\"Viloyatlar\", required=True)\n region = forms.ModelChoiceField(queryset=Uzbekiston_region.objects.all(), label=\"Tumanlar\", required=True)\n\n class Meta(UserCreationForm.Meta):\n model = User \n\n @transaction.atomic\n def save(self, commit=True):\n user = super().save(commit=False)\n user.is_active = True\n\n if commit:\n user.save()\n group = Group.objects.get(name=\"vacancies\")\n user.groups.add(group)\n\n user.first_name = self.cleaned_data.get('first_name')\n user.last_name = self.cleaned_data.get('last_name')\n user.email = self.cleaned_data.get('email')\n user.username = self.cleaned_data.get('username')\n user.password1 = self.cleaned_data.get('password1')\n user.password2 = self.cleaned_data.get('password2')\n user.save()\n useradminpage = Vacancies.objects.create(user=user)\n useradminpage.phone_numer=self.cleaned_data.get('phone_numer') \n useradminpage.name_company=self.cleaned_data.get('name_company') \n useradminpage.provinces=self.cleaned_data.get('provinces') \n useradminpage.region=self.cleaned_data.get('region') \n useradminpage.choose=self.cleaned_data.get('choose') \n useradminpage.save()\n return user\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.fields['provinces'].widget.attrs.update({'class': 'form-control rounded-1'}) \n self.fields['region'].widget.attrs.update({'class': 'form-control rounded-1'})\n self.fields['choose'].widget.attrs.update({'class': 'form-control rounded-1'})\n\n# ↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓\n# ↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓\n\n\n\n\n\n\n\n\n\n\n\n\nSEX = (\n ('Мужской', 'Мужской'),\n ('Женский', 'Женский'),\n )\n\nWORKEX = (\n (\"Есть опыт работы\", \"Есть опыт работы\"),\n (\"Нет опыта работы\", \"Нет опыта работы\"),\n )\nMONEY = (\n ('USD', \"USD\"),\n ('RUB', \"RUB\"),\n (\"UZS\", \"UZS\"),\n )\n\nclass SummaryForm(forms.ModelForm):\n last_name = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': 'Фамилия'}\n ))\n first_name = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'autofocus': 'autofocus', 'placeholder': 'Имя'}\n ))\n phone_name = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': '+9 '}\n ))\n data_time = forms.DateField(required=True, widget=forms.DateInput(\n attrs = {'type': 'date', 'class':'form-control rounded-1'}\n ))\n\n sex = forms.ChoiceField(widget=forms.RadioSelect, choices=SEX)\n citizenship = forms.ModelMultipleChoiceField(queryset=Citizenship.objects.all(), widget=forms.CheckboxSelectMultiple(), required=False)\n workexper = forms.ChoiceField(widget=forms.RadioSelect, choices=WORKEX)\n \n moneys = forms.ChoiceField(required=False, choices=MONEY)\n speciali = forms.ModelMultipleChoiceField(queryset=Specialization.objects.all(), widget=forms.CheckboxSelectMultiple(), required=False)\n\n thesalary = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': ''}\n ))\n carobjec = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': ''}\n ))\n key_skills = forms.CharField(required=False, widget=forms.Textarea(\n attrs = {'class':'form-control col-md-12', 'type':'text','style':'height: 100px', 'placeholder': \"\"}\n ))\n\n class Meta:\n model = Summary\n fields = ('last_name', 'first_name', 'phone_name', 'provinces', 'region', 'data_time', 'sex', 'citizenship', 'workexper', 'thesalary', 'carobjec', 'key_skills', 'moneys', 'speciali', 'profess', 'thchopr', 'image1', 'image2', 'image3', 'image4', 'image5', 'image6')\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.fields['provinces'].widget.attrs.update({'class': 'form-control rounded-1'}) \n self.fields['region'].widget.attrs.update({'class': 'form-control rounded-1'})\n self.fields['moneys'].widget.attrs.update({'class': 'form-control rounded-1'})\n self.fields['profess'].widget.attrs.update({'class': 'form-control rounded-1'})\n self.fields['thchopr'].widget.attrs.update({'class': 'form-control rounded-1'})\n \n self.fields['image1'].widget.attrs.update({'class': 'form-control rounded-1'})\n self.fields['image2'].widget.attrs.update({'class': 'form-control rounded-1'})\n self.fields['image3'].widget.attrs.update({'class': 'form-control rounded-1'})\n self.fields['image4'].widget.attrs.update({'class': 'form-control rounded-1'})\n self.fields['image5'].widget.attrs.update({'class': 'form-control rounded-1'})\n self.fields['image6'].widget.attrs.update({'class': 'form-control rounded-1'})\n\n\n\n\nclass WorkexperienceForm(forms.ModelForm):\n begwork = forms.DateField(required=False, widget=forms.DateInput(attrs={'type': 'date', 'class': 'form-control rounded-1 formset-field'}))\n finishw = forms.DateField(required=False, widget=forms.DateInput(attrs={'type': 'date', 'class': 'form-control rounded-1 formset-field'}))\n organiza = forms.CharField(required=False, widget=forms.TextInput(attrs={'class': 'form-control rounded-1 formset-field'}))\n position = forms.CharField(required=False, widget=forms.TextInput(attrs={'class': 'form-control rounded-1 formset-field'}))\n workplaceres = forms.CharField(required=False, widget=forms.Textarea(attrs={'class': 'form-control rounded-1 formset-field', 'style':'height: 50px'}))\n\n class Meta:\n model = Workexperience\n\n fields = [\n 'begwork',\n 'finishw',\n 'organiza',\n 'position',\n 'workplaceres',\n ]\n\n # widgets = {\n # 'begwork': forms.DateInput(attrs={'type': 'date', 'class': 'formset-field'}),\n # 'finishw': forms.DateInput(attrs={'type': 'date', 'class': 'formset-field'}),\n # 'organiza': forms.TextInput(attrs={'class': 'formset-field'}),\n # 'position': forms.TextInput(attrs={'class': 'formset-field'}),\n # 'workplaceres': forms.Textarea(attrs={'class': 'formset-field'}),\n # }\n\n\n\n\n\n\n\n\n\n\n\n\n\nCHOOSEM = (\n ('до налогов', \"до налогов\"),\n (\"в руках\", \"в руках\"),\n )\n\nEMPLOYMENT = (\n (\"Полная занятость\", \"Полная занятость\"),\n (\"Неполная занятость\", \"Неполная занятость\"),\n (\"Проектная работа или разовое задание\", \"Проектная работа или разовое задание\"),\n (\"Волонтерство\", \"Волонтерство\"),\n (\"Практика\", \"Практика\"),\n )\n\nWORKHO = (\n (\"Полный день\", \"Полный день\"),\n (\"Сменная работа\", \"Сменная работа\"),\n (\"По таблицах\", \"По таблицах\"),\n (\"Долгая работа\", \"Долгая работа\"),\n )\n\nclass VacanciessForm(forms.ModelForm):\n job_title = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'autofocus': 'autofocus', 'placeholder': ''}\n ))\n job_code = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': ''}\n ))\n thesalaryfrom = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': 'от'}\n ))\n thesalaryto = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': 'до'}\n ))\n choosemo = forms.ChoiceField(widget=forms.RadioSelect, choices=CHOOSEM)\n geolocation = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': ''}\n ))\n emptype = forms.ChoiceField(widget=forms.RadioSelect, choices=EMPLOYMENT)\n worhou = forms.ChoiceField(widget=forms.RadioSelect, choices=WORKHO)\n name = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': ''}\n ))\n phone = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', 'placeholder': '+9'}\n ))\n comment = forms.CharField(required=True, widget=forms.TextInput(\n attrs = {'class':'form-control rounded-1', \"placeholder\": \"\"}\n ))\n\n class Meta:\n model = Vacanciess\n fields = (\n 'job_title', 'job_code', 'profess', 'thchopr', 'description', 'moneys',\n 'thesalaryfrom', 'thesalaryto', 'choosemo', 'provinces', 'region', 'geolocation', 'emptype',\n 'worhou', 'name', 'phone', 'comment'\n )\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.fields['profess'].widget.attrs.update({'class': 'form-control rounded-1'}) \n self.fields['thchopr'].widget.attrs.update({'class': 'form-control rounded-1'}) \n self.fields['moneys'].widget.attrs.update({'class': 'form-control rounded-1'})\n self.fields['provinces'].widget.attrs.update({'class': 'form-control rounded-1'}) \n self.fields['region'].widget.attrs.update({'class': 'form-control rounded-1'}) \n \n","sub_path":"accounts/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":17659,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"504111481","text":"###########################################################################\n# \n# Copyright 2019 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n###########################################################################\n\nimport base64\nimport jsonpickle\nfrom time import sleep\n\nfrom django.core.management.base import BaseCommand, CommandError\nfrom django.conf import settings\nfrom django.utils import encoding\n\nfrom starthinker.util.project import project\nfrom starthinker.util.auth import get_profile\nfrom starthinker.util.auth.wrapper import CredentialsUserWrapper\nfrom starthinker.util.auth.storage import _credentials_storage_service\nfrom starthinker_ui.account.models import Account\n\nLEGACY_BUCKET_PREFIX = settings.UI_PROJECT.split(':', 1)[-1] # remove domain: part\n\ndef legacy_credentials_path(identifier):\n return '%s:ui/%s.json' % (LEGACY_BUCKET_PREFIX + \"-starthinker-users\", identifier)\n\ndef legacy_auth_decode(value):\n return jsonpickle.decode(base64.b64decode(encoding.smart_bytes(value)))\n\ndef legacy_credentails_get(cloud_path):\n service = _credentials_storage_service()\n bucket, filename = cloud_path.split(':',1)\n data = service.objects().get_media(bucket=bucket, object=filename).execute()\n return legacy_auth_decode(data)\n\n\nclass Command(BaseCommand):\n help = 'Migrate credentials to new format.'\n\n def add_arguments(self, parser):\n parser.add_argument(\n '--test',\n action='store_true',\n dest='test',\n default=False,\n help='Test conversion.',\n )\n\n def handle(self, *args, **kwargs):\n \n # loop through accounts\n for account in Account.objects.all():\n print('CONVERTING', account.email)\n\n try:\n # load legacy credentials\n credentials = legacy_credentails_get(legacy_credentials_path(account.identifier))\n\n # convert to new format\n new_credentials = CredentialsUserWrapper(credentials)\n\n # save new credentials\n account.set_credentials(new_credentials)\n\n if kwargs['test']:\n project.initialize(_user=account.get_credentials_path())\n profile = get_profile()\n print(profile)\n exit()\n\n except Exception as e:\n print(str(e))\n","sub_path":"starthinker_ui/account/management/commands/credentials_migrate.py","file_name":"credentials_migrate.py","file_ext":"py","file_size_in_byte":2716,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"15883018","text":"#!/usr/bin/python3\n# user Github API to list 10 commits from specific user in specific repo\nimport requests\nimport sys\n\n\nif __name__ == \"__main__\":\n # user Github API to list 10 commits from specific user in specific repo\n try:\n repo = sys.argv[1]\n owner = sys.argv[2]\n url = 'https://api.github.com/repos/{}/{}/commits'.format(owner, repo)\n r = requests.get(url)\n r_json = r.json()\n length = len(r_json)\n if length > 10:\n length = 10\n for i in range(length):\n print(\"{}: {}\".format(r_json[i].get('sha'),\n r_json[i].get('commit')\n .get('author').get('name')))\n except:\n print(\"None\")\n","sub_path":"0x11-python-network_1/100-github_commits.py","file_name":"100-github_commits.py","file_ext":"py","file_size_in_byte":744,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"160929882","text":"#!/usr/bin/env python\n#\n# SPDX-License-Identifier: Apache-2.0\n# Copyright Contributors to the OpenTimelineIO project\n\nimport opentimelineio as otio\nimport opentimelineio.test_utils as otio_test_utils\n\nimport unittest\n\n\nclass OpenTimeTypeSerializerTest(unittest.TestCase):\n\n def test_serialize_time(self):\n rt = otio.opentime.RationalTime(15, 24)\n encoded = otio.adapters.otio_json.write_to_string(rt)\n decoded = otio.adapters.otio_json.read_from_string(encoded)\n self.assertEqual(rt, decoded)\n\n rt_dur = otio.opentime.RationalTime(10, 20)\n tr = otio.opentime.TimeRange(rt, rt_dur)\n encoded = otio.adapters.otio_json.write_to_string(tr)\n decoded = otio.adapters.otio_json.read_from_string(encoded)\n self.assertEqual(tr, decoded)\n\n tt = otio.opentime.TimeTransform(rt, scale=1.5)\n encoded = otio.adapters.otio_json.write_to_string(tt)\n decoded = otio.adapters.otio_json.read_from_string(encoded)\n self.assertEqual(tt, decoded)\n\n\nclass SerializableObjTest(unittest.TestCase, otio_test_utils.OTIOAssertions):\n def test_cons(self):\n so = otio.core.SerializableObjectWithMetadata()\n so.metadata['foo'] = 'bar'\n self.assertEqual(so.metadata['foo'], 'bar')\n\n def test_update(self):\n so = otio.core.SerializableObjectWithMetadata()\n so.metadata.update({\"foo\": \"bar\"})\n self.assertEqual(so.metadata[\"foo\"], \"bar\")\n so_2 = otio.core.SerializableObjectWithMetadata()\n so_2.metadata[\"foo\"] = \"not bar\"\n so.metadata.update(so_2.metadata)\n self.assertEqual(so.metadata[\"foo\"], \"not bar\")\n\n def test_copy_lib(self):\n so = otio.core.SerializableObjectWithMetadata()\n so.metadata[\"meta_data\"] = {\"foo\": \"bar\"}\n\n import copy\n\n # shallow copy is an error\n with self.assertRaises(ValueError):\n so_cp = copy.copy(so)\n\n # deep copy\n so_cp = copy.deepcopy(so)\n self.assertIsOTIOEquivalentTo(so, so_cp)\n\n so_cp.metadata[\"foo\"] = \"bar\"\n self.assertNotEqual(so, so_cp)\n\n def test_copy_subclass(self):\n @otio.core.register_type\n class Foo(otio.core.SerializableObjectWithMetadata):\n _serializable_label = \"Foof.1\"\n\n foo = Foo()\n foo.metadata[\"meta_data\"] = {\"foo\": \"bar\"}\n\n import copy\n\n with self.assertRaises(ValueError):\n foo_copy = copy.copy(foo)\n\n foo_copy = copy.deepcopy(foo)\n\n self.assertEqual(Foo, type(foo_copy))\n\n def test_schema_versioning(self):\n @otio.core.register_type\n class FakeThing(otio.core.SerializableObject):\n _serializable_label = \"Stuff.1\"\n foo_two = otio.core.serializable_field(\"foo_2\", doc=\"test\")\n ft = FakeThing()\n\n self.assertEqual(ft.schema_name(), \"Stuff\")\n self.assertEqual(ft.schema_version(), 1)\n\n with self.assertRaises(otio.exceptions.UnsupportedSchemaError):\n otio.core.instance_from_schema(\n \"Stuff\",\n 2,\n {\"foo\": \"bar\"}\n )\n\n ft = otio.core.instance_from_schema(\"Stuff\", 1, {\"foo\": \"bar\"})\n self.assertEqual(ft._dynamic_fields['foo'], \"bar\")\n\n @otio.core.register_type\n class FakeThing(otio.core.SerializableObject):\n _serializable_label = \"NewStuff.4\"\n foo_two = otio.core.serializable_field(\"foo_2\")\n\n @otio.core.upgrade_function_for(FakeThing, 2)\n def upgrade_one_to_two(_data_dict):\n return {\"foo_2\": _data_dict[\"foo\"]}\n\n @otio.core.upgrade_function_for(FakeThing, 3)\n def upgrade_one_to_two_three(_data_dict):\n return {\"foo_3\": _data_dict[\"foo_2\"]}\n\n ft = otio.core.instance_from_schema(\"NewStuff\", 1, {\"foo\": \"bar\"})\n self.assertEqual(ft._dynamic_fields['foo_3'], \"bar\")\n\n ft = otio.core.instance_from_schema(\"NewStuff\", 3, {\"foo_2\": \"bar\"})\n self.assertEqual(ft._dynamic_fields['foo_3'], \"bar\")\n\n ft = otio.core.instance_from_schema(\"NewStuff\", 4, {\"foo_3\": \"bar\"})\n self.assertEqual(ft._dynamic_fields['foo_3'], \"bar\")\n\n def test_equality(self):\n o1 = otio.core.SerializableObject()\n o2 = otio.core.SerializableObject()\n self.assertTrue(o1 is not o2)\n self.assertTrue(o1.is_equivalent_to(o2))\n self.assertIsOTIOEquivalentTo(o1, o2)\n\n def test_equivalence_symmetry(self):\n def test_equivalence(A, B, msg):\n self.assertTrue(A.is_equivalent_to(B), \"{}: A ~= B\".format(msg))\n self.assertTrue(B.is_equivalent_to(A), \"{}: B ~= A\".format(msg))\n\n def test_difference(A, B, msg):\n self.assertFalse(A.is_equivalent_to(B), \"{}: A ~= B\".format(msg))\n self.assertFalse(B.is_equivalent_to(A), \"{}: B ~= A\".format(msg))\n\n A = otio.core.Composable()\n B = otio.core.Composable()\n test_equivalence(A, B, \"blank objects\")\n\n A.metadata[\"key\"] = {\"a\": 0}\n test_difference(A, B, \"A has different metadata\")\n\n B.metadata[\"key\"] = {\"a\": 0}\n test_equivalence(A, B, \"add metadata to B\")\n\n A.metadata[\"key\"][\"sub-key\"] = 1\n test_difference(A, B, \"Add dict within A with specific metadata\")\n\n def test_truthiness(self):\n o = otio.core.SerializableObject()\n self.assertTrue(o)\n\n def test_instancing_without_instancing_support(self):\n o = otio.core.SerializableObjectWithMetadata()\n c = otio.core.SerializableObjectWithMetadata()\n o.metadata[\"child1\"] = c\n o.metadata[\"child2\"] = c\n self.assertTrue(o.metadata[\"child1\"] is o.metadata[\"child2\"])\n\n oCopy = o.clone()\n # Note: If we ever enable INSTANCING_SUPPORT in the C++ code,\n # then this will (and should) fail\n self.assertTrue(oCopy.metadata[\"child1\"] is not oCopy.metadata[\"child2\"])\n\n def test_cycle_detection(self):\n o = otio.core.SerializableObjectWithMetadata()\n o.metadata[\"myself\"] = o\n\n # Note: If we ever enable INSTANCING_SUPPORT in the C++ code,\n # then modify the code below to be:\n # oCopy = o.clone()\n # self.assertTrue(oCopy is oCopy.metadata[\"myself\"])\n with self.assertRaises(ValueError):\n o.clone()\n\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"tests/test_serializable_object.py","file_name":"test_serializable_object.py","file_ext":"py","file_size_in_byte":6334,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"624683307","text":"import os\nimport sys\nfrom io import BytesIO\nimport numpy as np\nfrom PIL import Image\nimport cv2\n\nimport tensorflow as tf\nimport time\n\n\nworkspace = '/home/admin/workspace/models/workspace/seg/'\n\nmodel_dir = workspace + '035.512.saliency255/' # 'deconv.075.aug/'\nres_dir = model_dir + 'results/'\nif not os.path.exists(res_dir):\n os.makedirs(res_dir)\n\ndata_dir = workspace + 'test_images/'\nlistpath = data_dir + 'list.txt'\n\nFROZEN_GRAPH_NAME = model_dir + 'deploy/deploy_graph.pb'\nINPUT_SIZE = 512\nPADDING = True\nUSE_HEATMAP = True\n\nclass SgmtModel(object):\n \"\"\"Class to load deeplab model and run inference.\"\"\"\n\n INPUT_TENSOR_NAME = 'image:0'\n OUTPUT_TENSOR_NAME = 'heatmap:0'\n\n def __init__(self):\n \"\"\"Creates and loads pretrained deeplab model.\"\"\"\n self.graph = tf.Graph()\n graph_def = None\n with tf.gfile.GFile(FROZEN_GRAPH_NAME, \"rb\") as f:\n print(FROZEN_GRAPH_NAME)\n graph_def = tf.GraphDef().FromString(f.read())\n\n if graph_def is None:\n raise RuntimeError('Cannot find inference graph in tar archive.')\n\n with self.graph.as_default():\n tf.import_graph_def(graph_def, name='')\n\n gpu_options = tf.GPUOptions(per_process_gpu_memory_fraction=0.1)\n\n self.sess = tf.Session(graph=self.graph, config=tf.ConfigProto(gpu_options=gpu_options))\n\n def run(self, image):\n \"\"\"Runs inference on a single image.\n\n Args:\n image: A PIL.Image object, raw input image.\n\n Returns:\n resized_image: RGB image resized from original input image.\n seg_map: Segmentation map of `resized_image`.\n \"\"\"\n width, height = image.size\n resize_ratio = 1.0 * INPUT_SIZE / max(width, height)\n target_size = (int(resize_ratio * width), int(resize_ratio * height))\n resized_image = image.convert('RGB').resize(target_size, Image.ANTIALIAS)\n\n batch_size = 10\n the_input = [np.asarray(resized_image)]\n the_input = np.repeat(the_input, batch_size, 0)\n\n heatmap = self.sess.run(\n self.OUTPUT_TENSOR_NAME,\n feed_dict={self.INPUT_TENSOR_NAME: the_input})\n\n t1 = time.time()\n heatmap = self.sess.run(\n self.OUTPUT_TENSOR_NAME,\n feed_dict={self.INPUT_TENSOR_NAME: the_input})\n t2 = time.time()\n total = (t2 - t1) * 1000\n\n heatmap = heatmap[0, :, :, :]\n\n return heatmap, total\n\n\ndef infer_one(image, use_heatmap=True):\n # image preprocessing\n# img = Image.open(image_path)\n\n# padding = True\n width, height = image.size\n\n if PADDING:\n # padding\n large_one = max(width, height)\n \n scale = float(INPUT_SIZE) / float(large_one)\n \n new_width = 0\n new_height = 0\n if width >= height:\n new_width = INPUT_SIZE\n new_height = int(height * scale)\n else:\n new_height = INPUT_SIZE\n new_width = int(width * scale)\n \n image = image.resize((new_width, new_height), Image.ANTIALIAS)\n \n delta_w = INPUT_SIZE - new_width\n delta_h = INPUT_SIZE - new_height\n top, bottom = delta_h / 2, delta_h - delta_h / 2\n left, right = delta_w / 2, delta_w - delta_w / 2\n color = [127, 127, 127]\n img_array = np.array(image)\n img_array = cv2.copyMakeBorder(img_array, top, bottom, left, right, cv2.BORDER_CONSTANT,\n value=color)\n \n image = Image.fromarray(np.uint8(img_array))\n\n else:\n # resize\n image = image.resize((INPUT_SIZE, INPUT_SIZE), resample = Image.BILINEAR)\n img_array = np.array(image)\n\n # run model\n model = SgmtModel()\n heatmap, running_time = model.run(image)\n print(type(heatmap))\n heatmap = np.float32(heatmap) / 255.0\n if not use_heatmap:\n heatmap = np.where(heatmap > 0.5, 1, 0)\n\n # post processing\n if PADDING:\n # padding\n embed_array = img_array\n embed_array = np.multiply(img_array, heatmap) \n \n # get results\n embed_crop = embed_array[delta_h / 2 : new_height + delta_h / 2, delta_w / 2 : new_width + delta_w / 2]\n embed_crop = Image.fromarray(np.uint8(embed_crop))\n # embed_crop.save('data/embed_tf.png')\n \n heatmap_array = np.squeeze(heatmap)\n heatmap_array = heatmap_array[delta_h / 2 : new_height + delta_h / 2, delta_w / 2 : new_width + delta_w / 2] * 255\n heatmap_crop = Image.fromarray(np.uint8(heatmap_array))\n # heatmap_crop.save('data/heatmap_tf.png')\n\n else:\n # resize\n embed_array = img_array\n embed_array = np.multiply(img_array, heatmap)\n embed_crop = Image.fromarray(np.uint8(embed_array))\n# embed_crop = embed_crop.resize((width, height), resample = Image.BILINEAR)\n\n heatmap_array = np.squeeze(heatmap)\n heatmap_array = heatmap_array * 255\n heatmap_crop = Image.fromarray(np.uint8(heatmap_array))\n# heatmap_crop = heatmap_crop.resize((width, height), resample = Image.BILINEAR)\n\n return embed_crop, heatmap_crop, running_time\n\n\n# now start inferring\n# with open(listpath) as f:\n# lines = f.readlines()\n# lines = [x.strip('\\n') for x in lines] \n# print(lines)\n\n# for filename in lines:\n# filename_root = os.path.splitext(filename)[0]\n#\n# image = Image.open(data_dir + filename)\n# embed_crop, heatmap_crop, running_time = infer_one(image, USE_HEATMAP)\n# \n# image.save(res_dir + filename)\n# embed_crop.save(res_dir + filename_root + '.embed_tf.png')\n# heatmap_crop.save(res_dir + filename_root + '.heatmap_tf.png')\n# print('Time consumed on ', filename, ': ', running_time, ' ms.')\n\n\nif __name__ == \"__main__\":\n if len(sys.argv) < 3:\n print(\"inference \")\n exit(1)\n image = Image.open(sys.argv[1])\n embed_crop, heapmap_crop, running_time = infer_one(image, USE_HEATMAP)\n \n heapmap_crop.save(sys.argv[2])\n \n\n","sub_path":"workspace/seg/infer_utils/inference.py","file_name":"inference.py","file_ext":"py","file_size_in_byte":5546,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"548388023","text":"name = input('Enter your name: ')\nage = input('Enter your age: ')\n\nyear = 2018 - int(age) + 100\nmessage = name + ' will be 100 years old in ' + str(year) + '\\n'\nprint(message)\n\nnum = input('Enter a number: ')\nmessage = message * int(num)\n\nprint(message)","sub_path":"a.py","file_name":"a.py","file_ext":"py","file_size_in_byte":253,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"65809664","text":"from utils import TwitterStreamer\n\n# Twitter Bot Spotter\n# ******************FOLLOW THESE STEPS******************\n\n# *****BEFORE STARTING MAKE SURE YOU HAVE FILLED YOUR API KEYS INTO constants.py*****\n\n# STEP 1\n# Run \"step_1.py\" and stop it after 1 minute\n\n# STEP 2\n# In \"step_2.json\" replace the ',' at the beginning with '[' , then add ']' at the very end\n\n# STEP 3\n# Run \"step_3.py\"\n\n# STEP 4\n# Find your new Plotly chart in the \"step_4\" directory\n\n# ******************DELETE step_2.json BEFORE TRYING AGAIN******************\n\n\nif __name__ == '__main__':\n # try changing the keyword\n keyword = ['clinton']\n fetched_tweets_filename = \"step_2.json\"\n twitter_streamer = TwitterStreamer()\n twitter_streamer.stream_tweets(fetched_tweets_filename, keyword)\n\n","sub_path":"step_1.py","file_name":"step_1.py","file_ext":"py","file_size_in_byte":860,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"514982489","text":"import graphene\nfrom db.solutions import solutions as list_of_solutions\n\nfrom flask import (\n Flask,\n request,\n jsonify,\n render_template,\n redirect,\n make_response,\n session\n)\n\nfrom flask_graphql import GraphQLView\n\nfrom graphene_sqlalchemy import (\n SQLAlchemyObjectType,\n SQLAlchemyConnectionField\n)\n\nfrom app import app, db\n\nfrom core import (\n security,\n helpers,\n middleware\n)\n\nfrom core.models import (\n Owner,\n Paste,\n User,\n Audit\n)\n\nfrom version import VERSION\n\n# SQLAlchemy Types\nclass UserObject(SQLAlchemyObjectType):\n class Meta:\n model = User\n interfaces = (graphene.relay.Node, )\n\nclass PasteObject(SQLAlchemyObjectType):\n p_id = graphene.String(source='id')\n class Meta:\n model = Paste\n interfaces = (graphene.relay.Node, )\n\nclass OwnerObject(SQLAlchemyObjectType):\n class Meta:\n model = Owner\n interfaces = (graphene.relay.Node, )\n\nclass CreatePaste(graphene.Mutation):\n title = graphene.String()\n content = graphene.String()\n public = graphene.Boolean()\n paste = graphene.Field(lambda:PasteObject)\n burn = graphene.Boolean()\n\n class Arguments:\n title = graphene.String()\n content = graphene.String()\n public = graphene.Boolean(required=False, default_value=True)\n burn = graphene.Boolean(required=False, default_value=False)\n\n def mutate(self, info, title, content, public, burn):\n owner = Owner.query.filter_by(name='DVGAUser').first()\n\n paste_obj = Paste.create_paste(\n title=title,\n content=content, public=public, burn=burn,\n owner_id=owner.id, owner=owner, ip_addr=request.remote_addr,\n user_agent=request.headers.get('User-Agent', '')\n )\n\n Audit.create_audit_entry(gqloperation=helpers.get_opname(info.operation))\n\n return CreatePaste(paste=paste_obj)\n\nclass DeletePaste(graphene.Mutation):\n ok = graphene.Boolean()\n\n class Arguments:\n title = graphene.String()\n\n def mutate(self, info, title):\n Paste.query.filter_by(title=title).delete()\n db.session.commit()\n\n Audit.create_audit_entry(gqloperation=helpers.get_opname(info.operation))\n\n return DeletePaste(ok=True)\n\nclass UploadPaste(graphene.Mutation):\n content = graphene.String()\n filename = graphene.String()\n\n class Arguments:\n content = graphene.String(required=True)\n filename = graphene.String(required=True)\n\n result = graphene.String()\n\n def mutate(self, info, filename, content):\n result = helpers.save_file(filename, content)\n owner = Owner.query.filter_by(name='DVGAUser').first()\n\n Paste.create_paste(\n title='Imported Paste from File - {}'.format(helpers.generate_uuid()),\n content=content, public=False, burn=False,\n owner_id=owner.id, owner=owner, ip_addr=request.remote_addr,\n user_agent=request.headers.get('User-Agent', '')\n )\n\n Audit.create_audit_entry(gqloperation=helpers.get_opname(info.operation))\n\n return UploadPaste(result=result)\n\nclass ImportPaste(graphene.Mutation):\n result = graphene.String()\n\n class Arguments:\n host = graphene.String(required=True)\n port = graphene.Int(required=False)\n path = graphene.String(required=True)\n scheme = graphene.String(required=True)\n\n def mutate(self, info, host='pastebin.com', port=443, path='/', scheme=\"http\"):\n url = security.strip_dangerous_characters(f\"{scheme}://{host}:{port}{path}\")\n cmd = helpers.run_cmd(f'curl --insecure {url}')\n\n owner = Owner.query.filter_by(name='DVGAUser').first()\n Paste.create_paste(\n title='Imported Paste from URL - {}'.format(helpers.generate_uuid()),\n content=cmd, public=False, burn=False,\n owner_id=owner.id, owner=owner, ip_addr=request.remote_addr,\n user_agent=request.headers.get('User-Agent', '')\n )\n\n Audit.create_audit_entry(gqloperation=helpers.get_opname(info.operation))\n\n return ImportPaste(result=cmd)\n\nclass Mutations(graphene.ObjectType):\n create_paste = CreatePaste.Field()\n delete_paste = DeletePaste.Field()\n upload_paste = UploadPaste.Field()\n import_paste = ImportPaste.Field()\n\nclass Query(graphene.ObjectType):\n node = graphene.relay.Node.Field()\n pastes = graphene.List(PasteObject, public=graphene.Boolean())\n paste = graphene.Field(PasteObject, p_id=graphene.String())\n system_update = graphene.String()\n system_diagnostics = graphene.String(username=graphene.String(), password=graphene.String(), cmd=graphene.String())\n system_health = graphene.String()\n read_and_burn = graphene.Field(PasteObject, p_id=graphene.Int())\n\n def resolve_pastes(self, info, public=False):\n query = PasteObject.get_query(info)\n Audit.create_audit_entry(gqloperation=helpers.get_opname(info.operation))\n return query.filter_by(public=public, burn=False).order_by(Paste.id.desc())\n\n def resolve_paste(self, info, p_id):\n query = PasteObject.get_query(info)\n Audit.create_audit_entry(gqloperation=helpers.get_opname(info.operation))\n return query.filter_by(id=p_id, burn=False).first()\n\n def resolve_system_update(self, info):\n security.simulate_load()\n Audit.create_audit_entry(gqloperation=helpers.get_opname(info.operation))\n return 'no updates available'\n\n def resolve_system_diagnostics(self, info, username, password, cmd='whoami'):\n q = User.query.filter_by(username='admin').first()\n real_passw = q.password\n res, msg = security.check_creds(username, password, real_passw)\n Audit.create_audit_entry(gqloperation=helpers.get_opname(info.operation))\n if res:\n output = f'{cmd}: command not found'\n if security.allowed_cmds(cmd):\n output = helpers.run_cmd(cmd)\n return output\n return msg\n\n def resolve_read_and_burn(self, info, p_id):\n result = Paste.query.filter_by(id=p_id, burn=True).first()\n Paste.query.filter_by(id=p_id, burn=True).delete()\n db.session.commit()\n Audit.create_audit_entry(gqloperation=helpers.get_opname(info.operation))\n return result\n\n def resolve_system_health(self, info):\n Audit.create_audit_entry(gqloperation=helpers.get_opname(info.operation))\n return 'System Load: {}'.format(\n helpers.run_cmd(\"uptime | awk '{print $10, $11, $12}'\")\n )\n\n\n@app.route('/')\ndef index():\n resp = make_response(render_template('index.html'))\n resp.set_cookie(\"env\", \"Z3JhcGhpcWw6ZGlzYWJsZQ==\")\n return resp\n\n@app.route('/about')\ndef about():\n return render_template(\"about.html\")\n\n@app.route('/solutions')\ndef solutions():\n return render_template(\"solutions.html\", solutions=list_of_solutions)\n\n@app.route('/create_paste')\ndef create_paste():\n return render_template(\"paste.html\", page=\"create_paste\")\n\n@app.route('/import_paste')\ndef import_paste():\n return render_template(\"paste.html\", page=\"import_paste\")\n\n@app.route('/upload_paste')\ndef upload_paste():\n return render_template(\"paste.html\", page=\"upload_paste\")\n\n@app.route('/my_pastes')\ndef my_paste():\n return render_template(\"paste.html\", page=\"my_pastes\")\n\n@app.route('/public_pastes')\ndef public_paste():\n return render_template(\"paste.html\", page=\"public_pastes\")\n\n@app.route('/audit')\ndef audit():\n audit = Audit.query.all()\n return render_template(\"audit.html\", audit=audit)\n\n\n@app.route('/start_over')\ndef start_over():\n msg = \"Restored to default state.\"\n res = helpers.initialize()\n\n if 'done' not in res:\n msg=\"Could not restore to default state.\"\n\n return render_template('index.html', msg=msg)\n\n@app.route('/difficulty/')\ndef difficulty(level):\n if level in ('easy', 'hard'):\n message = f'Changed difficulty level to {level.capitalize()}'\n else:\n message = 'Level must be Beginner or Expert.'\n level = 'easy'\n\n helpers.set_mode(level)\n\n return render_template('index.html', msg = message)\n\n\n@app.context_processor\ndef get_version():\n return dict(version=VERSION)\n\n@app.before_request\ndef set_difficulty():\n mode_header = request.headers.get('X-DVGA-MODE', None)\n if mode_header:\n if mode_header == 'Expert':\n helpers.set_mode('hard')\n else:\n helpers.set_mode('easy')\n else:\n if session.get('difficulty') == None:\n helpers.set_mode('easy')\n\nschema = graphene.Schema(query=Query, mutation = Mutations)\n\ngql_middlew = [\n middleware.CostProtectionMiddleware(),\n middleware.DepthProtectionMiddleware(),\n middleware.IntrospectionMiddleware(),\n middleware.processMiddleware()\n]\n\nigql_middlew = [\n middleware.IGQLProtectionMiddleware()\n]\n\napp.add_url_rule('/graphql', view_func=GraphQLView.as_view(\n 'graphql',\n schema=schema,\n middleware=gql_middlew,\n batch=True\n))\n\napp.add_url_rule('/graphiql', view_func=GraphQLView.as_view(\n 'graphiql',\n schema = schema,\n graphiql = True,\n middleware = igql_middlew,\n batch=True\n))\n\n\n","sub_path":"Damn-Vulnerable-GraphQL-Application-master-SigSci-Agent/core/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":8644,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"394209635","text":"from tqdm import tqdm\nimport torch\nfrom modelfile import Model, Config\nimport torch.optim as optim\nfrom torch.utils.data import DataLoader\nfrom data_utils import load_data, Dataset\nimport os\n\n\nglove_vec_size = 300\nhidden_size = 100\n\ndata_dir = '../../data'\n\nlr = 0.002\t\t\t\t\t# learning rate\nDELTA = 0.0000000001\t\t# to avoid 0 in log\nweight_decay = 0.005\t\t# L2 regularization term\n\nmax_plen = 128\t\t# Max passage length\nmax_qlen = 16\t\t# Max query length\nnum_para = 10\t\t# Number of passages\n\nn_epochs = 100\t\t# Number of epochs\n\n\ndef train(n_epochs):\n\t\n\tmodel.train()\n\tfor epoch in tqdm(range(n_epochs)):\n\n\t\ttrain_loss = 0\n\t\t\n\t\tfor query_id, e_q, e_d, qseq_len, seq_len, labels in tqdm(training_generator):\n\t\t\t\n\t\t\tinputs = {\n\t\t\t\t\t'e_q' : e_q.to(device),\n\t\t\t\t\t'e_d' : e_d.to(device),\n\t\t\t\t\t'qseq_len' : qseq_len,\n\t\t\t\t\t'seq_len' : seq_len.to(device)\n\t\t\t}\n\n\t\t\toptimizer.zero_grad()\n\t\t\ts_d_norm = model(**inputs)\t\n\t\t\t# shape of s_d_norm - [batch_size, num_para]\n\n\t\t\tbatch_size = labels.shape[0]\n\t\t\t\n\t\t\tlabels = labels.to(device)\n\n\t\t\t# computing loss\n\t\t\tl1 = labels * torch.log(s_d_norm + DELTA)\n\t\t\tl2 = (1 - labels) * torch.log(1 - s_d_norm + DELTA)\n\t\t\tl_de = torch.add(l1,l2)\n\t\t\tloss = -torch.mean(l_de)\n\n\n\t\t\ttrain_loss += loss * batch_size\n\t\t\tloss.backward()\n\t\t\toptimizer.step()\n\t\t\n\n\t\tavg_loss = train_loss/len(train_data)\n\n\t\tprint('training loss = ',avg_loss.item(),' epoch = ',epoch+1)\n\t\t\n\t\ttorch.save({\n\t\t\t'model_state_dict': model.state_dict(),\n\t\t 'optimizer_state_dict': optimizer.state_dict(),\n\t\t}, 'checkpoints/saved_model_'+str(epoch)+'.pth')\n\n\n\ttorch.save({\n\t\t\t'model_state_dict': model.state_dict(),\n\t\t 'optimizer_state_dict': optimizer.state_dict(),\n\t\t}, 'checkpoints/saved_model.pth')\n\n\n\ntrain_data = load_data(data_dir + '/msmarco/train_v2.1.json')\n\nprint('Done loading Training data.')\n\n\nparams = {'batch_size': 128,\n 'shuffle': True,\n 'num_workers': 20,\n\t \t 'pin_memory': True}\n\n\ntraining_set = Dataset(data, max_plen, max_qlen, glove_vec_size, data_dir)\ntraining_generator = DataLoader(training_set, **params)\n\n\ndevice = torch.device('cpu')\n\ncuda = torch.cuda.is_available()\nif(cuda):\n\tdevice = torch.device('cuda')\n\n\nconfig = Config(glove_vec_size, hidden_size, max_plen, max_qlen, num_para, device)\n\nmodel = Model(config)\nif(cuda):\n\tmodel = model.to(device)\n\noptimizer = optim.Adam(model.parameters(), lr=lr, weight_decay=weight_decay)\n\nif not os.path.exists('checkpoints'):\n\tos.makedirs('checkpoints')\n\ntrain()","sub_path":"passage_ranking_extractive_summarization/passage_ranking/train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":2446,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"128029222","text":"#!/usr/bin/env python3\n\nimport os\nimport sys\nimport operator\nimport re\n\ndef usage():\n return '''\nThis script extracts common German surnames from phone book data collected in 2002. To use, first\ndownload, as .html files, the following three wiki pages (each batch contains 1000 names):\n\nhttp://wiki-de.genealogy.net/Die_1000_h%C3%A4ufigsten_Familiennamen_in_Deutschland\nhttp://wiki-de.genealogy.net/Die_2000_h%C3%A4ufigsten_Familiennamen_in_Deutschland\nhttp://wiki-de.genealogy.net/Die_3000_h%C3%A4ufigsten_Familiennamen_in_Deutschland\n\nPut those into a single directory and point it to this script:\n\n%s genealogy_de_html_dir ../data/german_surnames.txt\n\ngerman_surnames.txt will include one line per name, ordered by rank.\n...\n ''' % sys.argv[0]\n\nEXTRACT_RANK_AND_TOKEN_RE = re.compile(r'>([\\s\\d.]*)]*\\(Familienname\\)[^>]*>([\\w\\s]+)')\n\nHAS_UMLAUTS_RE = re.compile(r'[äöüß]')\n\ndef parse_wiki_tokens(html_doc_str):\n results = []\n num_tokens = 0\n\n for rank_and_token_match in EXTRACT_RANK_AND_TOKEN_RE.finditer(html_doc_str):\n num_tokens += 1\n rank_match = rank_and_token_match.group(1)\n if rank_match:\n rank = int(rank_match.replace('.', ''))\n else:\n rank = num_tokens\n token = rank_and_token_match.group(2)\n new_records = [(normalized, rank) for normalized in resolve_umlauts(token)]\n results.extend(new_records)\n\n # Each batch has 1000 entries.\n assert 1000 == num_tokens\n return results\n\ndef resolve_umlauts(token):\n normalized = token.lower().replace(' ', '')\n if not HAS_UMLAUTS_RE.search(normalized):\n return [normalized]\n else:\n normalized_ae = (normalized\n .replace('ä', 'ae')\n .replace('ö', 'oe')\n .replace('ü', 'ue')\n .replace('ß', 'ss'))\n normalized_a = (normalized\n .replace('ä', 'a')\n .replace('ö', 'o')\n .replace('ü', 'u')\n .replace('ß', 's'))\n return [normalized, normalized_a, normalized_ae]\n\n\ndef main(genealogy_de_html_root, output_filename):\n token_rank = [] # list of pairs\n for filename in os.listdir(genealogy_de_html_root):\n path = os.path.join(genealogy_de_html_root, filename)\n with open(path, 'r', encoding='utf8') as f:\n token_rank.extend(parse_wiki_tokens(f.read()))\n token_rank.sort(key=operator.itemgetter(1))\n with open(output_filename, 'w', encoding='utf8') as f:\n for surname, _ in token_rank:\n f.write('{}\\n'.format(surname))\n\nif __name__ == '__main__':\n if len(sys.argv) != 3:\n print(usage())\n else:\n main(*sys.argv[1:])\n sys.exit(0)\n","sub_path":"data-scripts/count_genealogy_de.py","file_name":"count_genealogy_de.py","file_ext":"py","file_size_in_byte":2687,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"559260402","text":"import logging\nfrom sanic import Blueprint\nfrom sanic.response import json\nfrom sanic.exceptions import InvalidUsage, NotFound\nfrom tao.models.product import Product, Label, Order\nfrom tao.settings import PRODUCTS_LINE_LIMIT\nfrom tao.utils import jsonify\nfrom bson.objectid import ObjectId\n\n\nproduct_bp = Blueprint('product')\n\n\n@product_bp.get('/api/v1/products')\nasync def get_products(request):\n \"\"\"获取商品列表\"\"\"\n results = [record async for record in Product.find({'flag': True})]\n return json(jsonify({'products': results, 'total': len(results)}))\n\n\n@product_bp.get('/api/v1/all_products')\nasync def get_all_products(request):\n \"\"\"获取所有商品列表\"\"\"\n products = [record async for record in Product.find({})]\n for i in products:\n label = await Label.find_one({'_id': ObjectId(i.get('label', None))})\n i.update({'label': label['name']})\n return json(jsonify({'products': products, 'total': await Product.count_documents({})}))\n\n\n@product_bp.post('/api/v1/order/')\nasync def post_order(request, user_id):\n \"\"\"提交商品订单\"\"\"\n if not request.json:\n raise InvalidUsage('not json request!')\n product = request.json.get('product')\n num = request.json.get('num')\n user_id = user_id\n order = await Order.create(product, user_id, num)\n return json(jsonify({'id': order.inserted_id}))\n\n\n@product_bp.get('/api/v1/user_orders')\nasync def get_all_order(request):\n \"\"\"获取某个用户所有订单\"\"\"\n user_id = request.json.get('id')\n orders = [record async for record in Order.find({'id': ObjectId(user_id)})]\n return json(jsonify({'products': orders, 'total': await Product.count_documents({})}))\n\n\n@product_bp.get('/api/v1/alll_orders')\nasync def get_all_order(request):\n \"\"\"获取用户所有订单\"\"\"\n user_id = request.json.get('id')\n orders = [record async for record in Order.find({})]\n return json(jsonify({'products': orders, 'total': await Product.count_documents({})}))\n\n@product_bp.put('/api/v1/loadProduct')\nasync def update_product(request):\n \"\"\"商品上下架\"\"\"\n if not request.json:\n raise InvalidUsage('not json request!')\n flag = request.json.get('flag')\n name = request.json.get('name')\n logging.info(flag)\n result = await Product.update_one({'name': name}, {'$set': {'flag': flag}})\n logging.info(result)\n return json(jsonify({'_id': flag}))\n\n\n@product_bp.put('/api/v1/product')\nasync def update_product(request):\n \"\"\"商品更新\"\"\"\n if not request.json:\n raise InvalidUsage('not json request!')\n name = request.json.get('name')\n stock = request.json.get('stock')\n price = request.json.get('price')\n label = request.json.get('label')\n product = await Product.update_one(\n {'_id': id},\n {'$set':\n {\n 'name': name,\n 'stock': stock,\n 'price': price,\n 'label': label\n }\n }\n )\n return json(jsonify({'id': product.inserted_id}))\n\n\n@product_bp.delete('/api/v1/product')\nasync def delete_product(request):\n if not request.json:\n raise InvalidUsage('not json request!')\n name = request.json.get('name')\n await Product.delete_one({'name': name})\n return json(jsonify({'name': name}))\n\n\n@product_bp.post('/api/v1/product')\nasync def post_product(request):\n \"\"\"商品录入\"\"\"\n if not request.json:\n raise InvalidUsage('not json request!')\n name = request.json.get('name')\n stock = request.json.get('stock')\n price = request.json.get('price')\n label = request.json.get('label')\n product = await Product.create(name, stock, price, label)\n return json(jsonify({'id': product.inserted_id}))\n\n\n@product_bp.get('/api/v1/labels')\nasync def get_label(request):\n \"\"\"获取商品类型\"\"\"\n labels = [record async for record in Label.find({})]\n return json(jsonify({'labels': labels}))\n\n\n@product_bp.post('api/v1/label')\nasync def add_label(request):\n \"\"\"添加商品类型\"\"\"\n name = request.json.get('name')\n # await Label.delete_many({'name': name})\n result = await Label.find_one({'name': name})\n logging.info(result)\n logging.info(not result)\n if not result:\n await Label.create(name)\n return json(jsonify({'result': 'success'}))\n raise InvalidUsage('分类已存在')\n","sub_path":"src/tao/web/api/products.py","file_name":"products.py","file_ext":"py","file_size_in_byte":4334,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"310964626","text":"\n\"\"\"\nThis is a main script to reproduce results in the paper 'Predicting United States policy outcomes\nwith Random Forests', by Shawn McGuire and Charles Delahunt.\n \nIt run a single setup (ie a hyperparameter and feature set combo). It will loop through each Policy\nDomain that is specified (or just once, if all Policy Domains combined is specified). For each\nPolicy Domain, it will train a Random Forest (or xgBoost) model, and optionally a Logistic model\nand a Neural Net model, on the cases within that Policy Domain. For each model, it repeats the\ntraining for the specified number of runs to collect accuracy statistics. Many plots and console\nprintouts are optional. Results are saved to .csv file. \n \nCopyright (c) 2020 Charles B. Delahunt. delahunt@uw.edu\nMIT License\n\n\"\"\" \n\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport xgboost as xgb\nfrom sklearn import metrics\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.linear_model import LinearRegression\nfrom sklearn.neural_network import MLPClassifier\nfrom sklearn import datasets\n\n# import support functions:\nfrom support_functions import plotFeatureStatsFunction, generateSetupStringFunction, \\\ngenerateTrainTestSplitFunction, parseDataframeFunction, calculateAccuracyStatsOnTestSetFunction, \\\ncalculatePrintAndSaveModelStatsOverAllRunsFunction\n\n\"\"\"\nNote re plot window location:\n1. To print plots in separate windows, run the following on the command line:\n%matplotlib qt\n2. To revert back to plots appearing in console:\n%matplotlib inline\n\"\"\"\n#%%\n\n\"\"\"\nUSER ENTRIES:\nThese are divided into entries you might modify, and entries (created for various experiments) that\nyou are unlikely to modify. So in most cases only the first set of entries will be relevant.\n\"\"\"\n\n''' 1. Parameters that might be modified: '''\nnumRuns = 3 # train this many models, to get accuracy statistics\nsaveResultsStub = 'resultsOneSetup_27july2020' # becomes part of filename\n\nstartYearForTest = 2005 # 2005 -> random train/test split from all data.\n# 1997 -> test set is all samples >= 1997 (retrodiction)\nrunLogisticRegressionFlag = True # True -> run logistic regression (as in Gilen's model) in\n# addition to Random Forest. Random Forest is always run.\noneModelPerPolicyDomainFlag = False # To select most salient IGs for each policy domain, use True.\n#--------------------------------------------\n# Interest group features:\nuseIGNAFlag = False # True -> use netIGA as feature. False -> do not use.\n# Some of the next few conditions have logical dependencies, eg they are mutually exclusive. Such\n# cases are called out before training begins.\nuseIndividualInterestGroupsFlag = True # Must be True for the next three flags to be relevant.\nuseAllInterestGroupsFlag = False\nuseRfChosenInterestGroupsFlag = True # True -> use RF-chosen, False -> use\n# logistic-chosen.\nuseShortInterestGroupListFlag = False # True -> use short list, False AND useAllIGsList == False ->\n# use medium list.\n# A list of Interest Groups and their column numbers are given at the end of this script.\n#-------------------------------------\n# policy domain feature:\nusePolicyDomainsFlag = False # True -> use policy domains as features.\n# NOTE: this script does not allow for using Policy Areas as features (this is\n# an artifact, and could be changed).\n#---------------------------------------------\n# Economic class features:\nuse90PercentFlag = True\nuse50PercentFlag = False\nuse10PercentFlag = False\nuse90Minus10PercentFlag = False\n#----------------------------------------------------\n# Output and plotting flags:\nprintDifferenceInRfChosenVsLogChosenFeaturesFlag = True # Table 4 in paper. NOTE: To reproduce\n# these results, use ONLY IGs as features, ie set both usePolicyDomainsFlag and\n# use90PercentFlag = False.\nshowFeaturePlotsFlag = True # Happens once per setup. This prints to console (and text file) RF\n# feature importance, IG-outcome correlations, and IG at-bats, and shows a bar plot of RF feature\n# importance. To see the most salient IGs for each policy domain, set this = True and\n# oneModelPerPolicyDomainFlag = True.\nnumFeaturesToPlot = 44 # 44 -> all IGs and also P90.\nshowAccuracyPlotsFlag = True # Plots accuracy results of each run separately. So if numRuns > 1,\n# set = False to prevent figure overload.\n\n# Outputs less likely to be used:\nprintToConsoleFlag = True # Only used in plotResultsFunction( )\nprintLogisticBetasFlag = True # To examine which features logistic regression weights.\nshowFeatureImportanceVsAtBatsScatterplotFlag = False # Print scatterplots of importance score vs\n# at-bats. This is voided if 'printLogisticBetasFlag' == False.\n\nprintTestCasesPerRunFlag = True\nprintExtraSetupDetailsFlag = True\nplotAccuraciesByThresholdsFlag = False # ! LOTS OF FIGURES (numRuns x numSetups x numModelTypes)\n# ---------- END OF ENTRIES MOST LIKELY TO BE MODIFIED -----------------------\n#%%\n''' 2. Parameters that can likely be left alone: '''\n\nlogisticPenalty = 'l2' # 'none', 'l1', 'l2'. 'l2' is the default. l1 and l2 return similar top\n# features and scores. 'none' returns wacky feature rankings. So sticking with 'l2' is good.\n\nuseBalancedModelFlag = True # whether to use balanced mode when training.\n\n# Choose whether to use Permutation or Gini impurity for RF feature selection:\nusePermutationFeatureRankingFlag = False # False -> use Gini impurity for choosing RF features.\n# True -> use permutation method for choosing RF features, and substitute these IGs in for the\n# usual RF-chosen SHORT list (ie only a short list is defined for permutation features). Gini\n# features work best.\nuseJointInsteadOfRFIGListsFlag = False # A one-off, can be ignored. Using combined (union of\n# rf-chosen and log-chosen) features gives worse results than RF-chosen features.\n\ntrainFractionDefault = 0.67\nignoreGunCasesFlag = False\nignoreMiscCasesFlag = False # these are cases that do not fit into the five main policy domains\n# (ie 'Miscellaneous').\nxgBoostFlag = False # True -> use xgBoost instead of RF. False -> standard RF.\nrunNeuralNetFlag = False # True -> train a NN in addition to a RF.\n\n# Flags to focus on cases where rich and poor disagree:\ndisagreementThreshold = 0.05 # To distinguish when 90 and 10 really disagree.\nrestrictTrainToHighDisagreementFlag = False # True -> only train on cases where |90-10| > 0.05.\nrestrictTestToHighDisagreementFlag = False # True -> only test on cases where |90-10| > 0.05.\n\n# Flags about how thresholds (for test set use) are picked on training sets:\nuseSimpleMaxForThreshFlag = True # False -> use a Bayesian-flavored version. See def\n# calculateAccuracyStatsOnTestSetFunction( )\nuncertaintyValue = 0.02 # used when 'useSimpleMaxForThreshFlag' = False.\n\nweightFeatureOutcomeCorrelationsByAtBatsOnly = True # True -> feature correlation with outcomes is\n# normalized by number of at-bats, not number of total cases.\nnumInLinspace = 101 # For roc and accuracy curves.\n#-------------------------------------------------------------------\n# Specialized model parameters (for NN, RF, and xgBoost):\n# RF AND xgboost parameters:\nmaxDepth = 4 # In general try 4. Try 13 when using all interest\n# groups on all data (ie one model for combined domains).\nnumTrees = 200\n# XG Boost parameters:\nnumRoundsBoosting = 50 # Number of rounds of boosting.\netaForBoosting = 1\n# NeuralNet params:\nhiddenLayers = (50, 50) # Try (50,50) for all cases.\n\n#---------------------------------------------------------------\n''' 3. Interest group subsets, that were determined to be most salient by\ntraining models and doing feature selection on individual Policy Domains.\nNote: Larger gaps between entries in a dict value denote different tiers of\nimportance (eg most important group vs 2nd most important group). '''\n\n# 1. RF-chosen via Gini impurity:\n# Medium subset:\nrfChosenInterestGroupsToKeepList = \\\n{'ECYN':[36, 38, 51, 33], 'FPYN':[29, 20, 52, 24, 46, 14, 33, 13],\n 'SWYN':[12, 49, 42, 47, 53, 40, 31],'RLYN':[45, 26, 22, 40, 39, 51],\n 'GNYN':[44], 'MISC':[25, 13, 24, 18, 46, 23, 26]}\n# Smaller subset:\nrfChosenInterestGroupsToKeepListShort = \\\n{'ECYN':[36, 38], 'FPYN':[29, 20, 52, 24], 'SWYN':[12, 49],'RLYN':[26, 45, 22],\n 'GNYN':[44], 'MISC':[25]}\n# 2. RF-chosen via Permutation method. The permutation method was unstable, giving different feature\n# sets depending on whether P90 was included as a feature. There are two choices, one of which is\n# commented out.\n# Either using only IGs (no P90):\nrfPermutationChosenIGsToKeepListShort = \\\n{'ECYN':[37, 36, 33],'FPYN':[29, 20], 'SWYN':[12, 26, 47],'RLYN':[45],\n 'GNYN':[44], 'MISC':[23, 25, 46]}\n# Or using IGs and P90:\n# rfPermutationChosenIGsToKeepListShort = \\\n#{'ECYN':[37, 25, 19, 47], 'FPYN':[29, 20, 13], 'SWYN':[12, 49 ,26], 'RLYN':[45], 'GNYN':[],\n# 'MISC':[23, 46]}\n# 3. Logistic-chosen:\n# Medium subset:\nlogisticChosenInterestGroupsToKeepList = \\\n{'ECYN':[47, 37, 31, 32], 'FPYN':[37, 46, 29, 52, 14, 24], 'SWYN':[49, 23, 13, 54, 15, 36, 40],\n 'RLYN':[45, 31, 16, 34, 42, 26], 'GNYN':[44], 'MISC':[23, 35, 36, 17, 51, 45,44, 22]}\n# Smaller subset:\nlogisticChosenInterestGroupsToKeepListShort = \\\n{'ECYN':[47, 37],'FPYN':[37, 46, 29, 52], 'SWYN':[49, 23, 13], 'RLYN':[45, 31, 16], 'GNYN':[44],\n 'MISC':[23, 35]}\n# 4. Combined RF- and logistic-chosen IGs (gives significantly worse results than RF-chosen):\n# Medium subset:\njointInterestGroupsToKeepList = \\\n{'ECYN':[36, 38, 51, 33, 47, 37, 31, 32], 'FPYN':[29, 20, 52, 24, 46, 14, 33, 13, 37, 24],\n 'SWYN':[12, 49, 42, 47, 53, 40, 31, 23, 13, 54, 15, 36],\n 'RLYN':[45, 26, 22, 40, 39, 51, 31, 16, 34, 42], 'GNYN':[44],\n 'MISC':[25, 13, 24, 18, 46, 23, 26, 35, 36, 17, 51, 45, 44, 22]}\n# Short subset:\njointInterestGroupsToKeepListShort = \\\n{'ECYN':[36, 38, 47, 37],'FPYN':[29, 20, 52, 24, 37, 46], 'SWYN':[12, 49, 23, 13],\n 'RLYN':[26, 45, 22, 45, 31, 16], 'GNYN':[44], 'MISC':[25, 23]}\n \ndataFolder = 'data/'\n\n\"\"\" END USER ENTRIES \"\"\"\n\n#%%\n\"\"\" MAIN: \"\"\"\n\n''' First some preparation:'''\n\n# Load the main dataset:\ndataFull = pd.read_csv(dataFolder + 'gilens_data_sm_copy.csv')\ndataFull = dataFull.iloc[0:1836] # All rows after 1835 are nan.\n\npolicyDomainList = ['ECYN', 'FPYN', 'SWYN', 'RLYN', 'GNYN', 'MISC'] # MISC is defined a couple\n# lines down. It contains all cases not in the 5 defined policy groups.\nallInterestGroupIndices = np.arange(12, 55) # Used if selecting a subset of IGs.\n\n# Add a column to the dataframe to mark cases in MISC:\ndataFull['MISC'] = np.round(-1*(dataFull.GNYN + dataFull.FPYN + dataFull.SWYN\n + dataFull.RLYN + dataFull.ECYN - 1))\n# Make a dataframe for saving results:\ncolumnNames = ['modelType', 'balanced', 'policyDomain', 'startYearForTest', 'use90', 'use50',\n 'use10', 'use90Minus10', 'useIGNA', 'useIndividualIGs', 'useAllIGs',\n 'useRfChosenMediumIGs', 'useRfChosenShortIGs', 'useLogChosenMediumIGs',\n 'useLogChosenShortIGs', 'usePolicyDomains', 'useOnly90Minus10Disagreements',\n 'trainFraction', 'disagreementThreshold', 'maxDepth', 'hiddenLayers',\n 'IGSubsetUsed', 'numRuns', 'rawAccTestMeans', 'rawAccTestMedians', 'rawAccTestStds',\n 'balAccTestMeans', 'balAccTestMedians', 'balAccStds']\nresultsDataFrame = pd.DataFrame(columns=columnNames)\n\n# Complete the names of the files to save:\nbalStr = 'UNbalanced'\nif useBalancedModelFlag:\n balStr = 'balanced'\ncombinedStr = 'combinedDomains'\nif oneModelPerPolicyDomainFlag:\n combinedStr = 'oneModelPerDomain'\nyearStr = str(startYearForTest)\n\nif useJointInsteadOfRFIGListsFlag:\n rfChosenInterestGroupsToKeepList = jointInterestGroupsToKeepList\n rfChosenInterestGroupsToKeepListShort = jointInterestGroupsToKeepListShort\nif usePermutationFeatureRankingFlag:\n rfChosenInterestGroupsToKeepListShort = \\\n rfPermutationChosenIGsToKeepListShort # only SHORT list is modified\n\nsaveDataframeResultsFilename = \\\n'resultsDataframe_' + combinedStr + '_' + balStr + '_' + yearStr + '_' \\\n+ saveResultsStub + '.csv' # save results to dataframe\nsaveStringResultsFilename = \\\n'resultsAsStrings_' + combinedStr + '_' + balStr + '_' + yearStr + '_' \\\n+ saveResultsStub + '.txt' # save results as textfile (same as console output)\n\nfid = open(saveStringResultsFilename, 'a')\n\n# correct a dependent flag if necessary:\nif oneModelPerPolicyDomainFlag:\n usePolicyDomainsList = [False]\n\n# Initialize arrays to hold model results from each setup. The second index should match or exceed\n# the number of setups being run:\nrfRawAccArray = np.zeros([numRuns, 4, len(policyDomainList)])\nrfBalAccArray = np.zeros([numRuns, 4, len(policyDomainList)])\nrfAucArray = np.zeros([numRuns, 4, len(policyDomainList)])\nlogRawAccArray = np.zeros([numRuns, 4, len(policyDomainList)])\nlogBalAccArray = np.zeros([numRuns, 4, len(policyDomainList)])\nlogAucArray = np.zeros([numRuns, 4, len(policyDomainList)])\n\n# Some setup conditions cause duplication due to logical dependencies. List some conditions to\n# trigger skipping such setups:\ncondition1 = useAllInterestGroupsFlag and (useRfChosenInterestGroupsFlag\n or useShortInterestGroupListFlag)\n# Using all interest groups is mutually exclusive with using only subsets.\ncondition2 = oneModelPerPolicyDomainFlag and usePolicyDomainsFlag # These are mutually exclusive\n# conditions.\ncondition3 = not useIndividualInterestGroupsFlag \\\nand (useAllInterestGroupsFlag or useRfChosenInterestGroupsFlag\n or useShortInterestGroupListFlag) # The three conditions in parentheses can be active only if\n# 'useIndividualInterestGroupsFlag' == True.\nif condition1:\n print('Note: Use either all interest groups or just subsets as features, but not both.')\nif condition2:\n print('Note: Running one model per policy domain makes Policy Domains a meaningless' \\\n + ' feature, though resulting models are still valid.')\nif condition3:\n print(\" Note: 'useIndividualInterestGroupsFlag' must be True for other interest group flags\"\n + \" to be meaningful. \")\n\n#%%\n# Define some variables:\n\n# 1. use the specified interest group list:\nif useRfChosenInterestGroupsFlag:\n if useShortInterestGroupListFlag:\n interestGroupsToKeepList = rfChosenInterestGroupsToKeepListShort\n else:\n interestGroupsToKeepList = rfChosenInterestGroupsToKeepList\nelse:\n if useShortInterestGroupListFlag:\n interestGroupsToKeepList = logisticChosenInterestGroupsToKeepListShort\n else:\n interestGroupsToKeepList = logisticChosenInterestGroupsToKeepList\n\nif useBalancedModelFlag:\n balanceStr = 'balanced'\nelse:\n balanceStr = None\n\nif startYearForTest < 2003: # Case: retrodiction\n trainFraction = 1 # Use all pre-cutoff date cases for training\nelse:\n trainFraction = trainFractionDefault # < 1, to divide the full dataset into Train/Test.\n\n# 2. Book-keeping: combine all these interest groups for use in one big model:\na = list(interestGroupsToKeepList.values())\nb = []\nfor i in a:\n for j in i:\n b.append(j)\ninterestGroupsToKeepGroupedByDomain = b\ncombinedInterestGroupsToKeep = np.unique(b)\nif useAllInterestGroupsFlag:\n combinedInterestGroupsToKeep = allInterestGroupIndices\nif not useIndividualInterestGroupsFlag:\n combinedInterestGroupsToKeep = []\n#%%\n# store setup params for future use:\nsetupParams = \\\n {'numRuns':numRuns, 'startYearForTest':startYearForTest, 'trainFraction':trainFraction,\n 'oneModelPerPolicyDomainFlag':oneModelPerPolicyDomainFlag, 'ignoreGunCasesFlag':ignoreGunCasesFlag,\n 'ignoreMiscCasesFlag':ignoreMiscCasesFlag, 'xgBoostFlag':xgBoostFlag, 'usePolicyDomainsFlag':\n usePolicyDomainsFlag, 'use90PercentFlag':use90PercentFlag, 'use50PercentFlag':use50PercentFlag,\n 'use10PercentFlag':use10PercentFlag, 'use90Minus10PercentFlag':use90Minus10PercentFlag,\n 'useIndividualInterestGroupsFlag':useIndividualInterestGroupsFlag, 'useAllInterestGroupsFlag':\n useAllInterestGroupsFlag, 'useIGNAFlag':useIGNAFlag, 'restrictTrainToHighDisagreementFlag':\n restrictTrainToHighDisagreementFlag, 'restrictTestToHighDisagreementFlag':\n restrictTestToHighDisagreementFlag, 'maxDepth':maxDepth, 'numTrees':numTrees, 'numRoundsBoosting':\n numRoundsBoosting, 'etaForBoosting':etaForBoosting, 'hiddenLayers':hiddenLayers,\n 'weightFeatureOutcomeCorrelationsByAtBatsOnly':weightFeatureOutcomeCorrelationsByAtBatsOnly,\n 'useBalancedModelFlag':useBalancedModelFlag, 'useSimpleMaxForThreshFlag':useSimpleMaxForThreshFlag,\n 'uncertaintyValue':uncertaintyValue, 'useRfChosenInterestGroupsFlag':useRfChosenInterestGroupsFlag,\n 'useShortInterestGroupListFlag':useShortInterestGroupListFlag, 'disagreementThreshold':\n disagreementThreshold, 'policyDomainList':policyDomainList, 'showAccuracyPlotsFlag':\n showAccuracyPlotsFlag, 'printToConsoleFlag':printToConsoleFlag, 'plotAccuraciesByThresholdsFlag':\n plotAccuraciesByThresholdsFlag, 'useOnly90Minus10Disagreements':restrictTestToHighDisagreementFlag,\n 'allInterestGroupIndices': allInterestGroupIndices, 'interestGroupsToKeepGroupedByDomain':\n interestGroupsToKeepGroupedByDomain, 'combinedInterestGroupsToKeep':combinedInterestGroupsToKeep}\n\n#%%\n# print out a summary of setup details for the model about to run:\nsetupStr = generateSetupStringFunction(setupParams)\nprint()\nprint(setupStr)\nfid.write('' + '\\n')\nfid.write(setupStr + '\\n')\n\n#%%\n# Prep for looping through models (either one model for combined policy domains\n# or one model per policy domain):\nif oneModelPerPolicyDomainFlag:\n numToLoop = len(policyDomainList) # To prevent repeated plots\nelse:\n numToLoop = 1 # If one model only for all domains\n\n# Initialize arrays to store results from all runs and all models:\na = np.zeros((numRuns, numToLoop))\nrandomForestBalAcc = np.copy(a)\nrandomForestRawAcc = np.copy(a)\nrandomForestAuc = np.copy(a)\nrandomForestSens = np.copy(a)\nrandomForestSpec = np.copy(a)\nrandomForestPrecision = np.copy(a)\nrandomForestRecall = np.copy(a)\nrandomForestF1 = np.copy(a)\nxgBoostBalAcc = np.copy(a)\nxgBoostRawAcc = np.copy(a)\nxgBoostAuc = np.copy(a)\nxgBoostSens = np.copy(a)\nxgBoostSpec = np.copy(a)\nxgBoostPrecision = np.copy(a)\nxgBoostRecall = np.copy(a)\nxgBoostF1 = np.copy(a)\nxgBoostF1Raw = np.copy(a)\nlogisticBalAcc = np.copy(a)\nlogisticRawAcc = np.copy(a)\nlogisticAuc = np.copy(a)\nlogisticSens = np.copy(a)\nlogisticSpec = np.copy(a)\nlogisticPrecision = np.copy(a)\nlogisticRecall = np.copy(a)\nlogisticF1 = np.copy(a)\nnNetBalAcc = np.copy(a)\nnNetRawAcc = np.copy(a)\nnNetAuc = np.copy(a)\nnNetSens = np.copy(a)\nnNetSpec = np.copy(a)\nnNetPrecision = np.copy(a)\nnNetRecall = np.copy(a)\nnNetF1 = np.copy(a)\nacceptTestResultsMatrix = np.ones(a.shape, dtype=bool) # To mark invalid train/test splits\n\n#%%\n# If this is the first Setup, generate train/test splits for each run-ind pair. For each run, the\n# same split will get used for all model types.\ndataTrainBooleanArray = \\\n[[None for j in range(numToLoop)] for i in range(numRuns)] # Makes an array numRuns x numToLoop.\ndataTestBooleanArray = \\\n[[None for j in range(numToLoop)] for i in range(numRuns)]\n\nfor ind in range(numToLoop):\n for run in range(numRuns):\n if oneModelPerPolicyDomainFlag:\n policyDomainsToKeep = policyDomainList[ind]\n else:\n policyDomainsToKeep = policyDomainList\n setupParams['policyDomainsToKeep'] = policyDomainsToKeep # Temporary, to make the\n # domain-specific train/test splits here. We reassign this later.\n dataTrainBoolean, dataTestBoolean = \\\n generateTrainTestSplitFunction(dataFull, setupParams)\n dataTrainBooleanArray[run][ind] = dataTrainBoolean\n dataTestBooleanArray[run][ind] = dataTestBoolean\n\n#%%\n''' Preparation is done. Now run the model(s):'''\n\nfor ind in range(numToLoop): \n # 'ind' indexes the policy domains we are running over. This is *not* the Setup, which is the \n # selection of all parameters including features). Either ind = 0 (if one model for policy \n # combined domains) or ind = 0 (economic), 1 (foreign), 2 (social welfare), 3 (religion),\n # 4 (guns), or 5 (misc). Note that for each 'ind', we may train multiple models (certainly\n # RF or xgBoost, and maybe Logistic and NN).\n\n # Initialize some matrices for this model:\n featureScoresAllRuns = np.zeros([100, numRuns]) # Overshoot number of features (since currently\n # unknown) when initializing.\n featureOutcomeCorrelationsAllRuns = np.zeros([100, numRuns])\n lregCoeffs = np.zeros([100, numRuns]) # To hold logistic beta coeffs.\n numAtBatsAllRuns = np.zeros([100, numRuns])\n numTrainCasesAllRuns = np.zeros([numRuns, 1])\n numTestCasesAllRuns = np.zeros([numRuns, 1])\n acceptTestResultsAllRuns = np.ones([numRuns, 1], dtype=bool) # This will mark whether a run had\n # Both pass and fail cases in the test set. Make a flag for extracting the cases in this\n # domain. Do this here since it applies to all runs:\n if oneModelPerPolicyDomainFlag:\n policyDomainsToKeep = policyDomainList[ind] # Case: one policy domain.\n else:\n policyDomainsToKeep = str(policyDomainList) # Only for print to console.\n# policyDomainsToKeep = policyDomainsToKeep.replace('[','')\n# policyDomainsToKeep = policyDomainsToKeep.replace(']','')\n# policyDomainsToKeep = policyDomainsToKeep.replace(\"'\",'')\n\n # Define the set of interest groups to use as features:\n if useAllInterestGroupsFlag:\n interestGroupsToKeep = allInterestGroupIndices\n else: # Use a strict subset of interest groups:\n if oneModelPerPolicyDomainFlag:\n interestGroupsToKeep = interestGroupsToKeepList[policyDomainsToKeep]\n else: # Case: only one model, so combine all relevant interest groups:\n interestGroupsToKeep = combinedInterestGroupsToKeep\n if not useIndividualInterestGroupsFlag:\n interestGroupsToKeep = []\n setupParams['policyDomainsToKeep'] = policyDomainsToKeep\n setupParams['interestGroupsToKeep'] = interestGroupsToKeep\n\n #%% Do many runs and save results to the above arrays:\n for run in range(numRuns):\n #%%\n \"\"\" Load and parse data\"\"\"\n\n # load the relevant train/test split:\n dataTrainBoolean = dataTrainBooleanArray[run][ind]\n dataTestBoolean = dataTestBooleanArray[run][ind]\n#\n # Create a dataframe with the correct columns:\n data = parseDataframeFunction(dataFull, setupParams)\n\n # Create feature and label arrays:\n dataTrain = data.loc[dataTrainBoolean]\n dataTest = data.loc[dataTestBoolean]\n\n train_features = dataTrain.drop('Binary Outcome', axis=1).values\n train_labels = dataTrain['Binary Outcome'].values\n test_features = dataTest.drop('Binary Outcome', axis=1).values\n test_labels = dataTest['Binary Outcome'].values\n\n # misc book-keeping:\n temp = dataTrain.drop('Binary Outcome', axis=1) # Used as an intermediate var here; has to\n # be a dataframe\n featureNames = list(temp.head(0))\n numFeaturesToPlot = min(numFeaturesToPlot, len(featureNames))\n # Reduce this param value if the original value exceeds the total number of features.\n\n # Check num of cases of each type:\n numPosTestCases = np.sum(test_labels == 1)\n numNegTestCases = np.sum(test_labels == 0)\n numNegTrainCases = np.sum(train_labels == 0)\n numPosTrainCases = np.sum(train_labels == 1)\n numTestCasesAllRuns[run] = numPosTestCases + numNegTestCases\n acceptTestResultsAllRuns[run] = numPosTestCases > 0 and numNegTestCases > 0 # False if\n # there are 0 pos or 0 neg test cases.\n\n # Print warnings re bad train/test splits:\n if numPosTestCases == 0:\n print(policyDomainsToKeep + ', run ' + str(run)\n + ' Test has 0 pos cases. Excluding this run.')\n fid.write(policyDomainsToKeep + ', run ' + str(run)\n + ' Test has 0 pos cases. Excluding this run.' + '\\n')\n if numNegTestCases == 0:\n print(policyDomainsToKeep + ', run ' + str(run)\n + ' Test has 0 neg cases. Excluding this run.')\n fid.write(policyDomainsToKeep + ', run ' + str(run)\n + ' Test has 0 neg cases. Excluding this run.' + '\\n')\n\n # Record number of at-bats in test set, per interest group:\n for i in range(len(featureNames)):\n if 'pred' in featureNames[i]:\n numAtBatsAllRuns[i, run] = \\\n np.sum(np.abs(dataTest[featureNames[i]].values - 0.5) > 0.1) # For voter\n # preferences, consider 0.4 to 0.6 as 'neutral'.\n else:\n numAtBatsAllRuns[i, run] = np.sum(dataTest[featureNames[i]].values != 0)\n\n #%%\n ''' Always train a random forest model '''\n\n ''' 1. standard random forest: '''\n if not xgBoostFlag:\n forestType = 'Random forest'\n modelStr = 'randomForest' # Different capitalization is artifact.\n\n # Instantiate model and train\n clf = RandomForestClassifier(n_estimators=numTrees, max_depth=maxDepth,\n class_weight=balanceStr)\n\n # Train the model for use on test set\n clf.fit(train_features, train_labels)\n # Run test set through model:\n testProbScores = clf.predict_proba(test_features) # This returns two columns of\n # probabiliies\n trainProbScores = clf.predict_proba(train_features) # For setting thresholds to apply\n # to test set\n\n # Record feature importances:\n if usePermutationFeatureRankingFlag:\n # CAUTION: This method is very noisy:\n # 1. We'll need the original model balanced acc:\n # a. Either use variable threshold:\n fullModelAccStats = calculateAccuracyStatsOnTestSetFunction \\\n (trainProbScores, testProbScores, train_labels, test_labels, setupParams, modelStr,\n fid)\n # b. Or fixed 0.5 threshold:\n testYhat = clf.predict(test_features)\n tp = np.sum(np.logical_and(testYhat == 1, test_labels == 1))\n fn = np.sum(np.logical_and(testYhat == 0, test_labels == 1))\n tn = np.sum(np.logical_and(testYhat == 0, test_labels == 0))\n fp = np.sum(np.logical_and(testYhat == 1, test_labels == 0))\n fullTestBalAcc = 0.5*(tp/(tp+fn) + tn/(tn + fp))\n\n # 2. Now mix up the values of each feature in turn, then calculate the loss in\n # accuracy:\n for i in range(len(featureNames)):\n X = test_features.copy()\n np.random.shuffle(X[:, i])\n # goes with (a):\n thisPermutedSetProbScores = clf.predict_proba(X)\n thisAccStats = \\\n calculateAccuracyStatsOnTestSetFunction(trainProbScores,\n thisPermutedSetProbScores, train_labels,\n test_labels, setupParams, modelStr,\n fid)\n featureScoresAllRuns[i, run] = \\\n (fullModelAccStats['balancedAccTest'] - thisAccStats['balancedAccTest']) \\\n / fullModelAccStats['balancedAccTest']\n\n else: # Just use the built-in gini impurity\n featureScoresAllRuns[range(len(featureNames)), run] = clf.feature_importances_\n\n # Eliminate surplus rows:\n featureScoresAllRuns = featureScoresAllRuns[range(len(featureNames)), :]\n\n ''' 2. xgboost: '''\n if xgBoostFlag:\n forestType = 'xgBoost'\n modelStr = 'xgBoost'\n\n datasets.dump_svmlight_file(train_features, train_labels, 'xgbTrain.txt')\n xgbTrain = xgb.DMatrix('xgbTrain.txt', silent=True)\n datasets.dump_svmlight_file(test_features, test_labels, 'xgbTest.txt')\n xgbTest = xgb.DMatrix('xgbTest.txt', silent=True)\n\n # define cost:\n if useBalancedModelFlag:\n scale_pos_weight = numNegTrainCases/numPosTrainCases\n else:\n scale_pos_weight = 1\n\n xgbParam = {'max_depth':maxDepth, 'eta':etaForBoosting, 'objective':'binary:logistic',\n 'booster':'gblinear', 'scale_pos_weight': scale_pos_weight}\n bst = xgb.train(xgbParam, xgbTrain, numRoundsBoosting)\n # make predictions:\n xgbTestOutput = bst.predict(xgbTest)\n testProbScores = np.zeros((len(xgbTestOutput), 2))\n testProbScores[:, 1] = xgbTestOutput\n testProbScores[:, 0] = 1 - xgbTestOutput\n # predictions on training set, to find a threshold:\n xgbTrainOutput = bst.predict(xgbTrain)\n trainProbScores = np.zeros((len(xgbTrainOutput), 2))\n trainProbScores[:, 1] = xgbTrainOutput\n trainProbScores[:, 0] = 1 - xgbTrainOutput\n # End of xgBoost-specific section.\n\n # Calculate test set accuracy: ('aS' = accuracy stats). This function includes possible roc,\n # balAcc, sens, spec plots.\n aS = calculateAccuracyStatsOnTestSetFunction(trainProbScores, testProbScores, train_labels,\n test_labels, setupParams, modelStr, fid)\n\n # Save to numRuns array:\n if xgBoostFlag:\n xgBoostBalAcc[run, ind] = aS['balancedAccTest']\n xgBoostAuc[run, ind] = aS['aucScore']\n xgBoostSens[run, ind] = aS['sensTest']\n xgBoostSpec[run, ind] = aS['specTest']\n xgBoostPrecision[run, ind] = aS['precisionTest']\n xgBoostRecall[run, ind] = aS['sensTest']\n xgBoostF1[run, ind] = 2*aS['sensTest']*aS['specTest'] \\\n / (aS['sensTest'] + aS['specTest'])\n xgBoostRawAcc[run, ind] = aS['rawAccTest']\n else:\n randomForestBalAcc[run, ind] = aS['balancedAccTest']\n randomForestAuc[run, ind] = aS['aucScore']\n randomForestSens[run, ind] = aS['sensTest']\n randomForestSpec[run, ind] = aS['specTest']\n randomForestPrecision[run, ind] = aS['precisionTest']\n randomForestRecall[run, ind] = aS['sensTest']\n randomForestF1[run, ind] = 2*aS['sensTest']*aS['specTest'] \\\n / (aS['sensTest'] + aS['specTest'])\n randomForestRawAcc[run, ind] = aS['rawAccTest']\n\n #%%\n ''' Train a Neural Net '''\n\n if runNeuralNetFlag:\n neuralNet = MLPClassifier(hiddenLayers, activation='relu', max_iter=5000,\n early_stopping=True)\n neuralNet = neuralNet.fit(train_features, train_labels)\n testProbScoresNN = neuralNet.predict_proba(test_features)\n trainProbScoresNN = neuralNet.predict_proba(train_features) # For setting thresholds to\n # apply to test set\n\n # Calculate test set accuracy: ('aS' = accuracy stats). Includes possible roc, balAcc,\n # sens, spec plots.\n aS = calculateAccuracyStatsOnTestSetFunction(trainProbScoresNN, testProbScoresNN,\n train_labels, test_labels, setupParams,\n 'neuralNet', fid)\n\n # save to numRuns array:\n nNetBalAcc[run, ind] = aS['balancedAccTest']\n nNetRawAcc[run, ind] = aS['rawAccTest']\n nNetAuc[run, ind] = aS['aucScore']\n nNetSens[run, ind] = aS['sensTest']\n nNetSpec[run, ind] = aS['specTest']\n nNetPrecision[run, ind] = aS['precisionTest']\n nNetRecall[run, ind] = aS['sensTest']\n nNetF1[run, ind] = 2*aS['sensTest']*aS['specTest'] \\\n / (aS['sensTest'] + aS['specTest'])\n\n #%%\n ''' Train a logistic regression model (ie Gilens method) '''\n if runLogisticRegressionFlag:\n # Instantiate and train model:\n if logisticPenalty == 'none':\n lreg = LogisticRegression(class_weight=balanceStr, penalty=logisticPenalty,\n solver='lbfgs') # C = 1, fit_intercept = True\n else:\n lreg = LogisticRegression(class_weight=balanceStr, penalty=logisticPenalty)\n\n # Create a feature scaler and fit it to the training data:\n scaler = StandardScaler()\n trainFeaturesScaled = scaler.fit_transform(train_features)\n testFeaturesScaled = scaler.transform(test_features)\n\n lreg = lreg.fit(trainFeaturesScaled, train_labels)\n\n # Run test set througgh model:\n testProbScoresLreg = lreg.predict_proba(testFeaturesScaled)\n trainProbScoresLreg = lreg.predict_proba(trainFeaturesScaled)\n\n # Calculate test set accuracy: ('aS' = accuracy stats). Includes possible roc, balAcc,\n # sens, spec plots.\n aS = calculateAccuracyStatsOnTestSetFunction(trainProbScoresLreg, testProbScoresLreg,\n train_labels, test_labels, setupParams,\n 'logistic', fid)\n\n # Save to numRuns array:\n logisticBalAcc[run, ind] = aS['balancedAccTest']\n logisticRawAcc[run, ind] = aS['rawAccTest']\n logisticAuc[run, ind] = aS['aucScore']\n logisticSens[run, ind] = aS['sensTest']\n logisticSpec[run, ind] = aS['specTest']\n logisticPrecision[run, ind] = aS['precisionTest']\n logisticRecall[run, ind] = aS['sensTest']\n logisticF1[run, ind] = 2*aS['sensTest']*aS['specTest'] \\\n / (aS['sensTest'] + aS['specTest'])\n\n # Save betas:\n betas = lreg.coef_[0]\n lregCoeffs[range(len(betas)), run] = betas.transpose()\n\n #%%\n # See whether each interest group's correlation with the test set outcomes is positive or\n # negative. Note: this calculation will not be meaningful for policy domain features.\n for i in range(len(featureNames)):\n temp = np.copy(test_labels)\n temp[np.where(temp == 0)[0]] = -1 # Change labels to -1 and 1 to allow inner product\n # with preferences.\n pref = test_features[:, i] # This feature's preferences.\n # For pred90 etc, set the middle region (0.4 to 0.6) to 0 and scale to [-2, 2] to match\n # interest groups.\n if 'pred' in featureNames[i]:\n pref = (pref - 0.5)*4\n pref = np.maximum(pref, -2) # Only needed if multiplier > 4.\n pref = np.minimum(pref, 2)\n pref[np.where(np.logical_and(pref > -0.4, pref < 0.4))[0]] = 0\n numberAtBats = np.sum(pref != 0) # For weighting correlation score by at-bats only\n # rather than all cases\n if weightFeatureOutcomeCorrelationsByAtBatsOnly:\n denom = 2*numberAtBats\n else:\n denom = 2*len(test_labels)\n featureOutcomeCorrelationsAllRuns[i, run] = sum(np.multiply(pref, temp)) / denom\n\n ''' End of loop through numRuns '''\n\n # Save the RF accuracies to the 3-D array: (the second dimension is a carry-over (deadwood here)\n # from the script that loops over many models + feature sets).\n rfRawAccArray[:, 0, ind] = (randomForestRawAcc[:, ind]).flatten()\n rfBalAccArray[:, 0, ind] = (randomForestBalAcc[:, ind]).flatten()\n rfAucArray[:, 0, ind] = (randomForestAuc[:, ind]).flatten()\n # save the logistic accuracies similarly:\n logRawAccArray[:, 0, ind] = (logisticRawAcc[:, ind]).flatten()\n logBalAccArray[:, 0, ind] = (logisticBalAcc[:, ind]).flatten()\n logAucArray[:, 0, ind] = (logisticAuc[:, ind]).flatten()\n\n # Print number of test cases to console:\n testCasesMean = np.round(np.mean(numTestCasesAllRuns), 1)\n testCasesStd = np.round(np.std(numTestCasesAllRuns), 1)\n if printTestCasesPerRunFlag:\n print(str(testCasesMean) + ' +/-' + str(testCasesStd) + ' test cases per run')\n fid.write(str(testCasesMean) + ' +/-' + str(testCasesStd) + ' test cases per run' + '\\n')\n # 3. Feature importance and feature-outcome correlation plots:\n if showFeaturePlotsFlag and not xgBoostFlag: # Because we do not have xgBoost feature scores.\n # Cut off the surplus placeholders in featureScores:\n featureScoresAllRuns = featureScoresAllRuns[range(len(featureNames)), :]\n plotFeatureStatsFunction(featureScoresAllRuns, featureOutcomeCorrelationsAllRuns,\n numAtBatsAllRuns, featureNames, str(policyDomainsToKeep),\n numFeaturesToPlot, testCasesMean, fid)\n\n # 4. Print out logistic betas, mean and std:\n lregCoeffs = lregCoeffs[range(len(featureNames))]\n if runLogisticRegressionFlag and printLogisticBetasFlag:\n betaMeans = np.round(np.mean(lregCoeffs, axis=1), 3)\n betaStds = np.round(np.std(lregCoeffs, axis=1), 3)\n betaFoms = np.round(np.divide(np.abs(betaMeans), betaStds), 3) # Use abs(mean)/std as FoM.\n # remove nans:\n betaFoms[np.where(np.isnan(betaFoms))[0]] = 0\n sortedBetaFomIndices = np.flip(np.argsort(betaFoms))\n sortedBetaFoms = betaFoms[sortedBetaFomIndices]\n print('')\n print('logistic regression betas for ' + str(policyDomainsToKeep) + ':')\n for i in range(numFeaturesToPlot):\n print(str(sortedBetaFoms[i]) + '= mu/sigma, ' + str(betaMeans[sortedBetaFomIndices[i]]) \\\n + ' +/- ' + str(betaStds[sortedBetaFomIndices[i]]) + ' '\n + featureNames[sortedBetaFomIndices[i]])\n fid.write('' + '\\n')\n fid.write('logistic regression betas for ' + str(policyDomainsToKeep) + ':' + '\\n')\n for i in range(numFeaturesToPlot):\n fid.write(str(sortedBetaFoms[i]) + '= mu/sigma, ' \\\n + str(betaMeans[sortedBetaFomIndices[i]]) + ' +/- ' \\\n + str(betaStds[sortedBetaFomIndices[i]]) + ' ' \\\n + featureNames[sortedBetaFomIndices[i]] + '\\n')\n\n # Scatterplot logistic mean feature score and also RF feature score vs number of at-bats for\n # each IG:\n # We need numAtBats:\n atBatMean = np.mean(numAtBatsAllRuns, axis=1)\n atBatMean = atBatMean[:len(betaMeans)]\n # scale the importance scores so they are comparable:\n scaledRfFeatureMeans = np.mean(featureScoresAllRuns, axis=1)\n scaledRfFeatureMeans = scaledRfFeatureMeans / np.max(scaledRfFeatureMeans)\n scaledBetaMeans = np.abs(betaMeans) / np.max(np.abs(betaMeans))\n\n if startYearForTest >= 2003 and not xgBoostFlag: # Since we have no xgBoost feature scores.\n # Calculate linear fits, to plot these:\n r = scaledRfFeatureMeans.reshape(-1, 1) # To make shape [n, 1] instead of [n,]\n lo = scaledBetaMeans.reshape(-1, 1)\n ab = atBatMean.reshape(-1, 1)\n # use only IGs with >0 at-bats:\n nonZeroAtBatInds = np.where(ab > 1)[0]\n r = r[nonZeroAtBatInds]\n lo = lo[nonZeroAtBatInds]\n ab = ab[nonZeroAtBatInds]\n\n lin = LinearRegression()\n lin = lin.fit(ab, r)\n mRF = lin.coef_[0]\n bRF = lin.intercept_[0]\n\n # calculate R^2 scores\n r2RF = metrics.r2_score(r, bRF + mRF*ab)\n r2RF = np.round(r2RF, 2)\n\n lin = lin.fit(ab, lo)\n mLog = lin.coef_[0]\n bLog = lin.intercept_[0]\n # calculate R^2 scores\n r2Log = metrics.r2_score(lo, bLog + mLog*ab)\n r2Log = np.round(r2Log, 2)\n\n if showFeatureImportanceVsAtBatsScatterplotFlag:\n plt.figure()\n plt.plot(atBatMean, scaledBetaMeans, 'r.', markersize=12, label='Logistic')\n plt.plot(atBatMean, scaledRfFeatureMeans, 'b.', markersize=12, label='RF')\n plt.legend(loc='lower right', prop={'size':12, 'weight':'bold'})\n plt.xlabel('# at-bats', fontsize=14, fontweight='bold')\n plt.ylabel('importance score', fontsize=14, fontweight='bold')\n plt.title(str(policyDomainsToKeep) +' feature importances, ' + '. RF R^2 = ' \\\n + str(r2RF) + ', logistic R^2 = ' + str(r2Log),\n fontsize=14, fontweight='bold')\n plt.xticks(size=12, weight='bold')\n plt.yticks(size=12, weight='bold')\n plt.xlim([0, testCasesMean])\n plt.ylim([0, 1.05])\n plt.grid(b=False)\n # Plot linear fits:\n xAxisVals = [0, int(np.max(ab))]\n plt.plot(xAxisVals, bLog + mLog*xAxisVals, 'r:')\n plt.plot(xAxisVals, bRF + mRF*xAxisVals, 'b:')\n plt.show()\n\n # Populate acceptableTestResultsMatrix for this policy domain:\n acceptTestResultsMatrix[:, ind] = acceptTestResultsAllRuns.transpose()\n\n''' End of loop through policy domains '''\n\n#%%\n''' print out statistics of results to console: '''\n\n# For each model type, calc stats and put them into a string:\n# Note that *BalAcc, *Auc etc have one column per model. If oneModelPerPolicyDomainFlag == True,\n# then col 0 to 4 = ECYN, FPYN, SWYN, RLYN, GNYN.\n\n# 1. Random Forest XOR xgBoost:\nif not xgBoostFlag:\n accResultDict = {'rawAcc':randomForestRawAcc, 'balAcc':randomForestBalAcc,\n 'AUC': randomForestAuc, 'prec': randomForestPrecision,\n 'recall': randomForestRecall, 'f1': randomForestF1}\nelse:\n accResultDict = {'rawAcc':xgBoostRawAcc, 'balAcc':xgBoostBalAcc, 'AUC': xgBoostAuc,\n 'prec': xgBoostPrecision, 'recall': xgBoostRecall, 'f1': xgBoostF1}\n\nif xgBoostFlag:\n typeStr = 'xgBoost'\nelse:\n typeStr = 'random forest'\nif useBalancedModelFlag:\n typeStr = typeStr + '_balanced'\nelse:\n typeStr = typeStr + '_UNbalanced'\n\nresultsDataFrame = \\\ncalculatePrintAndSaveModelStatsOverAllRunsFunction(accResultDict, acceptTestResultsMatrix, typeStr,\n setupParams, resultsDataFrame, fid)\n\n# 2. logistic:\nif runLogisticRegressionFlag:\n accResultDict = {'rawAcc':logisticRawAcc, 'balAcc':logisticBalAcc, 'AUC': logisticAuc, 'prec':\n logisticPrecision, 'recall': logisticRecall, 'f1': logisticF1}\n\n typeStr = 'logistic'\n if useBalancedModelFlag:\n typeStr = typeStr + '_balanced'\n else:\n typeStr = typeStr + '_UNbalanced'\n\n resultsDataFrame = \\\n calculatePrintAndSaveModelStatsOverAllRunsFunction(accResultDict, acceptTestResultsMatrix,\n typeStr, setupParams, resultsDataFrame, fid)\n # Also print out difference of RF and logistic:\n rfRa = randomForestRawAcc\n rfBa = randomForestBalAcc\n rfAuc = randomForestAuc\n if xgBoostFlag:\n rfRa = xgBoostRawAcc\n rfBa = xgBoostBalAcc\n rfAuc = xgBoostAuc\n accResultDict = {'rawAcc': rfRa - logisticRawAcc, 'balAcc': rfBa - logisticBalAcc,\n 'AUC': rfAuc - logisticAuc}\n typeStr = 'RF minus logistic'\n resultsDataFrame = \\\n calculatePrintAndSaveModelStatsOverAllRunsFunction(accResultDict, acceptTestResultsMatrix,\n typeStr, setupParams, resultsDataFrame, fid)\n# 3. Neural net:\nif runNeuralNetFlag:\n accResultDict = {'rawAcc':nNetRawAcc, 'balAcc':nNetBalAcc, 'AUC': nNetAuc,\n 'prec': nNetPrecision, 'recall': nNetRecall, 'f1': nNetF1}\n\n typeStr = 'neuralNet'\n if useBalancedModelFlag:\n typeStr = typeStr + '_balanced'\n else:\n typeStr = typeStr + '_UNbalanced'\n\n resultsDataFrame = \\\n calculatePrintAndSaveModelStatsOverAllRunsFunction(accResultDict, acceptTestResultsMatrix,\n typeStr, setupParams, resultsDataFrame, fid)\n\nresultsDataFrame.to_csv(saveDataframeResultsFilename)\n\nfid.close()\n\n#%%\n'''\nInterest groups column number key:\n\nAARP = 12\nAFL CIO = 13\nAirlines = 14\nAmerican Bankers Association = 15\nAmerican Council of Life Insurance = 16\nAmerican Farm Bureau Federation = 17\nAmerican Federation of State County and Municipal Employees = 18\nAmerican Hospital Association = 19\nAmerican Israel Public Affairs Committee = 20\nAmerican Legion = 21\nAmerican Medical Association = 22\nAssociation of Trial Lawyers = 23\nAutomobile companies = 24\nChamber of Commerce = 25\nChristian Coalition = 26\nComputer software and hardware = 27\nCredit Union National Association = 28\nDefense contractors = 29\nElectric companies = 30\nHealth Insurance Association = 31\nIndependent Insurance Agents of America = 32\nInternational Brotherhood of Teamsters = 33\nMotion Picture Association of America = 34\nNational Association of Broadcasters = 35\nNational Association of Home Builders = 36\nNational Association of Manufacturers = 37\nNational Association of Realtors = 38\nNational Beer Wholesalers Association = 39\nNational Education Association = 40\nNational Federation of Independent Business = 41\nNational Governors Association = 42\national Restaurant Association = 43\nNational Rifle Association = 44\nNational Right to Life Committee = 45\nOil Companies = 46\nPharmaceutical Research Manufacturers = 47\nRecording Industry Association = 48\nSecurities and investment companies = 49\nTelephone companies = 50\nTobbaco companies = 51\nUnited Auto Workers union = 52\nUniversities = 53\nVeterans of Foreign Wars of the US = 54\n '''\n'''\nMIT license:\nPermission is hereby granted, free of charge, to any person obtaining a copy of this software and\nassociated documentation files (the \"Software\"), to deal in the Software without restriction,\nincluding without limitation the rights to use, copy, modify, merge, publish, distribute,\nsublicense, and/or sell copies of the Software, and to permit persons to whom the Software is\nfurnished to do so, subject to the following conditions:\nThe above copyright notice and this permission notice shall be included in all copies or\nsubstantial portions of the Software.\nTHE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT\nNOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND\nNONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM,\nDAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT\nOF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.\n'''","sub_path":"run_models_main.py","file_name":"run_models_main.py","file_ext":"py","file_size_in_byte":47100,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"441845467","text":"import psycopg2\r\nimport json\r\n\r\n\r\nclass ArgumentsManager:\r\n\r\n @staticmethod\r\n def get_str(sys_args, arg_id):\r\n \"\"\"\r\n\r\n :param sys_args: Array of string. I would REALLY appreciate if you use the sys.argv here. Really.\r\n :param arg_id: String. Example \"-linkid\".\r\n :return:\r\n \"\"\"\r\n\r\n # basic check\r\n if (sys_args is None) or (arg_id is None):\r\n return None\r\n\r\n if (arg_id in sys_args) and (sys_args.index(arg_id) < (len(sys_args) - 1)):\r\n return sys_args[sys_args.index(arg_id) + 1]\r\n else:\r\n return None\r\n\r\n @staticmethod\r\n def get_int(sys_args, arg_id):\r\n \"\"\"\r\n\r\n :param sys_args:\r\n :param arg_id:\r\n :return:\r\n \"\"\"\r\n\r\n arg_value = ArgumentsManager.get_str(sys_args, arg_id)\r\n if arg_value is not None:\r\n try:\r\n return int(ArgumentsManager.get_str(sys_args, arg_id))\r\n except ValueError:\r\n print(\"Argument '{0}' is not an integer ({1}).\".format(arg_id, arg_value))\r\n return None\r\n else:\r\n return None\r\n\r\n @staticmethod\r\n def get_flt(sys_args, arg_id):\r\n \"\"\"\r\n\r\n :param sys_args:\r\n :param arg_id:\r\n :return:\r\n \"\"\"\r\n\r\n arg_value = ArgumentsManager.get_str(sys_args, arg_id)\r\n if arg_value is not None:\r\n try:\r\n return float(ArgumentsManager.get_str(sys_args, arg_id))\r\n except ValueError:\r\n print(\"Argument '{0}' is not an float ({1}).\".format(arg_id, arg_value))\r\n return None\r\n else:\r\n return None\r\n\r\n def __init__(self):\r\n return\r\n\r\n\r\nclass DatabaseManager:\r\n\r\n @staticmethod\r\n def open_database_connection(dbc_json_file_path):\r\n \"\"\"\r\n\r\n :param dbc_json_file_path:\r\n :return:\r\n \"\"\"\r\n\r\n # read database connection file\r\n with open(dbc_json_file_path) as dbc_json_file:\r\n dbc_json_data = json.load(dbc_json_file)\r\n\r\n db_conn = psycopg2.connect(\"dbname={0} user={1} password={2} host={3} port={4}\".format(\r\n dbc_json_data['database'], dbc_json_data['username'], dbc_json_data['password'], dbc_json_data['host'],\r\n dbc_json_data['port']))\r\n\r\n print(\"Connecting to '{0}' as '{1}'.\".format(dbc_json_data['database'], dbc_json_data['username']))\r\n\r\n return db_conn\r\n\r\n def __init__(self):\r\n return\r\n","sub_path":"backend/server_crons/def_lib.py","file_name":"def_lib.py","file_ext":"py","file_size_in_byte":2496,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"553856553","text":"# Copyright (c) 2020 University of Chicago\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or\n# implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom oslo_config import cfg\n\nfrom blazar.manager import exceptions as manager_exceptions\nfrom blazar.utils.openstack import base\nfrom oslo_log import log as logging\nfrom zunclient import client as zun_client\nfrom zunclient import exceptions as zun_exception\n\n\nzun_opts = [\n cfg.StrOpt(\n 'zun_api_version',\n default='1',\n help='Zun API version'),\n cfg.StrOpt(\n 'zun_api_microversion',\n default='1.22',\n help='Zun API microversion'),\n cfg.StrOpt(\n 'endpoint_override',\n help='Zun endpoint URL to use')\n]\n\nCONF = cfg.CONF\nCONF.register_opts(zun_opts, group='zun')\n\nLOG = logging.getLogger(__name__)\n\n\nclass BlazarZunClient(object):\n \"\"\"Client class for Zun service.\"\"\"\n\n def __init__(self, **kwargs):\n client_kwargs = base.client_kwargs(**kwargs)\n client_kwargs.setdefault('os_zun_api_version',\n CONF.zun.zun_api_microversion)\n self.zun = zun_client.Client(\n CONF.zun.zun_api_version, **client_kwargs)\n\n def __getattr__(self, attr):\n return getattr(self.zun, attr)\n\n\nclass ZunClientWrapper(object):\n @property\n def zun(self):\n zun = BlazarZunClient(endpoint_override=CONF.zun.endpoint_override)\n return zun\n\n\nclass ZunInventory(BlazarZunClient):\n def get_host_details(self, host):\n \"\"\"Get Zun capabilities of a single host\n\n :param host: UUID or name of zun compute node\n :return: Dict of capabilities or raise HostNotFound\n \"\"\"\n try:\n host = self.zun.hosts.get(host)\n except (zun_exception.NotFound, zun_exception.BadRequest):\n host_ids = []\n for h in self.zun.hosts.list():\n if h.hostname == host:\n host_ids.append(h.uuid)\n if len(host_ids) == 0:\n raise manager_exceptions.HostNotFound(host=host)\n elif len(host_ids) > 1:\n raise manager_exceptions.MultipleHostsFound(host=host)\n else:\n host = self.zun.hosts.get(host_ids[0])\n\n return {'id': host.uuid,\n 'name': host.hostname,\n 'containers': self.zun.containers.list(host=host.hostname)\n }\n","sub_path":"blazar/utils/openstack/zun.py","file_name":"zun.py","file_ext":"py","file_size_in_byte":2836,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"583201794","text":"# Copyright (c) 2021 Christian Corsica. All Rights Reserved.\n\nimport os\nimport subprocess\nfrom piper.core.dcc import DCC\n\n\nclass Maya(DCC):\n\n def __init__(self):\n super(Maya, self).__init__()\n self.registry_path = 'SOFTWARE\\\\Autodesk\\\\Maya'\n self.registry_exclude = 'Capabilities'\n self.registry_install_key = 'MAYA_INSTALL_LOCATION'\n self.registry_install_path = 'SOFTWARE\\\\Autodesk\\\\Maya\\\\{}\\\\Setup\\\\InstallPath'.format(self.version_replace)\n self.relative_python_path = ['bin', 'mayapy.exe']\n self.relative_batch_path = ['bin', 'mayabatch.exe']\n\n def runPythonBatch(self, version, command):\n batch_path = self.getBatchPath(version)\n full_command = '{0} -noAutoloadPlugins -command \"python(\"\"{1}\"\")\"'.format(batch_path, command)\n subprocess.call(full_command)\n\n def runPythonBatches(self, command):\n for version in self.getVersions():\n self.runPythonBatch(version, command)\n\n def onBeforeInstalling(self):\n os.environ['MAYA_SKIP_USERSETUP_PY'] = '1'\n os.environ['PYMEL_SKIP_MEL_INIT'] = '1'\n","sub_path":"piper/core/maya_dcc.py","file_name":"maya_dcc.py","file_ext":"py","file_size_in_byte":1108,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"176494243","text":"#!/usr/bin/env python3\n\n# Some tools for reporting time of functions and the like.\n#\nimport datetime\nimport time\nfrom functools import wraps\n\n\ndef timethis(func):\n ''' Report execution time of function; function result not maintained.'''\n @wraps(func)\n def wrapper(*args, **kwargs):\n start = time.time()\n result = func(*args, **kwargs)\n end = time.time()\n return end-start\n return wrapper\n\n\ndef logthis(func, filename=None, mode='a'):\n ''' Write function results to output filename; results returned.'''\n @wraps(func)\n def wrapper(*args, **kwargs):\n if not filename:\n filename = datetime.date.today() + '-' + func.__name__\n result = func(*args, **kwargs)\n with open(filename, mode) as f:\n print(result, file=f)\n return result\n return wrapper","sub_path":"03/logtools.py","file_name":"logtools.py","file_ext":"py","file_size_in_byte":843,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"481236660","text":"#======================================================================\n# ====================================\n# README file for Skew Correction component\n# ====================================\n\n# Filename : ocrd-anyBaseOCR-deskew.py\n\n# Author: Syed Saqib Bukhari, Mohammad Mohsin Reza, Md. Ajraf Rakib\n# Responsible: Syed Saqib Bukhari, Mohammad Mohsin Reza, Md. Ajraf Rakib\n# Contact Email: Saqib.Bukhari@dfki.de, Mohammad_mohsin.reza@dfki.de, Md_ajraf.rakib@dfki.de\n# Note: \n# 1) this work has been done in DFKI, Kaiserslautern, Germany.\n# 2) The parameters values are read from ocrd-anyBaseOCR-parameter.json file. The values can be changed in that file.\n# 3) The command line IO usage is based on \"OCR-D\" project guidelines (https://ocr-d.github.io/). A sample image file (samples/becker_quaestio_1586_00013.tif) and mets.xml (work_dir/mets.xml) are provided. The sequence of operations is: binarization, deskewing, cropping and dewarping (or can also be: binarization, dewarping, deskewing, and cropping; depends upon use-case).\n\n# *********** Method Behaviour ********************\n# This function takes a document image as input and do the skew correction of that document.\n# *********** Method Behaviour ********************\n\n# *********** LICENSE ********************\n# License: ocropus-nlbin.py (from https://github.com/tmbdev/ocropy/) contains both functionalities: binarization and skew correction. \n# This method (ocrd-anyBaseOCR-deskew.py) only contains the skew correction functionality of ocropus-nlbin.py. \n# It still has the same licenses as ocropus-nlbin, i.e Apache 2.0 (the ocropy license details are pasted below).\n# This file is dependend on ocrolib library which comes from https://github.com/tmbdev/ocropy/. \n\n#Copyright 2014 Thomas M. Breuel\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n# http://www.apache.org/licenses/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n#*********** LICENSE ********************\n#======================================================================\n#!/usr/bin/env python\n\nfrom __future__ import print_function\nfrom pylab import *\nfrom numpy.ctypeslib import ndpointer\nimport argparse,os,os.path,glob\nfrom scipy.ndimage import filters,interpolation,morphology,measurements\nfrom scipy import stats\nimport multiprocessing\nimport ocrolib\nimport json\nfrom xml.dom import minidom\n\nparser = argparse.ArgumentParser(\"\"\"\nImage deskewing using non-linear processing.\n \n python ocrd-anyBaseOCR-deskew.py -m (mets input file path) -I (input-file-grp name) -O (output-file-grp name) -w (Working directory)\n\nThis is a compute-intensive deskew method that works on degraded and historical book pages.\n\"\"\")\n\nparser.add_argument('-p','--parameter',type=str,help=\"Parameter file location\")\nparser.add_argument('-e','--escale',type=float,help='scale for estimating a mask over the text region, default: %(default)s')\nparser.add_argument('-t','--threshold',type=float,help='threshold, determines lightness, default: %(default)s')\nparser.add_argument('-b','--bignore',type=float,help='ignore this much of the border for threshold estimation, default: %(default)s')\nparser.add_argument('-ms','--maxskew',type=float,help='skew angle estimation parameters (degrees), default: %(default)s')\nparser.add_argument('--skewsteps',type=int,help='steps for skew angle estimation (per degree), default: %(default)s')\nparser.add_argument('--debug',type=float,help='display intermediate results, default: %(default)s')\nparser.add_argument('--lo',type=float,help='percentile for black estimation, default: %(default)s')\nparser.add_argument('--hi',type=float,help='percentile for white estimation, default: %(default)s')\nparser.add_argument('-Q','--parallel',type=int)\nparser.add_argument('-O','--Output',default=None,help=\"output directory\")\nparser.add_argument('-w','--work',type=str,help=\"Working directory location\", default=\".\")\nparser.add_argument('-I','--Input',default=None,help=\"Input directory\")\nparser.add_argument('-m','--mets',default=None,help=\"METs input file\")\nparser.add_argument('-o','--OutputMets',default=None,help=\"METs output file\")\nparser.add_argument('-g','--group',default=None,help=\"METs image group id\")\n\nargs = parser.parse_args()\n\n## Read parameter values from json file\nif args.parameter:\n if not os.path.exists(args.parameter):\n print(\"Error : Parameter file does not exists.\")\n sys.exit(0)\n else:\n param = json.load(open(args.parameter))\nelse:\n if not os.path.exists('ocrd-anyBaseOCR-parameter.json'):\n print(\"Error : Parameter file does not exists.\")\n sys.exit(0)\n else:\n param = json.load(open('ocrd-anyBaseOCR-parameter.json'))\n\nargs.bignore = param[\"anyBaseOCR\"][\"deskew\"][\"bignore\"]\nargs.escale = param[\"anyBaseOCR\"][\"deskew\"][\"escale\"]\nargs.threshold = param[\"anyBaseOCR\"][\"deskew\"][\"threshold\"]\nargs.lo = param[\"anyBaseOCR\"][\"deskew\"][\"lo\"]\nargs.hi = param[\"anyBaseOCR\"][\"deskew\"][\"hi\"]\nargs.maxskew = param[\"anyBaseOCR\"][\"deskew\"][\"maxskew\"]\nargs.skewsteps = param[\"anyBaseOCR\"][\"deskew\"][\"skewsteps\"]\nargs.debug = param[\"anyBaseOCR\"][\"deskew\"][\"debug\"]\nargs.parallel = param[\"anyBaseOCR\"][\"deskew\"][\"parallel\"]\n### End to read parameters\n\ndef parseXML(fpath):\n input_files=[]\n xmldoc = minidom.parse(fpath)\n nodes = xmldoc.getElementsByTagName('mets:fileGrp')\n for attr in nodes:\n if attr.attributes['USE'].value==args.Input:\n childNodes = attr.getElementsByTagName('mets:FLocat')\n for f in childNodes:\n input_files.append(f.attributes['xlink:href'].value)\n return input_files\n\ndef write_to_xml(fpath):\n xmldoc = minidom.parse(args.mets)\n subRoot = xmldoc.createElement('mets:fileGrp')\n subRoot.setAttribute('USE', args.Output)\n\n for f in fpath:\n #basefile = os.path.splitext(os.path.splitext(os.path.basename(f))[0])[0]\n basefile = ocrolib.allsplitext(os.path.basename(f))[0]\n child = xmldoc.createElement('mets:file')\n child.setAttribute('ID', 'DESKEW_'+basefile)\n child.setAttribute('GROUPID', 'P_' + basefile)\n child.setAttribute('MIMETYPE', \"image/png\")\n\n subChild = xmldoc.createElement('mets:FLocat')\n subChild.setAttribute('LOCTYPE', \"URL\")\n subChild.setAttribute('xlink:href', f)\n\n #xmldoc.getElementsByTagName('mets:file')[0].appendChild(subChild);\n subRoot.appendChild(child)\n child.appendChild(subChild)\n\n #subRoot.appendChild(child)\n xmldoc.getElementsByTagName('mets:fileSec')[0].appendChild(subRoot);\n\n if not args.OutputMets:\n metsFileSave = open(os.path.join(args.work, os.path.basename(args.mets)), \"w\")\n else:\n metsFileSave = open(os.path.join(args.work, args.OutputMets if args.OutputMets.endswith(\".xml\") else args.OutputMets+'.xml'), \"w\")\n metsFileSave.write(xmldoc.toxml()) \n\n#args.files = ocrolib.glob_all(args.files)\n\ndef print_info(*objs):\n print(\"INFO: \", *objs, file=sys.stdout)\n\ndef estimate_skew_angle(image,angles):\n estimates = []\n for a in angles:\n v = mean(interpolation.rotate(image,a,order=0,mode='constant'),axis=1)\n v = var(v)\n estimates.append((v,a))\n if args.debug>0:\n plot([y for x,y in estimates],[x for x,y in estimates])\n ginput(1,args.debug)\n _,a = max(estimates)\n return a\n\ndef deskew(fpath, job):\n base,_ = ocrolib.allsplitext(fpath)\n basefile = ocrolib.allsplitext(os.path.basename(fpath))[0]\n\n if args.parallel<2: print_info(\"=== %s %-3d\" % (fpath, job))\n raw = ocrolib.read_image_gray(fpath)\n\n flat = raw\n # estimate skew angle and rotate\n if args.maxskew>0:\n if args.parallel<2: print_info(\"estimating skew angle\")\n d0,d1 = flat.shape\n o0,o1 = int(args.bignore*d0),int(args.bignore*d1)\n flat = amax(flat)-flat\n flat -= amin(flat)\n est = flat[o0:d0-o0,o1:d1-o1]\n ma = args.maxskew\n ms = int(2*args.maxskew*args.skewsteps)\n angle = estimate_skew_angle(est,linspace(-ma,ma,ms+1))\n flat = interpolation.rotate(flat,angle,mode='constant',reshape=0)\n flat = amax(flat)-flat\n else:\n angle = 0\n\n # estimate low and high thresholds\n if args.parallel<2: print_info(\"estimating thresholds\")\n d0,d1 = flat.shape\n o0,o1 = int(args.bignore*d0),int(args.bignore*d1)\n est = flat[o0:d0-o0,o1:d1-o1]\n if args.escale>0:\n # by default, we use only regions that contain\n # significant variance; this makes the percentile\n # based low and high estimates more reliable\n e = args.escale\n v = est-filters.gaussian_filter(est,e*20.0)\n v = filters.gaussian_filter(v**2,e*20.0)**0.5\n v = (v>0.3*amax(v))\n v = morphology.binary_dilation(v,structure=ones((int(e*50),1)))\n v = morphology.binary_dilation(v,structure=ones((1,int(e*50))))\n if args.debug>0: imshow(v); ginput(1,args.debug)\n est = est[v]\n lo = stats.scoreatpercentile(est.ravel(),args.lo)\n hi = stats.scoreatpercentile(est.ravel(),args.hi)\n # rescale the image to get the gray scale image\n if args.parallel<2: print_info(\"rescaling\")\n flat -= lo\n flat /= (hi-lo)\n flat = clip(flat,0,1)\n if args.debug>0: imshow(flat,vmin=0,vmax=1); ginput(1,args.debug)\n bin = 1*(flat>args.threshold)\n\n # output the normalized grayscale and the thresholded images\n print_info(\"%s lo-hi (%.2f %.2f) angle %4.1f\" % (basefile, lo, hi, angle))\n if args.parallel<2: print_info(\"writing\")\n ocrolib.write_image_binary(base+\".ds.png\",bin)\n return base+\".ds.png\"\n\n# mendatory parameter check\nif not args.mets or not args.Input or not args.Output or not args.work:\n parser.print_help()\n print(\"Example: python ocrd-anyBaseOCR-deskew.py -m (mets input file path) -I (input-file-grp name) -O (output-file-grp name) -w (Working directory)\")\n sys.exit(0)\n\nif args.work:\n if not os.path.exists(args.work):\n os.mkdir(args.work)\n\nfiles = parseXML(args.mets)\nfname=[]\nfor i, f in enumerate(files):\n fname.append(deskew(str(f),i+1))\nwrite_to_xml(fname)\n","sub_path":"ocrd-anyBaseOCR-deskew.py","file_name":"ocrd-anyBaseOCR-deskew.py","file_ext":"py","file_size_in_byte":10461,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"139914334","text":"\"\"\" Parallel Jaccard/Tanimoto coefficient module \"\"\"\nimport csv\nimport math\nimport time\nimport re\nimport multiprocessing as mp\n\n\ndef open_file():\n \"\"\"Opens a tsv file and returns a list with the data.\"\"\"\n chemicals_list = []\n with open(\"ZINC_chemicals.tsv\") as file:\n reader = csv.reader(file, delimiter=\"\\t\", quotechar='\"')\n for row in reader:\n row_aux = (row[1], row[3])\n chemicals_list.append(row_aux)\n return chemicals_list\n\n\ndef analyze_string(chemical_compound):\n \"\"\"Analyzes the string of chemical compound's elements \\\n and returns the analyzed data in a dictionary.\"\"\"\n chemical_compound = \"\".join(re.findall(\"[a-zA-Z@]+\", chemical_compound))\n analyzed_string = {}\n for character in chemical_compound:\n if character in analyzed_string:\n new_value = analyzed_string.get(character) + 1\n analyzed_string.update({character: new_value})\n else:\n analyzed_string.update({character: 1})\n if '@' in analyzed_string:\n analyzed_string.update({'@': 1})\n return analyzed_string\n\n\ndef get_number_common_elements(chemical_a, chemical_b):\n \"\"\"Returns the number of common elements between chemical \\\n compound a and chemical compound b.\"\"\"\n number_elements = 0\n for key_a in chemical_a:\n if key_a in chemical_b:\n number_elements += min(chemical_a.get(key_a), chemical_b.get(key_a))\n return number_elements\n\n\ndef get_number_chemical_elements(chemical_compound):\n \"\"\"Returns the number of elements in a chemical compound.\"\"\"\n number_elements = 0\n for key in chemical_compound:\n number_elements += chemical_compound.get(key)\n return number_elements\n\n\ndef get_jac_tan_coefficient(chemical_a, chemical_b):\n \"\"\"Returns the coefficient of Jaccard/Tanimoto between two chemical compounds.\"\"\"\n letters_a = analyze_string(chemical_a)\n letters_b = analyze_string(chemical_b)\n elements_a = get_number_chemical_elements(letters_a)\n elements_b = get_number_chemical_elements(letters_b)\n common_elements = get_number_common_elements(letters_a, letters_b)\n return round((common_elements / (elements_a + elements_b - common_elements)), 2)\n\n\ndef get_pivots(chemicals_length, number_processors):\n \"\"\"Calculates the pivots to divide the chemicals between the threads.\"\"\"\n pivots_list = []\n pivots_list.append(0)\n for i in range(number_processors - 1, 0, -1):\n pivots_list.append(\n int(round(chemicals_length - (math.sqrt(i / number_processors) * chemicals_length))))\n pivots_list.append(chemicals_length - 1)\n return pivots_list\n\n\ndef fill_compared_list(chemicals_list, pivot_min, pivot_max, queues_list):\n \"\"\"Fills a list with the comparison of the chemical compounds between two pivots in the list.\"\"\"\n compared_chemicals_list = []\n for i in range(pivot_min, pivot_max):\n for j in range(i + 1, len(chemicals_list)):\n coef = get_jac_tan_coefficient(chemicals_list[i][1], chemicals_list[j][1])\n row = chemicals_list[i][0] + \"\\t\" + chemicals_list[j][0] + \"\\t\" + str(coef)\n compared_chemicals_list.append(row)\n queues_list.put(compared_chemicals_list)\n\n\ndef start_join_all(compared_chemicals_list, processes_list, queues_list):\n \"\"\"Starts and joins all the threads in the list.\"\"\"\n for process in processes_list:\n process.start()\n for i in range(len(processes_list)):\n compared_chemicals_list[i] = queues_list[i].get()\n processes_list[i].join()\n\n\ndef write_file(number_threads, compared_chemicals_list, total_time):\n \"\"\"Writes a tsv file with the compared chemicals.\"\"\"\n with open(\"chem_sim_total_Python_Processes.tsv\", \"w\") as record_file:\n record_file.write(\"Chem_ID_1\\tChem_ID_2\\tTanimoto_similarity\\n\")\n for i in range(number_threads):\n print(len(compared_chemicals_list[i]))\n for chemical in compared_chemicals_list[i]:\n record_file.write(chemical + \"\\n\")\n record_file.write(\"Total time = \" + str(total_time) + \" [s]\\n\")\n\n\ndef print_to_console(number_threads, compared_chemicals_list, total_time):\n \"\"\"Prints in console the compared chemicals.\"\"\"\n counter = 0\n for i in range(number_threads):\n for chemical in compared_chemicals_list[i]:\n counter += 1\n print(chemical)\n print(\"Total Elements = \" + str(counter))\n print(\"Total time = \" + str(total_time) + \" [s]\")\n\n\nif __name__ == \"__main__\":\n NUMBER_THREADS = mp.cpu_count()\n CHEMICALS_LIST = open_file()\n PIVOTS_LIST = get_pivots(len(CHEMICALS_LIST), NUMBER_THREADS)\n\n START_TIME = time.time()\n compared_chemicals = []\n queues = []\n processes = []\n for index in range(NUMBER_THREADS):\n compared_chemicals.append([])\n queues.append(mp.Queue())\n processes.append(mp.Process(target=fill_compared_list, args=(\n CHEMICALS_LIST, PIVOTS_LIST[index],\n PIVOTS_LIST[index + 1],\n queues[index], )))\n start_join_all(compared_chemicals, processes, queues)\n\n END_TIME = time.time()\n TOTAL_TIME = END_TIME - START_TIME\n write_file(NUMBER_THREADS, compared_chemicals, TOTAL_TIME)\n #print_to_console(NUMBER_THREADS, compared_chemicals, TOTAL_TIME)\n","sub_path":"jaccard_tanimoto_process.py","file_name":"jaccard_tanimoto_process.py","file_ext":"py","file_size_in_byte":5273,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"275984330","text":"import timeit\nimport random\n\n\ndef one_two():\n return random.randrange(1,3)\n\n# Basic Test\n'''\nlength = 60000\ntreshold = 0.05\ncount = Counter(one_two_three() for _ in range(length)) # past here student's function\n\nprint(1 in count, '1 is present')\nprint(2 in count, '2 is present')\nprint(3 in count, '3 is present')\nprint(abs(count[1] / length - 1 / 3) < treshold, '1 count is nearly equal 1/3')\nprint(abs(count[2] / length - 1 / 3) < treshold, '2 count is nearly equal 1/3')\nprint(abs(count[3] / length - 1 / 3) < treshold, '3 count is nearly equal 1/3')\n'''\n\nccquiel_setup = '''\nfrom __main__ import ccquiel_day30\n'''\n\ndef ccquiel_day30():\n t = one_two() + one_two()\n if t == 2: return 1\n if t == 4: return 2\n if t == 3:\n if one_two() == 1: return 3\n else: return ccquiel_day30()\n\nTEST_CODE_ccquiel = '''\nresult = ccquiel_day30()\n'''\n\ncharlie_ang_setup = '''\nfrom __main__ import charlie_ang_day30\n'''\n\ndef charlie_ang_day30():\n while True:\n if one_two() == 1:\n return one_two()\n elif one_two() == 1:\n return 3\n\nTEST_CODE_charlie_ang = '''\nresult = charlie_ang_day30()\n'''\n\ndiana_henninger_setup = '''\nfrom __main__ import diana_henninger_day30\n'''\n\ndef diana_henninger_day30():\n if one_two()==1:\n if one_two()==1: return 1\n else: return 2\n else:\n if one_two()== 1: return 3\n else: return diana_henninger_day30()\n\nTEST_CODE_diana_henninger = '''\nresult = diana_henninger_day30()\n'''\n\nJens_setup = '''\nfrom __main__ import Jens_day30\n'''\n\ndef Jens_day30():\n while True:\n r_list = [i for i in range(1, 4) if one_two() == 1]\n if len(r_list) == 1:\n return r_list[0]\n elif len(r_list) == 2:\n return r_list[0] if one_two() == 1 else r_list[1]\n\nTEST_CODE_Jens = '''\nresult = Jens_day30()\n'''\n\nKurt_Hinderer_setup = '''\nfrom __main__ import Kurt_Hinderer_day30\n'''\n\ndef Kurt_Hinderer_day30():\n repeat = False\n while not repeat:\n if one_two() == 1:\n if one_two() == 1:\n return 1\n else:\n return 2\n else:\n if one_two() == 1:\n return 3\n\nTEST_CODE_Kurt_Hinderer = '''\nresult = Kurt_Hinderer_day30()\n'''\n\nOleksandra_Chmel_setup = '''\nfrom __main__ import Oleksandra_Chmel_day30\n'''\n\ndef Oleksandra_Chmel_day30():\n one = one_two()\n two = one_two()\n if one == 1 and two == 1:\n return 1\n if one == 2 and two == 2:\n return 2\n if one == 2 and two == 1:\n return 3\n return Oleksandra_Chmel_day30()\n\nTEST_CODE_Oleksandra_Chmel = '''\nresult = Oleksandra_Chmel_day30()\n'''\n\nTushar_Jain_setup = '''\nfrom __main__ import Tushar_Jain_day30\n'''\n\ndef Tushar_Jain_day30():\n while True:\n if one_two() == 1:\n return one_two()\n elif one_two() == 1:\n return 3\n\nTEST_CODE_Tushar_Jain = '''\nresult = Tushar_Jain_day30()\n'''\n\nYang_setup = '''\nfrom __main__ import Yang_day30\n'''\n\ndef Yang_day30(): \n a,b = one_two(),one_two()+1\n if a*b==4: \n return 1 \n elif a*b<4: \n return a*b\n else: \n return Yang_day30() \n\nTEST_CODE_Yang = '''\nresult = Yang_day30()\n'''\n\nprint(\"Time for ccquiel test code: \" + str(timeit.timeit(stmt=TEST_CODE_ccquiel, setup=ccquiel_setup, number=100000)) + \" seconds\")\nprint(\"Time for charlie_ang test code: \" + str(timeit.timeit(stmt=TEST_CODE_charlie_ang, setup=charlie_ang_setup, number=100000)) + \" seconds\")\nprint(\"Time for diana_henninger test code: \" + str(timeit.timeit(stmt=TEST_CODE_diana_henninger, setup=diana_henninger_setup, number=100000)) + \" seconds\")\nprint(\"Time for Jens test code: \" + str(timeit.timeit(stmt=TEST_CODE_Jens, setup=Jens_setup, number=100000)) + \" seconds\")\nprint(\"Time for Kurt_Hinderer test code: \" + str(timeit.timeit(stmt=TEST_CODE_Kurt_Hinderer, setup=Kurt_Hinderer_setup, number=100000)) + \" seconds\")\nprint(\"Time for Oleksandra_Chmel test code: \" + str(timeit.timeit(stmt=TEST_CODE_Oleksandra_Chmel, setup=Oleksandra_Chmel_setup, number=100000)) + \" seconds\")\nprint(\"Time for Tushar_Jain test code: \" + str(timeit.timeit(stmt=TEST_CODE_Tushar_Jain, setup=Tushar_Jain_setup, number=100000)) + \" seconds\")\nprint(\"Time for Yang test code: \" + str(timeit.timeit(stmt=TEST_CODE_Yang, setup=Yang_setup, number=100000)) + \" seconds\")\n","sub_path":"day30/day30.py","file_name":"day30.py","file_ext":"py","file_size_in_byte":4308,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"312839932","text":"# coding=utf-8\n\nfrom pyexcel_xls import get_data\nimport sqlite3\n\nraw = get_data('2_教室代码.xls')\ndel raw[0]\n\nconn = sqlite3.connect('schedule.db')\ncur = conn.cursor()\n\ncur.execute('CREATE TABLE Classrooms(Id text, Name text, seats int)')\ncur.executemany('INSERT INTO Classrooms VALUES(?, ?, ?)', raw)\nprint(cur.execute('select id from Classrooms').fetchall())\n\nconn.commit()\nconn.close()\n","sub_path":"gen2.py","file_name":"gen2.py","file_ext":"py","file_size_in_byte":393,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"394128462","text":"import time\n\ndef FCcheck(new_x,new_y,un_visited):\n # print('x,y',new_x,new_y)\n length = len(un_visited)\n for i in range(0,length):\n node = un_visited[i]\n new_domain = []\n for y in node.domain:\n x = node.index\n # print('a',x,y,new_x,new_y)\n if x == new_x or y == new_y or (abs(x - new_x) == abs(y - new_y)):\n # print(\"a\")\n continue\n else:\n # print('b')\n new_domain.append(y)\n if len(new_domain) == 0:\n return \"DWO\"\n node.domain = new_domain\n # print(un_visited[1].domain,un_visited[2].domain,un_visited[3].domain)\n return True\n\ndef FC(node,N,is_visited,un_visited):\n\n # print('id,domain,solution',node.index,node.domain,node.solution)\n new_x = node.index\n domain = node.domain\n for i in domain:\n new_y = i\n constraintsOK = True\n for k in range(len(is_visited)):\n x = is_visited[k].solution[0]\n y = is_visited[k].solution[1]\n # print(\"bbb\",x,y,new_x,new_y)\n if x == new_x or y == new_y or (abs(x - new_x) == abs(y - new_y)):\n constraintsOK = False\n break\n if constraintsOK == False:\n continue\n save_domain = {}\n for k in range(len(un_visited)):\n save_tmp = [x for x in un_visited[k].domain]\n save_domain[un_visited[k].index] = save_tmp\n\n DWOoccurred = False\n if FCcheck(new_x,new_y,un_visited) == \"DWO\":\n DWOoccurred = True\n # print(\"new:\",new_x,new_y,constraintsOK,DWOoccurred)\n # print(un_visited[1].domain,un_visited[2].domain,un_visited[3].domain)\n if DWOoccurred == False:\n node.solution = (new_x,new_y)\n is_visited.append(node)\n if len(is_visited) >= N:\n return node\n min_node_index = 0\n min_domain_length = len(un_visited[0].domain)\n for k in range(len(un_visited)):\n if len(un_visited[k].domain) < min_domain_length:\n min_node_index = k\n\n newNode = un_visited.pop(min_node_index)\n newNode.parent = node\n dstNode = FC(newNode,N,is_visited,un_visited)\n if dstNode != None:\n return dstNode\n is_visited.pop(-1)\n un_visited.append(newNode)\n for k in range(len(un_visited)):\n index = un_visited[k].index\n un_visited[k].domain = [x for x in save_domain[index]]\n\n # un_visited.insert(0,node1)\n return None\n\n\ndef start_FC(N):\n\n un_visited = []\n for i in range(N):\n newNode = Queen(i,0,N,None)\n un_visited.append(newNode)\n\n for i in range(N):\n start_node = Queen(0,i,N,None)\n is_visited = []\n # is_visited.append(start_node)\n un_visited.pop(0)\n dst_node = FC(start_node,N,is_visited,un_visited)\n if dst_node != None:\n return dst_node\n\n return None\n\nclass Queen():\n def __init__(self,x,y,N,parent):\n self.domain = [x for x in range(N)]\n self.index = x\n self.solution = (x,y)\n self.parent = parent\n\ndef main():\n N = int(input('Please input N = '))\n start = time.clock()\n dst_node = start_FC(N)\n\n if dst_node == None:\n print(\"No solution\")\n else:\n solution = []\n while dst_node:\n solution.append(dst_node.solution)\n dst_node = dst_node.parent\n solution = list(reversed(solution))\n print(\"Solution: \",solution)\n\n end = time.clock()\n print(\"Forwardchecking run time : \",float(end - start),' s')\n\n queen_map = []\n for i in range(N):\n tmp = ['.'] * N\n queen_map.append(tmp)\n for i in range(len(solution)):\n queen_map[solution[i][0]][solution[i][1]] = 'X'\n\n for i in range(N):\n print(' ',' '.join(queen_map[i]))\n\nif __name__ == '__main__':\n main()","sub_path":"FC_Tree_MRV.py","file_name":"FC_Tree_MRV.py","file_ext":"py","file_size_in_byte":3970,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"628834104","text":"def maior(x, y):\n if x>y:\n return x\n else:\n return y\n\ndef menor(x, y):\n if x>y:\n return y\n else:\n return x\n\ndef listamaior(lista):\n if len(lista)==0:\n return none\n else:\n n= lista[0]\n for aux in lista:\n if aux>n:\n n= aux\n return n\n\ndef listamenor(lista):\n if len(lista)==0:\n return none\n else:\n n= lista[0]\n for aux in lista:\n if auxi:\n cont= cont+1\n return cont\n\ndef lista12(lista1, lista2):\n n1= lista1[0]\n n2= lista2[0]\n for aux in lista1:\n if aux> n1:\n n1= aux\n for ind in lista2:\n if ind> n2:\n n2= ind\n if n1> n2:\n return lista1\n else:\n return lista2\n\ndef listacomp(lista1, lista2):\n n1= lista1[0]\n n2= lista2[0]\n cont= 0\n for aux in lista1:\n if aux> n1:\n n1= aux\n for ind in lista2:\n if ind> n2:\n n2= ind\n\n if n1> n2:\n cont= cont+3\n elif n1==n2:\n cont= cont+1\n return cont\n\ndef impoupar(x):\n if (x%2)==0:\n x= True\n else:\n x= False\n return x\n\ndef listaparouimpar(lista):\n listapar= []\n listaimpar= []\n for aux in lista:\n if aux%2==0:\n listapar.append(aux)\n else:\n listaimpar.append(aux)\n return listapar.sort()\n return listaimpar.sort()\n\n \n","sub_path":"Pythons/trabalho sala.py","file_name":"trabalho sala.py","file_ext":"py","file_size_in_byte":1534,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"178068860","text":"\nfrom keras.utils import plot_model\nfrom keras.models import Sequential\nfrom keras.layers import Dense, LSTM, InputLayer, Bidirectional, TimeDistributed, Embedding, Activation\nfrom keras.optimizers import Adam\nfrom keras.preprocessing.sequence import pad_sequences\nfrom sklearn.model_selection import train_test_split\nfrom keras import backend as K\nimport numpy as np\nfrom nltk.corpus import brown\nfrom keras.callbacks import EarlyStopping\nearly_stopping = EarlyStopping(patience=3)\nimport matplotlib.pyplot as plt\n\n\n\ndef to_categorical(sequences, categories):\n cat_sequences = []\n for s in sequences:\n cats = []\n for item in s:\n cats.append(np.zeros(categories))\n cats[-1][item] = 1.0\n cat_sequences.append(cats)\n return np.array(cat_sequences)\n\n\ndef logits_to_tokens(sequences, index):\n token_sequences = []\n for categorical_sequence in sequences:\n token_sequence = []\n for categorical in categorical_sequence:\n token_sequence.append(index[np.argmax(categorical)])\n\n token_sequences.append(token_sequence)\n\n return token_sequences\n\n\ndef ignore_class_accuracy(to_ignore=0):\n def ignore_accuracy(y_true, y_pred):\n y_true_class = K.argmax(y_true, axis=-1)\n y_pred_class = K.argmax(y_pred, axis=-1)\n\n ignore_mask = K.cast(K.not_equal(y_pred_class, to_ignore), 'int32')\n matches = K.cast(K.equal(y_true_class, y_pred_class), 'int32') * ignore_mask\n accuracy = K.sum(matches) / K.maximum(K.sum(ignore_mask), 1)\n return accuracy\n\n return ignore_accuracy\n\n\nif __name__ == '__main__':\n\n tagged_sentences = brown.tagged_sents(tagset='universal')\n sentences, sentence_tags = [], []\n for tagged_sentence in tagged_sentences:\n sentence, tags = zip(*tagged_sentence)\n sentences.append(np.array(sentence))\n sentence_tags.append(np.array(tags))\n\n (train_sentences,\n test_sentences,\n train_tags,\n test_tags) = train_test_split(sentences, sentence_tags, test_size=0.2)\n\n words, tags = set([]), set([])\n\n for s in train_sentences:\n for w in s:\n words.add(w.lower())\n\n for ts in train_tags:\n for t in ts:\n tags.add(t)\n\n word2index = {w: i + 2 for i, w in enumerate(list(words))}\n word2index['-PAD-'] = 0 # The special value used for padding\n word2index['-OOV-'] = 1 # The special value used for OOVs\n\n tag2index = {t: i + 1 for i, t in enumerate(list(tags))}\n tag2index['-PAD-'] = 0 # The special value used to padding\n\n train_sentences_X, test_sentences_X, train_tags_y, test_tags_y = [], [], [], []\n\n for s in train_sentences:\n s_int = []\n for w in s:\n try:\n s_int.append(word2index[w.lower()])\n except KeyError:\n s_int.append(word2index['-OOV-'])\n\n train_sentences_X.append(s_int)\n\n for s in test_sentences:\n s_int = []\n for w in s:\n try:\n s_int.append(word2index[w.lower()])\n except KeyError:\n s_int.append(word2index['-OOV-'])\n\n test_sentences_X.append(s_int)\n\n for s in train_tags:\n train_tags_y.append([tag2index[t] for t in s])\n\n for s in test_tags:\n test_tags_y.append([tag2index[t] for t in s])\n\n MAX_LENGTH = len(max(train_sentences_X, key=len))\n\n train_sentences_X = pad_sequences(train_sentences_X, maxlen=MAX_LENGTH, padding='post')\n test_sentences_X = pad_sequences(test_sentences_X, maxlen=MAX_LENGTH, padding='post')\n train_tags_y = pad_sequences(train_tags_y, maxlen=MAX_LENGTH, padding='post')\n test_tags_y = pad_sequences(test_tags_y, maxlen=MAX_LENGTH, padding='post')\n\n cat_train_tags_y = to_categorical(train_tags_y, len(tag2index))\n model = Sequential()\n model.add(InputLayer(input_shape=(MAX_LENGTH,)))\n model.add(Embedding(len(word2index), 128))\n model.add(Dense(len(tag2index)))\n model.add(Dense(len(tag2index)))\n model.add(Activation('softmax'))\n\n model.compile(loss='categorical_crossentropy',\n optimizer=Adam(0.001),\n metrics=['accuracy', ignore_class_accuracy(0)])\n\n model.summary()\n\n hist = model.fit(train_sentences_X, to_categorical(train_tags_y, len(tag2index)), batch_size=2048, epochs=3, validation_split=0.2,callbacks=[early_stopping])\n\n score = model.evaluate(test_sentences_X, to_categorical(test_tags_y, len(tag2index)), verbose=0)\n print(model.metrics_names)\n\n print('model loss: {} acc: {} ignore_accuracy : {}'.format(score[0], score[1],score[2]))\n plot_model(model, to_file='tmp/dnn_model_structure.png', show_shapes=True)\n\n fig, loss_ax = plt.subplots()\n\n acc_ax = loss_ax.twinx()\n\n loss_ax.plot(hist.history['loss'], 'y', label='train loss')\n loss_ax.plot(hist.history['val_loss'], 'r', label='val loss')\n\n acc_ax.plot(hist.history['acc'], 'b', label='train acc')\n acc_ax.plot(hist.history['val_acc'], 'g', label='val acc')\n\n loss_ax.set_xlabel('epoch')\n loss_ax.set_ylabel('loss')\n acc_ax.set_ylabel('accuray')\n\n loss_ax.legend(loc='upper left')\n acc_ax.legend(loc='lower left')\n plt.savefig(\"tmp/dnn_tagger_hist.png\",dpi=300)\n\n","sub_path":"dnn_tagger.py","file_name":"dnn_tagger.py","file_ext":"py","file_size_in_byte":5210,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"66412202","text":"import subprocess\r\nimport os\r\nimport sys\r\nimport shutil\r\n\r\nstudioMode = (\"--studio\" in sys.argv) or (\"-s\" in sys.argv) # Uninstall from studio directory\r\ncleanup = (\"--cleanup\" in sys.argv) or (\"-c\" in sys.argv) # Cleanup files\r\n\r\nimport installUtil\r\ninstallUtil.setStudio(studioMode)\r\n\r\nif installUtil.verifyAndGoto():\r\n\tprint(\"Found install directory\")\r\n\r\nwFiles = installUtil.getWrapperFiles(_allFiles=True)\r\nfor i, fileName in enumerate(installUtil.getInstallFiles(_allFiles=True)):\r\n\twFileName = wFiles[i]\r\n\tif os.path.isfile(wFileName):\r\n\t\tif os.path.isfile(fileName):\r\n\t\t\tsubprocess.run(\"attrib -r \"+fileName)\r\n\t\t\tos.remove(fileName)\r\n\t\tos.rename(wFileName, fileName)\r\n\t\tprint(\"Deleted compiled wrapper executable \"+fileName+\" and renamed real executable to original\")\r\n\telse:\r\n\t\tprint(\"Skipping \"+fileName+\" because \"+wFileName+\" doesn't exist\")\r\n\r\ninstallerPaths = [\"command-window.cmd\", \"Icons\", \"Ahk2Exe\", \"Wrapper.ahk\", \"help.pyw\", \"install.py\", \"config.py\", \"__pycache__\", \"installUtil.py\", \"uninstall.py\"]\r\nif cleanup:\r\n\tif installUtil.isInVersionFolder():\r\n\t\tdoCleanup = input(\"Are you sure you want to cleanup *all* installed files? This includes your custom python files and the Wrapper.ahk file, so be sure to back them up if they contain important code! [Y/N]\")\r\n\t\tif doCleanup.lower() == \"y\":\r\n\t\t\tfor path in installUtil.getInstallFiles(_includeExt=False, _allFiles=True):\r\n\t\t\t\tif os.path.isfile(path+\".py\"):\r\n\t\t\t\t\tos.remove(path+\".py\")\r\n\t\t\tfor path in os.listdir(os.getcwd()):\r\n\t\t\t\tif path in installerPaths:\r\n\t\t\t\t\tif os.path.isdir(path):\r\n\t\t\t\t\t\tshutil.rmtree(path)\r\n\t\t\t\t\tif os.path.isfile(path):\r\n\t\t\t\t\t\tos.remove(path)\r\n\t\t\tprint(\"Cleaned up files!\")\r\n\t\t\tsys.exit()\r\n\t\telse:\r\n\t\t\tprint(\"Did not cleanup files: Canceled\")\r\n\telse:\r\n\t\tprint(\"Did not clean up files: Not in a valid install directory\")","sub_path":"Wrapper/uninstall.py","file_name":"uninstall.py","file_ext":"py","file_size_in_byte":1817,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"366369695","text":"from django.core.urlresolvers import reverse\nfrom edx_proctoring.api import get_all_exams_for_course\nfrom student.models import CourseEnrollment\nfrom xmodule.modulestore.django import modulestore\nfrom enrollment.serializers import CourseEnrollmentSerializer\nfrom openedx.core.djangoapps.course_groups.models import CourseUserGroup\nfrom opaque_keys.edx.keys import CourseKey, UsageKey\n\n\nVERIFIED = 'verified'\n\n\ndef get_course_enrollments(user_id=None, **kwargs):\n \"\"\"\n Retrieve a list representing all aggregated data for a user's course enrollments.\n Construct a representation of all course enrollment data for a specific user.\n Args:\n user_id (str): The name of the user to retrieve course enrollment information for.\n Returns:\n A serializable list of dictionaries of all aggregated enrollment data for a user.\n \"\"\"\n qset = CourseEnrollment.objects.filter(is_active=True, **kwargs)\n if user_id is not None:\n qset = qset.filter(user__username=user_id)\n qset = qset.order_by('created')\n return CourseEnrollmentSerializer(qset).data # pylint: disable=no-member\n\n\ndef get_user_proctored_exams(username, request):\n enrollments = CourseEnrollment.objects.filter(is_active=True,\n user__username=username)\n system = request.data.get('system')\n result = {}\n for enrollment in enrollments:\n course = enrollment.course\n try:\n course_id = str(course.id)\n except AttributeError:\n continue\n\n cohorts = CourseUserGroup.objects.filter(\n course_id=enrollment.course_id,\n users__username=username,\n group_type=CourseUserGroup.COHORT,\n name=VERIFIED\n )\n\n if course_id not in result and cohorts.exists():\n proctoring_service = modulestore().get_course(CourseKey.from_string(course_id)).proctoring_service\n if system and system != proctoring_service:\n continue\n result[course_id] = {\n \"id\": course_id,\n \"name\": course.display_name,\n \"uri\": request.build_absolute_uri(\n reverse('course_structure_api:v0:detail',\n kwargs={'course_id': course_id})),\n \"image_url\": course.course_image_url,\n \"start\": course.start,\n \"end\": course.end,\n \"system\": proctoring_service,\n 'exams': []\n }\n exams = get_all_exams_for_course(course_id=course.id)\n for exam in exams:\n if exam['is_proctored']:\n item_id = UsageKey.from_string(exam['content_id'])\n item = modulestore().get_item(item_id)\n exam['visible_to_staff_only'] = item.visible_to_staff_only\n oldest = None\n due_dates = []\n for vertical in item.get_children():\n if vertical.due:\n due_dates.append(vertical.due)\n if due_dates:\n oldest = min(due_dates)\n exam['deadline'] = oldest\n exam['start'] = item.start\n result[course_id]['exams'].append(exam)\n result = {key: value for key, value in result.items() if\n len(value['exams']) > 0}\n return result\n\n","sub_path":"open_edx_api_extension/data.py","file_name":"data.py","file_ext":"py","file_size_in_byte":3447,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"68529202","text":"#!/usr/bin/python3\nimport json, math, telepot\nfrom uptime import uptime\nfrom subprocess import call, check_output\n\n#enter your json path below with constant and bot information\nautht_path = '/home/auth.json'\n\ndef temp():\n res = check_output('vcgencmd measure_temp', shell=True)\n return(res.decode()).replace('temp=',' ').replace(\"'C\",\"°C🌡\")\n\ndef cpu():\n disk = check_output(\"top -bn1 | grep load | awk '{printf \\\"%.2f\\\", $(NF-2)}'\", shell=True)\n return(str(disk)).replace(\"b\",\"\").replace(\"'\",\"\")\n\ndef mem():\n mem = check_output(\"free -m | awk 'NR==2{printf \\\"%s/%sMB %.2f%%\\\", $3,$2,$3*100/$2 }'\", shell=True)\n return(str(mem)).replace(\"b\",\"\").replace(\"'\",\"\")\n\ndef disk():\n disk = check_output(\"df -h | awk '$NF==\\\"/\\\"{printf \\\"%.1f/%.1fGB %s\\\", $3,$2,$5}'\", shell=True)\n return(str(disk)).replace(\"b\",\"\").replace(\"'\",\"\")\n\ndef uptm():\n if math.floor(uptime() / 60) < 2:\n tm = (str(math.floor(uptime() / 60)) + ' minute')\n\n elif math.floor(uptime() / 60) >= 2 and math.floor(uptime() / 60) < 60:\n tm = (str(math.floor(uptime() / 60)) + ' minutes')\n\n elif math.floor(uptime() / 60) >= 60 and math.floor(uptime() / 60) < 120:\n tm = (str(math.floor(uptime() / 3600)) + ' hour')\n\n elif math.floor(uptime() / 60) >= 120 and math.floor(uptime() / 86400) < 1:\n tm = (str(math.floor(uptime() / 3600)) + ' hours')\n\n elif math.floor(uptime() / 86400) >= 1 and math.floor(uptime() / 86400) < 2:\n tm = (str(math.floor(uptime() / 86400)) + ' day')\n\n else:\n tm = (str(math.floor(uptime() / 86400)) + ' days')\n\n return(tm)\n\ndef porta():\n call('curl -4 ifconfig.co/port/22 > /home/pi/porta.txt', shell=True)\n with open('/home/pi/porta.txt') as file:\n ifconfig = json.load(file)\n ip = ifconfig['ip']\n\n if ifconfig['reachable'] == True:\n estado = '\\n*Firewall:* Disable ⛔'\n else:\n estado = '\\n*Firewall:* Enable ✅'\n\n call('rm /home/pi/porta.txt &', shell=True)\n\n return(ip, estado)\n\n\nip, estado = porta()\n\nwith open(autht_path) as file:\n secrets = json.load(file)\n\n bot = telepot.Bot(secrets['tg_tk_bot'])\n tg_destiny = secrets['tg_id_to']\n\nbot.sendMessage(tg_destiny, '*CPU Temp:*' + temp() + '*Uptime:* ' \\\n\t\t\t\t + uptm() + \"\\n*IP:* \" + ip + '\\n*CPU Load:* ' + cpu() + '\\n*Mem:* '\\\n + mem() + '\\n*Disk:* ' + disk() + estado, parse_mode='Markdown')","sub_path":"status.py","file_name":"status.py","file_ext":"py","file_size_in_byte":2423,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"80500785","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport os\nfrom cluster import Cluster\n\n\ndef gjf_head(fname):\n gjf_head_list = ['%mem=1GB', '%nprocl=1', '%nprocs=4', '%%chk=%s.chk' % fname,\n '# pw91pw91/6-31g* iop(3/33=1,5/33=2) nosymm', '', fname, '', '0 1', '']\n return '\\n'.join(gjf_head_list)\n\n\ndef write_gjf(idx, title, cluster):\n if not os.path.exists(str(idx)):\n os.mkdir(str(idx))\n with open('%d/%s.gjf' % (idx, title), 'w') as fout:\n fout.write(gjf_head(title))\n fout.write(cluster.to_gjf())\n fout.write('\\n') # add a \\n just in case\n\ncenter = Cluster('center', fname='center.xsd')\nneighbors = Cluster('neighbors', fname='neighbor.xsd')\n\nfor i, mlcl in enumerate(neighbors.molecules):\n neighbor = Cluster('neighbor%d' % i, molecules=[mlcl])\n write_gjf(i+1, 'mol1', center)\n write_gjf(i+1, 'mol2', neighbor)\n write_gjf(i+1, '2mol', center + neighbor)\n with open('%d/H.inp' % (i+1), 'w') as hout:\n hout.write('%d\\n%d\\n' % (center.homo, neighbor.homo))\n with open('%d/L.inp' % (i+1), 'w') as lout:\n lout.write('%d\\n%d\\n' % (center.lumo, neighbor.lumo))\n\n# write a cluster file just for debugging\n(center + neighbors).to_xyz(True)\n","sub_path":"TransferIntergral/build_dimer.py","file_name":"build_dimer.py","file_ext":"py","file_size_in_byte":1231,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"649995635","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Jun 13 14:52:36 2020\n\n@author: tonedogga\n\"\"\"\nimport numpy as np\nimport pandas as pd\nimport datetime as dt\nfrom pandas.plotting import scatter_matrix\n\nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\nimport matplotlib.cm as cm\nimport seaborn as sns\n#print(\"matplotlib:\",mpl.__version__)\n\nmpl.rc('axes', labelsize=14)\nmpl.rc('xtick', labelsize=12)\nmpl.rc('ytick', labelsize=12)\n\n\nwwjamsxls=\"Ww_jams_scan_data_300520.xlsx\"\n\n\n\n# self.start_point=0\n \n \ndef add_dates_back(df,all_dates):\n return pd.concat((df,all_dates_df),axis=1) #,on='ww_scan_week',how='right')\n\n \n \n \n#column_list=list([\"coles_scan_week\",\"bb_total_units\",\"bb_promo_disc\",\"sd_total_units\",\"sd_promo_disc\",\"bm_total_units\",\"bm_promo_disc\"]) \n \ncol_dict=dict({0:\"WW_scan_week\",\n 1:\"BB_off_promo_sales\",\n 2:\"BB_on_promo_sales\",\n 3:\"SD_off_promo_sales\",\n 4:\"SD_on_promo_sales\",\n 5:\"BM_off_promo_sales\",\n 6:\"BM_on_promo_sales\"})\n \ndf=pd.read_excel(wwjamsxls,-1,skiprows=[0,1,2]).T.reset_index() #,header=[0,1,2]) #,skip_rows=0) #[column_list] #,names=column_list) #,sheet_name=\"AttacheBI_sales_trans\",use_cols=range(0,16),verbose=True) # -1 means all rows #print(df)\n#print(\"before\",df)\ndf = df.rename(col_dict,axis='index').T\n\ndf['WW_scan_week']=pd.to_datetime(df['WW_scan_week'],format=\"%d/%m/%y\")\n#df['coles_scan_week']=df[\"date\"] #.strftime(\"%Y-%m-%d\")\ndf.fillna(0.0,inplace=True)\ndf.drop_duplicates(keep='first', inplace=True)\n#df.replace(0.0, np.nan, inplace=True)\n#print(\"after\",df)\n\ndf=df.sort_values(by=['WW_scan_week'], ascending=True)\ndf=df.set_index('WW_scan_week') \ndf=df.astype(np.float32) #,inplace=True)\ndf['weekno']= np.arange(len(df))\nprint(\"final\",df,df.T)\n\ndf['BB_on_promo']=(df['BB_on_promo_sales']>0.0)\ndf['SD_on_promo']=(df['SD_on_promo_sales']>0.0)\ndf['BM_on_promo']=(df['BM_on_promo_sales']>0.0)\n\ndf['BB_total_sales']=df['BB_off_promo_sales']+df['BB_on_promo_sales']\ndf['SD_total_sales']=df['SD_off_promo_sales']+df['SD_on_promo_sales']\ndf['BM_total_sales']=df['BM_off_promo_sales']+df['BM_on_promo_sales']\n\n\n\ndf.replace(0.0, np.nan, inplace=True)\n\n#sns.lmplot(x='weekno',y='BB_total_sales',data=df,col='SD_on_promo',hue='BM_on_promo') #,fit_reg=True,robust=True,legend=True) \n#sns.lmplot(x='weekno',y='BB_total_sales',data=df,col='BM_on_promo',hue='SD_on_promo') #,fit_reg=True,robust=True,legend=True) \nsns.lmplot(x='weekno',y='BB_total_sales',data=df,col='SD_on_promo',hue='BB_on_promo') #,fit_reg=True,robust=True,legend=True) \nsns.lmplot(x='weekno',y='BB_total_sales',data=df,col='BM_on_promo',hue='BB_on_promo')\n\nsns.lmplot(x='weekno',y='BB_total_sales',data=df,col='BM_on_promo',hue='SD_on_promo') #,fit_reg=True,robust=True,legend=True) \nsns.lmplot(x='weekno',y='BB_total_sales',data=df,col='SD_on_promo',hue='BM_on_promo')\n\n","sub_path":"brand_index_WW_v2_00.py","file_name":"brand_index_WW_v2_00.py","file_ext":"py","file_size_in_byte":2941,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"618654590","text":"#!/usr/bin/env python\n#-*- coding:utf-8 -*-\n\n# importing a useful library -- feel free to add any others you find necessary\nimport hashlib\nimport string\n\n# this will work if you place this script in your writeup folder\nwordlist = open(\"../probable-v2-top1575.txt\", 'r')\nhashfile = open(\"../hashes\", 'r')\n\n# a string equal to 'abcdefghijklmnopqrstuvwxyz'.\nsalts = string.ascii_lowercase\n\nwords = []\nfor word in wordlist:\n words.append(word.rstrip())\n\nhashes = []\nfor h in hashfile:\n hashes.append(h.rstrip())\n\nfor salt in salts:\n for word in words:\n salted = str(salt + word).encode('utf-8')\n computed_hash = hashlib.sha512(salted).hexdigest()\n for h in hashes:\n if h == computed_hash:\n print(\"---------------------------\")\n print(\"Match found!\")\n print(\"Hash: \" + h)\n print(\"Salt: \" + salt)\n print(\"Password: \" + word)\n print(\"---------------------------\")\n","sub_path":"week/9/writeup/part1.py","file_name":"part1.py","file_ext":"py","file_size_in_byte":988,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"491071646","text":"import taichi as ti\nimport os\nimport sys\nimport functools\n\nsys.path.append(os.path.join(ti.core.get_repo_dir(), 'tests', 'python'))\n\nfrom fuse_test_template import template_fuse_dense_x2y2z, \\\n template_fuse_reduction\n\n\n# Note: this is a short-term solution. In the long run we need to think about how to reuse pytest\ndef benchmark_async(func):\n @functools.wraps(func)\n def body():\n for arch in [ti.cpu, ti.cuda]:\n for async_mode in [True, False]:\n os.environ['TI_CURRENT_BENCHMARK'] = func.__name__\n ti.init(arch=arch, async_mode=async_mode)\n func()\n\n return body\n\n\n@benchmark_async\ndef fuse_dense_x2y2z():\n template_fuse_dense_x2y2z(size=100 * 1024**2,\n repeat=10,\n benchmark_repeat=10,\n benchmark=True)\n\n\n@benchmark_async\ndef fuse_reduction():\n template_fuse_reduction(size=100 * 1024**2,\n repeat=10,\n benchmark_repeat=10,\n benchmark=True)\n\n\n@benchmark_async\ndef fill_1d():\n a = ti.field(dtype=ti.f32, shape=100 * 1024**2)\n\n @ti.kernel\n def fill():\n for i in a:\n a[i] = 1.0\n\n ti.benchmark(fill, repeat=100)\n\n\n@benchmark_async\ndef fill_scalar():\n a = ti.field(dtype=ti.f32, shape=())\n\n @ti.kernel\n def fill():\n a[None] = 1.0\n\n ti.benchmark(fill, repeat=1000)\n\n\n@benchmark_async\ndef sparse_numpy():\n a = ti.field(dtype=ti.f32)\n b = ti.field(dtype=ti.f32)\n\n block_count = 64\n block_size = 32\n # a, b always share the same sparsity\n ti.root.pointer(ti.ij, block_count).dense(ti.ij, block_size).place(a, b)\n\n @ti.kernel\n def initialize():\n for i, j in ti.ndrange(block_count * block_size,\n block_count * block_size):\n if (i // block_size + j // block_size) % 4 == 0:\n a[i, j] = i + j\n\n @ti.kernel\n def saxpy(x: ti.template(), y: ti.template(), alpha: ti.f32):\n for i, j in x:\n y[i, j] = alpha * x[i, j] + y[i, j]\n\n def task():\n initialize()\n saxpy(a, b, 2)\n saxpy(b, a, 1.1)\n saxpy(b, a, 1.1)\n saxpy(a, b, 1.1)\n saxpy(a, b, 1.1)\n saxpy(a, b, 1.1)\n\n ti.benchmark(task, repeat=100)\n\n\nwith_autodiff = False # For some reason autodiff crashes with async.\nif with_autodiff:\n\n @benchmark_async\n def autodiff():\n\n n = 1024**2 * 10\n\n a = ti.field(dtype=ti.f32, shape=n, needs_grad=True)\n b = ti.field(dtype=ti.f32, shape=n)\n loss = ti.field(dtype=ti.f32, shape=(), needs_grad=True)\n\n @ti.kernel\n def compute_loss():\n for i in a:\n loss[None] += a[i]\n\n @ti.kernel\n def accumulate_grad():\n for i in a:\n b[i] += a.grad[i]\n\n def task():\n with ti.Tape(loss=loss):\n # The forward kernel of compute_loss should be completely eliminated (except for the last one)\n compute_loss()\n\n accumulate_grad()\n\n ti.benchmark(task, repeat=100)\n\n\n@benchmark_async\ndef stencil_reduction():\n a = ti.field(dtype=ti.f32)\n b = ti.field(dtype=ti.f32)\n total = ti.field(dtype=ti.f32, shape=())\n\n block_count = 1024\n block_size = 1024\n # a, b always share the same sparsity\n ti.root.pointer(ti.i, block_count).dense(ti.i, block_size).place(a, b)\n\n @ti.kernel\n def initialize():\n for i in range(block_size, (block_size - 1) * block_count):\n a[i] = i\n\n @ti.kernel\n def stencil():\n for i in a:\n b[i] = a[i - 1] + a[i] + a[i + 1]\n\n @ti.kernel\n def reduce():\n for i in a:\n total[None] += b[i]\n\n def task():\n for i in range(2):\n initialize()\n stencil()\n reduce()\n\n ti.benchmark(task, repeat=100)\n\n\n# TODO: add mpm_breakdown\n","sub_path":"benchmarks/async_cases.py","file_name":"async_cases.py","file_ext":"py","file_size_in_byte":3962,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"306227387","text":"#!/usr/bin/env python3\n# controller.py\n# author: Wrytus, Flosean\n#\n# purpose: The center controller to get the messages received from\n# nlu_server and responsible for controlling the devices\n\nimport os\nimport requests\nimport json\nimport time\nimport random\nimport subprocess\n#from device.DHT import Adafruit_DHT\n\ndef make_requests():\n req = {\n 'device': DEVICE_ID,\n 'event_type': 'request_cmd'\n }\n return requests.get('{}:{}'.format(SERVER_LOC, PORT), params=req)\n\ndef send_sensor():\n #temp, humidity = Adafruit_DHT.read_retry(Adafruit_DHT.DHT22, '18')\n out = subprocess.check_output(\"python2 /home/pi/Pistgut/RPi/device/DHT/examples/AdafruitDHT.py 22 18\", shell=True)\n temp, humidity = out.decode('utf8').strip()[1:-1].split(',')\n if float(temp) >= 30:\n ret = subprocess.call(\"python3 /home/pi/Pistgut/RPi/device/LED/LED_3_on.py\", shell=True)\n ret = subprocess.call(\"python3 /home/pi/Pistgut/RPi/device/LED/LED_1_off.py\", shell=True)\n ret = subprocess.call(\"python3 /home/pi/Pistgut/RPi/device/LED/LED_2_off.py\", shell=True)\n else:\n ret = subprocess.call(\"python3 /home/pi/Pistgut/RPi/device/LED/LED_3_off.py\", shell=True)\n\n print(temp, humidity)\n ack_status = {\n 'device_id': DEVICE_ID,\n 'event_type': 'sensor',\n 'temp': temp,\n 'humidity': humidity\n }\n ack = requests.post('{}:{}'.format(SERVER_LOC, PORT), data=ack_status)\n print('send sensor')\n\ndef switch(obj):\n cmd_type = obj['intent']\n\n ### Target: light\n if cmd_type == 'light':\n target_led = obj['command']['object']\n if target_led == 'all':\n if obj['command']['action'] == 'turn_on':\n ret = subprocess.call(\"python3 /home/pi/Pistgut/RPi/device/LED/LED_1_on.py\", shell=True)\n ret = subprocess.call(\"python3 /home/pi/Pistgut/RPi/device/LED/LED_2_on.py\", shell=True)\n elif obj['command']['action'] == 'turn_off':\n ret = subprocess.call(\"python3 /home/pi/Pistgut/RPi/device/LED/LED_1_off.py\", shell=True)\n ret = subprocess.call(\"python3 /home/pi/Pistgut/RPi/device/LED/LED_2_off.py\", shell=True)\n else:\n if obj['command']['action'] == 'turn_on':\n ret = subprocess.call(\"python3 /home/pi/Pistgut/RPi/device/LED/{}_on.py\".format(target_led), shell=True)\n elif obj['command']['action'] == 'turn_off':\n ret = subprocess.call(\"python3 /home/pi/Pistgut/RPi/device/LED/{}_off.py\".format(target_led), shell=True)\n\n ### Target: music\n elif cmd_type == 'music':\n music_list = os.listdir(MUSIC_PATH)\n song = obj['command']['song']\n global NOW_PLAYING\n if obj['command']['action'] == 'play':\n song_name = song['name']\n song_artist = song['artist']\n target = '周杰倫_七里香.mp3'\n for f in music_list:\n if song_name in f and song_artist in f:\n target = f\n print(target)\n\n args = []\n args.append('mpg321')\n args.append(os.path.join(MUSIC_PATH, target))\n print(args)\n\n if NOW_PLAYING is not None:\n NOW_PLAYING.kill()\n\n NOW_PLAYING = subprocess.Popen(args)\n print(NOW_PLAYING.pid)\n elif obj['command']['action'] == 'stop':\n NOW_PLAYING.kill()\n\ndef main():\n global CNT\n sensor_time_interval = 0\n while CNT > 0:\n print('timestamp: {}'.format(CNT))\n res = make_requests()\n print(res.content)\n if len(res.text) == 0:\n pass\n elif res.status_code == 200 and len(res.text) != 0:\n ack_status = {\n 'device_id': DEVICE_ID,\n 'event_type': 'ack',\n 'command_id': json.loads(res.text)['command_id'],\n 'status': 'success'\n }\n ack = requests.post('{}:{}'.format(SERVER_LOC, PORT), data=ack_status)\n switch(json.loads(res.text))\n else:\n ack_status = {\n 'event_type': 'error',\n 'details': 'requests_get_error'\n }\n ack = requests.post('{}:{}'.format(SERVER_LOC, PORT), data=ack_status)\n\n if sensor_time_interval % (SENSOR_TIME_LIMIT * 2) == 0:\n send_sensor()\n sensor_time_interval = 0\n sensor_time_interval += 1\n\n CNT -= 1\n time.sleep(0.5)\n\n\nif __name__ == '__main__':\n SERVER_LOC = 'http://52.197.239.243'\n PORT = '6789'\n DEVICE_ID = 'A'\n MUSIC_PATH = '/home/pi/Pistgut/RPi/device/Music'\n NOW_PLAYING = None\n CNT = 1e10\n SENSOR_TIME_LIMIT = 3\n\n main()\n","sub_path":"RPi/controller.py","file_name":"controller.py","file_ext":"py","file_size_in_byte":4696,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"231820503","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# @Time : 2019/11/15 21:23\n# @Author : smyyan & ghoskno\n# @File : brybt.py\n# @Software: PyCharm\n\nimport time\nimport os\nimport re\nimport pickle\nfrom io import BytesIO\nimport platform\nfrom contextlib import ContextDecorator\nfrom PIL import Image\nimport requests\nfrom requests.cookies import RequestsCookieJar\nfrom bs4 import BeautifulSoup\n\nfrom decaptcha import DeCaptcha\n\n# ##################需要配置的变量###################\n_username = '用户名'\n_passwd = '密码'\n_transmission_user_pw = 'user:passwd' # transmission的用户名和密码,按照格式填入\n_windows_download_path = './torrent' # windows测试下载种子路径\n_linux_download_path = '' # linux服务器下载种子的路径\n_torrent_infos = './torrent.pkl' # 种子信息保存文件路径\nmax_torrent = 20 # 最大种子数\nsearch_time = 120 # 轮询种子时间,默认120秒\n# ##################################################\n_decaptcha_model = 'captcha_classifier.pkl' # 验证码识别模型\n_cookies_save_path = 'ByrbtCookies.pickle' # cookies保存路径\n\n# 判断平台\nosName = platform.system()\nif osName == 'Windows':\n osName = 'Windows'\nelif osName == 'Linux':\n osName = 'Linux'\nelse:\n raise Exception('not support this system : {}'.format(osName))\n\n# 常量\n_BASE_URL = 'https://bt.byr.cn/'\n_tag_map = {\n 'free': '免费',\n 'twoup': '2x上传',\n 'twoupfree': '免费&2x上传',\n 'halfdown': '50%下载',\n 'twouphalfdown': '50%下载&2x上传',\n 'thirtypercent': '30%下载',\n}\n_cat_map = {\n '电影': 'movie',\n '剧集': 'episode',\n '动漫': 'anime',\n '音乐': 'music',\n '综艺': 'show',\n '游戏': 'game',\n '软件': 'software',\n '资料': 'material',\n '体育': 'sport',\n '记录': 'documentary',\n}\n\n\n# 全局变量\ndownload_path = None\nbyrbt_cookies = None\n\nif osName == 'Windows':\n download_path = os.path.abspath(_windows_download_path)\nelif osName == 'Linux':\n download_path = os.path.abspath(_linux_download_path)\nelse:\n raise Exception('not support system! {}'.format(osName))\n\ndecaptcha = DeCaptcha()\ndecaptcha.load_model(_decaptcha_model)\n\nold_torrent = list()\nif os.path.exists(_torrent_infos):\n old_torrent = pickle.load(open(_torrent_infos, 'rb'))\n\n\ndef get_url(url):\n return _BASE_URL + url\n\n\ndef login():\n url = get_url('login.php')\n headers = {\n 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.113 Safari/537.36'}\n\n session = requests.session()\n for i in range(5):\n login_content = session.get(url)\n login_soup = BeautifulSoup(login_content.text, 'lxml')\n\n img_url = _BASE_URL + login_soup.select('#nav_block > form > table > tr:nth-of-type(3) img')[0].attrs['src']\n img_file = Image.open(BytesIO(session.get(img_url).content))\n\n captcha_text = decaptcha.decode(img_file)\n\n login_res = session.post(get_url('takelogin.php'), headers=headers,\n data=dict(username=_username, password=_passwd, imagestring=captcha_text,\n imagehash=img_url.split('=')[-1]))\n if '最近消息' in login_res.text:\n cookies = {}\n for k, v in session.cookies.items():\n cookies[k] = v\n\n with open(_cookies_save_path, 'wb') as f:\n pickle.dump(cookies, f)\n return cookies\n\n time.sleep(1)\n\n raise Exception('Cat not get Cookies!')\n\n\ndef load_cookie():\n global byrbt_cookies\n if os.path.exists(_cookies_save_path):\n print('find {}, loading cookies'.format(_cookies_save_path))\n read_path = open(_cookies_save_path, 'rb')\n byrbt_cookies = pickle.load(read_path)\n else:\n print('not find {}, get cookies...'.format(_cookies_save_path))\n byrbt_cookies = login()\n\n return byrbt_cookies\n\n\ndef _get_tag(tag):\n try:\n if tag == '':\n return ''\n else:\n tag = tag.split('_')[0]\n\n return _tag_map[tag]\n except KeyError:\n return ''\n\n\ndef _get_torrent_info(table):\n assert isinstance(table, list)\n torrent_infos = list()\n for item in table:\n torrent_info = dict()\n tds = item.select('td')\n cat = tds[0].select('img')[0].attrs['title']\n main_td = tds[1].select('table > tr > td')[0]\n href = main_td.select('a')[0].attrs['href']\n seed_id = re.findall(r'id=(\\d+)&', href)[0]\n title = main_td.text\n title = title.split('\\n')\n if len(title) == 2:\n sub_title = title[1]\n title = title[0]\n else:\n sub_title = ''\n title = title[0]\n\n tags = set([font.attrs['class'][0] for font in main_td.select('b > font') if 'class' in font.attrs.keys()])\n if '' in tags:\n tags.remove('')\n\n is_seeding = len(main_td.select('img[src=\"pic/seeding.png\"]')) > 0\n is_finished = len(main_td.select('img[src=\"pic/finished.png\"]')) > 0\n\n is_hot = False\n if 'hot' in tags:\n is_hot = True\n tags.remove('hot')\n is_new = False\n if 'new' in tags:\n is_new = True\n tags.remove('new')\n is_recommended = False\n if 'recommended' in tags:\n is_recommended = True\n tags.remove('recommended')\n\n if 'class' in tds[1].select('table > tr')[0].attrs.keys():\n tag = _get_tag(tds[1].select('table > tr')[0].attrs['class'][0])\n else:\n tag = ''\n\n file_size = tds[6].text.split('\\n')\n\n seeding = int(tds[7].text) if tds[7].text.isdigit() else -1\n\n downloading = int(tds[8].text) if tds[8].text.isdigit() else -1\n\n finished = int(tds[9].text) if tds[9].text.isdigit() else -1\n\n torrent_info['cat'] = cat\n torrent_info['is_hot'] = is_hot\n torrent_info['tag'] = tag\n torrent_info['is_seeding'] = is_seeding\n torrent_info['is_finished'] = is_finished\n torrent_info['seed_id'] = seed_id\n torrent_info['title'] = title\n torrent_info['sub_title'] = sub_title\n torrent_info['seeding'] = seeding\n torrent_info['downloading'] = downloading\n torrent_info['finished'] = finished\n torrent_info['file_size'] = file_size\n torrent_info['is_new'] = is_new\n torrent_info['is_recommended'] = is_recommended\n torrent_infos.append(torrent_info)\n\n return torrent_infos\n\n\ndef get_torrent(torrent_infos, tags):\n free_infos = list()\n for torrent_info in torrent_infos:\n if torrent_info['tag'] in tags:\n free_infos.append(torrent_info)\n\n return free_infos\n\n\ndef get_ok_torrent(torrent_infos):\n ok_infos = list()\n if len(torrent_infos) >= 20: # 遇到free或者免费种子太多了,择优选取\n print('ok种子过多,怀疑free了。。。')\n for torrent_info in torrent_infos:\n if torrent_info['seed_id'] in old_torrent:\n continue\n if 'GB' not in torrent_info['file_size'][0]:\n continue\n if torrent_info['seeding'] <= 0 or torrent_info['downloading'] < 0:\n continue\n if torrent_info['seeding'] != 0 and float(torrent_info['downloading']) / float(\n torrent_info['seeding']) < 20:\n continue\n file_size = torrent_info['file_size'][0]\n file_size = file_size.replace('GB', '')\n file_size = float(file_size.strip())\n if file_size < 20.0:\n continue\n ok_infos.append(torrent_info)\n else:\n for torrent_info in torrent_infos:\n if torrent_info['seed_id'] in old_torrent:\n continue\n if 'GB' not in torrent_info['file_size'][0]:\n continue\n if torrent_info['seeding'] <= 0 or torrent_info['downloading'] < 0:\n continue\n if torrent_info['seeding'] != 0 and float(torrent_info['downloading']) / float(torrent_info['seeding']) < \\\n 0.6:\n continue\n\n ok_infos.append(torrent_info)\n\n return ok_infos\n\n\ndef download_torrent(op_str):\n cookie_jar = RequestsCookieJar()\n for k, v in byrbt_cookies.items():\n cookie_jar[k] = v\n headers = {\n 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.113 Safari/537.36'}\n\n id_re = re.findall(r'dl (\\d+)', op_str, re.I)\n if len(id_re) == 0:\n print('no such torrent')\n return\n torrent_id = id_re[0]\n\n download_url = 'download.php?id={}'.format(torrent_id)\n download_url = get_url(download_url)\n try:\n torrent = requests.get(download_url, cookies=cookie_jar, headers=headers)\n torrent_file_name = str(\n torrent.headers['Content-Disposition'].split(';')[1].strip().split('=')[-1][1:-1].encode('ascii',\n 'ignore').decode(\n 'ascii')).replace(' ', '#')\n with open(os.path.join(download_path, torrent_file_name), 'wb') as f:\n f.write(torrent.content)\n\n except:\n print('login failed!')\n return False\n\n index = 20\n while index > 0:\n if os.path.exists(os.path.join(download_path, torrent_file_name)):\n if osName == 'Linux':\n torrent_file_path = os.path.join(download_path, torrent_file_name)\n cmd_str = \"transmission-remote -n '{}' -a {}\".format(_transmission_user_pw, torrent_file_path)\n ret_val = os.system(cmd_str)\n if ret_val != 0:\n print('script `{}` returns {}'.format(cmd_str, ret_val))\n print('下载失败')\n else:\n print('下载成功')\n else:\n print('下载成功')\n return\n else:\n time.sleep(0.5)\n index = index - 1\n print('下载失败')\n return\n\n\ndef execCmd(cmd):\n r = os.popen(cmd)\n text = r.read()\n r.close()\n return text\n\n\ndef op_help():\n return \"\"\"\n byrbt bot: a bot that handles basic usage of bt.byr.cn\n usage:\n 1. main - run main program\n\n 2. download - download and start torrent file\n i.e. dl $id\n $id - torrent id, acquired by `ls` or `se`\n\n 3. list torrent status - list the torrent files status, merely call `transmission-remote -l`\n i.e. tls\n\n 4. remove torrent - remove specific torrent job, merely call `transmission-remote -t $id -r`\n i.e. trm $torrent_id\n\n 5. refresh - refresh cookies\n 6. help - print this message\n 7. exit\n \"\"\"\n\n\ndef list_torrent():\n os.system('transmission-remote -n \"{}\" -l'.format(_transmission_user_pw))\n\n\ndef get_info(text):\n text = text.split('\\n')\n sum_to = text[-2]\n text = text[1:-2]\n text_s = list()\n for t in text:\n ts = t.split()\n torrent = dict()\n torrent['id'] = ts[0]\n torrent['done'] = ts[1]\n torrent['size'] = ts[2] + ts[3]\n if 'GB' not in torrent['size']:\n torrent['size'] = '1GB'\n torrent['name'] = ts[-1]\n text_s.append(torrent)\n sum_to = sum_to.split()\n sum_size = sum_to[1] + sum_to[2]\n if 'GB' in sum_size or 'TB' in sum_size:\n pass\n else:\n sum_size = '1GB'\n\n return text_s, sum_size\n\n\ndef remove_torrent(op_str):\n id_re = re.findall(r'trm (\\d+)', op_str, re.I)\n if len(id_re) == 0:\n print('no such torrent id')\n return\n id_str = id_re[0]\n id_str = str(id_str)\n\n text = execCmd('transmission-remote -n \"{}\" -l'.format(_transmission_user_pw))\n text_s, sum_size = get_info(text)\n flag = False\n for to_info in text_s:\n if to_info['id'] == id_str:\n res = execCmd('transmission-remote -n \"{}\" -t {} --remove-and-delete'.format(_transmission_user_pw, id_str))\n if \"success\" not in res:\n print('remove torrent fail:')\n for k, v in to_info.items():\n print('{} : {}'.format(k, v))\n if os.path.exists(os.path.join(download_path, to_info['name'])):\n cmd_str = 'rm -rf {}'.format(os.path.join(download_path, to_info['name']))\n ret_val = os.system(cmd_str)\n if ret_val != 0:\n print('script `{}` returns {}'.format(cmd_str, ret_val))\n print(\n 'remove torrent from transmission-daemon success, but cat not remove it from disk!')\n else:\n print('remove torrent from transmission-daemon success!')\n else:\n print('remove torrent from transmission-daemon success!')\n flag = True\n break\n\n if flag is False:\n print('cat find this torrent id in torrent list, please use cmd \"tls\" ')\n\n\nclass TorrentBot(ContextDecorator):\n def __init__(self):\n super(TorrentBot, self).__init__()\n self.torrent_url = get_url('torrents.php')\n self.cookie_jar = RequestsCookieJar()\n for k, v in byrbt_cookies.items():\n self.cookie_jar[k] = v\n self.headers = {\n 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.113 Safari/537.36'}\n self.tags = ['免费', '免费&2x上传']\n\n def __enter__(self):\n print('启动byrbt_bot!')\n return self\n\n def __exit__(self, exc_type, exc_val, exc_tb):\n print('退出')\n print('保存数据')\n pickle.dump(old_torrent, open(_torrent_infos, 'wb'), protocol=2)\n\n def remove(self):\n text = execCmd('transmission-remote -n \"{}\" -l'.format(_transmission_user_pw))\n text_s, sum_size = get_info(text)\n if len(text_s) <= max_torrent:\n return\n torrent_len = len(text_s)\n while torrent_len > max_torrent:\n text_s.sort(key=lambda x: int(x['id'].strip(\"*\")), reverse=False)\n remove_torrent_info = text_s.pop(0)\n res = execCmd('transmission-remote -n \"{}\" -t {} --remove-and-delete'.format(_transmission_user_pw,\n remove_torrent_info['id']))\n if \"success\" not in res:\n print('remove torrent fail:')\n for k, v in remove_torrent_info.items():\n print('{} : {}'.format(k, v))\n\n if os.path.exists(os.path.join(download_path, remove_torrent_info['name'])):\n cmd_str = 'rm -rf {}'.format(os.path.join(download_path, remove_torrent_info['name']))\n ret_val = os.system(cmd_str)\n if ret_val != 0:\n print('script `{}` returns {}'.format(cmd_str, ret_val))\n print('remove torrent from transmission-daemon success, but cat not remove it from disk!')\n else:\n print('remove {} from transmission-daemon success!'.format(remove_torrent_info['name']))\n else:\n print('remove {} from transmission-daemon success!'.format(remove_torrent_info['name']))\n torrent_len = torrent_len - 1\n\n def download(self, torrent_id):\n global byrbt_cookies\n download_url = 'download.php?id={}'.format(torrent_id)\n download_url = get_url(download_url)\n torrent_file_name = None\n for i in range(5):\n try:\n torrent = requests.get(download_url, cookies=self.cookie_jar, headers=self.headers)\n torrent_file_name = str(\n torrent.headers['Content-Disposition'].split(';')[1].strip().split('=')[-1][1:-1].encode('ascii',\n 'ignore').decode(\n 'ascii')).replace(' ', '#')\n print(torrent_file_name)\n with open(os.path.join(download_path, torrent_file_name), 'wb') as f:\n f.write(torrent.content)\n break\n\n except:\n print('login failed')\n byrbt_cookies = load_cookie()\n self.__init__()\n continue\n\n index = 20\n while index > 0:\n if torrent_file_name is not None and os.path.exists(os.path.join(download_path, torrent_file_name)):\n if osName == 'Linux':\n torrent_file_path = os.path.join(download_path, torrent_file_name)\n cmd_str = 'transmission-remote -n \"{}\" -a {}'.format(_transmission_user_pw, torrent_file_path)\n ret_val = os.system(cmd_str)\n if ret_val != 0:\n print('script `{}` returns {}'.format(cmd_str, ret_val))\n return True\n else:\n print('添加种子: {}'.format(torrent_file_name))\n\n old_torrent.append(torrent_id)\n else:\n pass\n return True\n else:\n time.sleep(0.5)\n index = index - 1\n\n return True\n\n def start(self):\n global byrbt_cookies\n while True:\n print('扫描种子列表')\n try:\n torrents_soup = BeautifulSoup(\n requests.get(self.torrent_url, cookies=self.cookie_jar, headers=self.headers).content)\n torrent_table = torrents_soup.select('.torrents > form > tr')[1:]\n pass\n except:\n byrbt_cookies = load_cookie()\n self.__init__()\n continue\n torrent_infos = _get_torrent_info(torrent_table)\n\n free_infos = get_torrent(torrent_infos, self.tags)\n print('种子列表:')\n for i, info in enumerate(free_infos):\n print('{} : {} {} {}'.format(i, info['seed_id'], info['file_size'], info['title']))\n ok_torrent = get_ok_torrent(free_infos)\n print('可用种子:')\n for i, info in enumerate(ok_torrent):\n print('{} : {} {} {}'.format(i, info['seed_id'], info['file_size'], info['title']))\n for torrent in ok_torrent:\n if self.download(torrent['seed_id']) is False:\n print('{} download fail'.format(torrent['title']))\n continue\n self.remove()\n time.sleep(search_time)\n\n\ndef main():\n with TorrentBot() as byrbt_bot:\n byrbt_bot.start()\n\n\nif __name__ == '__main__':\n byrbt_cookies = load_cookie()\n\n print(op_help())\n while True:\n action_str = input()\n if action_str == 'refresh':\n print('refresh cookie by login!')\n byrbt_cookies = login()\n elif action_str == 'exit':\n break\n elif action_str == 'help':\n print(op_help())\n elif action_str == 'main':\n main()\n elif action_str.startswith('dl'):\n download_torrent(action_str)\n elif action_str.startswith('tls'):\n list_torrent()\n elif action_str.startswith('trm'):\n remove_torrent(action_str)\n else:\n print('invalid operation')\n print(op_help())\n","sub_path":"byrbt.py","file_name":"byrbt.py","file_ext":"py","file_size_in_byte":19555,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"146441259","text":"import logging\n\nfrom django.test import TestCase\n\nfrom translations.models import Translation\nfrom translations.models.factories import (\n ProjectFactory,\n LanguageFactory,\n)\nfrom translations.utils.sync_project import sync_project\n\n\nclass SyncProjectTestCase(TestCase):\n def setUp(self):\n logging.disable(logging.CRITICAL)\n\n english = LanguageFactory(name=\"English\", language_code=\"en\")\n swedish = LanguageFactory(name=\"Swedish\", language_code=\"sv\")\n\n self.project = ProjectFactory(\n name=\"Test Git Project\",\n repository_name=\"zetkin/translators-interface\",\n locale_files_path=\"backend/translations/utils/tests/mock_files/sync_project\",\n languages=(english, swedish),\n )\n\n def test_create_translations_from_git(self):\n sync_project(self.project)\n # Check 8 Swedish translations created\n swedish_translations = Translation.objects.filter(language__language_code=\"sv\")\n self.assertEqual(len(swedish_translations), 8)\n # Check 8 English translations created\n english_translations = Translation.objects.filter(language__language_code=\"en\")\n self.assertEqual(len(english_translations), 8)\n\n # Check a couple dotpath\n english_home_page_header_title = Translation.objects.get(\n language__language_code=\"en\",\n file_path=\"./home_page/en.yaml\",\n object_path=\"header.title\",\n )\n self.assertEqual(\n english_home_page_header_title.dotpath, \"home_page.header.title\"\n )\n self.assertEqual(english_home_page_header_title.text, \"Edit translations here\")\n\n swedish_home_page_header_title = Translation.objects.get(\n language__language_code=\"sv\",\n file_path=\"./home_page/sv.yaml\",\n object_path=\"header.title\",\n )\n self.assertEqual(\n swedish_home_page_header_title.dotpath, \"home_page.header.title\"\n )\n self.assertEqual(\n swedish_home_page_header_title.text, \"Redigera översättningar här\"\n )\n","sub_path":"backend/translations/utils/tests/test_sync_project.py","file_name":"test_sync_project.py","file_ext":"py","file_size_in_byte":2098,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"54032231","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Aug 4 09:00:28 2016\n\n@author: matthieu\n\"\"\"\n\nimport re\n\nfileName = '/data/contracteo/extractContract/data/raw/CGG_CDI_GUILLOUET.txt'\n\nwith open(fileName,'r',encoding='utf-8') as f:\n strCt = f.read()\n f.close()\n\n\nstructCt = []\n\n#### white spaces, line breaks, tabulations\nwhiteSpace = re.compile('\\s')\nwhitePos = []\nfor wh in whiteSpace.finditer(strCt):\n whitePos.append([wh.span()[0],wh.span()[1]])\n \nfirstElt = whitePos[0][0]\nwhitePos1 = []\nii = 0\nfor iWh in range(0,len(whitePos)-1):\n if whitePos[iWh][1]!=whitePos[iWh+1][0]:\n whitePos1.append([firstElt, whitePos[iWh][1]])\n firstElt = whitePos[iWh+1][0]\n \nwhitePos1.append([firstElt, whitePos[-1][1]])\n\nif whitePos1[0][0]==0:\n firstElt = whitePos[0][1]\n iWhite = 1\nelse:\n firstElt = 0\n iWhite = 0\n\nfor iWh in range(iWhite,len(whitePos1)):\n word = strCt[firstElt:whitePos1[iWh][0]]\n structCt.append([firstElt, whitePos1[iWh][0], word])\n firstElt = whitePos1[iWh][1]\n\nif firstElt=word[0] and begin=word[0] and aposPos=len(structCt):\n word = structCt[iW]\n \n if word[2].isdigit() and not inDigit:\n inDigit = True\n beginDigit = iW\n iW += 1\n \n elif inDigit and not word[2].isdigit():\n begin = structCt[beginDigit][0]\n end = structCt[iW-1][1]\n newElt = [begin, end, strCt[begin:end], 'DIGIT']\n for iDelete in range(beginDigit,iW):\n structCt.pop(beginDigit)\n structCt.insert(beginDigit,newElt)\n inDigit = False\n iW -= (iW-beginDigit)-1\n \n else:\n iW += 1\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n","sub_path":"tokenizer.py","file_name":"tokenizer.py","file_ext":"py","file_size_in_byte":5066,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"316261833","text":"'''\nCreated on 10 de Jan de 2012\n\n@author: ricardoosorio\n'''\n\n# --------------------------------------------\n# created by Ricardo Osorio\n# update 11/10/2011\n# --------------------------------------------\nfrom libDB import *\nfrom libXML import *\nfrom xml.dom.minidom import parseString,Document\nfrom encoding import smart_str\n\nclass dbFenotipos(object):\n \"\"\"dbFenotipos Class. Used to manage fenotipos\"\"\"\n \n fen_name = None\n fen_w = None\n fen_n = None\n fen_country = None\n fen_wineregion = None\n fen_vineyard = None\n fen_samplingsite = None\n fen_grapevar = None\n fen_ScAAT1_1 = None\n fen_ScAAT1_2 = None\n fen_ScAAT2_1 = None\n fen_ScAAT2_2 = None\n fen_ScAAT3_1 = None\n fen_ScAAT3_2 = None\n fen_ScAAT4_1 = None\n fen_ScAAT4_2 = None\n fen_ScAAT5_1 = None\n fen_ScAAT5_2 = None\n fen_ScAAT6_1 = None\n fen_ScAAT6_2 = None\n fen_C4_1 = None\n fen_C4_2 = None\n fen_C5_1 = None\n fen_C5_2 = None\n fen_C11_1 = None\n fen_C11_2 = None\n fen_YPL009c_1 = None\n fen_YPL009c_2 = None\n fen_YOR267c_1 = None\n fen_YOR267c_2 = None\n\n def __init__(self, *args):\n \"\"\"dbfenotipos is able to create a new fenotipo.\n\n STANDARD OPTIONS TO CREATE\n\n name\n\n \"\"\"\n if len(args) == 1:\n if self.create(args[0]) :\n dg(2, \"Loaded.\")\n \n def create(self, _fen_name):\n if self.isFenotipo(_fen_name) :\n dg(2, \"create() | Cannot create fenotipo. Fenotipo \" + _fen_name + \" already exists.\")\n else:\n self.fen_name = _fen_name\n database = db()\n database.sql(\"INSERT INTO FENOTIPOS (COLNUM) VALUES ('\" + _fen_name + \"' )\")\n self.fetch(_fen_name)\n dg(2, \"Fenotipo \" + _fen_name + \" created.\")\n\n def getname(self):\n if self.fen_name == None:\n dg(3, \"getfenotipo | No fenotipo loaded.\")\n return \"\"\n else :\n return self.fen_name\n # -- menu Properties\n \n def isFenotipo(self, _fen_name):\n database = db()\n row = database.query(\"SELECT * FROM FENOTIPOS WHERE (COLNUM='\" + _fen_name + \"')\")\n dg(2, \"isFenotipo | ROW: \" + str(row))\n return row != []\n\n def similars(self,argument):\n database = db()\n arg = str(argument)\n \n query = \"SELECT * FROM FENOTIPOS WHERE (COLNUM LIKE '%\" + arg + \"%')\"\n\n allrows = database.query(query)\n if allrows == [] :\n dg(2, \"Fenotipo \" + arg + \" does not exist.\")\n return None\n else :\n return allrows\n\n \n def load(self,row):\n self.fen_name = row[1]\n self.fen_w = row[2]\n self.fen_n = row[3]\n self.fen_country = row[4]\n self.fen_wineregion = row[5]\n self.fen_vineyard = row[6]\n self.fen_samplingsite = row[7]\n self.fen_grapevar = row[8]\n self.fen_ScAAT1_1 = row[9]\n self.fen_ScAAT1_2 = row[10]\n self.fen_ScAAT2_1 = row[11]\n self.fen_ScAAT2_2 = row[12]\n self.fen_ScAAT3_1 = row[13]\n self.fen_ScAAT3_2 = row[14]\n self.fen_ScAAT4_1 = row[15]\n self.fen_ScAAT4_2 = row[16]\n self.fen_ScAAT5_1 = row[17]\n self.fen_ScAAT5_2 = row[18]\n self.fen_ScAAT6_1 = row[19]\n self.fen_ScAAT6_2 = row[20]\n self.fen_C4_1 = row[21]\n self.fen_C4_2 = row[22]\n self.fen_C5_1 = row[23]\n self.fen_C5_2 = row[24]\n self.fen_C11_1 = row[25]\n self.fen_C11_2 = row[26]\n self.fen_YPL009c_1 = row[27]\n self.fen_YPL009c_2 = row[28]\n self.fen_YOR267c_1 = row[29]\n self.fen_YOR267c_2 = row[30]\n \n \n def fetch(self, argument):\n database = db()\n arg = str(argument)\n \n query = \"SELECT * FROM FENOTIPOS WHERE (COLNUM='\" + arg + \"')\"\n\n allrows = database.query(query)\n if allrows == [] :\n dg(2, \"Fenotipo \" + arg + \" does not exist.\")\n return False\n else :\n row=allrows[0]\n self.fen_name = row[1]\n self.fen_w = row[2]\n self.fen_n = row[3]\n self.fen_country = row[4]\n self.fen_wineregion = row[5]\n self.fen_vineyard = row[6]\n self.fen_samplingsite = row[7]\n self.fen_grapevar = row[8]\n self.fen_ScAAT1_1 = row[9]\n self.fen_ScAAT1_2 = row[10]\n self.fen_ScAAT2_1 = row[11]\n self.fen_ScAAT2_2 = row[12]\n self.fen_ScAAT3_1 = row[13]\n self.fen_ScAAT3_2 = row[14]\n self.fen_ScAAT4_1 = row[15]\n self.fen_ScAAT4_2 = row[16]\n self.fen_ScAAT5_1 = row[17]\n self.fen_ScAAT5_2 = row[18]\n self.fen_ScAAT6_1 = row[19]\n self.fen_ScAAT6_2 = row[20]\n self.fen_C4_1 = row[21]\n self.fen_C4_2 = row[22]\n self.fen_C5_1 = row[23]\n self.fen_C5_2 = row[24]\n self.fen_C11_1 = row[25]\n self.fen_C11_2 = row[26]\n self.fen_YPL009c_1 = row[27]\n self.fen_YPL009c_2 = row[28]\n self.fen_YOR267c_1 = row[29]\n self.fen_YOR267c_2 = row[30]\n\n dg(2, \"Fenotipo \" + self.fen_name + \" loaded.\")\n return True\n\n\n def allcountries_repr(self):\n \n \n database = db()\n xml=dbXML(\"FENOTIPO\")\n xml.ident(\"COUNTRIES\")\n allrows = database.query(\"SELECT DISTINCT COUNTRY FROM FENOTIPOS\")\n for row in allrows:\n xml.add(\"COUNTRY\",str(row[0]))\n xml.unident()\n return str(xml)\n\n\n\n def public_repr(self):\n if (self.fen_name == None):\n xml=dbXML(\"FENOTIPO\",\"No Fenotipo Loaded\")\n return str(xml)\n else :\n \n xml=dbXML(\"FENOTIPO\")\n xml.add(\"COLNUM\",self.fen_name)\n xml.add(\"COUNTRY\",self.fen_country)\n xml.add(\"WINEREGION\",self.fen_wineregion)\n xml.add(\"VINEYARD\",self.fen_vineyard)\n xml.add(\"GRAPEVAR\",self.fen_grapevar)\n xml.add(\"SCAAT1-1\",self.fen_ScAAT1_1)\n xml.add(\"SCAAT1-2\",self.fen_ScAAT1_2)\n \n return str(xml)\n \n# -- Representation of dbUser itself in XML format\n# \n# \n# ...\n# \n \n def __repr__(self):\n if (self.fen_name == None):\n xml=dbXML(\"FENOTIPO\",\"No Fenotipo Loaded\")\n return str(xml)\n else :\n \n xml=dbXML(\"FENOTIPO\")\n xml.add(\"COLNUM\",self.fen_name)\n xml.add(\"W\",smart_str(self.fen_w))\n xml.add(\"N\",smart_str(self.fen_n))\n xml.add(\"COUNTRY\",self.fen_country)\n xml.add(\"WINEREGION\",self.fen_wineregion)\n xml.add(\"VINEYARD\",self.fen_vineyard)\n xml.add(\"SAMPLINGSITE\",self.fen_samplingsite)\n xml.add(\"GRAPEVAR\",self.fen_grapevar)\n xml.add(\"SCAAT1-1\",self.fen_ScAAT1_1)\n xml.add(\"SCAAT1-2\",self.fen_ScAAT1_2)\n xml.add(\"SCAAT2-1\",self.fen_ScAAT2_1)\n xml.add(\"SCAAT2-2\",self.fen_ScAAT2_2)\n xml.add(\"SCAAT3-1\",self.fen_ScAAT3_1)\n xml.add(\"SCAAT3-2\",self.fen_ScAAT3_2)\n xml.add(\"SCAAT4-1\",self.fen_ScAAT4_1)\n xml.add(\"SCAAT4-2\",self.fen_ScAAT4_2)\n xml.add(\"SCAAT5-1\",self.fen_ScAAT5_1)\n xml.add(\"SCAAT5-2\",self.fen_ScAAT5_2)\n xml.add(\"SCAAT6-1\",self.fen_ScAAT6_1)\n xml.add(\"SCAAT6-2\",self.fen_ScAAT6_2)\n xml.add(\"C4-1\",self.fen_C4_1)\n xml.add(\"C4-2\",self.fen_C4_2)\n xml.add(\"C5-1\",self.fen_C5_1)\n xml.add(\"C5-2\",self.fen_C5_2)\n xml.add(\"C11-1\",self.fen_C11_2)\n xml.add(\"C11-2\",self.fen_C11_1)\n xml.add(\"YPL009C-1\",self.fen_YPL009c_1)\n xml.add(\"YPL009C-2\",self.fen_YPL009c_2)\n xml.add(\"YOR267C-1\",self.fen_YOR267c_1)\n xml.add(\"YOR267C-2\",self.fen_YOR267c_2)\n \n return str(xml)\n\n# Attributes\n \n name = property(getname)\n public=property(public_repr)\n allcountries=property(allcountries_repr)\n\n# Classes\n \n isFenotipo = classmethod(isFenotipo) \n","sub_path":"src/webserver/libFenotipos.py","file_name":"libFenotipos.py","file_ext":"py","file_size_in_byte":8238,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"571492907","text":"from tkinter import *\r\nfrom playerHome import playerHome\r\nfrom playerVisiting import playerVisiting\r\nfrom soccerBall import soccerBall\r\nimport math\r\nimport random\r\n####################################\r\n# customize these functions\r\n####################################\r\n\r\ndef init(data):\r\n # load data.xyz as appropriate\r\n data.mode = \"startScreen\"\r\n data.playerHomeIndex = 2\r\n data.playerVisitingIndex = 0\r\n data.picture = PhotoImage(file = \"newSoccer pitch.png\")\r\n data.Bayern = PhotoImage(file = \"Bayern.png\")\r\n data.Bayern = data.Bayern.subsample(2,2)\r\n data.realMadrid = PhotoImage(file = \"realMadrid.png\")\r\n data.realMadrid = data.realMadrid.subsample(3,3)\r\n data.Barcelona = PhotoImage(file = \"Barcelona.png\")\r\n data.Barcelona = data.Barcelona.subsample(5,5)\r\n data.Juventus = PhotoImage(file = \"Juventus.png\")\r\n data.Juventus = data.Juventus.subsample(4,4)\r\n data.picture = data.picture.subsample(2,2)\r\n data.leftPostX = 187.5\r\n data.rightPostX = 312.5\r\n data.intLeftPostX = int(data.leftPostX)\r\n data.intRightPostX = int(data.rightPostX)\r\n data.theFixedPoint = random.randint(data.intLeftPostX + 10,data.intRightPostX - 10)\r\n data.leftBoarder = 80\r\n data.rightBoarder = 420\r\n data.lowerBoarder = 30\r\n data.upperBoarder = 470\r\n data.soccerBall = soccerBall(data.width/2,data.height/2) \r\n data.soccerMovingState = False\r\n #data.collided means when the soccer ball collides with the visiting player\r\n data.collided = False\r\n data.inStrategy = False\r\n data.inShootingProcess = False\r\n data.bouncedWall = False\r\n data.direction = 0\r\n data.bouncedDirection = 0\r\n data.collidedSpeed = 0\r\n data.ballSpeed = 1.5\r\n data.blueTeamScore = 0\r\n data.redTeamScore = 0\r\n data.playerVisitingSpeed = 0.2\r\n data.playerVisitingSupportingSpeed = 0.2\r\n data.startingPostionHome = [[200,300],[300,300],[250,450]]\r\n data.team1 = [playerHome(data.startingPostionHome[0][0],data.startingPostionHome[0][1]),playerHome(data.startingPostionHome[1][0],data.startingPostionHome[1][1]),playerHome(data.startingPostionHome[2][0],data.startingPostionHome[2][1])]\r\n\r\ndef mousePressed(event, data):\r\n if(data.mode == \"levelSelection\"):\r\n if(event.x > 200 and event.x < 400 and event.y > 300 and event.y < 350):\r\n data.playerVisitingSpeed = 0.05\r\n data.playerVisitingSupportingSpeed = 0.05\r\n data.mode = \"gameState\"\r\n data.team2 = [playerVisiting(200,200,data.playerVisitingSpeed,data.playerVisitingSupportingSpeed,data.theFixedPoint),playerVisiting(300,200,data.playerVisitingSpeed,data.playerVisitingSupportingSpeed,data.theFixedPoint),playerVisiting(250,50,data.playerVisitingSupportingSpeed,data.playerVisitingSupportingSpeed,data.theFixedPoint)]\r\n \r\n if(event.x > 200 and event.x < 400 and event.y > 375 and event.x < 425):\r\n data.playerVisitingSpeed = 0.1\r\n data.playerVisitingSupportingSpeed = 0.1\r\n data.mode = \"gameState\"\r\n data.team2 = [playerVisiting(200,200,data.playerVisitingSpeed,data.playerVisitingSupportingSpeed,data.theFixedPoint),playerVisiting(300,200,data.playerVisitingSpeed,data.playerVisitingSupportingSpeed,data.theFixedPoint),playerVisiting(250,50,data.playerVisitingSupportingSpeed,data.playerVisitingSupportingSpeed,data.theFixedPoint)]\r\n \r\n if(event.x > 200 and event.x < 400 and event.y > 450 and event.x < 500):\r\n data.playerVisitingSpeed = 0.5\r\n data.playerVisitingSupportingSpeed = 0.5\r\n data.mode = \"gameState\"\r\n data.team2 = [playerVisiting(200,200,data.playerVisitingSpeed,data.playerVisitingSupportingSpeed,data.theFixedPoint),playerVisiting(300,200,data.playerVisitingSpeed,data.playerVisitingSupportingSpeed,data.theFixedPoint),playerVisiting(250,50,data.playerVisitingSupportingSpeed,data.playerVisitingSupportingSpeed,data.theFixedPoint)]\r\n # use event.x and event.y\r\n pass\r\n\r\ndef keyPressed(event, data):\r\n # use event.char and event.keysym\r\n if(event.keysym == \"Up\"):\r\n data.team1[data.playerHomeIndex].move(\"up\")\r\n data.team1[data.playerHomeIndex].speed = 10\r\n if(event.keysym == \"Down\"):\r\n data.team1[data.playerHomeIndex].move(\"down\")\r\n data.team1[data.playerHomeIndex].speed = 10\r\n if(event.keysym == \"Left\"):\r\n data.team1[data.playerHomeIndex].move(\"left\")\r\n data.team1[data.playerHomeIndex].speed = 10\r\n if(event.keysym == \"Right\"):\r\n data.team1[data.playerHomeIndex].move(\"right\")\r\n data.team1[data.playerHomeIndex].speed = 10\r\n if(event.keysym == \"s\"):\r\n data.mode = \"levelSelection\"\r\n if(event.keysym == \"p\"):\r\n data.mode = \"startScreen\"\r\n pass\r\n \r\ndef smartAngleFunction(selfX,selfY,otherX, otherY):\r\n if(abs(otherX - selfX) < 0.5 and (otherY - selfY) < 0):\r\n return 90\r\n if(abs(otherX - selfX) < 0.5 and (otherY - selfY) > 0):\r\n return 270\r\n if((otherX - selfX) > 0 and abs(otherY - selfY) < 0.5):\r\n return 0\r\n if((otherX - selfX) < 0 and abs(otherY - selfY) < 0.5):\r\n return 180\r\n theta = math.atan(abs(otherY - selfY) / abs(otherX - selfX))\r\n theta = math.degrees(theta)\r\n if((otherX - selfX) > 0 and (otherY - selfY) < 0):\r\n return theta\r\n elif((otherX - selfX) < 0 and (otherY - selfY) < 0):\r\n return 180 - theta\r\n elif((otherX - selfX) < 0 and (otherY - selfY) > 0):\r\n return 180 + theta\r\n elif((otherX - selfX) > 0 and (otherY - selfY) > 0):\r\n return 0 - theta\r\n \r\ndef visitingTeamRespond(data):\r\n pass\r\n \r\n #Switching players based on who is the closest\r\n newList = []\r\n i = 0\r\n while(i < len(data.team2)):\r\n newList = newList + [((data.team2[i].x - data.soccerBall.x)**(2) + \r\n (data.team2[i].y - data.soccerBall.y) ** (2))**(0.5)]\r\n i = i + 1\r\n data.playerVisitingIndex = newList.index(min(newList))\r\n \r\n #Check if the visiting player collides with the home player, if the visiting player collides with the home player, make the visiting player retreat back in the same direction from where it comes from by a speed, the time duration only lasts for 1ms, after a millisecond, the ball speed gets set to 0. \r\n \r\n \r\n for i in data.team1:\r\n for j in data.team2:\r\n if(j.collideWithOtherTeamCheck(i)[0] == True):\r\n j.bounce(i)\r\n \r\n #Check if the visiting player collides with the visiting player, if the visiting player collides with the visiting player, make the visiting player bounce back\r\n if(data.team2[0].collideWithSameTeamCheck(data.team2[1])[0] == True):\r\n data.team2[0].bounce(1)\r\n if(data.team2[1].collideWithSameTeamCheck(data.team2[2])[0] == True):\r\n data.team2[1].bounce(2)\r\n if(data.team2[2].collideWithSameTeamCheck(data.team2[0])[0] == True):\r\n data.team2[2].bounce(0)\r\n \r\n \r\n #Make the visiting team player gravitates toward shooting position\r\n newDirection = data.team2[data.playerVisitingIndex].ballChasingDirection(data.soccerBall)\r\n if(data.inShootingProcess == False):\r\n data.team2[data.playerVisitingIndex].move(newDirection,data.team2[data.playerVisitingIndex].speed)\r\n if(data.team2[data.playerVisitingIndex].checkInFirePosition(data.soccerBall)):\r\n data.inShootingProcess = True\r\n \r\n #If the ball moves to the designated position, make a hit \r\n if(data.inShootingProcess == True):\r\n toTheBallDirection = data.team2[data.playerVisitingIndex].ballShootingDirectionWhenReady(data.soccerBall)\r\n # print(\"toTheBallDirection\",toTheBallDirection)\r\n # print(\"data.team2[data.playerVisitingIndex]\",data.team2[data.playerVisitingIndex])\r\n # print(\"data.team2[data.playerVisitingIndex].speed\",data.team2[data.playerVisitingIndex].speed)\r\n \r\n data.team2[data.playerVisitingIndex].move(toTheBallDirection,data.team2[data.playerVisitingIndex].speed)\r\n if(abs(data.ballSpeed-0)<0.04):\r\n data.inShootingProcess = False\r\n data.ballSpeed = 0\r\n #If one of the team2 player collides with the ball, make the ball move accordingly to the direction it collides\r\n if(data.team2[data.playerVisitingIndex].collideWithSoccerBallCheck(data.soccerBall)[0]):\r\n data.soccerMovingState = True\r\n data.direction = data.team2[data.playerVisitingIndex].collideWithSoccerBallCheck(data.soccerBall)[1]\r\n \r\n#Check if the soccer ball has scored\r\ndef checkScored(data):\r\n if(data.soccerBall.y < data.lowerBoarder and data.soccerBall.x > data.leftPostX and data.soccerBall.x < data.rightPostX):\r\n data.blueTeamScore = data.blueTeamScore + 1\r\n data.soccerBall.y = data.width/2\r\n data.soccerBall.x = data.height/2\r\n data.ballSpeed= 1.5\r\n reset(data)\r\n if(data.soccerBall.y > data.upperBoarder and data.soccerBall.x > data.leftPostX and data.soccerBall.x < data.rightPostX):\r\n data.redTeamScore = data.redTeamScore + 1\r\n data.soccerBall.y = data.width/2\r\n data.soccerBall.x = data.height/2\r\n data.ballSpeed= 1.5\r\n reset(data)\r\n\r\n#Reset Everything to the starting position\r\ndef reset(data):\r\n for i in data.team1:\r\n i.reset()\r\n for i in data.team2:\r\n i.reset()\r\n\r\ndef strategicMoveHome(data):\r\n if(data.playerHomeIndex == 0):\r\n \r\n if(data.team1[0].y < (data.height / 2) and \r\n data.team1[0].y >= (data.height / 4) and \r\n data.team1[0].x < (data.width) / 2):\r\n data.inStrategy = True\r\n data.team1[1].makingMoveUpperFieldRightSide(data.team1[0])\r\n if(data.team1[1].makingMoveUpperFieldRightSide(data.team1[0]) == True):\r\n data.inStrategy = False\r\n else:\r\n data.inStrategy = True\r\n \r\n if(data.team1[0].y < (data.height / 2) and \r\n data.team1[0].y >= (data.height / 4) and \r\n data.team1[0].x > (data.width) / 2):\r\n data.inStrategy = True\r\n data.team1[1].makingMoveUpperFieldLeftSide(data.team1[0])\r\n if(data.team1[1].makingMoveUpperFieldLeftSide(data.team1[0] == False)):\r\n data.inStrategy = False\r\n else:\r\n data.inStrategy = True\r\n \r\n if(data.team1[0].y > (data.height / 2) and \r\n data.team1[0].y <= (data.height / 4 * 3) and \r\n data.team1[0].x < (data.width) / 2):\r\n data.inStrategy = True\r\n data.team1[1].makingMoveLowerFieldRightSide(data.team1[0])\r\n if(data.team1[1].makingMoveLowerFieldRightSide(data.team1[0]) == True):\r\n data.inStrategy = False\r\n else:\r\n data.inStrategy = True\r\n \r\n if(data.team1[0].y > (data.height / 2) and \r\n data.team1[0].y <= (data.height / 4 * 3) and \r\n data.team1[0].x > (data.width) / 2):\r\n data.inStrategy = True\r\n data.team1[1].makingMoveLowerFieldLeftSide(data.team1[0])\r\n if(data.team1[1].makingMoveLowerFieldLeftSide(data.team1[0]) == True):\r\n data.inStrategy = False\r\n else:\r\n data.inStrategy = True\r\n \r\n if(data.playerHomeIndex == 1):\r\n \r\n if(data.team1[1].y < (data.height / 2) and \r\n data.team1[1].y >= (data.height / 4) and \r\n data.team1[1].x < (data.width) / 2):\r\n data.inStrategy = True\r\n data.team1[0].makingMoveUpperFieldRightSide(data.team1[1])\r\n if(data.team1[0].makingMoveUpperFieldRightSide(data.team1[1]) == True):\r\n data.inStrategy = False\r\n else:\r\n data.inStrategy = True\r\n \r\n if(data.team1[1].y < (data.height / 2) and \r\n data.team1[1].y >= (data.height / 4) and \r\n data.team1[1].x > (data.width) / 2):\r\n data.inStrategy = True\r\n data.team1[0].makingMoveUpperFieldLeftSide(data.team1[1])\r\n if(data.team1[0].makingMoveUpperFieldLeftSide(data.team1[1] == False)):\r\n data.inStrategy = False\r\n else:\r\n data.inStrategy = True\r\n \r\n if(data.team1[1].y > (data.height / 2) and \r\n data.team1[1].y <= (data.height / 4 * 3) and \r\n data.team1[1].x < (data.width) / 2):\r\n data.inStrategy = True\r\n data.team1[0].makingMoveLowerFieldRightSide(data.team1[1])\r\n if(data.team1[0].makingMoveLowerFieldRightSide(data.team1[1]) == True):\r\n data.inStrategy = False\r\n else:\r\n data.inStrategy = True\r\n \r\n if(data.team1[1].y > (data.height / 2) and \r\n data.team1[1].y <= (data.height / 4 * 3) and \r\n data.team1[1].x > (data.width) / 2):\r\n data.inStrategy = True\r\n data.team1[0].makingMoveLowerFieldLeftSide(data.team1[1])\r\n if(data.team1[0].makingMoveLowerFieldLeftSide(data.team1[1]) == True):\r\n data.inStrategy = False\r\n else:\r\n data.inStrategy = True\r\n \r\ndef strategicMoveVisiting(data):\r\n if(data.playerVisitingIndex == 0):\r\n if(data.team2[0].y > (data.height / 2) and\r\n data.team2[0].y <= (data.height / 4 * 3) and\r\n data.team2[0].x < (data.width / 2)):\r\n data.team2[1].makingMoveUpperFieldRightSide(data.team2[0])\r\n \r\n if(data.team2[0].y > (data.height / 2) and \r\n data.team2[0].y <= (data.height / 4 * 3) and \r\n data.team2[0].x > (data.width / 2)):\r\n data.team2[1].makingMoveUpperFieldLeftSide(data.team2[0])\r\n \r\n if(data.team2[0].y < (data.height / 2) and \r\n data.team2[0].y >= (data.height / 4) and\r\n data.team2[0].x < (data.width / 2)):\r\n data.team2[1].makingMoveLowerFieldRightSide(data.team2[0])\r\n \r\n \r\n \r\ndef timerFired(data):\r\n if(data.mode == \"gameState\"):\r\n if(data.blueTeamScore > 3):\r\n data.mode = \"blueTeamWinState\"\r\n if(data.redTeamScore > 3):\r\n data.mode = \"redTeamWinState\"\r\n # Check if the soccer has scored\r\n checkScored(data)\r\n \r\n #Make one of the visiting team player to chase the ball\r\n visitingTeamRespond(data)\r\n \r\n #Make the home team move strategically \r\n strategicMoveHome(data) \r\n \r\n #Make the oppposing team move strategically \r\n strategicMoveVisiting(data)\r\n \r\n #Normal moving state\r\n if(data.soccerMovingState == True):\r\n if(data.collided == False and data.bouncedWall == False):\r\n data.ballSpeed = data.ballSpeed - 0.01\r\n if(abs(data.ballSpeed-0)<0.02):\r\n data.soccerMovingState = False\r\n data.ballSpeed = 1.5\r\n data.soccerBall.reactToKick(data.direction,data.ballSpeed)\r\n \r\n #Check if the ball collides with the wall or not\r\n if(data.soccerBall.bouncingBoarder(data.direction,data.leftBoarder,data.rightBoarder,data.lowerBoarder,data.upperBoarder,data.bouncedDirection)[0] == True):\r\n data.bouncedWall = True\r\n if(data.bouncedWall == True) and (data.soccerMovingState == True):\r\n data.bouncedDirection = data.soccerBall.bouncingBoarder(data.direction,data.leftBoarder,data.rightBoarder,data.lowerBoarder,data.upperBoarder,data.bouncedDirection)[1]\r\n data.ballSpeed = data.ballSpeed - 0.01\r\n if(abs(data.ballSpeed - 0) < 0.02):\r\n data.soccerMovingState = False\r\n data.ballSpeed = 0\r\n data.bouncedWall = False\r\n data.soccerBall.reactToKick(data.bouncedDirection,data.ballSpeed)\r\n \r\n #Check if the ball collide with Visiting player\r\n # for i in data.team2:\r\n # if(i == data.playerHomeIndex):\r\n # continue\r\n # else:\r\n # if(data.soccerBall.collideWithVisitingTeamCheck(i) and data.soccerMovingState == True):\r\n # data.collided = True\r\n # data.collidedSpeed = -data.ballSpeed\r\n \r\n #Switching Control of the ball for home player\r\n # for i in range(len(data.team1)):\r\n # if(i == data.playerHomeIndex):\r\n # continue\r\n # else:\r\n # if(data.soccerBall.collideWithHomeTeamCheck(data.team1[i]) and data.soccerMovingState == True):\r\n # data.collided = True\r\n # data.collidedSpeed = -data.ballSpeed\r\n # data.playerHomeIndex = i\r\n \r\n #Switching Control of the ball for home player\r\n # if(data.inStrategy == False):\r\n newList = []\r\n i = 0\r\n while(i < len(data.team1)):\r\n newList = newList + [((data.team1[i].x - data.soccerBall.x)**(2) + \r\n (data.team1[i].y - data.soccerBall.y) ** (2))**(0.5)]\r\n i = i + 1\r\n data.playerHomeIndex = newList.index(min(newList))\r\n \r\n #When the control is swithced, try to make the other player accords with the player with the ball. \r\n \r\n #If the player with the ball is greater than half of the line and is in the left, make the accompanying player move in a slow speed up. Trying to go to the point where it is 30 degrees above the home player. \r\n strategicMoveHome(data)\r\n # strategicMoveVisiting(data)\r\n \r\n # if(data.team2[data.playerHomeIndex].y < (data.height / 4)):\r\n # data.team2[1].makingMoveInPenaltyArea()\r\n \r\n # if(data.playerHomeIndex == 1):\r\n # if(data.team1[1].y < (data.height / 2) and \r\n # data.team1[1].y >= (data.height /4 )and \r\n # data.team1[1].x < (data.width) / 2):\r\n # data.team1[0].makingMoveUpperFieldRightSide(data.team1[1])\r\n # if(data.team2[data.playerHomeIndex].y < (data.height / 4)):\r\n # data.team2[1].makingMoveInPenaltyArea()\r\n \r\n \r\n \r\n #This is the central control of how to react when collided\r\n if(data.soccerMovingState == True and data.collided == True):\r\n data.collidedSpeed = data.collidedSpeed + 0.07\r\n if(abs(data.collidedSpeed-0)<0.1):\r\n data.soccerMovingState = False\r\n data.collided = False\r\n data.collidedSpeed = 1.5\r\n data.soccerBall.reactToKick(data.team1[data.playerHomeIndex].collideWithSoccerBallCheck(data.soccerBall)[1],data.collidedSpeed)\r\n \r\n #Check if the player hits a soccer ball\r\n if(data.team1[data.playerHomeIndex].collideWithSoccerBallCheck(data.soccerBall)[0]):\r\n data.soccerMovingState = True\r\n data.direction = data.team1[data.playerHomeIndex].collideWithSoccerBallCheck(data.soccerBall)[1]\r\n \r\n #Check if the player hits a wall. If he hits a wall, makes him freeze\r\n data.team1[data.playerHomeIndex].collideWithWallCheck(data.leftBoarder,data.rightBoarder,data.lowerBoarder,data.upperBoarder)\r\n \r\n #Check if a player hits a visiting team player, if so, he can't move forward anymore. \r\n for i in data.team2:\r\n if(data.team1[data.playerHomeIndex].collideWithOtherTeamCheck(i)):\r\n data.team1[data.playerHomeIndex].move(data.team1[data.playerHomeIndex].collideWithOtherTeamCheck(i)[0])\r\n \r\n #Check if a player hits a home team player, if so, he can't move forward anymore\r\n for i in range(len(data.team1)):\r\n if(i == data.playerHomeIndex):\r\n continue\r\n else:\r\n if(data.team1[data.playerHomeIndex].collideWithSameTeamCheck(data.team1[i])):\r\n data.team1[data.playerHomeIndex].move(data.team1[data.playerHomeIndex].collideWithSameTeamCheck(data.team1[i])[0])\r\n \r\n\r\n \r\n \r\n \r\ndef draw(canvas,data):\r\n if(data.mode == \"blueTeamWinState\"):\r\n canvas.create_text(250,250,text = \"Congratulations!\",font = \"30\")\r\n if(data.mode == \"redTeamWinState\"):\r\n canvas.create_text(250,250,text = \"Tough Loss\",font = \"30\")\r\n if (data.mode == \"startScreen\"):\r\n canvas.create_rectangle(0,0,500,500,fill = \"green\")\r\n canvas.create_text(250,250,text = \"Exciting Soccer Game\",fill = \"orange\",\r\n font=\"Times 28 bold italic\")\r\n canvas.create_image(125,125,image = data.Barcelona)\r\n canvas.create_image(125,375,image = data.Juventus)\r\n canvas.create_image(375,125,image = data.realMadrid)\r\n canvas.create_image(375,375,image = data.Bayern)\r\n if(data.mode == \"gameState\"):\r\n canvas.create_image(data.width/2,data.height/2,image = data.picture)\r\n canvas.create_text(40,50,text = \"BlueTeam:%d\"%data.blueTeamScore,font = \"Times 12 italic\")\r\n canvas.create_text(40,70,text = \"RedTeam:%d\"%data.redTeamScore,font = \"Times 12 italic\")\r\n for i in data.team1:\r\n i.draw(canvas)\r\n for i in data.team2:\r\n i.draw(canvas)\r\n data.soccerBall.draw(canvas)\r\n \r\n if(data.mode == \"levelSelection\"):\r\n levelColor = \"indian red\"\r\n canvas.create_rectangle(0,0,500,500,fill = \"orange\")\r\n canvas.create_text(200,200,text = \"Choose Your Level\", font = \"Times 28 bold italic\",)\r\n canvas.create_rectangle(200,300,400,350,fill = levelColor,outline = levelColor)\r\n canvas.create_text(300,325,text = \"intramural\", font = \"Times 18\")\r\n canvas.create_rectangle(200,375,400,425,fill = levelColor,outline = levelColor)\r\n canvas.create_text(300,400,text = \"club\",font = \"Times 18\")\r\n canvas.create_rectangle(200,450,400,500,fill = levelColor,outline = levelColor)\r\n canvas.create_text(300,475,text = \"professional\",font = \"Times 18\")\r\n \r\n \r\n \r\ndef redrawAll(canvas, data):\r\n # draw in canvas\r\n draw(canvas, data)\r\n pass\r\n\r\n####################################\r\n# use the run function as-is\r\n#####################################\r\n\r\n#Important!!!This code Is directly cited from 15112 course note. \r\n#This is the link to the website. Thank you! https://www.cs.cmu.edu/~112/notes/notes-graphics.html\r\ndef run(width=300, height=300):\r\n def redrawAllWrapper(canvas, data):\r\n canvas.delete(ALL)\r\n canvas.create_rectangle(0, 0, data.width, data.height,\r\n fill='white', width=0)\r\n redrawAll(canvas, data)\r\n canvas.update() \r\n\r\n def mousePressedWrapper(event, canvas, data):\r\n mousePressed(event, data)\r\n redrawAllWrapper(canvas, data)\r\n\r\n def keyPressedWrapper(event, canvas, data):\r\n keyPressed(event, data)\r\n redrawAllWrapper(canvas, data)\r\n def timerFiredWrapper(canvas, data):\r\n timerFired(data)\r\n redrawAllWrapper(canvas, data)\r\n # pause, then call timerFired again\r\n canvas.after(data.timerDelay, timerFiredWrapper, canvas, data)\r\n # Set up data and call init\r\n class Struct(object): pass\r\n data = Struct()\r\n data.timerDelay = 1\r\n data.width = width\r\n data.height = height\r\n root = Tk()\r\n root.resizable(width=False, height=False) # prevents resizing window\r\n init(data)\r\n # create the root and the canvas\r\n canvas = Canvas(root, width=data.width, height=data.height)\r\n canvas.configure(bd=0, highlightthickness=0)\r\n canvas.pack()\r\n # set up events\r\n root.bind(\"\", lambda event:\r\n mousePressedWrapper(event, canvas, data))\r\n root.bind(\"\", lambda event:\r\n keyPressedWrapper(event, canvas, data))\r\n timerFiredWrapper(canvas, data)\r\n # and launch the app\r\n root.mainloop() # blocks until window is closed\r\n print(\"bye!\")\r\n\r\nrun(500, 500)","sub_path":"tkinter_soccer_game.py","file_name":"tkinter_soccer_game.py","file_ext":"py","file_size_in_byte":23955,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"550508325","text":"import sys\n\nfrom PyQt5.QtWidgets import QApplication, QLabel, QMainWindow, QPushButton, QToolTip\n\nfrom PyQt5 import QtGui\n\n\nclass Janela(QMainWindow):\n def __init__(self):\n super().__init__()\n\n self.topo = 100\n self.esquerda = 100\n self.largura = 800\n self.altura = 700\n self.titulo = \"Primeira Janela\"\n \n \n\n botao1 = QPushButton('Botão 1', self)\n botao1.move(250, 250) \n botao1.resize(50,50) \n botao1.setStyleSheet('QPushButton {background-color:#0FB329; font-size:10px}') \n botao1.clicked.connect(self.bota01_click) \n\n \n\n botao2 = QPushButton('Botão 2', self)\n botao2.move(350, 250) \n botao2.resize(50,50) \n botao2.setStyleSheet('QPushButton {background-color:#0FB329; font-size:10px}') \n botao2.clicked.connect(self.bota02_click) \n\n\n\n botao3 = QPushButton('Botão 3', self)\n botao3.move(450, 250) \n botao3.resize(50,50) \n botao3.setStyleSheet('QPushButton {background-color:#0FB329; font-size:10px}') \n botao3.clicked.connect(self.bota03_click) \n \n \n self.label_1 = QLabel(self)\n self.label_1.setText(\"LINGUAGEM DE PROGRAMAÇÃO\")\n self.label_1.move(220, 220)\n self.label_1.setStyleSheet(\"QLabel {color:red; font-size:25px}\")\n self.label_1.resize(400, 25)\n\n \n self.logo = QLabel(self)\n self.logo.move(500, 200)\n self.logo.resize(500, 600)\n self.CarregarJanela()\n\n def CarregarJanela(self):\n self.setGeometry(self.esquerda, self.topo, self.largura, self.altura)\n self.setWindowTitle(self.titulo)\n self.show()\n\n \n def bota01_click(self):\n self.label_1.setStyleSheet(\"QLabel {color:blue; font-size:25px}\")\n\n self.label_1.setText(\"Pyton\")\n self.logo.setPixmap(QtGui.QPixmap('python.png')) \n\n \n def bota02_click(self):\n self.label_1.setStyleSheet(\"QLabel {color:blue; font-size:25px}\")\n\n self.label_1.setText(\"Java\")\n self.logo.setPixmap(QtGui.QPixmap('java.png')) \n\n def bota03_click(self):\n self.label_1.setStyleSheet(\"QLabel {color:blue; font-size:25px}\")\n\n self.label_1.setText(\"C\")\n self.logo.setPixmap(QtGui.QPixmap('cc.png')) \n\n\n\naplicação = QApplication(sys.argv)\n\nj = Janela()\nsys.exit(aplicação.exec_())","sub_path":"Aula05/Praticas Aula 05/01.py","file_name":"01.py","file_ext":"py","file_size_in_byte":2352,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"538008567","text":"import sys\nimport numpy as np\ndef f(n,a_):\n a=[]\n for i,ai in enumerate(a_):\n a.append([1+i,ai])\n a.sort(key=lambda x:x[1],reverse=True)\n dp=np.zeros((n+1,n+1),np.int64) #dp[i,j]:i人目までを配置し終えていて、そのうちj人を右に置いた時の最大値。j<=i\n dp[1,1]=a[0][1]*(n-a[0][0])\n dp[1,0]=a[0][1]*(a[0][0]-1)\n for i in range(2,n+1):#i人目の幼児。a[i-1]\n dp[i,0]=dp[i-1,0]+a[i-1][1]*abs((a[i-1][0]-1-(i-1))) #i人目を左にする。その時左に溜まっているのはi-1人\n dp[i,i]=dp[i-1,i-1]+a[i-1][1]*abs((n-a[i-1][0]-(i-1))) #i人目を右にする。その時右に溜まっているのはi-1人\n #i人目を左に\n l=dp[i-1,1:i]+[a[i-1][1]*abs((a[i-1][0]-1-j))for j in range(i-2,-1,-1)] #j:左に溜まっている人数\n #i人目を右に\n r=dp[i-1,:i-1]+[a[i-1][1]*abs((n-a[i-1][0]-j)) for j in range(i-1)] #j:右に溜まっている人数\n dp[i,1:i]=np.maximum(l,r)\n return max(dp[n])\n\nif __name__ == '__main__':\n input = sys.stdin.readline\n n=int(input())\n a_=list(map(int,input().split()))\n print(f(n,a_))\n","sub_path":"Python_codes/p02709/s142304741.py","file_name":"s142304741.py","file_ext":"py","file_size_in_byte":1206,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"334285668","text":"from time import sleep\nimport psycopg2\nimport serial\nfrom scast_buffer import *\n\nclass SerialPort:\n def __init__(self, port_name):\n self.ser = serial.Serial(\n port = port_name,\n baudrate = 19200,\n parity = serial.PARITY_NONE,\n stopbits = serial.STOPBITS_ONE,\n bytesize = serial.EIGHTBITS)\n\n\n buffer = ScastBuffer()\n while 1:\n chars = \"\"\n while self.ser.inWaiting() > 0:\n chars += self.ser.read(1).encode('hex')\n if len(chars) > 0 and buffer.add(chars):\n self.process_alarms_and_relays()\n sleep(1)\n\n def process_alarms_and_relays(self):\n conn = psycopg2.connect(\"dbname=pyppm user=pyppm password=PyPPM\")\n cur = conn.cursor()\n cur.execute('SELECT * FROM \"GetRelayStates\"()')\n for record in cur:\n self.ser.write(\"AT+UCAST:%s,00%0.2X\\r\\n\" % (record[0], record[1]))\n conn.commit()\n cur.close()\n conn.close()\n\n","sub_path":"serial_port.py","file_name":"serial_port.py","file_ext":"py","file_size_in_byte":900,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"284472591","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Jul 23 15:47:31 2016\n\n@author: SRINIVAS\n\"\"\"\n\nimport numpy as np\nimport wavio\nimport sounddevice as sd\nfs= 44100\na=0\nduration=5\nprint(0)\nrecording = sd.rec(duration * fs, samplerate=fs, channels=2)\nsd.wait()\nprint(0)\n\nwavio.write('ine.wav',recording,rate=44100,sampwidth=3)","sub_path":"WAV.py","file_name":"WAV.py","file_ext":"py","file_size_in_byte":315,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"462330697","text":"# -*- coding=UTF-8 -*-\n\"\"\"Tray icon and menu. \"\"\"\n\nfrom __future__ import (absolute_import, division, print_function,\n unicode_literals)\n\nfrom Qt.QtGui import QCursor, QIcon\nfrom Qt.QtWidgets import QSystemTrayIcon\n\nfrom ..pathtools import module_path\nfrom .core import Menu\n\n\nclass Tray(QSystemTrayIcon):\n \"\"\"Tray icon with menu. \"\"\"\n\n instance = None\n initiated = False\n icon = module_path('assets', 'tray_icon.png')\n\n def __new__(cls):\n if cls.instance:\n return cls.instance\n return super(Tray, cls).__new__(cls)\n\n def __init__(self):\n if self.initiated:\n return\n\n super(Tray, self).__init__(QIcon(self.icon))\n\n # Menu.\n self.menu = Menu()\n self.setContextMenu(self.menu)\n\n # Signals.\n self.activated.connect(self.on_activated)\n\n self.initiated = True\n Tray.instance = self\n\n def on_activated(self, reason):\n if reason == self.Trigger:\n self.contextMenu().popup(QCursor.pos())\n\n @classmethod\n def _message(cls, title, text, seconds, icon):\n \"\"\"Show a traytip. \"\"\"\n\n tray = cls()\n tray.show()\n tray.showMessage(title, text, icon=icon, msecs=seconds * 1000)\n\n @classmethod\n def message(cls, title, text, seconds=3):\n \"\"\"Show a traytip with no icon. \"\"\"\n\n cls._message(title, text, seconds, QSystemTrayIcon.NoIcon)\n\n @classmethod\n def information(cls, title, text, seconds=3):\n \"\"\"Show a traytip with information icon. \"\"\"\n\n cls._message(title, text, seconds, QSystemTrayIcon.Information)\n\n @classmethod\n def warning(cls, title, text, seconds=3):\n \"\"\"Show a traytip with warning icon. \"\"\"\n\n cls._message(title, text, seconds, QSystemTrayIcon.Warning)\n\n @classmethod\n def critical(cls, title, text, seconds=3):\n \"\"\"Show a traytip with critical icon. \"\"\"\n\n cls._message(title, text, seconds, QSystemTrayIcon.Critical)\n","sub_path":"wlf/uitools/tray.py","file_name":"tray.py","file_ext":"py","file_size_in_byte":1998,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"444753494","text":"from collections import defaultdict\n\n\nclass Track:\n def __init__(self, nodes, tracks):\n self.nodes = {}\n self.tracks = {}\n\n tracks_by_node = defaultdict(lambda: [])\n\n for node in nodes:\n self.nodes[node.uuid] = node\n\n for track in tracks:\n self.tracks[track.uuid] = track\n\n for node in track.get_nodes():\n tracks_by_node[node].append(track)\n\n for node in tracks_by_node:\n if node not in self.nodes:\n raise AssertionError(node + ' not found in provided nodes')\n if len(tracks_by_node[node]) != 2:\n print('Warning: node has ' + str(len(tracks_by_node[node])) + ' connections')\n\n for track in tracks_by_node[node]:\n for other_track in tracks_by_node[node]:\n if track.uuid == other_track.uuid:\n continue\n\n track.add_connection(node, other_track)\n\n def get_updated_location(self, loc, offset):\n return self.tracks[loc.track_uuid].get_updated_location(loc, offset)\n\n def to_string(self):\n string = 'Track\\n'\n string += '--- NODES ---\\n\\n'\n\n for n in self.nodes.values():\n string += n.to_string() + '\\n'\n\n string += '\\n--- TRACKS --- \\n\\n'\n\n for t in self.tracks.values():\n string += t.to_string()\n\n return string\n\n def render(self, render):\n for track in self.tracks.values():\n geom = track.get_geometry()\n render.attachNewNode(geom)\n","sub_path":"src/layout/track.py","file_name":"track.py","file_ext":"py","file_size_in_byte":1569,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"263121949","text":"default_app_config = 'emitter.AppConfig'\n\nfrom django.apps import AppConfig, apps\n\nclass AppConfig(AppConfig): # Our app config class\n name = 'emitter'\n verbose_name = 'Emitter'\n\n def ready(self):\n for app in apps.app_configs.keys():\n try:\n __import__('%s.listeners' % app)\n except ImportError:\n pass\n if LISTENERS:\n print('\\n[listeners]')\n for listener in LISTENERS.keys():\n print('\\t*\\t%s' % listener)\n print('\\n')\n\nLISTENERS = {}\n\ndef addListener(*args, **kwargs):\n \"\"\"\n Create listener (decorator).\n \"\"\"\n def registerListener(*args, **kwargs):\n def wrap_function(fun):\n channel = args[0]\n LISTENERS[channel] = fun\n return wrap_function\n\n if len(args) == 1 and callable(args[0]):\n raise Exception('invalid listener: no args or kwargs')\n return registerListener(*args, **kwargs)\n\ndef callListener(channel, *args, **kwargs):\n return LISTENERS[channel](*args, **kwargs)","sub_path":"backend/emitter/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":1060,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"68803258","text":"# https://leetcode.com/problems/powx-n/\n\nclass Solution(object):\n def myPow(self, x, n):\n \"\"\"\n :type x: float\n :type n: int\n :rtype: float\n \"\"\"\n\n if not n:\n return 1\n if n < 0:\n return 1 / self.myPow(x, -n)\n if n % 2:\n return x * self.myPow(x, n - 1)\n return self.myPow(x * x, n / 2)\n\n\ndef main():\n solution = Solution()\n\n x = 2.2\n n = 2\n print(solution.myPow(x, n))\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"leetcode_py/50_pow_xn/Solution.py","file_name":"Solution.py","file_ext":"py","file_size_in_byte":518,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"89971509","text":"#!/usr/bin/env python2\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Feb 4 19:38:51 2018\n\n@author: vr308\n\"\"\"\nfrom mpl_toolkits.mplot3d import Axes3D \nimport matplotlib.pylab as plt\nimport scipy.interpolate as intp\nfrom matplotlib import cm\nimport random\nimport numpy as np\n\n#np.square(np.sin(X))*np.cos(Y) + Y*np.sin(X)\n#X*np.exp(-np.square(X) - np.square(Y))\n\ndef sample_surface(X,Y):\n \n return((np.subtract(X,1)*np.exp(-np.square(np.subtract(X,1)) - np.square(Y)) + np.cos(X)*np.exp(-np.square(np.add(Y,1)) - np.square(X)) + np.exp(-np.square(np.add(X,2)) - np.square(np.add(Y,-2)))))\n\nx = y = np.linspace(-4,4,1000)\nX, Y = np.meshgrid(x,y)\nZ = sample_surface(X,Y)\n\nx_sample = random.sample(set(x),100)\ny_sample = random.sample(set(y),100)\nz_sample = sample_surface(x_sample,y_sample) + 0.2*np.random.randn(100)\n\n# 3D rendering of the function along with scatter\nfig = plt.figure()\nax = fig.gca(projection='3d')\nax.scatter(x_sample,y_sample,z_sample, c='k', depthshade=False)\nax.plot_surface(X,Y,Z, cmap=cm.get_cmap('jet'), alpha=0.8)\n\n# 2D contour of the mathematical function along with scatter\n\nplt.figure()\nplt.contourf(X,Y,Z, cmap=cm.get_cmap('jet'))\nplt.plot(x_sample,y_sample,'ko')\n\n# Multivariate interpolation using rbf technique applied to scatter data\n\nrbf_interp = intp.Rbf(x_sample,y_sample,z_sample,smooth=0.2, epsilon=1)\nX_int, Y_int = np.meshgrid(np.linspace(-4,4,100), np.linspace(-4,4,100)) \nZ_interp = rbf_interp(X_int,Y_int)\n\n# Multivariate interpolation using nearest neighbour technique applied to scatter data\n\ninputs = np.array([x_sample,y_sample]).reshape(100,2)\noutputs = z_sample\nninterp = intp.NearestNDInterpolator(inputs, outputs)\nZ_interp = ninterp(X_int,Y_int)\n\n# Plot the interpolated surface\n\nfig = plt.figure()\nax = fig.gca(projection='3d')\nax.scatter(x_sample,y_sample,z_sample, c='k', depthshade=False)\nax.plot_surface(X_int,Y_int,Z_interp, cmap=cm.get_cmap('jet'), alpha=0.8)\n\n# Plot the contour \nplt.figure()\nplt.contourf(X,Y,Z, cmap=cm.get_cmap('jet'))\nplt.plot(x_sample,y_sample,'ko')\n\nplt.figure()\nplt.contourf(X_int,Y_int,Z_interp, cmap=cm.get_cmap('jet'))\nplt.plot(x_sample,y_sample,'ko')\n\n\n","sub_path":"Visualisation/surfaces.py","file_name":"surfaces.py","file_ext":"py","file_size_in_byte":2141,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"398916545","text":"from rest_framework.reverse import reverse\n\nCERTIFICATE_LIST_API_URL = reverse(\"certificate-list\")\n\n\ndef certificate_detail_api_url(certificate_id):\n return reverse('certificate-detail', args=[certificate_id])\n\n\ndef sample_certificate_payload(**params):\n defaults = {\n \"issuer\": \"test\",\n \"title\": \"test\",\n \"certificate_id\": \"test\",\n \"certificate_url\": \"http://test.com\",\n \"valid_until\": None,\n \"does_not_expire\": True\n }\n\n defaults.update(params)\n\n return defaults\n\n\ndef new_certificate_payload():\n payload = {\n \"issuer\": \"newissuer\",\n \"title\": \"newtitle\",\n \"certificate_id\": \"newid\",\n \"certificate_url\": \"http://newtest.com\",\n \"valid_until\": None,\n \"does_not_expire\": True\n }\n\n return payload\n","sub_path":"src/tests/utils/api/certificate.py","file_name":"certificate.py","file_ext":"py","file_size_in_byte":802,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"319761910","text":"from flask import Flask, render_template, flash, request\nfrom wtforms import Form, TextField, TextAreaField, validators, StringField, SubmitField\nfrom pymongo import MongoClient\nfrom flask import jsonify\n# pprint library is used to make the output look more pretty\nfrom pprint import pprint\n# connect to MongoDB, change the << MONGODB URL >> to reflect your own connection string\nmyclient = MongoClient(\"mongodb://10.0.0.13:27017/\")\nmydb = myclient[\"cloudapp\"]\nmycol = mydb[\"messages\"]\n# Issue the serverStatus command and print the results\n\n# App config.\nDEBUG = True\napp = Flask(__name__)\napp.config.from_object(__name__)\napp.config['SECRET_KEY'] = '7d441f27d441f27567d441f2b6176a'\n\n\nclass ReusableForm(Form):\n name = TextField('Message:', validators=[validators.required()])\n\n\n@app.route(\"/send\", methods=['GET', 'POST'])\ndef send():\n data = request.get_data()\n print(data)\n message = data.decode(\"utf-8\").split(\"=\")[1]\n result = mycol.insert_one({'message': message})\n form = ReusableForm(request.form)\n return render_template('hello.html', form=form)\n\n@app.route(\"/retreive\", methods=['GET', 'POST'])\ndef retreive():\n messages = {}\n count = 0\n for x in mycol.find():\n count += 1\n messages[str(count)] = x['message']\n return jsonify(messages)\n\n\nif __name__ == '__main__':\n app.run(host='0.0.0.0', port='50001')\n","sub_path":"rianaAssignment3Pt2/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":1365,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"388020150","text":"from typing import Optional\nfrom utils.measure import checker\n\n\nclass ListNode:\n def __init__(self, val=0, next=None):\n self.val = val\n self.next = next\n\n\ndef printl(l: ListNode) -> None:\n while l:\n print(l.val, end=' ')\n l = l.next\n print('')\n\n\ndef make(l: list[int]) -> ListNode | None:\n head = ListNode()\n curr = head\n\n for val in l:\n curr.next = ListNode(val)\n curr = curr.next\n\n return head.next\n\n\ndef unmake(ln: ListNode | None) -> list[int]:\n res = []\n while ln:\n res.append(ln.val)\n ln = ln.next\n return res\n\n\nclass Solution:\n\n def mergeTwoLists(self, list1: Optional[ListNode], list2: Optional[ListNode]) -> Optional[ListNode]:\n merged = ListNode()\n curr = merged\n\n while list1 and list2:\n if list1.val < list2.val:\n curr.next = list1\n list1 = list1.next\n else:\n curr.next = list2\n list2 = list2.next\n curr = curr.next\n\n curr.next = list1 or list2\n\n return merged.next\n\n def wrapped(self, list1: list[int], list2: list[int]) -> list[int]:\n return unmake(self.mergeTwoLists(make(list1), make(list2)))\n\n\nif __name__ == '__main__':\n with checker(Solution().wrapped, repeat=0, check_success=True) as c:\n c.check_2([1, 2, 4], [1, 3, 4], [1, 1, 2, 3, 4, 4])\n c.check_2([], [], [])\n c.check_2([], [0], [0])\n c.check_2([1, 2, 4, 6, 7, 8], [1, 3, 4], [1, 1, 2, 3, 4, 4, 6, 7, 8])\n","sub_path":"algorithms/l/studyplan/mergesortedlists.py","file_name":"mergesortedlists.py","file_ext":"py","file_size_in_byte":1538,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"585819886","text":"#!/usr/bin/env python3\n\nimport sys\nimport requests\nfrom urllib.parse import urlparse\nfrom bs4 import BeautifulSoup\n\n\ndef validate_url(url):\n \"Check if URL is valid.\"\n parsed_url = urlparse(url)\n if parsed_url.scheme and parsed_url.netloc:\n return True\n\n return False\n\n\ndef get_url_scheme(url):\n \"Get the scheme of a URL.\"\n parsed_url = urlparse(url)\n return parsed_url.scheme\n\n\ndef get_url_netloc(url):\n \"Extract the NETwork LOCality (netloc) out of a URL.\"\n parsed_url = urlparse(url)\n return parsed_url.netloc\n\n\ndef evaluate_cmd():\n \"\"\"\n Evaluate the command line arguments or exit with help printed if none\n given.\n \"\"\"\n url_arg = sys.argv[1]\n\n if len(sys.argv) <= 1:\n print(__doc__)\n sys.exit(0)\n\n if len(sys.argv) > 2:\n print(\"The script takes only 1 argument - a URL!\")\n\n if not validate_url(url_arg):\n print(\n \"\\n\\n\\nERROR: The URL '{}' {}!\\n\\n\\n\".format(\n url_arg,\n \"you have given is invalid\"\n )\n )\n sys.exit(1)\n\n return url_arg\n\n\ndef is_tag_visible(tag):\n \"Checks if a given element is visible in the article\"\n style = tag.attrs.get('style', False)\n if style and ('hidden' in style or 'display: none' in style or 'display:none' in style):\n return False\n\n parent = tag.parent\n if parent and not is_tag_visible(parent):\n return False\n\n return True\n\n\ndef remove_footer(content):\n \"Removes footer of an article, if such exists.\"\n footer = content.find('footer')\n\n if not footer:\n footer = content.find('div', class_='footer')\n\n footer.decompose()\n\n return content\n\n\ndef download_img(url, filename):\n \"Downloads image\"\n request = requests.get(url)\n if request.ok:\n with open(filename, 'wb') as img_file:\n img_file.write(request.content)\n\n\nif __name__ == \"__main__\":\n target_url = evaluate_cmd()\n response = requests.get(target_url)\n\n if not response.ok:\n print(\"\\n\\n\\nERROR: Status code: {}!\\n\\n\\n\".format(status))\n sys.exit(2)\n\n content = response.text\n content_parsed = BeautifulSoup(content, \"html.parser\")\n content_parsed = remove_footer(content_parsed)\n title = content_parsed.title.string\n print(title)\n\n articles = content_parsed.find_all('article')\n paragraphs = content_parsed.find_all('p')\n if articles:\n for article in articles:\n if(is_tag_visible(article)):\n print(article.prettify())\n else:\n for p in paragraphs:\n if(is_tag_visible(p)):\n print(p.prettify())\n\n for img in content_parsed.find_all('img'):\n img_src = img.get('src')\n if not validate_url(img_src):\n img_src = ''.join([\n get_url_scheme(target_url),\n \"://\",\n get_url_netloc(target_url),\n img_src\n ])\n if not validate_url(img_src):\n next\n download_img(img_src, \"./test.jpg\")\n break\n","sub_path":"fetcher.py","file_name":"fetcher.py","file_ext":"py","file_size_in_byte":3045,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"505152153","text":"from django.shortcuts import render\nfrom .Run_Requests import *\nfrom .Profs import *\n\n\n# Create your views here.\ndef parse_modle_1_request(user_input:dict):\n proof_list = []\n for key in user_input:\n print(user_input[key])\n proof_list.append(Prof(key, user_input[key][0], user_input[key][1]))\n return proof_list\n\nkeyslist=[]\nvalueslist=[]\nprofslist=[]\n\n\ndef get_scores(request):\n\t# if this is a POST request we need to process the form data\n\tif request.method == 'POST':\n #print('request is POST')\n #print(type(request.POST)) \n #print(type(request.POST.lists()))\n #print(request.POST.lists())\n #print(parse_modle_1_request(request.POST.dict()) \n for key, value in request.POST.lists():\n print(key)\n print(value)\n \n if key == 'psyhco':\n #continue\n psycho = (value[0], value[1], value[2], value[3])\n \n else:\n keyslist.append(key)\n valueslist.append(value)\n \n #print(keyslist)\n #print(valueslist)\n for i in valueslist:\n i.insert(0, keyslist[valueslist.index(i)])\n #print(valueslist)\n for i in valueslist:\n print(i)\n this_prof=Prof(i[0],i[1],i[2])\n profslist.append(this_prof)\n print(profslist)\n print(psycho)\n all_results=main(profslist,psycho)\n \n \n return render(request, 'module_3.html', {\n'Tech_avg_bagrut_score': all_results[0].bagrot_score,\n'Tech_optimal_bagrut_score':all_results[0].optimal_bagrot_regex,\n'Tech_sechem_score':all_results[0].sechem_score,\n'Tech_accepted_majors':all_results[0].accepted_majors,\n'huji_avg_bagrut_score':all_results[1].bagrut_score,\n'huji_accepted_majors':all_results[1].accepted_str,\n'TLV_average_bagrut_score':all_results[2].bagrut_score,\n'TLV_sechem_score':all_results[2].sechem,\n'TLV_accepted_majors':all_results[2].accepted_str,\n'BGU_bagrut_score':all_results[3].bagrot_score,\n'BGU_scheme_score':all_results[3].scheme_score,\n#'':,\n})","sub_path":"Tziunim_server/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":2220,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"138384113","text":"#-*-coding:utf-8-*-\n# 计算大小写字母的个数\ndef string_test(s):\n d = {\"UPPER _CASE\":0, \"LOWER_CASE\":0} # 用字典,键值对的形式分别定义大小写字母的个数,大写:个数;小写:个数\n for c in s: # 遍历s中的每一个字母,\n if c.isupper(): # 如果该字母是大写的\n d[\"UPPER _CASE\"] += 1 #大写字母的个数加一\n elif c.islower():# 如果该字母是小写的\n d[\"LOWER_CASE\"] += 1 # 小写字母个数加一\n # else:\n # pass\n print(\"Original String :\", s)\n print(\"No.of Upper case characters :\", d[\"UPPER _CASE\"])\n print(\"No.of Lower case Characters :\", d[\"LOWER_CASE\"])\n\nstring_test(\"I am a Student.\")# 计算大小写字母的个数\n","sub_path":"function/exercise/7.py","file_name":"7.py","file_ext":"py","file_size_in_byte":761,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"277638774","text":"import appuifw\nimport btsocket as socket\nimport e32\n\nclass BTReader:\n\tdef connect(self):\n\t\tglobal sock\n\t\tarduino_addr='00:19:a4:02:44:2a' #add your arduino BT adress here\n\t\tsock=socket.socket(socket.AF_BT, socket.SOCK_STREAM)\n\t\ttarget=(arduino_addr,1) # serial connection to arduino BT\n\t\tsock.connect(target)\n\t\t\n\tdef readline(self):\n\t\tline=[]\n\t\twhile 1:\n\t\t\tch=self.sock.recv(1)\n\t\t\tif(ch==\"A\"):\n\t\t\t\tbreak\n\t\t\tline.append(ch)\n\t\treturn ''.join(line)\n\tdef writechar(ch):\n\t\t# ch=\"1\";\n\t\tself.sock.send(ch);\n\tdef close(self):\n\t\tself.sock.close();\n\nbt=BTReader()\nbt.connect()\n\nmode = appuifw.query(u\"MODE\",\"number\")\nred_val = appuifw.query(u\"RED\",\"number\")\ngreen_val = appuifw.query(u\"GREEN\",\"number\")\nblue_val = appuifw.query(u\"BLUE\",\"number\")\ndelay_val = appuifw.query(u\"Delay\",\"number\")\n\nm_str = \"m\" + str(mode)\nsock.send(m_str)\n\nr_str = \"r\" + chr(int(red_val))\nsock.send(r_str)\n\ng_str = \"g\" + chr(int(green_val))\nsock.send(g_str)\n\nb_str = \"b\" + chr(int(blue_val))\nsock.send(b_str)\n\nd_str = \"d\" + chr(int(delay_val))\nsock.send(d_str)","sub_path":"GUI/Mobile/SymbainS60/RGB_led_controller.py","file_name":"RGB_led_controller.py","file_ext":"py","file_size_in_byte":1029,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"377528016","text":"'''\nAuthor: your name\nDate: 2021-05-23 11:03:59\nLastEditTime: 2021-05-23 11:15:47\nLastEditors: Please set LastEditors\nDescription: In User Settings Edit\nFilePath: \\Study\\MagePython\\Chapter5\\5-2-5.py\n'''\nimport datetime\nimport time\n\ndef logger(fn):\n def wrap(*args,**kwargs):\n # before enhance function \n print(\"arg={},kwarge={}\".format(args,kwargs))\n start = datetime.datetime.now()\n ret = fn(*args,**kwargs)\n #after enhance function\n delta= datetime.datetime.now()-start\n print(\"function {} took {}s.\".format(fn.__name__,delta.total_seconds()))\n return ret\n return wrap\n\n@logger\ndef add(x,y):\n print(\"===call add =========================\")\n time.sleep(6)\n return x + y\n\nprint(add(4,y=7))\n ","sub_path":"MagePython/Chapter5/5-2-5.py","file_name":"5-2-5.py","file_ext":"py","file_size_in_byte":769,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"607182763","text":"import math\n\ndef miniWindow(arr):\n low, high = 0, len(arr) - 1\n while low < len(arr) - 1 and arr[low] <= arr[low + 1]:\n low += 1\n if low == len(arr):\n return 0\n while high > 0 and arr[high] >= arr[high - 1]:\n high -= 1\n if high == 0:\n return len(arr)\n \n minNum = math.inf\n maxNum = -math.inf\n for k in range(low, high + 1):\n minNum = min(minNum, arr[k])\n maxNum = max(maxNum, arr[k])\n \n while low > 0 and arr[low - 1] > minNum:\n low -= 1\n while high < len(arr) - 1 and arr[high + 1] < maxNum:\n high += 1\n\n return high - low + 1\n\ndef main():\n print(miniWindow([1,2,5,3,7,10,9,12]))\n\nmain()","sub_path":"Algorithm/src/twopointers/C3_M_MiniWindowSort/C3_M_MiniWindowSort.py","file_name":"C3_M_MiniWindowSort.py","file_ext":"py","file_size_in_byte":626,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"383398264","text":"import argparse\n\nimport numpy as np\nimport torch\nfrom torch import optim\nfrom torchvision import utils\nfrom tqdm import tqdm\n\nfrom model import Glow\nfrom samplers import memory_mnist, memory_fashion, celeba, ffhq_5, cifar_horses_40, ffhq_50, cifar_horses_20, cifar_horses_80, mnist_30, mnist_gan_all, mnist_pad, cifar_horses_20_top, cifar_horses_40_top, cifar_horses_20_top_small_lr, cifar_horses_40_top_small_lr, arrows_small, arrows_big, cifar_20_picked_inds_2, cifar_40_picked_inds_2, cifar_20_picked_inds_3, cifar_40_picked_inds_3\nfrom utils import (\n net_args,\n calc_z_shapes,\n calc_loss,\n string_args,\n)\n\nparser = net_args(argparse.ArgumentParser(description=\"Glow trainer\"))\n\n\ndef train(args, model, optimizer):\n if args.dataset == \"mnist\":\n dataset_f = memory_mnist\n elif args.dataset == \"fashion_mnist\":\n dataset_f = memory_fashion\n elif args.dataset == \"celeba\":\n dataset_f = celeba\n elif args.dataset == \"ffhq_gan_32\":\n dataset_f = ffhq_gan_32\n elif args.dataset == \"cifar_horses_40\":\n dataset_f = cifar_horses_40\n elif args.dataset == \"ffhq_50\":\n dataset_f = ffhq_50\n elif args.dataset == \"cifar_horses_20\":\n dataset_f = cifar_horses_20\n elif args.dataset == \"cifar_horses_80\":\n dataset_f = cifar_horses_80\n elif args.dataset == \"mnist_30\":\n dataset_f = mnist_30\n elif args.dataset == \"mnist_gan_all\":\n dataset_f = mnist_gan_all\n elif args.dataset == \"mnist_pad\":\n dataset_f = mnist_pad\n elif args.dataset == \"cifar_horses_20_top\":\n dataset_f = cifar_horses_20_top\n elif args.dataset == \"cifar_horses_40_top\":\n dataset_f = cifar_horses_40_top\n elif args.dataset == \"cifar_horses_20_top_small_lr\":\n dataset_f = cifar_horses_20_top_small_lr\n elif args.dataset == \"cifar_horses_40_top_small_lr\":\n dataset_f = cifar_horses_40_top_small_lr\n elif args.dataset == \"arrows_small\":\n dataset_f = arrows_small\n elif args.dataset == \"arrows_big\":\n dataset_f = arrows_big\n elif args.dataset == \"cifar_20_picked_inds_2\":\n dataset_f = cifar_20_picked_inds_2\n elif args.dataset == \"cifar_40_picked_inds_2\":\n dataset_f = cifar_40_picked_inds_2\n elif args.dataset == \"cifar_40_picked_inds_3\":\n dataset_f = cifar_40_picked_inds_3\n elif args.dataset == \"cifar_20_picked_inds_3\":\n dataset_f = cifar_20_picked_inds_3\n else:\n raise ValueError(\"Unknown dataset:\", args.dataset)\n\n repr_args = string_args(args)\n n_bins = 2.0 ** args.n_bits\n\n z_sample = []\n z_shapes = calc_z_shapes(args.n_channels, args.img_size, args.n_flow, args.n_block)\n for z in z_shapes:\n z_new = torch.randn(args.n_sample, *z) * args.temp\n z_sample.append(z_new.to(device))\n\n epoch_losses = []\n f_train_loss = open(f\"losses/losses_train_{repr_args}_.txt\", \"a\", buffering=1)\n f_test_loss = open(f\"losses/losses_test_{repr_args}_.txt\", \"a\", buffering=1)\n\n last_model_path = f\"checkpoint/model_{repr_args}_last_.pt\"\n try:\n model.load_state_dict(torch.load(last_model_path))\n model.eval()\n f_epoch = open(f\"checkpoint/last_epoch_{repr_args}.txt\", \"r\", buffering=1)\n epoch_n = int(f_epoch.readline().strip())\n f_epoch.close()\n except FileNotFoundError:\n print(\"Training the model from scratch.\")\n epoch_n = 0\n\n with tqdm(range(epoch_n, args.epochs + epoch_n)) as pbar:\n for i in pbar:\n repr_args = string_args(args)\n train_loader, val_loader, train_val_loader = dataset_f(\n args.batch, args.img_size, args.n_channels\n )\n train_losses = []\n for image in train_loader:\n if isinstance(image, list):\n image = image[0]\n optimizer.zero_grad()\n image = image.to(device)\n noisy_image = image\n if args.tr_dq:\n noisy_image += torch.rand_like(image) / n_bins\n noisy_image += torch.randn_like(image) * args.delta\n log_p, logdet, _ = model(noisy_image)\n\n logdet = logdet.mean()\n loss, log_p, log_det = calc_loss(\n log_p, logdet, args.img_size, n_bins, args.n_channels\n )\n loss.backward()\n optimizer.step()\n train_losses.append(loss.item())\n current_train_loss = np.mean(train_losses)\n print(f\"{current_train_loss},{args.delta},{i + 1}\", file=f_train_loss)\n with torch.no_grad():\n utils.save_image(\n model.reverse(z_sample).cpu().data,\n f\"sample/sample_{repr_args}_{str(i + 1).zfill(6)}.png\",\n normalize=True,\n nrow=10,\n range=(-0.5, 0.5),\n )\n losses = []\n logdets = []\n logps = []\n for image in val_loader:\n if isinstance(image, list):\n image = image[0]\n image = image.to(device)\n log_p, logdet, _ = model(image)\n logdet = logdet.mean()\n loss, log_p, log_det = calc_loss(\n log_p, logdet, args.img_size, n_bins, args.n_channels\n )\n losses.append(loss.item())\n logdets.append(log_det.item())\n logps.append(log_p.item())\n pbar.set_description(\n f\"Loss: {np.mean(losses):.5f}; logP: {np.mean(logps):.5f}; logdet: {np.mean(logdets):.5f}; delta: {args.delta:.5f}\"\n )\n current_loss = np.mean(losses)\n print(f\"{current_loss},{args.delta},{i + 1}\", file=f_test_loss)\n epoch_losses.append(current_loss)\n # early stopping\n if len(epoch_losses) >= 20 and epoch_losses[-20] < min(epoch_losses[-19:]):\n break\n '''\n too much space\n if (i + 1) % 5 == 0:\n torch.save(\n model.state_dict(), f\"checkpoint/model_{repr_args}_{i + 1}_.pt\"\n )\n '''\n torch.save(model.state_dict(), last_model_path)\n f_epoch = open(\n f\"checkpoint/last_epoch_{repr_args}.txt\", \"w\", buffering=1\n )\n f_epoch.write(str(i + 1))\n f_epoch.close()\n\n f_ll = open(f\"ll/ll_{repr_args}_{i + 1}.txt\", \"w\")\n train_loader, val_loader, train_val_loader = dataset_f(\n args.batch, args.img_size, args.n_channels\n )\n train_val_loader = iter(train_val_loader)\n for image_val in val_loader:\n image = image_val\n if isinstance(image, list):\n image = image[0]\n image = image.to(device)\n log_p_val, logdet_val, _ = model(image)\n\n image = next(train_val_loader)\n if isinstance(image, list):\n image = image[0]\n image = image.to(device)\n log_p_train_val, logdet_train_val, _ = model(image)\n\n for (\n lpv,\n ldv,\n lptv,\n ldtv,\n ) in zip(log_p_val, logdet_val, log_p_train_val, logdet_train_val):\n print(\n args.delta,\n lpv.item(),\n ldv.item(),\n lptv.item(),\n ldtv.item(),\n file=f_ll,\n )\n f_ll.close()\n f_train_loss.close()\n f_test_loss.close()\n\n\nif __name__ == \"__main__\":\n args = parser.parse_args()\n print(string_args(args))\n device = args.device\n model = Glow(\n args.n_channels,\n args.n_flow,\n args.n_block,\n affine=args.affine,\n conv_lu=not args.no_lu,\n )\n model = model.to(device)\n optimizer = optim.Adam(model.parameters(), lr=args.lr)\n train(args, model, optimizer)\n","sub_path":"train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":8404,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"650911909","text":"import cv2\nimport cv2.cv as cv\n\n\nidc_hl = int(62)\nidc_sl = int(100)\nidc_vl = int(50)\n\nidc_hh = int(82)\nidc_sh = int(200)\nidc_vh = int(200)\n\nidc_green_range = (cv.Scalar(idc_hl,idc_sl,idc_vl), cv.Scalar(idc_hh,idc_sh,idc_vh))\n\nridc_hl = int(172)\nridc_sl = int(180)\nridc_vl = int(150)\n\nridc_hh = int(7)\nridc_sh = int(220)\nridc_vh = int(254)\n\nidc_red_range = (cv.Scalar(ridc_hl,ridc_sl,ridc_vl), cv.Scalar(ridc_hh,ridc_sh,ridc_vh))\n\nyidc_hl = int(20)\nyidc_sl = int(133)\nyidc_vl = int(121)\n\nyidc_hh = int(27)\nyidc_sh = int(193)\nyidc_vh = int(243)\n\nidc_yellow_range = (cv.Scalar(yidc_hl,yidc_sl,yidc_vl), cv.Scalar(yidc_hh,yidc_sh,yidc_vh))\n\n# Reduce image size\ndef bgr_to_hsv(img):\n\t# Convert from BGR to HSV\n\thsv = cv.CreateImage(cv.GetSize(img), 8, 3)\n\tcv.CvtColor(img, hsv, cv.CV_BGR2HSV)\n\treturn hsv\n\ndef threshold_green_balls(img):\n\t# Threshold the img in hsv space for green\n\timg_thresh = cv.CreateImage(cv.GetSize(img), 8, 1)\n\tcv2.cv.InRangeS(img, idc_green_range[0], idc_green_range[1], img_thresh)\n\treturn img_thresh\n\ndef threshold_red_balls(img):\n\t# Threshold the img in hsv space for green\n\timg_thresh = cv.CreateImage(cv.GetSize(img), 8, 1)\n\tcv2.cv.InRangeS(img, idc_red_range[0], idc_red_range[1], img_thresh)\n\treturn img_thresh","sub_path":"b2/thalamic_modules.py","file_name":"thalamic_modules.py","file_ext":"py","file_size_in_byte":1237,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"394067658","text":"\"\"\"\n207. Course Schedule\n- Medium\n- DFS, BFS, Graph, Topological Sort\n- Link: https://leetcode.com/problems/course-schedule/\n\"\"\"\n\n\n# Solution 1: Topological Sort\n# Time: O(E + V), where V is the number of vertex and E is the number of edge | Space: O(E + V)\nclass GNode(object):\n def __init__(self):\n self.inDegrees = 0\n self.outNodes = []\n\n\nclass Solution:\n def canFinish(self, numCourses: int, prerequisites: List[List[int]]) -> bool:\n from collections import defaultdict, deque\n\n graph = defaultdict(GNode)\n\n # 建立 graph,計算 number of dependency\n totalDeps = 0\n for relation in prerequisites:\n nextCourse, prevCourse = relation[0], relation[1]\n graph[prevCourse].outNodes.append(nextCourse)\n graph[nextCourse].inDegrees += 1\n totalDeps += 1 # total dependencies\n\n # 紀錄起始節點 (不存在 indegrees 的 node)\n nodepCourses = deque()\n for idx, node in graph.items():\n if node.inDegrees == 0:\n nodepCourses.append(idx)\n\n # 從起始節點開始,針對其 outNodes 開始消除 edge/dependency\n removedEdges = 0\n while nodepCourses:\n course = nodepCourses.pop()\n\n for nextCourse in graph[course].outNodes:\n graph[nextCourse].inDegrees -= 1\n removedEdges += 1\n\n # 當 inDegrees 不等於 0 時,代表該處有迴圈的可能\n if graph[nextCourse].inDegrees == 0:\n nodepCourses.append(nextCourse)\n\n # 當消除的 edge 等同於一開始計算的 dependency 數量,代表該圖不存在 cycle\n if removedEdges == totalDeps:\n return True\n else:\n return False\n","sub_path":"leetcode/graph/207_course_schedule.py","file_name":"207_course_schedule.py","file_ext":"py","file_size_in_byte":1789,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"4562753","text":"from collections import Counter\nX = int(input())\nshoe_sizes = Counter(map(int, input().split()))\nN = int(input())\narr = []\nfor i in range(N):\n arr_demand = [int(a) for a in input().split()]\n arr.append(arr_demand)\n\nearn = 0\nfor i in arr:\n size = i[0]\n if shoe_sizes[size]:\n earn += i[1]\n shoe_sizes[size] -= 1\n\n# print(shoe_sizes)\n# print(arr)\nprint(earn)\n\n# similar but simpler one\nimport collections\nnumShoes = int(input())\nshoes = collections.Counter(map(int, input().split()))\nnumCust = int(input())\nincome = 0\nfor i in range(numCust):\n size, price = map(int, input().split())\n if shoes[size]:\n income += price\n shoes[size] -= 1\nprint(income)","sub_path":"HR/HR_Py_Coll_Cntr.py","file_name":"HR_Py_Coll_Cntr.py","file_ext":"py","file_size_in_byte":692,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"242323022","text":"from django.urls import path\n\nfrom . import views\n\nurlpatterns = [\n path(\"\", views.index, name=\"index\"),\n path(\"search/\", views.search, name=\"search\"),\n path(\"random/\", views.random_page, name=\"random_page\"),\n path(\"editor/\", views.editor, name=\"editor\"),\n path(\"wiki/\", views.wiki, name=\"entry-detail\")\n]\n","sub_path":"encyclopedia/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":332,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"240235504","text":"# -*- coding: utf-8 -*-\n# @Author: lidong\n# @Date: 2018-03-18 16:31:14\n# @Last Modified by: yulidong\n# @Last Modified time: 2018-09-21 23:32:28\n\nimport torch\nimport numpy as np\nimport torch.nn as nn\nimport torch.nn.functional as F\n\ndef l1_r(input, target, weight=None, size_average=True):\n relation=[]\n loss=nn.MSELoss()\n\n for i in range(3):\n target=torch.reshape(target,(input[i].shape))\n #print(target.shape)\n t=loss(input[i],target)\n #print(t.item())\n relation.append(t)\n\n return relation\ndef l1_a(input, target, weight=None, size_average=True):\n relation=[]\n loss=nn.MSELoss()\n\n for i in range(4):\n target=torch.reshape(target,(input[i].shape))\n #print(target.shape)\n t=torch.sqrt(loss(input[i],target))\n #print(t.item())\n relation.append(t)\n\n return relation\ndef log_r(input, target, weight=None, size_average=True):\n relation=[]\n d=[]\n out=[]\n target=torch.reshape(target,(input[0].shape))\n target=torch.log(target+1e-6)\n loss=nn.MSELoss()\n for i in range(3):\n # pre=input[i]\n # num=torch.sum(torch.where(pre>0,torch.ones_like(pre),torch.zeros_like(pre)))/torch.sum(torch.ones_like(pre))\n # print(num)\n input[i]=torch.log(input[i]+1e-6) \n relation.append(loss(input[i],target))\n #d.append(0.5*torch.pow(torch.sum(input[i]-target),2)/torch.pow(torch.sum(torch.ones_like(input[i])),2))\n #out.append(relation[i]-d[i])\n return relation \ndef cross_entropy2d(input, target, weight=None, size_average=True):\n n, c, h, w = input.size()\n #print(c,target.max().data.cpu().numpy())\n\n log_p = F.log_softmax(input, dim=1)\n log_p = log_p.transpose(1, 2).transpose(2, 3).contiguous().view(-1, c)\n log_p = log_p[target.view(n * h * w, 1).repeat(1, c) >= 0]\n log_p = log_p.view(-1, c)\n\n mask = target >= 0\n target = target[mask]\n #loss=log_p.sum()\n loss = F.nll_loss(log_p, target,\n weight=weight, size_average=False)\n #print(loss, mask.data.sum())\n if size_average:\n # print(mask.data.sum())\n loss /= mask.data.sum()\n # loss=loss/(950*540)\n return loss\ndef l1(input, target,mask, weight=None, size_average=True):\n # P1=mask[...,0].cuda(0)\n # P2=mask[...,3].cuda(0)\n one,zero=torch.ones(1).cuda(0),torch.zeros(1).cuda(0)\n mask=torch.where(input>0,one,zero)\n #print(torch.sum(mask)/(input.shape[0]*input.shape[1]))\n mask=torch.reshape(mask,(input.shape))\n target=torch.reshape(target,(input.shape))\n loss=nn.L1Loss(reduction='none')\n #loss=nn.MSELoss(reduction='none')\n #print(torch.sqrt(loss(input,target)).shape)\n relation=torch.sum(mask*loss(input,target))/torch.sum(mask)\n #print(torch.max(torch.where(input>zero,input,zero)),torch.min(torch.where(P1>zero,input,192*one)))\n #relation=torch.sum(torch.sqrt(mask*loss(input,target)))/torch.sum(mask)\n return relation\ndef l2(input, target, weight=None, size_average=True):\n target=torch.reshape(target,(input.shape))\n #print(input.shape)\n #print(target.shape)\n # num=torch.sum(torch.where(input==0,torch.ones_like(input),torch.zeros_like(input)))\n # positive=num/torch.sum(torch.ones_like(input))\n #print(positive.item())\n loss=nn.MSELoss()\n relation=loss(input,target)\n #mean=torch.abs(torch.mean(input)-torch.mean(target))\n #print(\"pre_depth:%.4f,ground_depth:%.4f\"%(torch.mean(input[1]).data.cpu().numpy().astype('float32'),torch.mean(target).data.cpu().numpy().astype('float32')))\n #output=relation+0.2*mean\n return relation\ndef log_loss(input, target, weight=None, size_average=True):\n # num=torch.sum(torch.where(input==0,torch.ones_like(input),torch.zeros_like(input)))\n # positive=num/torch.sum(torch.ones_like(input))\n # print(positive.item())\n target=torch.reshape(target,(input.shape))\n loss=nn.MSELoss() \n input=torch.log(input+1e-12) \n target=torch.log(target+1e-12) \n #relation=torch.sqrt(loss(input,target)) \n relation=loss(input,target) \n d=0.5*torch.pow(torch.sum(input-target),2)/torch.pow(torch.sum(torch.ones_like(input)),2)\n #relation=relation-d \n return relation\n\n # target=torch.reshape(target,(input.shape))\n # #loss=nn.MSELoss()\n # num=torch.sum(torch.where(input>0,torch.ones_like(input),torch.zeros_like(input)))\n # input=torch.log(torch.where(input>0,input,torch.ones_like(input)))\n # target=torch.log(torch.where(target>0,target,torch.ones_like(target)))\n # # #relation=torch.sqrt(loss(input,target))\n # relation=torch.sum(torch.pow(torch.where(input==0,input,input-target),2))/num\n # d=torch.pow(torch.sum(torch.where(input==0,input,input-target)),2)/torch.pow(num,2)*0.5\n # #positive=num/torch.sum(torch.ones_like(input))\n # #print(positive.item())\n # #-torch.sum(torch.where(input<0,input,torch.zeros_like(input)))/num\n # losses=relation+d\n # return losses\n\ndef log_l1(input, target, weight=None, size_average=True):\n l1loss=l1(input,target)\n logloss=log_loss(input,target)\n num=torch.sum(torch.where(input==0,torch.ones_like(input),torch.zeros_like(input)))\n positive=num/torch.sum(torch.ones_like(input))\n print(positive.item())\n loss=(1-positive)*logloss+positive*l1loss\n return loss\ndef l1_kitti(input, target, weight=None, size_average=True):\n zero=torch.zeros_like(input)\n target=torch.reshape(target,(input.shape))\n input=torch.where(target>0,input,zero)\n target=torch.where(target>0,target,zero)\n loss=nn.MSELoss(size_average=False) \n relation=loss(input,target)/torch.sum(torch.where(target>0,torch.ones_like(input),zero))\n return relation\ndef log_kitti(input, target, weight=None, size_average=True):\n zero=torch.zeros_like(input)\n target=torch.reshape(target,(input.shape))\n loss=nn.MSELoss(size_average=False) \n input=torch.where(target>0,torch.log(input),zero)\n target=torch.where(target>0,torch.log(target),zero)\n\n #relation=torch.sqrt(loss(input,target)) \n relation=loss(input,target)/torch.sum(torch.where(target>0,torch.ones_like(input),zero))\n d=0.5*torch.pow(torch.sum(input-target),2)/torch.pow(torch.sum(torch.where(target>0,torch.ones_like(input),zero)),2)\n \n return relation-d \n# def region(input,target,instance):\n# loss=0\n# lf=nn.MSELoss(size_average=False,reduce=False)\n# target=torch.reshape(target,(input.shape))\n# instance=torch.reshape(instance,(input.shape))\n# zero=torch.zeros_like(input)\n# one=torch.ones_like(input)\n# dis=lf(input,target)\n# for i in range(0,int(torch.max(instance).item()+1)):\n# input_region=torch.where(instance==i,input,zero)\n# ground_region=torch.where(instance==i,target,zero)\n# m=torch.max(ground_region)\n# if m==0:\n# continue\n# num=torch.sum(torch.where(instance==i,one,zero))\n# loss+=lf(input_region,ground_region)/num\n# # average=torch.sum(input_region)/num\n# # input_region=input_region-average\n# # input_region=torch.pow(input_region,2)\n# # var=torch.sum(input_region)/num\n# # loss+=0.5*var\n# loss=loss/torch.max(instance)\n# return loss\n\n\ndef region(input,target,instance):\n loss=0\n lf=nn.MSELoss(size_average=False,reduce=False)\n target=torch.reshape(target,(input.shape))\n # input=torch.log(input+1e-12) \n # target=torch.log(target+1e-12) \n instance=torch.reshape(instance,(input.shape))\n zero=torch.zeros_like(input)\n one=torch.ones_like(input)\n dis=lf(input,target)\n for i in range(1,int(torch.max(instance).item()+1)):\n dis_region=torch.where(instance==i,dis,zero)\n num=torch.sum(torch.where(instance==i,one,zero))\n average=torch.sum(dis_region)/num\n loss=loss+average\n #dis_region=torch.where(instance==i,dis_region-average,zero)\n # var=0.1*torch.sqrt(torch.sum(torch.pow(dis_region,2))/num)/average\n # loss=loss+var\n loss=loss/(torch.max(instance))\n return loss\n\ndef region_log(input,target,instance):\n loss=0\n lf=nn.MSELoss(size_average=False,reduce=False)\n target=torch.reshape(target,(input.shape))\n input=torch.log(input+1e-6) \n target=torch.log(target+1e-6) \n instance=torch.reshape(instance,(input.shape))\n zero=torch.zeros_like(input)\n one=torch.ones_like(input)\n dis=lf(input,target)\n for i in range(1,int(torch.max(instance).item()+1)):\n dis_region=torch.where(instance==i,dis,zero)\n num=torch.sum(torch.where(instance==i,one,zero))\n average=torch.sum(dis_region)/num\n loss=loss+average\n # dis_region=torch.where(instance==i,dis_region-average,zero)\n # var=(torch.sum(torch.pow(dis_region,2))/num)/average\n # loss=loss+var\n loss=loss/(torch.max(instance))\n #print(torch.max(instance).item())\n return loss\n\n\ndef region_r(input,target,instance):\n loss=0\n relation=[]\n lf=nn.MSELoss(size_average=False,reduce=False)\n target=torch.reshape(target,(input[0].shape))\n \n target=torch.log(target+1e-6) \n instance=torch.reshape(instance,(input[0].shape))\n zero=torch.zeros_like(input[0])\n one=torch.ones_like(input[0])\n for i in range(3):\n input[i]=torch.log(input[i]+1e-6)\n dis=lf(input[i],target)\n for i in range(1,int(torch.max(instance).item()+1)):\n dis_region=torch.where(instance==i,dis,zero)\n num=torch.sum(torch.where(instance==i,one,zero))\n average=torch.sum(dis_region)/num\n loss=loss+average\n #print(torch.max(instance).item())\n relation.append(loss/(torch.max(instance)))\n loss=0\n return relation","sub_path":"back of code/CMF/cmf/loss-20181014143903.py","file_name":"loss-20181014143903.py","file_ext":"py","file_size_in_byte":9650,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"509003424","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nfrom decimal import Decimal\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('booking', '0012_auto_20150219_1918'),\n ]\n\n operations = [\n migrations.AlterField(\n model_name='userprofile',\n name='cash',\n field=models.DecimalField(\n default=Decimal('0.0'), max_digits=6, decimal_places=2),\n preserve_default=True,\n ),\n ]\n","sub_path":"Booking/booking/migrations/0013_auto_20150220_1519.py","file_name":"0013_auto_20150220_1519.py","file_ext":"py","file_size_in_byte":529,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"197122088","text":"\ndef z_algorithm(s):\n \"\"\"\n Given a string S of length n, the Z Algorithm produces an array Z where Z[i] is the length of the longest substring\n starting from S[i] which is also a prefix of S, i.e. the maximum k such that S[j] = S[i+j] or all 0<=j R):\n L = R = i\n while R < n and s[R-L] == s[R]:\n R += 1\n z[i] = R - L\n R -= 1\n else:\n k = i - L\n if z[k] < R - i + 1:\n z[i] = z[k]\n else:\n L = i\n while R < n and s[R-L] == s[R]:\n R += 1\n z[i] = R - L\n R -= 1\n return z\n","sub_path":"alg/z_algorithm.py","file_name":"z_algorithm.py","file_ext":"py","file_size_in_byte":948,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"610959116","text":"import string\n\ndef filter_words(words, skip_words):\n \"\"\"This function takes a list of words and returns a copy of the list from\n which all words provided in the list skip_words have been removed.\n For example:\n \"\"\"\n initial_input = words\n new_input = []\n\n for s in initial_input:\n valid = True\n for s_w in skip_words:\n if(s == s_w):\n valid = False\n if (valid):\n new_input = new_input + [s]\n return new_input\n\n\n\n \ndef remove_punct(text):\n \"\"\"This function is used to remove all punctuation\n marks from a string. Spaces do not count as punctuation and should\n not be removed. The funcion takes a string and returns a new string\n which does not contain any puctuation. For example:\n \"\"\"\n no_punct = \"\"\n for char in text:\n if not (char in string.punctuation):\n no_punct = no_punct + char\n\n return no_punct\n\n\ndef normalise_input(user_input):\n \"\"\"This function removes all punctuation from the string and converts it to\n lower case. It then splits the string into a list of words (also removing\n any extra spaces between words) and further removes all \"unimportant\"\n words from the list of words using the filter_words() function. The\n resulting list of \"important\" words is returned. For example:\n \"\"\"\n # Remove punctuation and convert to lower case\n no_punct = remove_punct(user_input).lower()\n word_temp = \"\"\n word_list = []\n\n for ch in no_punct:\n if(ch != ' '):\n word_temp = word_temp + ch\n elif(ch == ' '):\n if (word_temp != \"\"):\n word_list = word_list + [word_temp]\n word_temp = \"\"\n\n if (word_temp != \"\"):\n word_list = word_list + [word_temp]\n \n word_list = filter_words(word_list, skip_words)\n\n return word_list\n\n","sub_path":"gameparser.py","file_name":"gameparser.py","file_ext":"py","file_size_in_byte":1859,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"366648755","text":"from tkinter import *\r\nfrom PIL import Image, ImageTk\r\nimport requests, json, io, socket, threading\r\n\r\nclass Login:\r\n def __init__(self):\r\n self.root = Tk()\r\n self.root.title(\"Chat - Login\")\r\n #self.root.iconbitmap(\"icon.ico\")\r\n self.root.resizable(0, 0)\r\n self.root.configure(bg=\"#505050\", pady=20, padx=20)\r\n\r\n self.clLogin = Frame(self.root, bg=\"#505050\")\r\n self.clLogin.pack()\r\n\r\n self.login = ImageTk.PhotoImage(Image.open(\"Login.png\"))\r\n self.lLogin = Label(self.clLogin, font=(\"Verdana\", \"30\"), bg=\"#505050\", fg=\"#FFFFFF\", image=self.login, height=100)\r\n self.lLogin.pack()\r\n\r\n self.autenticado = Frame(self.root, bg=\"#505050\")\r\n self.autenticado.pack()\r\n\r\n self.lautenticar = Label(self.autenticado, bg=\"#505050\", fg=\"#FFFFFF\")\r\n self.lautenticar.pack()\r\n\r\n self.ccredenciais = Frame(self.root, bg=\"#505050\")\r\n self.ccredenciais.pack()\r\n\r\n self.cllcredenciais = Frame(self.ccredenciais, bg=\"#505050\")\r\n self.cllcredenciais.pack(side=LEFT)\r\n\r\n self.lmatricula = Label(self.cllcredenciais, font=(\"Verdana\", \"18\"), bg=\"#505050\", fg=\"#FFFFFF\")\r\n self.lmatricula[\"text\"] = \"Matrícula:\"\r\n self.lmatricula.pack()\r\n\r\n self.lSenha = Label(self.cllcredenciais, font=(\"Verdana\", \"18\"), bg=\"#505050\", fg=\"#FFFFFF\")\r\n self.lSenha[\"text\"] = \"Senha:\"\r\n self.lSenha.pack()\r\n\r\n self.ceecredenciais = Frame(self.ccredenciais, bg=\"#505050\", pady=5)\r\n self.ceecredenciais.pack()\r\n\r\n self.matricula = Entry(self.ceecredenciais, width=20, bg=\"#707070\", borderwidth=0, font=(\"Verdana\", \"15\"), insertbackground=\"#FFFFFF\", fg=\"white\")\r\n self.matricula.bind(\"\", self.autenticacao)\r\n self.matricula.pack()\r\n\r\n self.senha = Entry(self.ceecredenciais, width=20, bg=\"#707070\", borderwidth=0, show=\"●\", font=(\"Verdana\", \"15\"), insertbackground=\"#FFFFFF\", fg=\"white\")\r\n self.senha.bind(\"\", self.autenticacao)\r\n self.senha.pack(pady=(8,0))\r\n\r\n self.cAutenticar = Frame(self.root, bg=\"#505050\")\r\n self.cAutenticar.pack(pady=(15,0))\r\n\r\n self.autenticar = Button(self.cAutenticar, width=15, bg=\"#707070\", borderwidth=0, font=(\"Verdana\", \"18\"), fg=\"white\")\r\n self.autenticar[\"text\"] = \"Autenticar\"\r\n self.autenticar.bind(\"\", self.autenticacao)\r\n self.autenticar.pack()\r\n\r\n self.root.mainloop()\r\n\r\n def token(self):\r\n matricula = self.matricula.get()\r\n senha = self.senha.get()\r\n url_token = \"https://suap.ifrn.edu.br/api/v2/autenticacao/token/\"\r\n credenciais = {\r\n \"username\":matricula,\r\n \"password\":senha\r\n }\r\n response = requests.post(url_token, data=credenciais)\r\n if response.status_code == 200:\r\n tokenDir = json.loads(response.content.decode(\"utf-8\"))\r\n return tokenDir[\"token\"]\r\n else:\r\n return None\r\n\r\n def autenticacao(self, event):\r\n if self.token() == None:\r\n self.matricula.delete(0,END)\r\n self.senha.delete(0,END)\r\n self.lautenticar[\"text\"] = \"Usuário e/ou senha inválidos!\"\r\n \r\n else:\r\n self.recuperaDados()\r\n\r\n def recuperaDados(self):\r\n cabecalho = {\"Authorization\":\"JWT {}\".format(self.token())}\r\n url_dados = \"https://suap.ifrn.edu.br/api/v2/minhas-informacoes/meus-dados/\"\r\n self.dadosDir = requests.get(url_dados, headers=cabecalho)\r\n self.dadosDir = json.loads(self.dadosDir.content.decode(\"utf-8\"))\r\n self.imagem = requests.get(\"https://suap.ifrn.edu.br\" + self.dadosDir[\"url_foto_75x100\"])\r\n self.imagem = self.imagem.content\r\n self.imagem = io.BytesIO(self.imagem)\r\n self.imagemopen = Image.open(self.imagem)\r\n self.foto = ImageTk.PhotoImage(self.imagemopen)\r\n self.reformular()\r\n\r\n def reformular(self):\r\n self.clLogin.destroy()\r\n self.lLogin.destroy()\r\n self.autenticado.destroy()\r\n self.lautenticar.destroy()\r\n self.ccredenciais.destroy()\r\n self.cllcredenciais.destroy()\r\n self.lmatricula.destroy()\r\n self.lSenha.destroy()\r\n self.ceecredenciais.destroy()\r\n self.matricula.destroy()\r\n self.senha.destroy()\r\n self.cAutenticar.destroy()\r\n self.autenticar.destroy()\r\n \r\n self.root.title(\"Chat - Cliente\")\r\n self.root.configure(bg=\"#404040\", padx=20)\r\n\r\n self.cliente()\r\n\r\n self.D = {}\r\n\r\n self.cCons = Frame(self.root, bg=\"#606060\")\r\n self.cCons.pack(side=LEFT, fill=BOTH)\r\n\r\n self.cMens = Frame(self.root, bg=\"#606060\", padx=10)\r\n\r\n self.cMenRecv = Frame(self.cMens, bg=\"#606060\", pady=10)\r\n self.cMenRecv.pack(fill=BOTH, expand=1)\r\n \r\n self.cMenstyle = Frame(self.cMenRecv, bg=\"#707070\", padx=4)\r\n self.cMenstyle.pack(fill=BOTH, expand=1)\r\n \r\n self.cRecv = Frame(self.cMenstyle, bg=\"#606060\")\r\n self.cRecv.pack(fill=BOTH, expand=1)\r\n\r\n self.cEnv = Frame(self.cMens, bg=\"#606060\", padx=10, pady=10)\r\n self.cEnv.pack(fill=BOTH)\r\n\r\n self.cIdentificacao = Frame(self.cCons, bg=\"#606060\", padx=5, pady=5)\r\n self.cIdentificacao.pack(fill=X)\r\n\r\n self.cIFoto = Frame(self.cIdentificacao, bg=\"#505050\", padx=3, pady=3)\r\n self.cIFoto.pack(side=LEFT)\r\n\r\n self.iIdentificacao = Label(self.cIFoto, image=self.foto, borderwidth=0)\r\n self.iIdentificacao.pack()\r\n\r\n self.ccIdentificacao = Frame(self.cIdentificacao)\r\n self.ccIdentificacao.pack(side=LEFT, fill=X)\r\n \r\n self.l1Identificacao = Label(self.ccIdentificacao, text=\"Nome:\", bg=\"#606060\", fg=\"#FFFFFF\", font=(\"Verdana\", \"8\", \"bold\"), anchor=W)\r\n self.l1Identificacao.pack(fill=X)\r\n\r\n self.l2Identificacao = Label(self.ccIdentificacao, text=\"Matrícula:\", bg=\"#606060\", fg=\"#FFFFFF\", font=(\"Verdana\", \"8\", \"bold\"), anchor=W)\r\n self.l2Identificacao.pack(fill=X)\r\n\r\n self.l3Identificacao = Label(self.ccIdentificacao, text=\"Email:\", bg=\"#606060\", fg=\"#FFFFFF\", font=(\"Verdana\", \"8\", \"bold\"), anchor=W)\r\n self.l3Identificacao.pack(fill=X)\r\n\r\n self.l4Identificacao = Label(self.ccIdentificacao, text=\"Curso:\", bg=\"#606060\", fg=\"#FFFFFF\", font=(\"Verdana\", \"8\", \"bold\"), anchor=W)\r\n self.l4Identificacao.pack(fill=X)\r\n\r\n self.l5Identificacao = Label(self.ccIdentificacao, text=\"Vínculo:\", bg=\"#606060\", fg=\"#FFFFFF\", font=(\"Verdana\", \"8\", \"bold\"), anchor=W)\r\n self.l5Identificacao.pack(fill=X)\r\n\r\n self.cccIdentificacao = Frame(self.cIdentificacao, pady=5, bg=\"#606060\")\r\n self.cccIdentificacao.pack(fill=X)\r\n\r\n self.ll1Identificacao = Label(self.cccIdentificacao, text=self.dadosDir[\"vinculo\"][\"nome\"], bg=\"#606060\", fg=\"#FFFFFF\", font=(\"Verdana\", \"8\"), anchor=W)\r\n self.ll1Identificacao.pack(fill=X)\r\n\r\n self.ll2Identificacao = Label(self.cccIdentificacao, text=self.dadosDir[\"vinculo\"][\"matricula\"], bg=\"#606060\", fg=\"#FFFFFF\", font=(\"Verdana\", \"8\"), anchor=W)\r\n self.ll2Identificacao.pack(fill=X)\r\n\r\n self.ll3Identificacao = Label(self.cccIdentificacao, text=self.dadosDir[\"email\"], bg=\"#606060\", fg=\"#FFFFFF\", font=(\"Verdana\", \"8\"), anchor=W)\r\n self.ll3Identificacao.pack(fill=X)\r\n\r\n self.ll4Identificacao = Label(self.cccIdentificacao, text=self.dadosDir[\"vinculo\"][\"curso\"], bg=\"#606060\", fg=\"#FFFFFF\", font=(\"Verdana\", \"8\"), anchor=W)\r\n self.ll4Identificacao.pack(fill=X)\r\n\r\n self.ll5Identificacao = Label(self.cccIdentificacao, text=self.dadosDir[\"tipo_vinculo\"], bg=\"#606060\", fg=\"#FFFFFF\", font=(\"Verdana\", \"8\"), anchor=W)\r\n self.ll5Identificacao.pack(fill=X)\r\n\r\n self.cIStyle = Frame(self.cCons, bg=\"#707070\", height=4, width=350)\r\n self.cIStyle.pack()\r\n \r\n self.listaconexao = Listbox(self.cCons, bg=\"#505050\", activestyle=\"none\", justify=CENTER,font=(\"Verdana\", \"20\"), fg=\"white\", borderwidth=0, selectbackground=\"#606060\", highlightthickness=0, state=\"disabled\")\r\n self.listaconexao.pack(fill=BOTH,expand=1)\r\n\r\n self.add = Button(self.cCons, text=\"Adicionar Amigo\", borderwidth=0, bg=\"#707070\", fg=\"#FFFFFF\", font=(\"Verdana\", \"18\"))\r\n self.add.bind(\"\", self.adicionarAmigo)\r\n self.add.pack(fill=X)\r\n \r\n self.smen = Scrollbar(self.cRecv)\r\n self.caixademensagens = Text(self.cRecv, state=\"disabled\", cursor=\"arrow\", font=(\"Verdana\", \"15\"), bg=\"#606060\", fg=\"#FFFFFF\", relief=\"flat\")\r\n self.smen.pack(side=RIGHT, fill=Y)\r\n self.caixademensagens.pack(fill=BOTH, expand=1)\r\n self.smen.config(command=self.caixademensagens.yview)\r\n self.caixademensagens.config(yscrollcommand=self.smen.set)\r\n\r\n self.caixadeenvio = Text(self.cEnv, height=1, insertbackground=\"#FFFFFF\", bg=\"#707070\", relief=\"flat\", font=(\"Verdana\", \"15\"), fg=\"white\")\r\n self.caixadeenvio.bind(\"\", self.enviandoMensagem)\r\n self.caixadeenvio.bind(\"\", self.aumentar)\r\n self.caixadeenvio.pack(fill=BOTH)\r\n\r\n def enviandoMensagem(self, event):\r\n msg = {\"flag\":\"MSG\", \"content\":None}\r\n content = self.caixadeenvio.get(0.0,INSERT)\r\n msg[\"content\"] = {\"matricula\":self.dadosDir[\"vinculo\"][\"matricula\"], \"msg\":content, \"grupo\":None}\r\n nome = self.listaconexao.get(self.listaconexao.curselection()[0])\r\n for matricula in self.D.keys():\r\n if self.D[matricula][\"nome\"] == nome:\r\n msg[\"content\"][\"grupo\"] = self.D[matricula][\"grupo\"]\r\n msg = json.dumps(msg)\r\n msg = msg.encode(\"utf-8\")\r\n self.s.send(msg)\r\n self.caixadeenvio.delete(0.0,END)\r\n self.caixadeenvio.configure(height=1)\r\n\r\n def recebendo(self):\r\n while True:\r\n msg = self.s.recv(512)\r\n msg = msg.decode(\"utf-8\")\r\n msg = json.loads(msg)\r\n if msg[\"rflag\"] == \"MSG\":\r\n self.printMSG(msg[\"rcontent\"])\r\n elif msg[\"rflag\"] == \"CST\":\r\n self.respAmigo(msg[\"rcontent\"])\r\n elif msg[\"rflag\"] == \"SLA1\":\r\n self.recebeGrupo(msg[\"rcontent\"])\r\n elif msg[\"rflag\"] == \"SLA2\":\r\n self.recebeMensagens(msg[\"rcontent\"])\r\n\r\n def printMSG(self, msg):\r\n msg[\"mensagem\"] = msg[\"mensagem\"].replace(\"\\n\", \"\\n\\t\")\r\n self.caixademensagens.configure(state=\"normal\")\r\n if msg[\"nome\"] == self.dadosDir[\"vinculo\"][\"nome\"]:\r\n if msg[\"mensagem\"][0] == \"\\n\":\r\n self.caixademensagens.insert(END, \"\\nVocê:\\n\" + str(msg[\"mensagem\"][1:]))\r\n else:\r\n msg[\"mensagem\"] = \"\\t\" + msg[\"mensagem\"]\r\n self.caixademensagens.insert(END, \"\\nVocê:\\n\" + str(msg[\"mensagem\"]))\r\n else:\r\n if msg[\"mensagem\"][0] == \"\\n\":\r\n self.caixademensagens.insert(END, \"\\n{}:\\n\".format(msg[\"nome\"]) + str(msg[\"mensagem\"][1:]))\r\n else:\r\n msg[\"mensagem\"] = \"\\t\" + msg[\"mensagem\"]\r\n self.caixademensagens.insert(END, \"\\n{}:\\n\".format(msg[\"nome\"]) + str(msg[\"mensagem\"]))\r\n self.caixademensagens.configure(state=\"disabled\")\r\n self.caixademensagens.see(\"end\")\r\n\r\n def respAmigo(self, resp):\r\n if resp[\"nome\"] == \"\" and resp[\"matricula\"] == \"\":\r\n self.lladd.configure(text=\"Usuário Não Encontrado!\")\r\n else:\r\n self.D[resp[\"matricula\"]] = {\"nome\":resp[\"nome\"], \"grupo\": None}\r\n self.listaconexao.configure(state=\"normal\")\r\n self.listaconexao.bind(\"<>\", self.selecionarAmigo)\r\n self.listaconexao.insert(END, resp[\"nome\"])\r\n self.add.destroy()\r\n\r\n def recebeGrupo(self, resp):\r\n self.D[resp[\"matricula_destino\"]][\"grupo\"]= resp[\"grupo\"]\r\n\r\n def recebeMensagens(self, resp):\r\n self.caixademensagens.configure(state=\"normal\")\r\n self.caixademensagens.delete(0.0,END)\r\n for mensagem in resp:\r\n mensagem[1] = mensagem[1].replace(\"\\n\", \"\\n\\t\")\r\n if mensagem[0] == self.dadosDir[\"vinculo\"][\"nome\"]:\r\n if mensagem[1][0] == \"\\n\":\r\n self.caixademensagens.insert(END, \"\\nVocê:\\n\" + str(mensagem[1][1:]))\r\n else:\r\n mensagem[1] = \"\\t\" + mensagem[1]\r\n self.caixademensagens.insert(END, \"\\nVocê:\\n\" + str(mensagem[1]))\r\n else:\r\n if mensagem[1][0] == \"\\n\":\r\n self.caixademensagens.insert(END, \"\\n{}:\\n\".format(mensagem[0]) + str(mensagem[1][1:]))\r\n else:\r\n mensagem[1] = \"\\t\" + mensagem[1]\r\n self.caixademensagens.insert(END, \"\\n{}:\\n\".format(mensagem[0]) + str(mensagem[1]))\r\n self.caixademensagens.configure(state=\"disabled\")\r\n self.caixademensagens.see(\"end\")\r\n \r\n def cliente(self):\r\n reg = {\"flag\":\"REG\", \"content\":None}\r\n content = {\"nome\":self.dadosDir[\"vinculo\"][\"nome\"], \"matricula\":self.dadosDir[\"vinculo\"][\"matricula\"]}\r\n reg[\"content\"] = content\r\n reg = json.dumps(reg)\r\n reg = reg.encode(\"utf-8\")\r\n \r\n host = \"localhost\"\r\n port = 5000\r\n self.s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\r\n self.s.connect((host, port))\r\n self.s.send(reg)\r\n threading.Thread(target=self.recebendo).start()\r\n\r\n def aumentar(self,event):\r\n self.caixadeenvio.configure(height=2)\r\n\r\n def adicionarAmigo(self, event):\r\n self.add = Toplevel()\r\n #self.add.iconbitmap(\"icon.ico\")\r\n self.add.title(\"Chat - Adicionar Amigo\")\r\n self.add.configure(bg=\"#505050\", padx = 20, pady=20)\r\n self.add.resizable(0, 0)\r\n\r\n cladd = Frame(self.add, bg=\"#505050\")\r\n cladd.pack()\r\n\r\n ladd = Label(cladd, font=(\"Verdana\", \"18\"), text=\"Adicionar Amigo\", bg=\"#505050\", fg=\"#FFFFFF\")\r\n ladd.pack()\r\n\r\n self.lladd = Label(cladd, font=(\"Verdana\"), text=\"\", bg=\"#505050\", fg=\"#FFFFFF\")\r\n self.lladd.pack()\r\n\r\n cadd = Frame(self.add, bg=\"#505050\")\r\n cadd.pack()\r\n\r\n cladd = Frame(cadd, bg=\"#505050\")\r\n cladd.pack(side=LEFT)\r\n\r\n lmatriculaadd = Label(cladd, font=(\"Verdana\", \"15\"), text=\"Matrícula:\", bg=\"#505050\", fg=\"#FFFFFF\")\r\n lmatriculaadd.pack(pady=(6,0))\r\n\r\n ceadd = Frame(cadd, bg=\"#505050\")\r\n ceadd.pack()\r\n\r\n self.ematriculaadd = Entry(ceadd, width=20, bg=\"#707070\", borderwidth=0, font=(\"Verdana\", \"15\"), insertbackground=\"#FFFFFF\", fg=\"white\")\r\n self.ematriculaadd.pack(pady=(10,0))\r\n\r\n cbadd = Frame(self.add, bg=\"#505050\")\r\n cbadd.pack()\r\n\r\n badd = Button(cbadd, bg=\"#707070\", fg=\"#FFFFFF\", font=(\"Verdana\", \"18\"), text=\"Adicionar\", borderwidth=0)\r\n badd.bind(\"\", self.consultarAmigo)\r\n badd.pack(pady=(20,0))\r\n\r\n def consultarAmigo(self, event):\r\n if self.ematriculaadd.get() != self.dadosDir[\"vinculo\"][\"matricula\"]:\r\n cst = {\"flag\":\"CST\", \"content\":self.ematriculaadd.get()}\r\n cst = json.dumps(cst)\r\n cst = cst.encode(\"utf-8\")\r\n self.s.send(cst)\r\n else:\r\n self.lladd.configure(text=\"Você Não Pode Se Adicionar!\")\r\n\r\n def selecionarAmigo(self, event):\r\n self.cMens.pack(side=LEFT, fill=BOTH, expand=1)\r\n nome = event.widget.get(event.widget.curselection()[0])\r\n for matricula in self.D.keys():\r\n if self.D[matricula][\"nome\"] == nome:\r\n if self.D[matricula][\"grupo\"] == None:\r\n sla = {\"flag\":\"SLA1\", \"content\":{\"matricula_origem\":self.dadosDir[\"vinculo\"][\"matricula\"], \"matricula_destino\":matricula}}\r\n sla = json.dumps(sla)\r\n sla = sla.encode(\"utf-8\")\r\n self.s.send(sla)\r\n else:\r\n sla = {\"flag\":\"SLA2\", \"content\":{\"matricula_origem\":self.dadosDir[\"vinculo\"][\"matricula\"], \"grupo\":self.D[matricula][\"grupo\"]}}\r\n sla = json.dumps(sla)\r\n sla = sla.encode(\"utf-8\")\r\n self.s.send(sla)\r\n break\r\n \r\nLogin()\r\n","sub_path":"aplicativo.py","file_name":"aplicativo.py","file_ext":"py","file_size_in_byte":16362,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"161328183","text":"from tkinter import *\r\nfrom PIL import ImageTk, Image\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nimport pandas as pd\r\nimport DataIn as DI\r\nfrom InputValidation import InputValidation as InpVal\r\nimport DataOut as DO\r\nimport StockOut as SO\r\n\r\n\r\n\r\n\r\nroot = Tk()\r\nroot.title(\"Customer Recommendation System\")\r\nroot.iconbitmap(\"D:/Study Material/Project/Final Draft/custom.ico\")\r\n\r\nFrontPage = LabelFrame(root)\r\n\r\n\r\n\r\ndef onClick(frame):\r\n label = Label(frame,text = \"Some button was clicked\",padx = 30, background = \"Black\",pady = 30)\r\n label.grid(row = 30, column = 0, columnspan = 2)\r\n # simple print() will print the text on the console \r\n # instead of the root widget window \r\n\r\n\r\ndef RemoveFrame(frame, Frame):\r\n\tFrame.grid()\r\n\tframe.grid_forget()\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\ndef InputCustomerFrame():\r\n\tFrontPage.grid_forget()\r\n\r\n\r\n\tdef SubmitButton():\r\n\t\tif NameInp.get() == \" \":\r\n\t\t\t\tErrorLabel = Label(Inpframe, text = \"Enter required fields' values.\",fg = \"red\", pady = 20)\r\n\t\t\t\tErrorLabel.grid(row = 5, column = 0, columnspan = 2)\r\n\t\t\t\treturn\r\n\t\tdata = {}\r\n\t\tdata[\"Name\"] = NameInp.get()\r\n\t\tdata[\"Age\"] = AgeInp.get()\r\n\t\tdata[\"Annual Income (k$)\"] = SalaryInp.get()\r\n\t\tdata[\"Genre\"] = GenderInp.get()\r\n\t\tdata[\"Spending Score (1-100)\"] = ScoreInp.get()\r\n\t\tprint( data)\t\r\n\t\tdata = DI.TakeCustomerGui(data)\r\n\t\tcstr = str(data[\"Cluster\"])\r\n\t\tcsnm = str(data[\"Name\"])\r\n\t\tmsg = csnm+\" belonged to Cluster \" + cstr\r\n\t\tprint(msg)\r\n\t\tClusterLabel = Label(Inpframe, text = msg)\r\n\t\tClusterLabel.grid(row = 8)\t\t\r\n\t\tGenderInp.delete(0,END)\r\n\t\tNameInp.delete(0,END)\r\n\t\tSalaryInp.delete(0,END)\r\n\t\tAgeInp.delete(0,END)\r\n\t\tScoreInp.delete(0,END)\r\n\r\n\r\n\t# create the frame\r\n\tglobal Inpframe\r\n\tInpframe = LabelFrame(root, text = \"Input Customer Data here:\", padx = 120, pady = 50, bd = 3, relief = SUNKEN)\r\n\t\r\n\tExitButton = Button(Inpframe, text= \"Back\", padx= 20,pady = 10, fg= \"Red\", command =lambda: RemoveFrame(Inpframe, FrontPage))\r\n\tSubmitButton = Button(Inpframe, text= \"Submit\", padx= 20,pady = 10, fg= \"Red\", command =SubmitButton)\r\n\r\n\r\n\tNameLabel = Label(Inpframe, text = \"Name:\", pady = 20)\r\n\tNameInp = Entry(Inpframe, width = 40, borderwidth= 3)\r\n\tNameInp.insert(10, \" \")\r\n\r\n\tAgeLabel = Label(Inpframe, text = \"Age:\", pady = 20)\r\n\tAgeInp = Entry(Inpframe, width = 40, borderwidth= 3)\r\n\tAgeInp.insert(10, \" \")\r\n\r\n\r\n\tSalaryLabel = Label(Inpframe, text = \"Salary:\", pady = 20)\r\n\tSalaryInp = Entry(Inpframe, width = 40, borderwidth= 3)\r\n\tSalaryInp.insert(10, \" \")\r\n\r\n\r\n\tGenderLabel = Label(Inpframe, text = \"Gender:\", pady = 20)\r\n\tGenderInp = Entry(Inpframe, width = 40, borderwidth= 3)\r\n\tGenderInp.insert(10, \" \")\r\n\r\n\r\n\r\n\tScoreLabel = Label(Inpframe, text = \"Spending Socre(1-100):\", pady = 20)\r\n\tScoreInp = Entry(Inpframe, width = 40, borderwidth= 3)\r\n\tScoreInp.insert(10, \" \")\r\n\r\n\r\n\t\r\n\t# show the InpFrame widgets\r\n\tNameLabel.grid(row = 0, column = 0)\r\n\tNameInp.grid(row = 0, column = 1)\r\n\tAgeLabel.grid(row = 1, column = 0)\r\n\tAgeInp.grid(row = 1, column = 1)\r\n\tSalaryLabel.grid(row = 2, column = 0)\r\n\tSalaryInp.grid(row = 2, column = 1)\r\n\tGenderLabel.grid(row = 3, column = 0)\r\n\tGenderInp.grid(row = 3, column = 1)\r\n\tScoreLabel.grid(row = 4, column = 0)\r\n\tScoreInp.grid(row = 4, column = 1)\r\n\t\r\n\r\n\r\n\r\n\t#show the frame\r\n\tInpframe.grid(row = 10, column = 0, columnspan = 2, pady = 40)\r\n\tExitButton.grid(row = 10, column = 1)\r\n\tSubmitButton.grid(row = 10, column = 0,padx = 30)\r\n\r\ni = 0\r\n\r\ndef InputBasketFrame():\r\n\tFrontPage.grid_forget()\r\n\tbasket = []\r\n\r\n\tdef AddButtonFun():\r\n\t\tglobal i\r\n\t\tx = ItemInp.get().lower()\r\n\t\tbasket.append(x)\r\n\t\tprint(basket)\t\r\n\t\tCurrentItemLabel = Label(Inpframe, text = ItemInp.get(), pady = 20)\r\n\t\tCurrentItemLabel.grid(row = 11+i, column = 0)\r\n\t\ti+=1\r\n\t\tItemInp.delete(0,END)\r\n\r\n\r\n\tdef SubmitButtonFun():\r\n\t\tAddButtonFun()\r\n\t\tDI.SaveBasketData(basket)\r\n\t\tglobal i\r\n\t\ti= 0\r\n\t\tRemoveFrame(Inpframe, FrontPage)\r\n\t\tInputBasketFrame()\r\n\r\n\r\n\r\n\r\n\t# create the frame\r\n\tglobal Inpframe\r\n\tInpframe = LabelFrame(root, text = \"Input Basket Items here:\", padx = 120, pady = 50, bd = 3, relief = SUNKEN)\r\n\t\r\n\tExitButton = Button(Inpframe, text= \"Finish\", padx= 20,pady = 10, fg= \"Red\", command =lambda: RemoveFrame(Inpframe, FrontPage))\r\n\tAddButton = Button(Inpframe, text= \"Add Item\", padx= 20,pady = 10, fg= \"Red\", command =AddButtonFun)\r\n\tSubmitButton = Button(Inpframe, text= \"Submit Basket\", padx= 20,pady = 10, fg= \"Red\", command =SubmitButtonFun)\r\n\r\n\r\n\tItemLabel = Label(Inpframe, text = \"Enter Item:\", pady = 20)\r\n\tItemInp = Entry(Inpframe, width = 40, borderwidth= 3)\r\n\tItemInp.insert(10, \" \")\r\n\r\n\r\n\r\n\r\n\t# show the InpFrame widgets\r\n\tItemLabel.grid(row = 0, column = 0)\r\n\tItemInp.grid(row = 0, column = 1)\r\n\t\r\n\r\n\r\n\t#show the frame\r\n\tInpframe.grid(row = 10, column = 0, columnspan = 2, pady = 40)\r\n\tExitButton.grid(row = 10, column = 2)\r\n\tAddButton.grid(row = 10, column = 0,padx = 30)\r\n\tSubmitButton.grid(row = 10, column = 1,padx = 30)\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\ndef ShowCustomerFrameDisplay():\r\n\tFrontPage.grid_forget()\r\n\tShowCustomerFrame = LabelFrame(root)\r\n\tShowLabel = Label(ShowCustomerFrame, text= \"Cutomer Data\",font = (\"Courier 14 bold\"), padx= 20,\r\n\t pady = 20, fg= \"Red\")\r\n\r\n\tInputLabel = Label(ShowCustomerFrame, text= \"How do you want the Data to be Filtered? :\", padx= 20,\r\n\t pady = 20, fg= \"Blue\")\r\n\r\n\r\n\t#Functions\r\n\tdef ClusterGraph():\r\n\t\tdataset = pd.read_csv('New Customer.csv')\r\n\t\tX = dataset.iloc[:, [2,3,4,5,6]].values\r\n\r\n\t\tfrom sklearn.cluster import KMeans\r\n\t\twcss = []\r\n\t\tfor i in range(1, 11):\r\n\t\t kmeans = KMeans(n_clusters = i, init = 'k-means++', random_state = 42)\r\n\t\t kmeans.fit(X)\r\n\t\t wcss.append(kmeans.inertia_)\r\n\r\n\t\tkmeans = KMeans(n_clusters = 5, init = 'k-means++', random_state = 42)\r\n\t\ty_kmeans = kmeans.fit_predict(X)\r\n\t\tprint(y_kmeans)\r\n\r\n\t\tplt.scatter(X[y_kmeans == 0, 3], X[y_kmeans == 0, 4], s = 100, c = 'red', label = 'Cluster 1')\r\n\t\tplt.scatter(X[y_kmeans == 1, 3], X[y_kmeans == 1, 4], s = 100, c = 'blue', label = 'Cluster 2')\r\n\t\tplt.scatter(X[y_kmeans == 2, 3], X[y_kmeans == 2, 4], s = 100, c = 'green', label = 'Cluster 3')\r\n\t\tplt.scatter(X[y_kmeans == 3, 3], X[y_kmeans == 3, 4], s = 100, c = 'cyan', label = 'Cluster 4')\r\n\t\tplt.scatter(X[y_kmeans == 4, 3], X[y_kmeans == 4, 4], s = 100, c = 'magenta', label = 'Cluster 5')\r\n\t\tplt.scatter(kmeans.cluster_centers_[:, 0], kmeans.cluster_centers_[:, 1], s = 300, c = 'yellow', label = 'Centroids')\r\n\t\tplt.title('Clusters of customers')\r\n\t\tplt.xlabel('Annual Income (k$)')\r\n\t\tplt.ylabel('Spending Score (1-100)')\r\n\t\tplt.legend()\r\n\t\tplt.show()\r\n\r\n\r\n\r\n\r\n\tdef ShowAllData():\r\n\t\tShowCustomerFrame.grid_forget()\r\n\t\tdataset = pd.read_csv(r'D:\\Study Material\\Project\\Final Draft\\New Customer.csv')\r\n\t\tx = max(dataset[\"CustomerID\"])\r\n\r\n\t\tOutputFrame = LabelFrame(root, text = \"Output Window\")\r\n\t\tprint(list(dataset.columns))\r\n\t\tItemFrame = LabelFrame(OutputFrame)\r\n\r\n\t\t#scroll = Scrollbar(OutputFrame, orient = 'vertical')\r\n\t\t#scroll.config(command = OutputFrame.xview)\r\n\t\tBackButton = Button(OutputFrame, text = \"Back\", command = lambda: RemoveFrame(OutputFrame, ShowCustomerFrame))\r\n\t\tBackButton.grid(row = 0, column = 0, padx = 7, pady = 7)\r\n\r\n\r\n\t\tHeadLable = Label(OutputFrame, text = \"Showing ALL CUSTOMERS Data\", relief = SUNKEN, border = 2, padx = 0, pady = 3, font = \"BOLD 28\")\r\n\t\tHeadLable.grid(row = 0, column = 1, padx = 7, pady = 7, columnspan = 20)\r\n\r\n\r\n\r\n\t\tcoln = 0\r\n\t\tfor i in list(dataset.columns):\r\n\t\t\tColLable = Label(ItemFrame, text = str(i), relief = SUNKEN, border = 0, padx = 0, pady = 3, font = \"BOLD\")\r\n\t\t\tColLable.grid(row = 1, column = coln, padx = 7, pady = 7)\r\n\t\t\tcoln = coln + 1\r\n\t\tOutputFrame.grid()\r\n\r\n\t\trown = 1\r\n\t\tcount = 0\r\n\t\tfor row in range(x):\t\r\n\t\t\tif count < 30:\r\n\t\t\t\tcount+=1\r\n\t\t\t\tfor col in range(len(list(dataset.columns))):\r\n\t\t\t\t\telement = dataset.iloc[row,col]\r\n\t\t\t\t\tEleLabel = Label(ItemFrame, text = str(element), relief = SUNKEN, border = 1, padx = 95)\r\n\t\t\t\t\tEleLabel.grid(row = row+2, column = col)\r\n\t\tItemFrame.grid(columnspan = 6)\r\n\t\tBackButton = Button(OutputFrame, text = \"Back\", command = lambda: RemoveFrame(OutputFrame, ShowCustomerFrame))\r\n\t\tBackButton.grid(row = 3, column = 3, padx = 7, pady = 7)\r\n\r\n\tdef GetCluster(): \r\n\t\tGetClusterFrame = LabelFrame(ShowCustomerFrame, text = \"Choose the Cluster\")\r\n\r\n\r\n\r\n\t\tdef ShowClusterCustomers(cls = 1):\r\n\t\t\tShowCustomerFrame.grid_forget()\r\n\t\t\tdataset = pd.read_csv(r'D:\\Study Material\\Project\\Final Draft\\New Customer.csv')\r\n\t\t\tx = max(dataset[\"CustomerID\"])\r\n\t\t\tOutputFrame = LabelFrame(root, text = \"Output Window\")\r\n\r\n\r\n\r\n\t\t\tBackButton = Button(OutputFrame, text = \"Back\", command = lambda: RemoveFrame(OutputFrame, ShowCustomerFrame))\r\n\t\t\tBackButton.grid(row = 0, column = 0, padx = 7, pady = 7)\r\n\r\n\t\t\tHeadLable = Label(OutputFrame, text = \"Showing Cluster \"+str(cls)+\" CUSTOMERS Data\", relief = SUNKEN, border = 2, padx = 0, pady = 3, font = \"BOLD 28\")\r\n\t\t\tHeadLable.grid(row = 0, column = 1, padx = 7, pady = 7, columnspan = 20)\r\n\r\n\r\n\r\n\r\n\t\t\tcoln = 0\r\n\t\t\tfor i in list(dataset.columns):\r\n\t\t\t\tColLable = Label(OutputFrame, text = str(i), relief = SUNKEN, border = 0, padx = 0, pady = 3, font = \"BOLD\")\r\n\t\t\t\tColLable.grid(row = 1, column = coln, padx = 7, pady = 7)\r\n\t\t\t\tcoln = coln + 1\r\n\t\t\tOutputFrame.grid(row = 0, pady = 30)\r\n\t\t\trown = 1\r\n\t\t\tfor col in range(len(list(dataset.columns))):\r\n\t\t\t\tfor row in range(x):\r\n\t\t\t\t\tif dataset.iloc[row,2] == cls:\r\n\t\t\t\t\t\telement = dataset.iloc[row,col]\r\n\t\t\t\t\t\tEleLabel = Label(OutputFrame, text = str(element), relief = SUNKEN, border = 1, padx = 95)\r\n\t\t\t\t\t\tEleLabel.grid(row = row+2, column = col)\r\n\r\n\r\n\r\n\r\n\r\n\r\n\t\tcls0 = Button(GetClusterFrame, text = \"Cluster 0\", command = lambda: ShowClusterCustomers(0))\r\n\t\tcls1 = Button(GetClusterFrame, text = \"Cluster 1\", command = lambda: ShowClusterCustomers(1))\r\n\t\tcls2 = Button(GetClusterFrame, text = \"Cluster 2\", command = lambda: ShowClusterCustomers(2))\r\n\t\tcls3 = Button(GetClusterFrame, text = \"Cluster 3\", command = lambda: ShowClusterCustomers(3))\r\n\t\tcls4 = Button(GetClusterFrame, text = \"Cluster 4\", command = lambda: ShowClusterCustomers(4))\r\n\r\n\t\tHead = Label(GetClusterFrame, text = \"Choose which cluster data is to be shown: \")\r\n\r\n\t\tHead.grid(row = 0, column = 0, columnspan = 2)\r\n\t\tcls0.grid(row = 1, column = 0)\r\n\t\tcls1.grid(row = 2, column = 0)\r\n\t\tcls2.grid(row = 3, column = 0)\r\n\t\tcls3.grid(row = 1, column = 1)\r\n\t\tcls4.grid(row = 2, column = 1)\r\n\t\tGetClusterFrame.grid(row= 5)\r\n\r\n\r\n\r\n\tdef GetOrder():\r\n\t\tGetOrderFrame = LabelFrame(ShowCustomerFrame, text = \"Choose the type of Ordering: \")\r\n\r\n\r\n\r\n\t\tdef ShowOrderedCustomers(cls = 1):\r\n\t\t\tShowCustomerFrame.grid_forget()\r\n\t\t\tdataset = pd.read_csv(r'D:\\Study Material\\Project\\Final Draft\\New Customer.csv')\r\n\t\t\tOutputFrame = LabelFrame(root, text = \"Output Window\")\r\n\r\n\r\n\r\n\t\t\tBackButton = Button(OutputFrame, text = \"Back\", command = lambda: RemoveFrame(OutputFrame, ShowCustomerFrame))\r\n\t\t\tBackButton.grid(row = 0, column = 0, padx = 7, pady = 7)\r\n\r\n\t\t\tHeadLable = Label(OutputFrame, text = \"Showing Customer in Order of Spending Score\", relief = SUNKEN, border = 2, padx = 0, pady = 3, font = \"BOLD 28\")\r\n\t\t\tHeadLable.grid(row = 0, column = 1, padx = 7, pady = 7, columnspan = 20)\r\n\r\n\r\n\r\n\r\n\t\t\tcoln = 0\r\n\t\t\tfor i in list(dataset.columns):\r\n\t\t\t\tColLable = Label(OutputFrame, text = str(i), relief = SUNKEN, border = 0, padx = 0, pady = 3, font = \"BOLD\")\r\n\t\t\t\tColLable.grid(row = 1, column = coln, padx = 7, pady = 7)\r\n\t\t\t\tcoln = coln + 1\r\n\t\t\tOutputFrame.grid(row = 0, pady = 30)\r\n\t\t\trown = 1\r\n\r\n\t\t\tdf = dataset.sort_values(\"Spending Score (1-100)\", ascending = cls)\r\n\t\t\tx = max(dataset[\"CustomerID\"])\r\n\r\n\t\t\tfor col in range(len(list(df.columns))):\r\n\t\t\t\tfor row in range(x):\r\n\t\t\t\t\telement = df.iloc[row,col]\r\n\t\t\t\t\tEleLabel = Label(OutputFrame, text = str(element), relief = SUNKEN, border = 1, padx = 95)\r\n\t\t\t\t\tEleLabel.grid(row = row+2, column = col)\r\n\r\n\r\n\r\n\r\n\r\n\t\tord0 = Button(GetOrderFrame, text = \"Ascending Order\", command = lambda: ShowOrderedCustomers(1))\r\n\t\tord1 = Button(GetOrderFrame, text = \"Descending Order\", command = lambda: ShowOrderedCustomers(0))\r\n\r\n\t\tHead = Label(GetOrderFrame, text = \"Choose in what Oreder Data is to be shown: \")\r\n\r\n\t\tHead.grid(row = 0, column = 0, columnspan = 2)\r\n\t\tord0.grid(row = 1, column = 0)\r\n\t\tord1.grid(row = 2, column = 0)\r\n\t\tGetOrderFrame.grid(row= 8)\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\t# Buttons\r\n\tAllCustomerButton = Button(ShowCustomerFrame, text = \"All Customer Data\", padx = 0, pady = 20, command = ShowAllData)\r\n\tClusterCustomerButton = Button(ShowCustomerFrame, text = \"Cluster Customer Data\", padx = 0, pady = 20, command = GetCluster)\r\n\tSortedCustomerButton = Button(ShowCustomerFrame, text = \"Sorted Customer Data\", padx = 0, pady = 20, command =GetOrder)\r\n\tBackButton = Button(ShowCustomerFrame, text = \"Go Back\", padx = 0, pady = 20, command =lambda: RemoveFrame(ShowCustomerFrame, FrontPage))\r\n\tClusterShowButton = Button(ShowCustomerFrame, text = \"Show Graph of All Customers \\nAccording to Clusters\", command = ClusterGraph, pady = 30)\r\n\r\n\r\n\r\n\r\n\r\n\t# put Customer Display page on the screen\r\n\tShowLabel.grid(row = 0, column = 0)\r\n\tInputLabel.grid(row = 1, column = 0)\r\n\tAllCustomerButton.grid(row = 2, column = 0, pady = 20)\r\n\tClusterCustomerButton.grid(row = 3, column = 0, pady = 20)\r\n\tSortedCustomerButton.grid(row = 7, column = 0, pady = 20)\r\n\t\r\n\tClusterShowButton.grid(row = 9, column = 0)\r\n\tBackButton.grid(row = 10, column = 0, pady = 20)\r\n\tShowCustomerFrame.grid()\r\n\t\r\n\r\n\r\n\r\ndef ShowItemFrameDisplay():\r\n\tFrontPage.grid_forget()\r\n\tShowItemFrame = LabelFrame(root)\r\n\tShowLabel = Label(ShowItemFrame, text= \"Stock Items Data\",font = (\"Courier 14 bold\"), padx= 20,\r\n\t pady = 20, fg= \"Red\")\r\n\r\n\tInputLabel = Label(ShowItemFrame, text= \"How do you want the Data to be Filtered? :\", padx= 20,\r\n\t pady = 20, fg= \"Blue\")\r\n\r\n\r\n\r\n\r\n\r\n\r\n\tdef GetOrder():\r\n\t\tGetOrderFrame = LabelFrame(ShowItemFrame, text = \"Choose the type of Ordering: \")\r\n\r\n\t\tdef SortedItems(ord):\r\n\t\t\tdataset = pd.read_csv(\"D:\\Study Material\\Project\\Final Draft\\Items Table.csv\")\r\n\t\t\tdataset = dataset.sort_values(\"Rating\", ascending = bool(ord))\r\n\r\n\t\t\tprint(\"\\nColumns:\\n\",list(dataset.columns),\"\\n\")\r\n\r\n\t\t\tShowItemFrame.grid_forget()\r\n\t\t\tx = dataset[\"Item\"].count()\r\n\r\n\t\t\tOutputFrame = LabelFrame(root, text = \"Output Window\")\r\n\t\t\tprint(list(dataset.columns))\r\n\r\n\t\t\t#scroll = Scrollbar(OutputFrame, orient = 'vertical')\r\n\t\t\t#scroll.config(command = OutputFrame.xview)\r\n\t\t\tBackButton = Button(OutputFrame, text = \"Back\", command = lambda: RemoveFrame(OutputFrame, ShowItemFrame))\r\n\t\t\tBackButton.grid(row = 0, column = 0)\r\n\r\n\r\n\t\t\tHeadLable = Label(OutputFrame, text = \"Showing ALL Stock Items by Rating\", relief = SUNKEN, border = 2, padx = 0, pady = 3, font = \"BOLD 28\")\r\n\t\t\tHeadLable.grid(row = 0, column = 1, padx = 7, pady = 7, columnspan = 2)\r\n\r\n\r\n\r\n\t\t\tcoln = 0\r\n\t\t\tfor i in list(dataset.columns):\r\n\t\t\t\tColLable = Label(OutputFrame, text = str(i), relief = SUNKEN, border = 0, padx = 0, pady = 3, font = \"BOLD\")\r\n\t\t\t\tColLable.grid(row = 1, column = coln, padx = 7, pady = 7)\r\n\t\t\t\tcoln = coln + 1\r\n\t\t\tOutputFrame.grid()\r\n\t\t\trown = 1\r\n\t\t\tfor col in range(len(list(dataset.columns))):\r\n\t\t\t\tfor row in range(x):\r\n\t\t\t\t\telement = dataset.iloc[row,col]\r\n\t\t\t\t\tEleLabel = Label(OutputFrame, text = str(element), relief = SUNKEN, border = 1, padx =140)\r\n\t\t\t\t\tEleLabel.grid(row = row+2, column = col)\r\n\r\n\r\n\r\n\r\n\r\n\r\n\t\tord0 = Button(GetOrderFrame, text = \"Ascending Order\", command = lambda: SortedItems(1))\r\n\t\tord1 = Button(GetOrderFrame, text = \"Descending Order\", command = lambda: SortedItems(0))\r\n\r\n\t\tHead = Label(GetOrderFrame, text = \"Choose in what Oreder Data is to be shown: \")\r\n\r\n\t\tHead.grid(row = 0, column = 0, columnspan = 2)\r\n\t\tord0.grid(row = 1, column = 0)\r\n\t\tord1.grid(row = 2, column = 0)\r\n\t\tGetOrderFrame.grid(row= 5)\r\n\r\n\r\n\r\n\r\n\r\n\tdef AllItems():\r\n\t\tdataset = pd.read_csv(\"D:\\Study Material\\Project\\Final Draft\\Items Table.csv\")\r\n\t\tprint(\"\\nColumns:\\n\",list(dataset.columns),\"\\n\")\r\n\r\n\t\tShowItemFrame.grid_forget()\r\n\t\tx = dataset[\"Item\"].count()\r\n\r\n\t\tOutputFrame = LabelFrame(root, text = \"Output Window\")\r\n\t\tprint(list(dataset.columns))\r\n\r\n\t\t#scroll = Scrollbar(OutputFrame, orient = 'vertical')\r\n\t\t#scroll.config(command = OutputFrame.xview)\r\n\t\tBackButton = Button(OutputFrame, text = \"Back\", command = lambda: RemoveFrame(OutputFrame, ShowItemFrame))\r\n\t\tBackButton.grid(row = 0, column = 0)\r\n\r\n\r\n\t\tHeadLable = Label(OutputFrame, text = \"Showing ALL Stock Items\", relief = SUNKEN, border = 2, padx = 0, pady = 3, font = \"BOLD 28\")\r\n\t\tHeadLable.grid(row = 0, column = 1, padx = 7, pady = 7, columnspan = 2)\r\n\r\n\r\n\r\n\t\tcoln = 0\r\n\t\tfor i in list(dataset.columns):\r\n\t\t\tColLable = Label(OutputFrame, text = str(i), relief = SUNKEN, border = 0, padx = 0, pady = 3, font = \"BOLD\")\r\n\t\t\tColLable.grid(row = 1, column = coln, padx = 7, pady = 7)\r\n\t\t\tcoln = coln + 1\r\n\t\tOutputFrame.grid()\r\n\t\trown = 1\r\n\t\tfor col in range(len(list(dataset.columns))):\r\n\t\t\tfor row in range(x):\r\n\t\t\t\telement = dataset.iloc[row,col]\r\n\t\t\t\tEleLabel = Label(OutputFrame, text = str(element), relief = SUNKEN, border = 1, padx =140)\r\n\t\t\t\tEleLabel.grid(row = row+2, column = col)\r\n\r\n\r\n\r\n\r\n\r\n\r\n\tdef Together():\r\n\t\tdataset = SO.Together()\r\n\t\tShowItemFrame.grid_forget()\r\n\t\tprint(dataset.columns)\r\n\t\tx = dataset[\"Support\"].count()\r\n\r\n\t\tOutputFrame = LabelFrame(root, text = \"Output Window\")\r\n\t\tprint(list(dataset.columns))\r\n\r\n\t\t#scroll = Scrollbar(OutputFrame, orient = 'vertical')\r\n\t\t#scroll.config(command = OutputFrame.xview)\r\n\t\tBackButton = Button(OutputFrame, text = \"Back\", command = lambda: RemoveFrame(OutputFrame, ShowItemFrame))\r\n\t\tBackButton.grid(row = 0, column = 0)\r\n\r\n\r\n\t\tHeadLable = Label(OutputFrame, text = \"Showing ALL Stock Items\", relief = SUNKEN, border = 2, padx = 0, pady = 3, font = \"BOLD 28\")\r\n\t\tHeadLable.grid(row = 0, column = 1, padx = 7, pady = 7, columnspan = 2)\r\n\r\n\r\n\r\n\t\tcoln = 0\r\n\t\tfor i in list(dataset.columns):\r\n\t\t\tColLable = Label(OutputFrame, text = str(i), relief = SUNKEN, border = 0, padx = 0, pady = 3, font = \"BOLD\")\r\n\t\t\tColLable.grid(row = 1, column = coln, padx = 7, pady = 7)\r\n\t\t\tcoln = coln + 1\r\n\t\tOutputFrame.grid()\r\n\t\trown = 1\r\n\t\tfor col in range(len(list(dataset.columns))):\r\n\t\t\tfor row in range(x):\r\n\t\t\t\telement = dataset.iloc[row,col]\r\n\t\t\t\tEleLabel = Label(OutputFrame, text = str(element), relief = SUNKEN, border = 1, padx =70)\r\n\t\t\t\tEleLabel.grid(row = row+2, column = col)\r\n\r\n\r\n\r\n\r\n\r\n\r\n\t# Buttons\r\n\tAllItemButton = Button(ShowItemFrame, text = \"All Items Data\", padx = 0, pady = 20, command = AllItems)\r\n\tTogetherButton = Button(ShowItemFrame, text = \"Items Bought Together\", padx = 0, pady = 20, command =Together)\r\n\tSortedItemButton = Button(ShowItemFrame, text = \"Rating-wise Sorted Data\", padx = 0, pady = 20, command =GetOrder)\r\n\tBackButton = Button(ShowItemFrame, text = \"Go Back\", padx = 0, pady = 20, command =lambda: RemoveFrame(ShowItemFrame, FrontPage))\r\n\r\n\r\n\t# put First page on the screen\r\n\tShowLabel.grid(row = 0, column = 0)\r\n\tInputLabel.grid(row = 1, column = 0)\r\n\tAllItemButton.grid(row = 2, column = 0, pady = 20)\r\n\tTogetherButton.grid(row = 3, column = 0, pady = 20)\r\n\tSortedItemButton.grid(row = 4, column = 0, pady = 20)\r\n\tBackButton.grid(row = 6, column = 0)\r\n\tShowItemFrame.grid()\r\n\r\n\r\n\r\ndef EditCustomerFrame():\r\n\tFrontPage.grid_forget()\r\n\tEditCustomerFrame = LabelFrame(root)\r\n\tEditLabel = Label(EditCustomerFrame, text= \"Enter Cutomer Data to search by:\",font = (\"Courier 14 bold\"), padx= 20,\r\n\t pady = 20, fg= \"Red\")\r\n\r\n\r\n\r\n\tNameLabel = Label(EditCustomerFrame, text = \"Enter Name:\", pady = 20)\r\n\tNameInp = Entry(EditCustomerFrame, width = 40, borderwidth= 3)\r\n\tNameInp.insert(10, \" \")\r\n\tNameSearch = Button(EditCustomerFrame, text = \"Search\", fg = \"Red\")\r\n\r\n\tOrLabel = Label(EditCustomerFrame, text = \"OR\")\r\n\r\n\tIDLabel = Label(EditCustomerFrame, text = \"Enter Name:\", pady = 20)\r\n\tIDInp = Entry(EditCustomerFrame, width = 40, borderwidth= 3)\r\n\tIDInp.insert(10, \" \")\r\n\tIDSearch = Button(EditCustomerFrame, text = \"Search\", fg = \"Red\")\r\n\r\n\tEditLabel.grid(row = 0, column = 0, columnspan = 2)\r\n\tNameLabel.grid(row = 1,column = 0)\r\n\tNameInp.grid(row = 1,column = 1)\r\n\tNameSearch.grid(row = 1,column = 2)\r\n\tOrLabel.grid(row = 2,column = 0, columnspan = 2)\r\n\tIDLabel.grid(row = 3,column = 0)\r\n\tIDInp.grid(row = 3,column = 1)\r\n\tIDSearch.grid(row = 3,column = 2)\r\n\tEditCustomerFrame.pack()\r\n\r\n\r\n\r\n\r\ndef EditBasketFrame():\r\n\tFrontPage.grid_forget()\r\n\tEditStockFrame = LabelFrame(root)\r\n\tEditLabel = Label(EditStockFrame, text= \"Enter Item Data to search by:\",font = (\"Courier 14 bold\"), padx= 20,\r\n\t pady = 20, fg= \"Red\")\r\n\r\n\r\n\r\n\tNameLabel = Label(EditStockFrame, text = \"Enter Name:\", pady = 20)\r\n\tNameInp = Entry(EditStockFrame, width = 40, borderwidth= 3)\r\n\tNameInp.insert(10, \" \")\r\n\tNameSearch = Button(EditStockFrame, text = \"Search\", fg = \"Red\")\r\n\r\n\tOrLabel = Label(EditStockFrame, text = \"OR\")\r\n\r\n\tIDLabel = Label(EditStockFrame, text = \"Enter ID:\", pady = 20)\r\n\tIDInp = Entry(EditStockFrame, width = 40, borderwidth= 3)\r\n\tIDInp.insert(10, \" \")\r\n\tIDSearch = Button(EditStockFrame, text = \"Search\", fg = \"Red\")\r\n\r\n\tEditLabel.grid(row = 0, column = 0, columnspan = 2)\r\n\tNameLabel.grid(row = 1,column = 0)\r\n\tNameInp.grid(row = 1,column = 1)\r\n\tNameSearch.grid(row = 1,column = 2)\r\n\tOrLabel.grid(row = 2,column = 0, columnspan = 2)\r\n\tIDLabel.grid(row = 3,column = 0)\r\n\tIDInp.grid(row = 3,column = 1)\r\n\tIDSearch.grid(row = 3,column = 2)\r\n\tEditStockFrame.pack()\r\n\r\n\r\n\r\n\r\n\r\n\r\ndef FrontPageFrame():\r\n\t# first page widgets\r\n\t# the operation buttons\r\n\tBasketLabel = Label(FrontPage, text= \"Customers\",font = (\"Courier 14 bold\"), padx= 20,\r\n\t pady = 20, fg= \"Red\")\r\n\r\n\tCustomerLabel = Label(FrontPage, text= \"Stock\",font = (\"Courier 14 bold\"), padx= 20,\r\n\t pady = 20, fg= \"Blue\")\r\n\r\n\t# greeting image\r\n\tGreetImg= ImageTk.PhotoImage(Image.open(\"file-delivery.png\"))\r\n\tGreetImgLabel = Label(FrontPage, image = GreetImg)\r\n\r\n\t# greeting text\r\n\tGreeting = Label(FrontPage,text = \"Welcome to Customer Recommendation System\", font = \"Courier 18 bold\",anchor = \"s\",padx = 10, pady =40)\r\n\r\n\t# Buttons\r\n\tCustomerDataSubmitButton = Button(FrontPage, text = \"Enter Customer Data\", padx = 0, pady = 20, command = InputCustomerFrame)\r\n\tBasketDataSubmitButton = Button(FrontPage, text = \"Enter Basket Data\", padx = 10, pady = 20,command = InputBasketFrame)\r\n\r\n\r\n\tCustomerDataShowButton = Button(FrontPage, text = \"Show Customer Data\", padx = 0, pady = 20, command = ShowCustomerFrameDisplay)\r\n\tBasketDataShowButton = Button(FrontPage, text = \"Show Basket Data\", padx = 10, pady = 20, command = ShowItemFrameDisplay)\r\n\r\n\r\n\tCustomerEditButton = Button(FrontPage, text = \"Edit Customer Data\", padx = 0, pady = 20, command = EditCustomerFrame)\r\n\tBasketEditButton = Button(FrontPage, text = \"Edit Basket Data\", padx = 10, pady = 20, command = EditBasketFrame)\r\n\r\n\r\n\r\n\t# put First page on the screen\r\n\tCustomerLabel.grid(row = 2, column = 1)\r\n\tBasketLabel.grid(row = 2, column = 0)\r\n\tGreeting.grid(row = 1, column = 0,columnspan = 2)\r\n\tGreetImgLabel.grid(row = 0 , column = 0, columnspan = 2)\r\n\tCustomerDataSubmitButton.grid(row = 3, column = 0, pady = 20)\r\n\tBasketDataSubmitButton.grid(row = 3, column = 1, pady = 20)\r\n\tCustomerDataShowButton.grid(row = 2, column = 0, pady = 20)\r\n\tBasketDataShowButton.grid(row = 2, column = 1, pady = 20)\r\n\tCustomerEditButton.grid(row = 4, column = 0, pady = 20)\r\n\tBasketEditButton.grid(row = 4, column = 1, pady = 20)\r\n\r\n\r\nFrontPageFrame()\r\n\r\n\r\nFrontPage.grid()\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n# something is wrong with the image int the front page frame so it is not showing\r\nGreetImg= ImageTk.PhotoImage(Image.open(\"file-delivery.png\"))\r\nGreetImgLabel = Label(FrontPage, image = GreetImg)\r\nGreetImgLabel.grid(row = 0,column = 0, columnspan = 2)\r\n\r\n\r\n\r\nroot.mainloop()\r\n","sub_path":"FinalProject.py","file_name":"FinalProject.py","file_ext":"py","file_size_in_byte":23633,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"146163834","text":"# 实现 int sqrt(int x) 函数\r\n#计算并返回 x 的平方根,其中 x 是非负整数。由于返回类型是整数,结果只保留整数的部分,小数部分将被舍去\r\n\r\ndef mySqrt(x):\r\n if x == 0:\r\n return 0\r\n begin = 1\r\n end = x\r\n while begin <= end:\r\n mid = begin + (end-begin)//2\r\n s = mid**2\r\n if s <= x <(mid+1)**2:\r\n return mid\r\n elif s > x:\r\n end = mid\r\n else:\r\n begin = mid + 1\r\n\r\nif __name__ == \"__main__\":\r\n x = 1\r\n print(mySqrt(x))\r\n","sub_path":"每日一题/69.x的平方根.py","file_name":"69.x的平方根.py","file_ext":"py","file_size_in_byte":553,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"438394612","text":"# -*- coding: utf-8 -*-\n\"\"\"\nTencentBlueKing is pleased to support the open source community by making 蓝鲸智云-权限中心(BlueKing-IAM) available.\nCopyright (C) 2017-2021 THL A29 Limited, a Tencent company. All rights reserved.\nLicensed under the MIT License (the \"License\"); you may not use this file except in compliance with the License.\nYou may obtain a copy of the License at http://opensource.org/licenses/MIT\nUnless required by applicable law or agreed to in writing, software distributed under the License is distributed on\nan \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the\nspecific language governing permissions and limitations under the License.\n\"\"\"\nfrom typing import List, Optional\n\nfrom django.db import transaction\n\nfrom backend.apps.policy.models import Policy as PolicyModel\nfrom backend.component import iam\nfrom backend.util.json import json_dumps\n\nfrom ..models import Policy, PolicyIDExpiredAt, Subject\nfrom .query import PolicyList, new_backend_policy_list_by_subject\n\n\nclass PolicyOperationService:\n def delete_by_ids(self, system_id: str, subject: Subject, policy_ids: List[int]):\n \"\"\"\n 删除指定policy_id的策略\n \"\"\"\n with transaction.atomic():\n self._delete_db_policies(system_id, subject, policy_ids)\n iam.delete_policies(system_id, subject.type, subject.id, policy_ids)\n\n def alter(\n self,\n system_id: str,\n subject: Subject,\n create_policies: Optional[List[Policy]] = None,\n update_policies: Optional[List[Policy]] = None,\n delete_policy_ids: Optional[List[int]] = None,\n ):\n \"\"\"\n 变更subject的Policies\n \"\"\"\n create_policies = create_policies or []\n update_policies = update_policies or []\n delete_policy_ids = delete_policy_ids or []\n\n with transaction.atomic():\n if create_policies:\n self._create_db_policies(system_id, subject, create_policies)\n\n if update_policies:\n self._update_db_policies(system_id, subject, update_policies)\n\n if delete_policy_ids:\n self._delete_db_policies(system_id, subject, delete_policy_ids)\n\n if create_policies or update_policies or delete_policy_ids:\n self._alter_backend_policies(system_id, subject, create_policies, update_policies, delete_policy_ids)\n\n if create_policies:\n self._sync_db_policy_id(system_id, subject)\n\n def _alter_backend_policies(\n self,\n system_id: str,\n subject: Subject,\n create_policies: List[Policy],\n update_policies: List[Policy],\n delete_policy_ids: List[int],\n ):\n \"\"\"\n 执行对policies的创建, 更新, 删除操作, 调用后端批量操作接口\n \"\"\"\n # 组装backend变更策略的数据\n backend_create_policies = [p.to_backend_dict() for p in create_policies]\n backend_update_policies = [p.to_backend_dict() for p in update_policies]\n\n return iam.alter_policies(\n system_id, subject.type, subject.id, backend_create_policies, backend_update_policies, delete_policy_ids\n )\n\n def _create_db_policies(self, system_id: str, subject: Subject, policies: List[Policy]) -> None:\n \"\"\"\n 创建新的策略\n \"\"\"\n db_policies = [p.to_db_model(system_id, subject) for p in policies]\n PolicyModel.objects.bulk_create(db_policies, batch_size=100)\n\n def _update_db_policies(self, system_id: str, subject: Subject, policies: List[Policy]) -> None:\n \"\"\"\n 更新已有的策略\n \"\"\"\n policy_list = PolicyList(policies)\n\n db_policies = PolicyModel.objects.filter(\n subject_id=subject.id, subject_type=subject.type, system_id=system_id, policy_id__in=policy_list.ids\n ).only(\"id\", \"action_id\")\n\n # 使用主键更新, 避免死锁\n for p in db_policies:\n update_policy = policy_list.get(p.action_id)\n if not update_policy:\n continue\n PolicyModel.objects.filter(id=p.id).update(\n _resources=json_dumps([rt.dict() for rt in update_policy.related_resource_types])\n )\n\n def _delete_db_policies(self, system_id: str, subject: Subject, policy_ids: List[int]):\n \"\"\"\n 删除db Policies\n \"\"\"\n PolicyModel.objects.filter(\n system_id=system_id, subject_type=subject.type, subject_id=subject.id, policy_id__in=policy_ids\n ).delete()\n\n def _sync_db_policy_id(self, system_id: str, subject: Subject) -> None:\n \"\"\"\n 同步SaaS-后端策略的policy_id\n \"\"\"\n db_policies = PolicyModel.objects.filter(\n system_id=system_id, subject_type=subject.type, subject_id=subject.id, policy_id=0\n ).defer(\"_resources\", \"_environment\")\n\n if len(db_policies) == 0:\n return\n\n backend_policy_list = new_backend_policy_list_by_subject(system_id, subject)\n for p in db_policies:\n backend_policy = backend_policy_list.get(p.action_id)\n if not backend_policy:\n continue\n p.policy_id = backend_policy.id\n\n PolicyModel.objects.bulk_update(db_policies, fields=[\"policy_id\"], batch_size=100)\n\n def renew(self, subject: Subject, thin_policies: List[PolicyIDExpiredAt]):\n \"\"\"\n 权策续期\n \"\"\"\n iam.update_policy_expired_at(subject.type, subject.id, [one.dict() for one in thin_policies])\n","sub_path":"saas/backend/service/policy/operation.py","file_name":"operation.py","file_ext":"py","file_size_in_byte":5592,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"287932716","text":"# uncompyle6 version 3.7.4\n# Python bytecode 2.5 (62131)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.linux-i686/egg/boduch/constant/priority.py\n# Compiled at: 2009-08-14 17:29:31\n\"\"\"This module defines constants related to handle priority.\"\"\"\nPRIORITY_CRITICAL = 40\nPRIORITY_MAJOR = 30\nPRIORITY_MINOR = 20\nPRIORITY_TRIVIAL = 10\n__all__ = [\n 'PRIORITY_CRITICAL', 'PRIORITY_MAJOR', 'PRIORITY_MINOR',\n 'PRIORITY_TRIVIAL']","sub_path":"pycfiles/boduch-0.2.1-py2.5/priority.py","file_name":"priority.py","file_ext":"py","file_size_in_byte":488,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"25113794","text":"__author__ = 'Maksymilian Mika'\r\n\r\nimport sys, pygame\r\nimport Display\r\n\r\npygame.init()\r\npygame.font.init()\r\nsize = width, height = 60*8, 60*8\r\nblack = 0, 0, 0\r\nscreen = pygame.display.set_mode(size)\r\n\r\nboard_display = Display.BoardDisplay()\r\n\r\nset = False\r\n\r\nwhile 1:\r\n for event in pygame.event.get():\r\n if event.type == pygame.QUIT: sys.exit()\r\n if event.type == pygame.MOUSEBUTTONDOWN:\r\n board_display.dragged(pygame.mouse.get_pos())\r\n if event.type == pygame.MOUSEBUTTONUP:\r\n board_display.dropped(pygame.mouse.get_pos())\r\n\r\n \r\n screen.fill(black)\r\n board_display.display(screen)\r\n pygame.display.flip()\r\n","sub_path":"Main.py","file_name":"Main.py","file_ext":"py","file_size_in_byte":703,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"570720086","text":"from argparse import Namespace\nfrom base64 import b64decode\nfrom functools import wraps\nfrom typing import Optional\n\nfrom flask import Response, request\nfrom flask.json import JSONEncoder\nfrom flask_restful import Resource, abort\nfrom werkzeug.exceptions import HTTPException\n\nfrom api.app import status\nfrom api.app.exc import APIError\nfrom config import config\nfrom db.exc import DBError\nfrom db.models import UserAccessToken\n\n__all__ = (\n 'login_required',\n 'APIResource',\n)\n\n\ndef login_required(func):\n @wraps(func)\n def check_authorization(*args, **kwargs):\n user = APIResource.get_user()\n if user is None:\n return abort(403)\n return func(user, *args, **kwargs)\n\n return check_authorization\n\n\nclass APIResource(Resource):\n\n @property\n def ip_address(self):\n return request.environ.get('REMOTE_ADDR', None)\n\n def dispatch_request(self, *args, **kwargs):\n common_processor = getattr(self, 'common', None)\n # noinspection PyBroadException\n try:\n if common_processor is not None:\n ret = common_processor(*args, **kwargs)\n if not isinstance(ret, (list, tuple)):\n args = (ret,)\n response = super().dispatch_request(*args)\n except HTTPException as exc:\n return self.error_response(exc)\n except DBError as exc:\n return self.error_response(exc, status.HTTP_400_BAD_REQUEST)\n except Exception:\n if config.DEBUG:\n raise\n return self.error_response(APIError())\n\n try:\n response, code = response\n except ValueError:\n code = status.HTTP_200_OK\n return response, code\n\n @classmethod\n def error_response(cls, exc, status_code=None):\n if status_code is None:\n status_code = exc.code\n description = exc.description\n data = dict(code=exc.__class__.__name__, description=description)\n data = JSONEncoder().encode(data)\n resp = Response(\n response=data,\n status=status_code,\n mimetype='application/json'\n )\n return resp\n\n @staticmethod\n def get_page(args: Namespace, default_page: int = 1):\n page = args.page\n return default_page if page is None or page <= 0 else page\n\n @staticmethod\n def get_per_page(args: Namespace, default_per_page: int = 10):\n per_page = args.per_page\n return default_per_page if per_page is None or per_page <= 0 else per_page\n\n @staticmethod\n def get_token() -> Optional[str]:\n authorization = request.headers.get('Authorization', '')\n if not authorization:\n return None\n\n try:\n bearer = authorization.split(' ')[1]\n except IndexError:\n return None\n\n try:\n return b64decode(bearer).decode()\n except (UnicodeEncodeError, ValueError):\n return None\n\n @staticmethod\n def get_user(token: Optional[str] = None):\n if token is None:\n token = APIResource.get_token()\n return UserAccessToken.get_user_by_token(token)\n","sub_path":"api/app/resources/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":3167,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"162078122","text":"from django import forms\nfrom .models import Tienda, Region, Ciudad,Producto,Lista\nfrom django.contrib.auth.forms import UserCreationForm\nfrom django.contrib.auth.models import User\nfrom django.core.exceptions import ValidationError\nfrom django.forms.models import inlineformset_factory\nfrom django.contrib.auth import get_user_model\nUser = get_user_model()\nclass TiendaForm(forms.ModelForm):\n\n class Meta:\n model = Tienda\n fields = ('nombre','sucursal','direccion','region','ciudad')\n \n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.fields['ciudad'].queryset = Ciudad.objects.none()\n\n if 'region' in self.data:\n try:\n region_id = int(self.data.get('region'))\n self.fields['ciudad'].queryset = Ciudad.objects.filter(region_id=region_id).order_by('nombre')\n except (ValueError, TypeError):\n pass # invalid input from the client; ignore and fallback to empty Ciudad queryset\n elif self.instance.pk:\n self.fields['ciudad'].queryset = self.instance.region.ciudad_set.order_by('nombre')\n \n\nclass ListaForm(forms.ModelForm):\n class Meta:\n model = Lista\n \n fields = ['nombre']\n\n \n def __init__(self, *args, **kwargs):\n super(ListaForm, self).__init__(*args, **kwargs)\n for field in iter(self.fields):\n self.fields[field].widget.attrs.update({\n 'class': 'form-control'\n })\n \nclass ProductoForm(forms.ModelForm):\n\n\n class Meta:\n model = Producto\n fields = ['nombre','costoPresupuestado','costoReal','notasAdicionales', 'estado','tienda','lista']\n\n def __init__(self, *args, **kwargs):\n super(ProductoForm, self).__init__(*args, **kwargs)\n for field in iter(self.fields):\n self.fields[field].widget.attrs.update({\n 'class': 'form-control'\n })\n\nclass DetalleProductoForm(forms.ModelForm):\n class Meta:\n model = Producto\n fields = ['nombre','costoPresupuestado','costoReal','notasAdicionales', 'estado','tienda']\n\n def __init__(self, *args, **kwargs):\n super(DetalleProductoForm, self).__init__(*args, **kwargs)\n for field in iter(self.fields):\n instance = getattr(self, 'nombre', '')\n if instance is not None:\n self.fields['nombre'].widget.attrs['readonly'] = True\n self.fields['costoPresupuestado'].widget.attrs['readonly'] = True\n self.fields['notasAdicionales'].widget.attrs['readonly'] = True\n self.fields['tienda'].widget.attrs['disabled'] = True\n \n def clean_nombre(self):\n instance = getattr(self, 'instance', None)\n if instance and instance.pk:\n return instance.nombre\n else:\n return self.cleaned_data['nombre']\n \n \n \n \nDetalleCompraFormSet = inlineformset_factory(Lista, Producto, form=ProductoForm, extra=5)\n\n\nDetalleCompraFormSetV2 = inlineformset_factory(Lista, Producto, form=DetalleProductoForm, extra=10)\n\n\n\n \n\n\n \n \n\n\n \n\n\n\n\n\n\n","sub_path":"blog/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":3189,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"190589281","text":"#!/usr/bin/python3\n\"\"\" start app\"\"\"\nfrom models import storage\nfrom api.v1.views import app_views\nfrom flask import Flask, jsonify\nfrom models import storage\nfrom os import environ\n\n\napp = Flask(__name__)\napp.register_blueprint(app_views)\n\n\n@app.teardown_appcontext\ndef close(self):\n \"\"\"call close methot\"\"\"\n storage.close()\n\n\n@app.errorhandler(404)\ndef error_not_found(e):\n return {\"error\": \"Not found\"}, 404\n\nif __name__ == \"__main__\":\n host = environ.get('HBNB_API_HOST', '0.0.0.0')\n port = environ.get('HBNB_API_PORT', '5000')\n app.run(host=host, port=port, threaded=True)\n","sub_path":"api/v1/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":595,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"313290091","text":"# Name: William W\n# Period 6\n# Dice Rolling Simulator\n\nimport random\n\nrollnums = [0, 0, 0, 0, 0, 0]\n\n# main code\nnumber = int(input(\"How many rolls? \"))\n\nfor i in range(number):\n roll = random.randint(1, 6)\n print(str(i + 1) + \": Rolled \" + str(roll))\n rollnums[roll - 1] += 1\n\nprint(\"Total Rolls: \" + str(number + 1))\n\nfor i in range(6):\n print(str(i + 1) + \": \" + str(rollnums[i]))\n\nprint(\"\\nPercentages:\")\nfor i in range(6):\n print(str(i + 1) + \": \" + str(rollnums[i] / 10 * 100))","sub_path":"dice.py","file_name":"dice.py","file_ext":"py","file_size_in_byte":498,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"614857575","text":"from functools import wraps\nfrom datetime import timedelta\nfrom time import sleep\n\n\ndef with_delay(total_delay: timedelta,\n sleep_delay: timedelta,\n to_catch=(AssertionError, ),\n strict=True):\n \"\"\"\n Description:\n Decorator for cycling execution of a function till it returns without throwing any error.\n It can be used for functions that should wait an input, variable change and so on.\n\n When original function is called, if it returns a result, it is returned from a decorator.\n If in same call any assertion that is listed in to_catch is raised, original function will be called\n once again in sleep_delay.\n After total_delay if error from to_catch does not disappear, actual error will be risen.\n If original function throws error not listed in to_catch list, it will be propagated outside the decorator.\n\n :param total_delay: Total delay to perform repeated function call.\n :param sleep_delay: Sleep time between function calls.\n :param to_catch: Exceptions to ignore to continue cycling.\n :param strict: If True than can raise any function exception not listed in to_catch list.\n If False, then all exception (based on Exception) will be suppressed.\n\n Usage:\n with_delay(timedelta(minutes=3), timedelta(seconds=30))\n def function():\n some_bool_arg = get_some_bool_arg()\n assert some_bool_arg is True\n\n In this case will be performed validation multiple times until some_bool_arg becomes True.\n Every 30 seconds and during 3 minutes function will be executed periodically until timeout.\n After timeout will try to execute function without catching an exception.\n \"\"\"\n def decorator(function):\n @wraps(function)\n def wrapper(*args, **kwargs):\n end_time = datetime.today() + total_delay\n while datetime.today() < end_time:\n try:\n return function(*args, **kwargs)\n except to_catch: # catch only selected exceptions\n sleep(sleep_delay.total_seconds())\n except Exception: # catch remaining exceptions\n if strict is True:\n raise\n else:\n sleep(sleep_delay.total_seconds())\n # execute original function without try-except block\n # as total_delay was passed\n return function(*args, **kwargs)\n return wrapper\n return decorator\n\n\nif __name__ == '__main__':\n import unittest\n from datetime import datetime\n\n class TestCase(unittest.TestCase):\n # todo mock sleep function\n def test_timeout(self):\n \"\"\"\n Expected sequence: f(), sleep(0.7), f(), sleep(0.7), f() raises AssertionError.\n \"\"\"\n @with_delay(timedelta(seconds=1), timedelta(seconds=0.7))\n def f():\n raise AssertionError\n start = datetime.today()\n with self.assertRaises(AssertionError):\n f()\n end = datetime.today()\n delta = end - start\n # expected delta: 0.7[s] * 2 = 1.4[s]\n # tolerance: 0.1[s]\n self.assertTrue(timedelta(seconds=1.3) <= delta <= timedelta(seconds=1.5))\n\n def test_no_timeout(self):\n \"\"\"\n Expected sequence: f(), sleep(0.7), f() returns None\n \"\"\"\n c = 0\n @with_delay(timedelta(seconds=1), timedelta(seconds=0.7))\n def f():\n nonlocal c\n c += 1\n if c <= 1: # first call raise AssertionError\n raise AssertionError\n # return None\n start = datetime.today()\n f()\n end = datetime.today()\n delta = end - start\n # expected delta: 0.7[s]\n # tolerance: 0.1[s]\n self.assertTrue(timedelta(seconds=0.6) <= delta <= timedelta(seconds=0.8))\n\n def test_no_delay(self):\n \"\"\"\n Expected sequence: f() returns None\n \"\"\"\n @with_delay(timedelta(seconds=1), timedelta(seconds=0.7))\n def f():\n pass\n start = datetime.today()\n f()\n end = datetime.today()\n delta = end - start\n # expected delta: 0[s]\n # tolerance: 0.1[s]\n self.assertTrue(delta <= timedelta(seconds=0.1))\n\n unittest.main()\n","sub_path":"util/with_delay.py","file_name":"with_delay.py","file_ext":"py","file_size_in_byte":4568,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"372450676","text":"import numpy as np\r\n\r\ndef Linear_Regression(x, y):\r\n \r\n x_mean = np.mean(x)\r\n y_mean = np.mean(y)\r\n\r\n sum_of_cross_deviations_YX = np.sum(x * y) - len(x) * y_mean * x_mean\r\n sum_of_squared_devations_X = np.sum(x**2) - len(x) * x_mean * x_mean\r\n\r\n beta1 = sum_of_cross_deviations_YX / sum_of_squared_devations_X\r\n beta0 = y_mean - beta1 * x_mean\r\n estimatedY = beta0 + beta1 * x\r\n \r\n return (estimatedY, beta0, beta1)\r\n \r\n\r\n","sub_path":"regression_function.py","file_name":"regression_function.py","file_ext":"py","file_size_in_byte":456,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"349790164","text":"#!/usr/bin/env python3\n\nimport argparse\nimport asyncio\nimport logging\n\nimport capnp\nimport calculator_capnp\n\n\nlogger = logging.getLogger(__name__)\nlogger.setLevel(logging.DEBUG)\n\n\nasync def evaluate_impl(expression, params=None):\n \"\"\"Implementation of CalculatorImpl::evaluate(), also shared by\n FunctionImpl::call(). In the latter case, `params` are the parameter\n values passed to the function; in the former case, `params` is just an\n empty list.\"\"\"\n\n which = expression.which()\n\n if which == \"literal\":\n return expression.literal\n elif which == \"previousResult\":\n return (await expression.previousResult.read()).value\n elif which == \"parameter\":\n assert expression.parameter < len(params)\n return params[expression.parameter]\n elif which == \"call\":\n call = expression.call\n func = call.function\n\n # Evaluate each parameter.\n paramPromises = [evaluate_impl(param, params) for param in call.params]\n vals = await asyncio.gather(*paramPromises)\n\n # When the parameters are complete, call the function.\n result = await func.call(vals)\n return result.value\n else:\n raise ValueError(\"Unknown expression type: \" + which)\n\n\nclass ValueImpl(calculator_capnp.Calculator.Value.Server):\n \"Simple implementation of the Calculator.Value Cap'n Proto interface.\"\n\n def __init__(self, value):\n self.value = value\n\n async def read(self, **kwargs):\n return self.value\n\n\nclass FunctionImpl(calculator_capnp.Calculator.Function.Server):\n\n \"\"\"Implementation of the Calculator.Function Cap'n Proto interface, where the\n function is defined by a Calculator.Expression.\"\"\"\n\n def __init__(self, paramCount, body):\n self.paramCount = paramCount\n self.body = body.as_builder()\n\n async def call(self, params, _context, **kwargs):\n \"\"\"Note that we're returning a Promise object here, and bypassing the\n helper functionality that normally sets the results struct from the\n returned object. Instead, we set _context.results directly inside of\n another promise\"\"\"\n\n assert len(params) == self.paramCount\n return await evaluate_impl(self.body, params)\n\n\nclass OperatorImpl(calculator_capnp.Calculator.Function.Server):\n\n \"\"\"Implementation of the Calculator.Function Cap'n Proto interface, wrapping\n basic binary arithmetic operators.\"\"\"\n\n def __init__(self, op):\n self.op = op\n\n async def call(self, params, **kwargs):\n assert len(params) == 2\n\n op = self.op\n\n if op == \"add\":\n return params[0] + params[1]\n elif op == \"subtract\":\n return params[0] - params[1]\n elif op == \"multiply\":\n return params[0] * params[1]\n elif op == \"divide\":\n return params[0] / params[1]\n else:\n raise ValueError(\"Unknown operator\")\n\n\nclass CalculatorImpl(calculator_capnp.Calculator.Server):\n \"Implementation of the Calculator Cap'n Proto interface.\"\n\n async def evaluate(self, expression, _context, **kwargs):\n return ValueImpl(await evaluate_impl(expression))\n\n async def defFunction(self, paramCount, body, _context, **kwargs):\n return FunctionImpl(paramCount, body)\n\n async def getOperator(self, op, **kwargs):\n return OperatorImpl(op)\n\n\nasync def new_connection(stream):\n await capnp.TwoPartyServer(stream, bootstrap=CalculatorImpl()).on_disconnect()\n\n\ndef parse_args():\n parser = argparse.ArgumentParser(\n usage=\"\"\"Runs the server bound to the given address/port ADDRESS. \"\"\"\n )\n\n parser.add_argument(\"address\", help=\"ADDRESS:PORT\")\n\n return parser.parse_args()\n\n\nasync def main():\n host, port = parse_args().address.split(\":\")\n server = await capnp.AsyncIoStream.create_server(new_connection, host, port)\n async with server:\n await server.serve_forever()\n\n\nif __name__ == \"__main__\":\n asyncio.run(main())\n","sub_path":"examples/async_calculator_server.py","file_name":"async_calculator_server.py","file_ext":"py","file_size_in_byte":3972,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"601285860","text":"import random\n\na = random.randint(1, 20)\nprint(\"请输入你要猜的数字\")\nb = 0\nwhile a != b:\n c = input()\n b = int(c)\n if b < a:\n print(\"你输入小了\")\n elif b > a:\n print(\"你输入大了\")\nelse:\n print(\"猜对了\")\n","sub_path":"0324/text01/baiyunfucecond02.py","file_name":"baiyunfucecond02.py","file_ext":"py","file_size_in_byte":254,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"256037209","text":"#!/usr/bin/python3\nimport requests\nfrom bs4 import BeautifulSoup as bs\nimport csv\nfrom pprint import pprint\nfrom config import headers as h\nfrom time import sleep\nimport psycopg2\nimport datetime\nimport json\nimport re\nfrom selenium import webdriver\n\n\nb = webdriver.Chrome()\nb.implicitly_wait(15)\n\n\ndef get_html(url):\n b.get(url)\n sleep(5)\n return b.page_source\n\n\n# Возвращает 6 категорий товаров\ndef get_store_menu(html):\n store_menu = []\n soup = bs(html, 'lxml')\n store_menu = soup.find('div', id='store.menu').find('nav', class_='navigation').find('ul').find_all('li')\n for store in store_menu:\n store = store.find('a').get('href')\n store_menu.append(store)\n return store_menu\n\n\n# Return list cards from page\ndef get_cards(html):\n list_cards = []\n soup = bs(html, 'lxml')\n cards = soup.find('div', class_='products wrapper grid products-grid').find('ol').find_all('li')\n for card in cards:\n card = card.find('a').get('href')\n print(card)\n list_cards.append(card)\n return list_cards\n\nlist_flex = []\ndef get_data(html):\n soup = bs(html, 'lxml')\n name = soup.find('div', class_='category-colors').find('div', class_='page-title-wrapper product').text.strip()\n print(name)\n specifications = soup.find('div', class_='specifications').text.strip().replace('\\n', '').replace(' ', ' ')\n print(specifications)\n available = soup.find('div', class_='specifications').find('a').get('href')\n list_flex.append(available)\n available_text = soup.find('div', class_='specifications').find('a').text.strip()\n print(available_text)\n overview = soup.find('div', class_='product attribute desktop overview').find('div').text.strip()\n print(overview)\n scheme_picture = soup.find('div', class_='product-info-cross-section-image desktop').find('img').get('src')\n print(scheme_picture)\n banner = soup.find('div', class_='usp-banner').find('div', class_='row').find('div').text.strip().replace('\\n', ' ')\n print(banner)\n\n main_pic = soup.find('div', class_='gallery-placeholder')\\\n .find('div', attrs={'data-gallery-role':'gallery'}).find('div', class_='fotorama__stage__shaft')\\\n .find('div').get('href')\n print(main_pic)\n\n\n\ndef main():\n headers = h\n '''\n url = 'https://oracdecor.ru/plintusy?p=2'\n get_cards(get_html(url, headers))\n '''\n urls = [\n 'https://oracdecor.ru/plintusy/sx138f_skirting',\n 'https://oracdecor.ru/karnizy/c391_cornice_moulding',\n 'https://oracdecor.ru/moldingi/p4020_panel_moulding',\n 'https://oracdecor.ru/skrytoe-osveschenie/c380_uplighter'\n ]\n for url in urls:\n get_data(get_html(url))\n print('---------')\n\n\nif __name__ == '__main__':\n main()","sub_path":"components/oracdecor/selenium_oradecor.py","file_name":"selenium_oradecor.py","file_ext":"py","file_size_in_byte":2778,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"179965747","text":"import logging\nimport time\nimport pandas as pd\nfrom ibapi.utils import iswrapper\nfrom ibapi.client import EClient\nfrom ibapi.wrapper import EWrapper\n# types\nfrom ibapi.common import * # @UnusedWildImport\nfrom ibapi.contract import * # @UnusedWildImport\n\n# https://stackoverflow.com/questions/41510945/interactive-brokers-obtain-historical-data-of-opt-midpoint-and-trades\n# https://groups.io/g/twsapi/topic/data_for_expired_contracts_no/4042776?p=\n\nclass TestApp(EWrapper, EClient):\n def __init__(self):\n EWrapper.__init__(self)\n EClient.__init__(self, wrapper=self)\n self.data = [] # Initialize variable to store candle\n self.contract = Contract()\n\n def nextValidId(self, orderId: int):\n # we can start now\n self.start()\n\n def start(self):\n self.historicalDataOperations_req()\n print(\"Executing requests ... finished\")\n\n def historicalDataOperations_req(self):\n self.contract.symbol = \"TQQQ\"\n self.contract.secType = \"OPT\"\n self.contract.exchange = \"SMART\"\n self.contract.currency = \"USD\"\n self.contract.lastTradeDateOrContractMonth = \"20210730\"\n self.contract.strike = 128\n self.contract.right = \"C\"\n self.contract.multiplier = \"100\"\n\n self.reqHistoricalData(4103, self.contract, '',\n \"2 D\", \"1 hour\", \"MIDPOINT\", 1, 1, False, [])\n\n # https://interactivebrokers.github.io/tws-api/historical_bars.html\n\n def historicalData(self, reqId: int, bar: BarData):\n self.data.append([reqId, bar])\n #print(\"HistoricalData. ReqId:\", reqId, \"BarData.\", bar)\n\n df = pd.DataFrame(self.data)\n print(df)\n df.to_csv('history.csv')\n self.disconnect()\n\ndef main():\n app = TestApp()\n app.connect(\"127.0.0.1\", port=7497, clientId=102)\n print(\"serverVersion:%s connectionTime:%s\" % (app.serverVersion(), app.twsConnectionTime()))\n app.run()\n\n\nif __name__ == \"__main__\":\n main()\n\n","sub_path":"options/historical_options_bars_clean.py","file_name":"historical_options_bars_clean.py","file_ext":"py","file_size_in_byte":1987,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"411953403","text":"from django import forms\nfrom django.forms.formsets import formset_factory\nfrom django.forms.models import modelformset_factory\nfrom .models import *\n\n\nclass AddIngredientsForm(forms.ModelForm):\n class Meta:\n model = Ingredients\n fields = ('ingredient_name',)\n\n\nAddIngredientsFormSet = formset_factory(AddIngredientsForm, extra=5)\n\nAddIngredientsInOrderFormSet = modelformset_factory(Ingredients, form=AddIngredientsForm, fields=('ingredient_name', 'count'))\n\n\nclass AddDishForm(forms.ModelForm):\n class Meta:\n model = Dish\n fields = ('dish_name',)\n exclude = ('ingredients', 'author')\n\n\nclass AddOrderForm(forms.ModelForm):\n class Meta:\n model = Order\n exclude = ('dish', 'author', 'ingredients')\n","sub_path":"mymenu/menu/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":756,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"58619612","text":"import json\nimport requests\nfrom parse import parse\nimport warnings\n\nheaders = {'User-Agent':'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:68.0) Gecko/20100101 Firefox/68.0'}\n\nwarnings.filterwarnings(\"ignore\")\napikey = \"\"\ndef get_message():\n\tr = requests.post(\"https://www.hackthebox.eu/api/shouts/get/initial/html/100?api_token=\"+apikey)\n\tjs = json.loads(r.content)\n\thtml = js['html']\n\tlastmsg = html[99]\n\tfor x in html:\n\t\tif 'https://www.hackthebox.eu/storage/avatars/' in x:\n\t\t\tparse(x)\n\nquite = True\n\ndef get_last_message(quite):\n\twhile True:\n\t\tr = requests.post(\"https://www.hackthebox.eu/api/shouts/get/initial/html/1?api_token=\"+apikey)\n\t\tjs = json.loads(r.content)\n\t\thtml = js['html'][0]\n\t\tif quite == True:\n\t\t\tif 'https://www.hackthebox.eu/storage/avatars/' in html:\n\t\t\t\tif html != lastmsg:\n\t\t\t\t\tparse(html)\n\t\t\t\t\tlastmsg = html\n\t\telse:\n\t\t\tif html != lastmsg:\n\t\t\t\tparse(html)\n\t\t\t\tlastmsg = html\n","sub_path":"chats.py","file_name":"chats.py","file_ext":"py","file_size_in_byte":901,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"540170182","text":"\"\"\"pytest configuration.\n\"\"\"\n\nfrom __future__ import print_function\nimport sys\nimport os.path\nimport datetime\nimport re\nfrom random import getrandbits\nimport zlib\nimport subprocess\nimport shutil\nimport tempfile\nimport logging\nfrom distutils.version import StrictVersion as Version\nimport pytest\nimport icat\nimport icat.config\ntry:\n from suds.sax.date import UtcTimezone\nexcept ImportError:\n UtcTimezone = None\n\n\n# Note that pytest captures stderr, so we won't see any logging by\n# default. But since Suds uses logging, it's better to still have\n# a well defined basic logging configuration in place.\nlogging.basicConfig(level=logging.INFO)\n\ntestdir = os.path.dirname(__file__)\n\n\n# ============================= helper ===============================\n\nif sys.version_info < (3, 0):\n def buf(seq):\n return buffer(bytearray(seq))\nelse:\n def buf(seq):\n return bytearray(seq)\n\nclass DummyDatafile(object):\n \"\"\"A dummy file with random content to be used for test upload.\n \"\"\"\n def __init__(self, directory, name, size, date=None):\n if date is not None:\n date = (date, date)\n self.name = name\n self.fname = os.path.join(directory, name)\n chunksize = 8192\n crc32 = 0\n with open(self.fname, 'wb') as f:\n while size > 0:\n if chunksize > size:\n chunksize = size\n chunk = buf(getrandbits(8) for _ in range(chunksize))\n size -= chunksize\n crc32 = zlib.crc32(chunk, crc32)\n f.write(chunk)\n if date:\n os.utime(self.fname, date)\n self.crc32 = \"%x\" % (crc32 & 0xffffffff)\n self.stat = os.stat(self.fname)\n self.size = self.stat.st_size\n if UtcTimezone:\n mtime = int(self.stat.st_mtime)\n self.mtime = datetime.datetime.fromtimestamp(mtime, UtcTimezone())\n else:\n self.mtime = None\n\nclass tmpSessionId:\n \"\"\"Temporarily switch to another sessionId in an ICAT client.\n \"\"\"\n def __init__(self, client, sessionId):\n self.client = client\n self.saveSessionId = client.sessionId\n self.sessionId = sessionId\n def __enter__(self):\n self.client.sessionId = self.sessionId\n return self.client\n def __exit__(self, type, value, tb):\n self.client.sessionId = self.saveSessionId\n\n\ndef gettestdata(fname):\n fname = os.path.join(testdir, \"data\", fname)\n assert os.path.isfile(fname)\n return fname\n\n\ndef getConfig(confSection=\"root\", **confArgs):\n \"\"\"Get the configuration, skip on ConfigError.\n \"\"\"\n confFile = os.path.join(testdir, \"data\", \"icat.cfg\")\n if not os.path.isfile(confFile):\n pytest.skip(\"no test ICAT server configured\")\n try:\n args = [\"-c\", confFile, \"-s\", confSection]\n conf = icat.config.Config(**confArgs).getconfig(args)\n conf.cmdargs = [\"-c\", conf.configFile[0], \"-s\", conf.configSection]\n return conf\n except icat.ConfigError as err:\n pytest.skip(err.message)\n\n\ndef get_icat_version():\n conf = getConfig(needlogin=False)\n client = icat.Client(conf.url, **conf.client_kwargs)\n return client.apiversion\n\n# ICAT server version we talk to. Ignore any errors from\n# get_icat_version(), if something fails (e.g. no server is configured\n# at all), set a dummy zero version number.\ntry:\n icat_version = get_icat_version()\nexcept:\n icat_version = Version(\"0.0\")\n\ndef require_icat_version(minversion, reason):\n if icat_version < minversion:\n pytest.skip(\"need ICAT server version %s or newer: %s\" \n % (minversion, reason))\n\n\ndef callscript(scriptname, args, stdin=None, stdout=None, stderr=None):\n script = os.path.join(testdir, \"scripts\", scriptname)\n cmd = [sys.executable, script] + args\n print(\"\\n>\", *cmd)\n subprocess.check_call(cmd, stdin=stdin, stdout=stdout, stderr=stderr)\n\n\nyaml_filter = (re.compile(r\"^# (Date|Service|ICAT-API|Generator): .*$\"),\n r\"# \\1: ###\")\nxml_filter = (re.compile(r\"^\\s*<(date|service|apiversion|generator)>.*\\1>$\"),\n r\" <\\1>###\\1>\")\n\ndef filter_file(infile, outfile, pattern, repl):\n \"\"\"Filter a text file.\n\n This may be needed to compare some test output file with\n predefined results, because some information in the file might not\n depend on the actual test but rather dynamically change with each\n call. Such as the header of a dump file that contains date and\n ICAT version.\n \"\"\"\n with open(infile, 'rt') as inf, open(outfile, 'wt') as outf:\n while True:\n l = inf.readline()\n if not l:\n break\n l = re.sub(pattern, repl, l)\n outf.write(l)\n\n\n# ============================ fixtures ==============================\n\n# Deliberately not using the 'tmpdir' fixture provided by pytest,\n# because it seem to use a predictable directory name in /tmp wich is\n# insecure.\n\nclass TmpDir(object):\n \"\"\"Provide a temporary directory.\n \"\"\"\n def __init__(self):\n self.dir = tempfile.mkdtemp(prefix=\"python-icat-test-\")\n def __del__(self):\n self.cleanup()\n def cleanup(self):\n if self.dir:\n shutil.rmtree(self.dir)\n self.dir = None\n\n@pytest.fixture(scope=\"session\")\ndef tmpdirsec(request):\n tmpdir = TmpDir()\n request.addfinalizer(tmpdir.cleanup)\n return tmpdir\n\n\n@pytest.fixture(scope=\"session\")\ndef standardConfig():\n return getConfig()\n\n\ntestcontent = gettestdata(\"icatdump.yaml\")\n\n@pytest.fixture(scope=\"session\")\ndef setupicat(standardConfig):\n require_icat_version(\"4.4.0\", \"need InvestigationGroup\")\n callscript(\"wipeicat.py\", standardConfig.cmdargs)\n args = standardConfig.cmdargs + [\"-f\", \"YAML\", \"-i\", testcontent]\n callscript(\"icatingest.py\", args)\n\n\n# ============================= hooks ================================\n\nclass DependencyItemStatus(object):\n \"\"\"Status of a test item in a dependency manager.\n \"\"\"\n\n Phases = ('setup', 'call', 'teardown')\n\n def __init__(self):\n self.results = { w:None for w in self.Phases }\n\n def __str__(self):\n l = [\"%s: %s\" % (w, self.results[w]) for w in self.Phases]\n return \"Status(%s)\" % \", \".join(l)\n\n def addResult(self, rep):\n self.results[rep.when] = rep.outcome\n\n def isSuccess(self):\n return list(self.results.values()) == ['passed', 'passed', 'passed']\n\nclass DependencyManager(object):\n \"\"\"Dependency manager, stores the results of tests.\n \"\"\"\n\n ScopeCls = {'module':pytest.Module, 'session':pytest.Session}\n\n @classmethod\n def getManager(cls, item, scope='module'):\n \"\"\"Get the DependencyManager object from the node at scope level.\n Create it, if not yet present.\n \"\"\"\n node = item.getparent(cls.ScopeCls[scope])\n if not hasattr(node, 'dependencyManager'):\n node.dependencyManager = cls()\n return node.dependencyManager\n\n def __init__(self):\n self.results = {}\n\n def addResult(self, item, marker, rep):\n name = marker.kwargs.get('name')\n if not name:\n name = item.name\n status = self.results.setdefault(name, DependencyItemStatus())\n status.addResult(rep)\n\n def checkDepend(self, depends):\n for i in depends:\n if not(i in self.results and self.results[i].isSuccess()):\n pytest.skip(\"depends on %s\" % i)\n\n\n@pytest.hookimpl(tryfirst=True, hookwrapper=True)\ndef pytest_runtest_makereport(item, call):\n \"\"\"Store the test outcome if this item is marked \"dependency\".\n \"\"\"\n outcome = yield\n marker = item.get_marker(\"dependency\")\n if marker is not None:\n rep = outcome.get_result()\n manager = DependencyManager.getManager(item)\n manager.addResult(item, marker, rep)\n\n\ndef pytest_runtest_setup(item):\n \"\"\"Check dependencies if this item is marked \"dependency\".\n Skip if any of the dependencies has not been run successfully.\n \"\"\"\n marker = item.get_marker(\"dependency\")\n if marker is not None:\n depends = marker.kwargs.get('depends')\n if depends:\n manager = DependencyManager.getManager(item)\n manager.checkDepend(depends)\n\n\ndef require_servertest():\n if not pytest.config.getoption(\"--servertests\"):\n pytest.skip(\"need --servertests option to run\")\n\ndef pytest_addoption(parser):\n parser.addoption(\"--servertests\", action=\"store_true\",\n help=\"run tests for testing the server.\")\n\n\ndef pytest_report_header(config):\n \"\"\"Add information on the icat package used in the tests.\n \"\"\"\n modpath = os.path.dirname(os.path.abspath(icat.__file__))\n if icat_version > \"0.0\":\n server = icat_version\n else:\n server = \"-\"\n return [ \"python-icat: %s (%s)\" % (icat.__version__, icat.__revision__), \n \" %s\" % (modpath),\n \"icat.server: %s\" % server]\n","sub_path":"tests/conftest.py","file_name":"conftest.py","file_ext":"py","file_size_in_byte":8948,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"71317362","text":"# -*- coding: utf-8 -*-\r\n\r\nimport requests\r\nimport pymysql\r\nimport json\r\nimport datetime\r\nimport time\r\nimport unittest\r\nfrom Common.readConfig import *\r\n\r\napi_host=global_var.api_host #测试地址:192.168.10.56\r\nheaders1=global_var.headers4 #发货宝18883612485\r\nheaders2=global_var.headers5 #小二18599937985\r\nheaders3=global_var.headers6 #运营后台18883612485\r\nheaders4=global_var.headers7 #直营系统18883612485\r\nheadersNull=global_var.headersNull #居家小二客户端18883612485\r\ngoods=global_var.goods\r\ndb=global_var.db #数据库 59\r\ndb1=global_var.db1 #数据库70\r\narg1=global_var.arg1 #断言文本\r\nPictureid = '5b4ef81bd423d40001f0195e' #图片id\r\n\r\n# 录单使用的产品信息\r\n\r\n\r\nclass testYiChangFlow(unittest.TestCase):\r\n def test_a(self):\r\n '''录单'''\r\n #发货宝录单接口\r\n url1 = \"http://\" + api_host + \"/ms-fahuobao-order/FhbOrder/saveOrder\"\r\n global i\r\n i = datetime.datetime.now()\r\n print(\"收件人姓名:异常流程测试\"+ str(i.month) + str(i.day)+'-'+str(i.hour)+'-'+str(i.minute))\r\n data1 = {\r\n \"businessNo\": \"BSTE02\",\r\n \"serviceNo\": \"FHB01\",\r\n \"orderWay\": \"1\",\r\n \"wokerUserName\": \"\",\r\n \"wokerPhone\": \"\",\r\n \"wokerPrice\": \"\",\r\n \"checked\": \"\",\r\n \"verfiyType\": \"\",\r\n \"goods\": goods,\r\n \"isElevator\": \"\",\r\n \"predictServiceDate\": \"\",\r\n \"predictDevliveryDate\": \"\",\r\n \"memo\": \"\",\r\n \"isArriva\": 1,\r\n \"boolCollection\": \"0\",\r\n \"collectionMoney\": \"\",\r\n \"collectionMemo\": \"\",\r\n \"allVolume\": \"1\",\r\n \"allWeight\": \"1\",\r\n \"allPackages\": \"1\",\r\n \"consigneeName\": \"异常流程测试\"+ str(i.month) + str(i.day)+'-'+str(i.hour)+'-'+str(i.minute),\r\n \"consigneePhone\": \"18883612485\",\r\n \"consigneeAddress\": \"231\",\r\n \"deliveryName\": \"23213\",\r\n \"deliveryPhone\": \"18883612485\",\r\n \"provinceNo\": \"430000\",\r\n \"province\": \"湖南省\",\r\n \"cityNo\": \"430100\",\r\n \"city\": \"长沙市\",\r\n \"districtNo\": \"430103\",\r\n \"district\": \"天心区\",\r\n \"deliveryProvinceNo\": \"\",\r\n \"deliveryProvince\": \"\",\r\n \"deliveryCityNo\": \"\",\r\n \"deliveryCity\": \"\",\r\n \"deliveryDistrictNo\": \"\",\r\n \"deliveryDistrict\": \"\",\r\n \"verifyOrderNo\": \"\"\r\n }\r\n\r\n request1 = requests.request(\"POST\", url=url1, data = json.dumps(data1) ,headers = headers1)\r\n print(\"录单:\" + request1.text)\r\n time.sleep(2)\r\n self.assertIn(arg1, request1.text, msg='测试field')\r\n\r\n def test_b(self):\r\n '''通过订单的收件人姓名查询出订单id、订单编号,将订单推到天心区并用185这个账号进行竞价'''\r\n i = datetime.datetime.now()\r\n consigne_name1 = \"异常流程测试\" + str(i.month) + str(i.day)+'-'+str(i.hour)+'-'+str(i.minute)\r\n\r\n # 通过订单的收件人姓名查询出订单id\r\n sql1 = \"SELECT fhb_order_id,order_no FROM fhb_order_consignee_info a inner join fhb_order b on a.fhb_order_id=b.id WHERE a.consigne_name = '\"+consigne_name1+\"' ORDER BY a.foundtime DESC\"\r\n\r\n # 使用cursor()方法获取操作游标\r\n cursor = db.cursor()\r\n # 执行SQL语句\r\n cursor.execute(sql1)\r\n # 获取所有记录列表\r\n results = cursor.fetchall()\r\n # print(results[0])\r\n # 有多个的情况,取第一个订单的id\r\n global orderid,orderno\r\n orderid = results[0]['fhb_order_id']\r\n orderno = results[0]['order_no']\r\n print(\"订单id:\" + orderid)\r\n print(\"订单编号:\" + orderno)\r\n db.close()\r\n # 将订单推到天心区并用185这个账号进行竞价\r\n url2 = \"http://\" + api_host + \"/ms-fahuobao-order/bidding/quoted-price\"\r\n data2 = {\r\n \"memo\":\"\",\r\n \"money\":\"200\",\r\n \"orderId\":orderid\r\n }\r\n\r\n request2 = requests.request(\"POST\", url=url2, data=json.dumps(data2), headers=headers2)\r\n print(\"竞价:\" + request2.text)\r\n self.assertIn(arg1, request2.text, msg='测试field')\r\n def test_c(self):\r\n # 通过师傅id查询竞价记录\r\n '''通过师傅id查询竞价记录'''\r\n db.connect()\r\n sql2 = \"SELECT id,fhb_order_id FROM fhb_order_bidding_log WHERE people_user_id ='0cacc658-dd29-40bb-9c69-b2e19677275f' and fhb_order_id= '\"+orderid+\"' ORDER BY foundtime DESC\"\r\n # 使用cursor()方法获取操作游标\r\n cursor2 = db.cursor()\r\n # 执行SQL语句\r\n cursor2.execute(sql2)\r\n # 获取所有记录列表\r\n results2 = cursor2.fetchall()\r\n # print(results[0])\r\n # 有多个的情况,取第一个订单的id\r\n global fhb_order_id,jingjiaid\r\n fhb_order_id=orderid\r\n jingjiaid = results2[0]['id']\r\n print(\"订单id:\" + fhb_order_id)\r\n print(\"竞价id:\" + jingjiaid)\r\n # db.close()\r\n def test_d(self):\r\n # 修改竞价金额为0.01\r\n '''修改竞价金额为0.01'''\r\n sql3 = \"UPDATE fhb_order_bidding_log set money = '0.01' where fhb_order_id = '\" + orderid + \"'\"\r\n print(sql3)\r\n cursor3 = db.cursor()\r\n # 执行SQL语句\r\n cursor3.execute(sql3)\r\n #MySQL的默认存储引擎就是InnoDB, 所以对数据库数据的操作会在事先分配的缓存中进行, 只有在commit之后, 数据库的数据才会改变\r\n db.commit()\r\n def test_e(self):\r\n # 选择接口\r\n '''选择师傅'''\r\n url3 = \"http://\" + api_host + \"/ms-fahuobao-order/FhbOrder/choice-pay?t=1531964865851&orderId=\"+fhb_order_id+\"&biddingLogId=\"+jingjiaid+\"\"\r\n request_yichang = requests.request(\"GET\", url=url3, headers=headers1)\r\n # print('选择中标师傅'+request_yichang.text)\r\n time.sleep(2)\r\n self.assertIn(arg1, request_yichang.text, msg='测试field')\r\n\r\n\r\n def test_f(self):\r\n # 支付接口,objectList为订单id\r\n '''支付'''\r\n time.sleep(5)\r\n url4 = \"http://\" + api_host + \"/ms-fahuobao-user/wallet/balance-pay\"\r\n data4 = {\"objectList\":[fhb_order_id],\"money\":0.01,\"password\":\"123456\"}\r\n request4 = requests.request(\"POST\", url=url4, data=json.dumps(data4), headers=headers1)\r\n print(\"支付:\" + request4.text)\r\n time.sleep(2)\r\n self.assertIn(arg1, request4.text, msg='测试field')\r\n\r\n\r\n # 通过发货宝订单编号查出scm订单id与订单编号\r\n def test_g(self):\r\n '''通过发货宝订单编号查出scm订单id与订单编号'''\r\n time.sleep(5)\r\n db1.connect()\r\n sql4 = \"select id,order_no from order_data where order_no='\"+orderno+\"'\"\r\n\r\n # 使用cursor()方法获取操作游标\r\n cursor3 = db1.cursor()\r\n # 执行SQL语句\r\n cursor3.execute(sql4)\r\n # 获取所有记录列表\r\n results3 = cursor3.fetchall()\r\n # print(results[0])\r\n # 有多个的情况,取第一个订单的id\r\n global scmorderid,scmorderno\r\n scmorderid = results3[0]['id']\r\n scmorderno=results3[0]['order_no']\r\n print(\"scm订单id:\" + scmorderid)\r\n print(\"scm订单编号:\" + scmorderno)\r\n\r\n # db2.close()\r\n def test_h(self):\r\n # 预约\r\n '''预约'''\r\n url5 = \"http://\" + api_host + \"/ms-fahuobao-order-data/appOrder/appointappoint-distributionOne-choose\"\r\n data5 = {\r\n \"branchUserId\": \"\",\r\n \"cause\": \"\",\r\n \"codeYT\": \"night\",\r\n \"ids\": [scmorderid],\r\n \"timeYT\": str(i.year) + \"-\" + str(i.month) + \"-\" + str(i.day)\r\n }\r\n # print(json.dumps(data5))\r\n request5 = requests.request(\"POST\", url=url5, data=json.dumps(data5), headers=headers2)\r\n print(\"预约:\" + request5.text)\r\n time.sleep(2)\r\n self.assertIn(arg1, request5.text, msg='测试field')\r\n def test_i(self):\r\n # 师傅发起异常\r\n '''师傅发起异常'''\r\n Pictureid='5b4ef81bd423d40001f0195e'\r\n url6 = \"http://\" + api_host + \"/ms-fahuobao-order/fhbAbnormal/saveAbnormal\"\r\n data6 = {\r\n \"trunkPicture\":[\r\n Pictureid\r\n ],\r\n \"isPickUp\":1,\r\n \"abnormalCode\":\"EXE02\",\r\n \"abnormalPicture\":[\r\n Pictureid\r\n ],\r\n \"abnormalMemo\":\"\",\r\n \"abnormalPacks\":\"1\",\r\n \"orderId\":scmorderid\r\n }\r\n\r\n request6 = requests.request(\"POST\", url=url6, data=json.dumps(data6), headers=headers2)\r\n print(\"师傅发起异常:\" + request6.text)\r\n time.sleep(2)\r\n self.assertIn(arg1, request6.text, msg='测试field')\r\n\r\n def test_j(self):\r\n # 通过订单id查出异常id\r\n '''通过订单id查出异常id'''\r\n time.sleep(5)\r\n db.connect()\r\n sql5 = \"SELECT id FROM fhb_order_abnormal WHERE order_id ='\"+orderid+\"' ORDER BY found_date DESC\"\r\n\r\n # 使用cursor()方法获取操作游标\r\n cursor5 = db.cursor()\r\n # 执行SQL语句\r\n cursor5.execute(sql5)\r\n # 获取所有记录列表\r\n results5 = cursor5.fetchall()\r\n # print(results[0])\r\n # 有多个的情况,取第一个订单的id\r\n global yichangid\r\n yichangid = results5[0]['id']\r\n print(\"异常id:\" + yichangid)\r\n def test_k(self):\r\n # 货主给出方案\r\n '''货主给出方案'''\r\n url7 = \"http://\" + api_host + \"/ms-fahuobao-order-abnormal/FhbOrderAbnormal/merProvideScheme\"\r\n data7 = {\r\n \"schemeDesc\":\"231321\",\r\n \"pic\":[\r\n Pictureid\r\n ],\r\n \"id\":yichangid\r\n }\r\n request7 = requests.request(\"POST\", url=url7, data=json.dumps(data7), headers=headers1)\r\n print(\"货主给出方案:\" + request7.text)\r\n time.sleep(2)\r\n self.assertIn(arg1, request7.text, msg='测试field')\r\n def test_l(self):\r\n # 师傅不同意货主方案(发起仲裁)\r\n '''师傅不同意货主方案(发起仲裁)'''\r\n url8 = \"http://\" + api_host + \"/ms-fahuobao-order-abnormal/FhbOrderAbnormal/workerApplyArbitration\"\r\n data8 = {\r\n \"pic\":[\r\n Pictureid\r\n ],\r\n \"schemeDesc\":\"321\",\r\n \"id\":yichangid\r\n }\r\n request8 = requests.request(\"POST\", url=url8, data=json.dumps(data8), headers=headers2)\r\n print(\"师傅不同意货主方案(发起仲裁):\" + request8.text)\r\n time.sleep(2)\r\n self.assertIn(arg1, request8.text, msg='测试field')\r\n def test_m(self):\r\n # 仲裁处理\r\n '''仲裁处理'''\r\n url9 = \"http://\" + api_host + \"/ms-fahuobao-order-abnormal/FhbOrderAbnormal/dealArbitration\"\r\n data9 ={\"schemeDesc\":\"3131\",\"refundAmount\":\"0.00\",\"pic\":[Pictureid],\"refundOrderState\":1,\"id\":yichangid}\r\n request9 = requests.request(\"POST\", url=url9, data=json.dumps(data9), headers=headers3)\r\n print(\"仲裁处理:\" + request9.text)\r\n time.sleep(2)\r\n self.assertIn(arg1, request9.text, msg='测试field')\r\n\r\n\r\nif __name__ == '__main__':\r\n unittest.main()\r\n","sub_path":"SaaSFlowTest/NewFlowTestCase/testYiChangNormalFlow.py","file_name":"testYiChangNormalFlow.py","file_ext":"py","file_size_in_byte":11772,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"532253287","text":"# -*- coding: utf-8 -*-\nimport random\nimport numpy as np\nimport tensorflow as tf\nfrom framework.algorithm.predictor import Predictor as Predictor_Base\nfrom algorithms.dqn.model import Model\n\n\nclass Predictor(Predictor_Base):\n def __init__(self, with_random=True):\n super().__init__(with_random)\n self.model = Model(with_random)\n self._init_sess()\n self.epsilon_start = 1.0\n self.epsilon_end = 0.01\n self.epsilon_decay_steps = 20000\n self.epsilons = np.linspace(self.epsilon_start,\n self.epsilon_end,\n self.epsilon_decay_steps)\n self.step = 0\n\n def get_value(self, feature):\n return self.sess.run(self.model.value,\n feed_dict={self.feature: [feature]})\n\n def process(self, feature):\n q_value, value = self.sess.run([self.model.q_value, self.model.value], \\\n feed_dict={self.feature: [feature]})\n return self._sample(q_value), value, None\n\n def _sample(self, q_value):\n self.step += 1\n epsilon = self.epsilons[min(self.step,\n self.epsilon_decay_steps - 1)]\n rand_float = float(random.uniform(0, self.epsilon_decay_steps)) / float(self.epsilon_decay_steps)\n if rand_float < epsilon and self.with_random:\n return random.randint(0, self.action_dim - 1)\n else:\n return np.argmax(q_value, axis=1)\n \n def _init_sess(self):\n cpu_num = 1\n config = tf.ConfigProto(device_count={\"CPU\": cpu_num}, inter_op_parallelism_threads=cpu_num, \\\n intra_op_parallelism_threads=cpu_num, log_device_placement=False)\n self.graph = tf.Graph()\n with self.graph.as_default():\n self.feature = tf.placeholder(shape=(None, self.state_dim[0], self.state_dim[1], self.state_dim[2]), \\\n name=\"feature\", dtype=np.uint8)\n self.model.inference(self.feature)\n self.init_saver = tf.train.Saver(tf.global_variables())\n self.sess = tf.Session(graph=self.graph, config=config)\n self.sess.run(tf.global_variables_initializer())\n","sub_path":"demonstrate/algorithms/dqn/predictor.py","file_name":"predictor.py","file_ext":"py","file_size_in_byte":2201,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"400267934","text":"import pygame, os\nfrom pygame import image\nfrom .drawable import Drawable\nfrom polybius.managers import FRAMES\n\nclass Animated(Drawable):\n \n def __init__(self, imageName, location):\n \n super().__init__(imageName, location, (0,0))\n \n self._frame = 0\n self._row = 0\n self._animationTimer = 0\n self._framesPerSecond = 5\n self._nFrames = 2\n \n self._animate = True\n \n\n \n def updateAnimation(self, ticks):\n if self._animate:\n self._animationTimer += ticks\n \n if self._animationTimer > 1 / self._framesPerSecond:\n \n self._frame += 1\n self._frame %= self._nFrames\n self._animationTimer -= 1 / self._framesPerSecond\n self._image = FRAMES.getFrame(self._imageName, (self._frame, self._row))\n self._defaultImage = self._image\n \n if self.isFlipped():\n self._image = pygame.transform.flip(self._image, True, False)\n if self.isScaled():\n self.scale(self._scaleValue)\n if self.isRotated():\n angle = self._rotation\n self.setRotation(0)\n self.rotate(angle)\n\n self._mask = pygame.mask.from_surface(self._image)\n \n def startAnimation(self):\n self._animate = True\n \n def stopAnimation(self):\n self._animate = False\n\n def setFramesInRow(self, frames):\n self._nFrames = frames\n\n def setRowOnSpriteSheet(self, row):\n self._row = row\n\n def setCurrentFrame(self, frame):\n self._frame = frame\n\n def setFPS(self, fps):\n self._framesPerSecond = fps\n\n def setAnimationTimer(self, animationTime):\n self._animationTimer = animationTime\n","sub_path":"polybius/graphics/basics/animated.py","file_name":"animated.py","file_ext":"py","file_size_in_byte":1744,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"286652441","text":"# Author: Izaak Neutelings (December 2018)\n# /shome/ytakahas/work/Leptoquark/CMSSW_9_4_4/src/PhysicsTools/NanoAODTools/NanoTreeProducer/leptonSF\n# HTT: https://github.com/CMS-HTT/LeptonEfficiencies\n# https://twiki.cern.ch/twiki/bin/view/CMS/MuonReferenceEffs2017\nfrom corrections import modulepath\nfrom ScaleFactorTool import ScaleFactor, ScaleFactorHTT, ScaleFactorEmb\n\n\nclass EmbeddingSFs:\n def __init__(self, year=2017):\n \"\"\"Load histograms from files.\"\"\"\n\n assert year in [\n 2016, 2017, 2018\n ], \"EmbeddingSFs: You must choose a year from: 2016, 2017, or 2018.\"\n if year == 2016:\n pathEmb = modulepath + \"/leptonEfficiencies/kit/inputs/2016/KIT/legacy_16_v1/\"\n pathEmbSel = modulepath + \"/leptonEfficiencies/kit/inputs/2016/KIT/embeddingselection/\"\n self.sftool_id = ScaleFactorEmb(\n pathEmb + \"muon_TP_Data_2016_Fits_ID_pt_eta_bins.root\",\n pathEmb + \"muon_TP_Embedding_2016_Fits_ID_pt_eta_bins.root\",\n 'ID',2016) # MediumID,\n self.sftool_iso = ScaleFactorEmb(\n pathEmb + \"muon_TP_Data_2016_Fits_Iso_pt_eta_bins.root\",\n pathEmb + \"muon_TP_Embedding_2016_Fits_Iso_pt_eta_bins.root\",\n \"Iso\",2016) # isolation\n self.sftool_trig = ScaleFactorEmb(\n pathEmb +\n \"muon_TP_Data_2016_Fits_Trg_pt_eta_bins.root\",\n pathEmb +\n \"muon_TP_Embedding_2016_Fits_Trg_pt_eta_bins.root\",\n 'Trg',2016)\n self.sftool_seltrig = ScaleFactorEmb(\n pathEmbSel + \"embeddingselection_TP_Data_2016_Fits_Trg8_pt_eta_bins.root\",\n pathEmbSel + \"embeddingselection_TP_Data_2016_Fits_Trg17_pt_eta_bins.root\",\n 'selection',2016)\n elif year == 2017:\n pathEmb = modulepath + \"/leptonEfficiencies/kit/inputs/2017/KIT/legacy/\"\n pathEmbSel = modulepath + \"/leptonEfficiencies/kit/inputs/2017/ICSF/2017/\"\n self.sftool_id = ScaleFactorEmb(\n pathEmb + \"muon_TP_Data_2017_Fits_ID_pt_eta_bins.root\",\n pathEmb + \"muon_TP_Embedding_2017_Fits_ID_pt_eta_bins.root\",\n 'ID',2017) # MediumID,\n self.sftool_iso = ScaleFactorEmb(\n pathEmb + \"muon_TP_Data_2017_Fits_Iso_pt_eta_bins.root\",\n pathEmb + \"muon_TP_Embedding_2017_Fits_Iso_pt_eta_bins.root\",\n \"Iso\",2017) # isolation\n self.sftool_trig = ScaleFactorEmb(\n pathEmb +\n \"muon_TP_Data_2017_Fits_Trg_IsoMu27_or_IsoMu24_pt_eta_bins.root\",\n pathEmb +\n \"muon_TP_Embedding_2017_Fits_Trg_IsoMu27_or_IsoMu24_pt_eta_bins.root\",\n 'Trg_IsoMu27_or_IsoMu24',2017)\n self.sftool_seltrig = ScaleFactorEmb(\n pathEmbSel + \"Mu8/muon_SFs.root\",\n pathEmbSel + \"Mu17/muon_SFs.root\",\n 'selection',2017)\n else:\n pathEmb = modulepath + \"/leptonEfficiencies/kit/inputs/2018/KIT/v18_2/\"\n pathEmbSel = modulepath + \"/leptonEfficiencies/kit/inputs/2018/KIT/2018/\"\n self.sftool_id = ScaleFactorEmb(\n pathEmb + \"muon_TP_Data_2018_Fits_ID_pt_eta_bins.root\",\n pathEmb + \"muon_TP_Embedding_2018_Fits_ID_pt_eta_bins.root\",\n 'ID',2018) # MediumID,\n self.sftool_iso = ScaleFactorEmb(\n pathEmb + \"muon_TP_Data_2018_Fits_Iso_pt_eta_bins.root\",\n pathEmb + \"muon_TP_Embedding_2018_Fits_Iso_pt_eta_bins.root\",\n \"Iso\",2018) # isolation\n self.sftool_trig = ScaleFactorEmb(\n pathEmb +\n \"muon_TP_Data_2018_Fits_Trg_IsoMu27_or_IsoMu24_pt_eta_bins.root\",\n pathEmb +\n \"muon_TP_Embedding_2018_Fits_Trg_IsoMu27_or_IsoMu24_pt_eta_bins.root\",\n 'Trg_IsoMu27_or_IsoMu24',2018)\n self.sftool_seltrig = ScaleFactorEmb(\n pathEmbSel + \"Mu8/muon_SFs.root\",\n pathEmbSel + \"Mu17/muon_SFs.root\", 'selection',2018)\n\n def getTriggerSF(self, pt, eta):\n \"\"\"Get SF for single muon trigger.\"\"\"\n return self.sftool_trig.getSF(pt, abs(eta))\n\n def getIsoSF(self, pt, eta):\n \"\"\"Get SF for muon Isolation.\"\"\"\n return self.sftool_trig.getSF(pt, abs(eta))\n\n def getIdSF(self, pt, eta):\n \"\"\"Get SF for muon identification \"\"\"\n return self.sftool_id.getSF(pt, abs(eta))\n\n def getEmbSelSF(self, pt_1, eta_1, pt_2, eta_2):\n \"\"\" Get SF for Embeddeded DoubleMuon Selection \"\"\"\n return self.sftool_seltrig.getSelectionSF(pt_1, abs(eta_1), pt_2,\n abs(eta_2))\n","sub_path":"corrections/EmbeddingSFs.py","file_name":"EmbeddingSFs.py","file_ext":"py","file_size_in_byte":4766,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"131559001","text":"\n\nfrom collections import deque\nfrom collections import OrderedDict\nimport os\nimport random\n\nimport numpy as np\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\nfrom torch.autograd import Variable\n\nfrom tools.base import BasePolicyOptimizer\n\nfrom IPython import embed\n\n\n# model\ndef normalized_columns_initializer(weights, std=1.0):\n out = torch.randn(weights.size())\n out *= std / torch.sqrt(out.pow(2).sum(1, keepdim=True))\n return out\n\n\ndef weights_init(m):\n classname = m.__class__.__name__\n if classname.find('Conv') != -1:\n weight_shape = list(m.weight.data.size())\n fan_in = np.prod(weight_shape[1:4])\n fan_out = np.prod(weight_shape[2:4]) * weight_shape[0]\n w_bound = np.sqrt(6. / (fan_in + fan_out))\n m.weight.data.uniform_(-w_bound, w_bound)\n m.bias.data.fill_(0)\n elif classname.find('Linear') != -1:\n weight_shape = list(m.weight.data.size())\n fan_in = weight_shape[1]\n fan_out = weight_shape[0]\n w_bound = np.sqrt(6. / (fan_in + fan_out))\n m.weight.data.uniform_(-w_bound, w_bound)\n m.bias.data.fill_(0)\n\n\nclass Mode(object):\n NORMAL = 0\n REWARD_PREDICTION = 1\n PIXEL_CONTROL = 2\n\n\nclass Model(torch.nn.Module):\n\n def __init__(self, args, env, n_chnnnels, n_states, n_actions):\n super(Model, self).__init__()\n self.eval_interval = args.eval_interval\n self.n_actions = env.n_actions\n self.n_outputs = env.n_outputs\n self.n_lstm_cells = 256\n\n # visual embedding\n self.conv1 = nn.Conv2d(env.ob_channels, 16, 6, stride=3, padding=0)\n # self.pool1 = nn.MaxPool2d(2)\n self.conv2 = nn.Conv2d(16, 32, 6, stride=3, padding=0)\n # self.pool2 = nn.MaxPool2d(2)\n # self.conv2_bn = nn.BatchNorm2d(16)\n self.conv3 = nn.Conv2d(32, 32, 3, stride=1, padding=0)\n # self.conv3_bn = nn.BatchNorm2d(32)\n self.conv4 = nn.Conv2d(32, 32, 3, stride=1, padding=0)\n # self.conv4_bn = nn.BatchNorm2d(32)\n # self.conv5 = nn.Conv2d(32, 32, 3, stride=1, padding=0)\n # self.conv6 = nn.Conv2d(32, 32, 3, stride=1, padding=0)\n # self.fc1 = nn.Linear(288, self.n_lstm_cells)\n self.fc1 = nn.Linear(512, self.n_lstm_cells)\n\n # recurrent processing\n self.lstm = nn.LSTMCell(self.n_lstm_cells + n_states + env.n_keys + 1,\n self.n_lstm_cells)\n\n # policy and baseline\n # num_outputs = n_actions\n self.critic = nn.Linear(self.n_lstm_cells, 1)\n # self.actor = nn.Linear(self.n_lstm_cells, n_actions)\n for out_idx in range(len(env.n_outputs)):\n setattr(self, 'actor{}'.format(out_idx),\n nn.Linear(self.n_lstm_cells, env.n_outputs[out_idx]))\n\n # reward prediction\n self.rp_fc1 = nn.Linear(self.n_lstm_cells * 3 + n_states, 128)\n self.reward = nn.Linear(128, 3) # positive, negative, neutral\n\n self.apply(weights_init)\n\n self.fc1.weight.data = normalized_columns_initializer(\n self.fc1.weight.data, 1.0)\n self.fc1.bias.data.fill_(0)\n\n self.lstm.weight_ih.data = normalized_columns_initializer(\n self.lstm.weight_ih.data, 1.0)\n self.lstm.bias_ih.data.fill_(0)\n self.lstm.weight_hh.data = normalized_columns_initializer(\n self.lstm.weight_hh.data, 1.0)\n self.lstm.bias_hh.data.fill_(0)\n\n self.critic.weight.data = normalized_columns_initializer(\n self.critic.weight.data, 1.0)\n self.critic.bias.data.fill_(0)\n\n for out_idx in range(len(env.n_outputs)):\n actor = getattr(self, 'actor{}'.format(out_idx))\n actor.weight.data = normalized_columns_initializer(\n actor.weight.data, 0.1)\n actor.bias.data.fill_(0)\n\n # self.actor.weight.data = normalized_columns_initializer(\n # self.actor.weight.data, 0.1)\n # self.actor.bias.data.fill_(0)\n\n self.rp_fc1.weight.data = normalized_columns_initializer(\n self.rp_fc1.weight.data, 1.0)\n self.rp_fc1.bias.data.fill_(0)\n\n self.reward.weight.data = normalized_columns_initializer(\n self.reward.weight.data, 1.0)\n self.reward.bias.data.fill_(0)\n self.train()\n\n # act 메소드\n # self._prev_action = torch.zeros(1, env.n_keys)\n # self._prev_r = torch.zeros(1, 1)\n\n # act buffer\n self.cx = np.zeros((1, self.n_lstm_cells))\n self.hx = np.zeros((1, self.n_lstm_cells))\n self.last_action = 0\n self.prev_a = np.zeros((1, env.n_keys))\n self.prev_r = np.zeros((1, 1))\n # self.value = None\n # self.probs = None\n self.state_value = None\n self.action_values = None\n\n def visual_embedding(self, obs):\n x = F.relu(self.conv1(obs))\n # x = self.pool1(x)\n\n # x = F.relu(self.conv2_bn(self.conv2(x)))\n # x = F.relu(self.conv3_bn(self.conv3(x)))\n # x = F.relu(self.conv4_bn(self.conv4(x)))\n\n x = F.relu(self.conv2(x))\n # x = self.pool2(x)\n\n x_ = self.conv3(x) + x[:, :, 1:-1, 1:-1]\n x = F.relu(x_)\n\n x = self.conv4(x) + x_[:, :, 1:-1, 1:-1]\n x = F.relu(x)\n # x = F.relu(self.conv5(x) + x[:, :, 1:-1, 1:-1])\n # x = F.relu(self.conv6(x) + x[:, :, 1:-1, 1:-1])\n\n x = x.view(x.size(0), -1)\n return self.fc1(x)\n\n def forward(self, inputs, mode=Mode.NORMAL):\n if mode == Mode.NORMAL:\n obs, state, (hx, cx), pa, pr = inputs\n x = self.visual_embedding(obs)\n x = torch.cat((x, state, pa, pr), dim=1)\n hx, cx = self.lstm(x, (hx, cx))\n x = hx\n\n value = self.critic(x)\n probs = list()\n for out_idx in range(len(self.n_outputs)):\n actor = getattr(self, 'actor{}'.format(out_idx))\n logit = actor(x)\n probs.append(F.softmax(logit, dim=1))\n probs = torch.einsum('bi,bj->bij', probs)\n probs = probs.reshape(-1, self.n_actions)\n log_probs = torch.log(probs)\n\n return value, probs, log_probs, (hx, cx)\n\n elif mode == Mode.REWARD_PREDICTION:\n ob1, ob2, ob3, state = inputs\n x1 = self.visual_embedding(ob1)\n x2 = self.visual_embedding(ob2)\n x3 = self.visual_embedding(ob3)\n x = torch.cat((x1, x2, x3, state), dim=1)\n x = F.relu(self.rp_fc1(x))\n return self.reward(x)\n\n elif mode == Mode.PIXEL_CONTROL:\n pass\n\n else:\n raise NotImplementedError\n\n def reset_buffer(self):\n self.cx.fill(0)\n self.hx.fill(0)\n self.prev_a.fill(0)\n self.prev_r.fill(0)\n\n def act(self, device, args, inputs, eval_game):\n with torch.no_grad():\n # obs, state, (hx, cx), pa, pr = inputs\n obs, state = inputs\n hx, cx = self.hx, self.cx\n pa, pr = self.prev_a, self.prev_r\n\n obs = torch.from_numpy(obs).unsqueeze(0).to(torch.float).to(device)\n state = torch.from_numpy(state).unsqueeze(\n 0).to(torch.float).to(device)\n hx = torch.from_numpy(hx).to(torch.float).to(device)\n cx = torch.from_numpy(cx).to(torch.float).to(device)\n pa = torch.from_numpy(pa).to(torch.float).to(device)\n pr = torch.from_numpy(pr).to(torch.float).to(device)\n inputs_pt = obs, state, (hx, cx), pa, pr\n\n value, prob, log_prob, (hx, cx) = self(inputs_pt, mode=Mode.NORMAL)\n prob_np = prob.cpu().numpy()\n log_prob_ = log_prob.cpu().numpy()\n self.hx = hx.cpu().numpy()[:]\n self.cx = cx.cpu().numpy()[:]\n\n action = np.random.choice(prob.shape[1], p=prob_np.squeeze(0))\n self.last_action = action\n action_log_prob = np.take(log_prob_, [1], axis=1)\n\n self.state_value = value.cpu().mean().item()\n self.action_values = prob.cpu().numpy()\n\n return action, action_log_prob\n\n def buffer_step(self, env):\n self.prev_a[:] = env.actions[self.last_action]\n self.prev_r[:] = env.reward\n\n def to_params(self):\n return [p.data.cpu().numpy() for p in self.parameters()]\n\n def from_solution(self, device, solution):\n for param, weight in zip(self.parameters(), solution.params):\n param.data = torch.from_numpy(weight).to(device)\n\n def save(self):\n pass\n\n def load(self):\n pass\n\n def is_eval(self, n_games):\n return n_games % self.eval_interval == 0\n\n\n# optimizer\nclass MetaOptimizer(BasePolicyOptimizer):\n\n @staticmethod\n def seed(seed):\n torch.manual_seed(seed)\n random.seed(seed)\n\n @staticmethod\n def make_model(device, args, env):\n n_channels = env.ob_channels\n n_states = env.state_dims\n n_actions = len(env.actions)\n model = Model(args, env, n_channels, n_states, n_actions)\n model.eval_interval = 1\n return model.to(device)\n\n def __init__(self, rank, device, args, env, model, mailbox):\n super(Impala, self).__init__(rank, device, args, env, model, mailbox)\n\n self.max_steps = args.max_steps\n\n height = args.ob_height\n width = args.ob_width\n len_lstm = model.n_lstm_cells\n self.n_batches = args.n_batches\n self.n_steps = args.n_steps\n\n self.episodes = [[] for _ in range(args.n_batches)]\n self.observations = torch.zeros(\n args.n_steps + 1, args.n_batches, env.ob_channels, height, width).to(device)\n self.states = torch.zeros(\n args.n_steps + 1, args.n_batches, env.state_dims).to(device)\n self.actions = torch.LongTensor(\n args.n_steps, args.n_batches).to(device)\n self.action_log_probs_mu = torch.zeros(\n args.n_steps, args.n_batches).to(device)\n\n self.hx = torch.zeros(\n args.n_steps + 1, args.n_batches, len_lstm).to(device)\n self.cx = torch.zeros(\n args.n_steps + 1, args.n_batches, len_lstm).to(device)\n self.prev_actions = torch.zeros(\n args.n_steps + 1, args.n_batches, env.n_keys).to(device)\n self.prev_rewards = torch.zeros(\n args.n_steps + 1, args.n_batches, 1).to(device)\n self.rewards = torch.zeros(args.n_steps, args.n_batches, 1).to(device)\n self.masks = torch.zeros(args.n_steps, args.n_batches, 1).to(device)\n\n self.optimizer = self.update_optimizer(args, model)\n self._loss_variable = 0\n self.loss_dict = OrderedDict(\n loss=0, value_loss=0, policy_loss=0, entropy_loss=0)\n self.value_dict = OrderedDict(value=0)\n\n # aux. tasks\n self.aux_tasks = list()\n if UnrealRewardPredction.get_coef(args) > 0.0:\n self.aux_tasks.append(UnrealRewardPredction(\n rank, device, args, env, model, mailbox))\n\n def update_optimizer(self, args, model):\n\n if args.optimizer == 'sgd':\n self.optimizer = optim.SGD(\n model.parameters(), lr=args.lr, momentum=args.momentum)\n elif args.optimizer == 'rmsprop':\n self.optimizer = optim.RMSprop(\n model.parameters(), lr=args.lr, momentum=args.momentum)\n elif args.optimizer == 'adam':\n self.optimizer = optim.Adam(model.parameters(), lr=args.lr)\n else:\n raise NotImplementedError\n\n return self.optimizer\n\n def ready(self):\n if min([len(self.episodes[env_id]) for env_id in range(self.n_batches)]) > self.n_steps:\n return True\n else:\n return False\n\n def put(self, episode):\n env_id = np.argmin([len(self.episodes[env_id])\n for env_id in range(self.n_batches)])\n # self.episodes[env_id] += episode[:]\n self.episodes[env_id] += episode\n\n for aux_task in self.aux_tasks:\n aux_task.put(episode)\n\n def step(self, device, args, env, model):\n height = args.ob_height\n width = args.ob_width\n len_lstm = model.n_lstm_cells\n self.n_batches = args.n_batches\n self.n_steps = args.n_steps\n\n # Section: Value, Policy Func\n reward_list = deque(maxlen=1000)\n for env_id in range(self.n_batches):\n for step in range(self.n_steps):\n # ob1, s1, hx1, cx1, action, a_u, reward, ob2, s2, info2, done = self.episodes[env_id].pop(0)\n sample = self.episodes[env_id].pop(0)\n ob1 = sample.get('ob1')\n s1 = sample.get('state1')\n hx1 = sample.get('hx')\n cx1 = sample.get('cx')\n action = sample.get('action')\n a_u = sample.get('action_log_prob')\n reward = sample.get('reward')\n ob2 = sample.get('ob2')\n s2 = sample.get('state2')\n info2 = sample.get('info')\n done = sample.get('done')\n\n ob1 = torch.from_numpy(ob1).to(device)\n s1 = torch.from_numpy(s1).to(device)\n hx1 = torch.from_numpy(hx1).to(device)\n cx1 = torch.from_numpy(cx1).to(device)\n reward_list.append(reward)\n\n self.observations[step, env_id] = ob1\n self.states[step, env_id] = s1\n self.actions[step, env_id] = action\n self.action_log_probs_mu[step, env_id] = torch.tensor(a_u)\n self.rewards[step, env_id] = reward\n self.masks[step, env_id] = 0.0 if done else 1.0\n self.hx[step + 1, env_id] = hx1 * self.masks[step, env_id]\n self.cx[step + 1, env_id] = cx1 * self.masks[step, env_id]\n self.prev_actions[step + 1, env_id,\n :] = torch.tensor(env.actions[action]).to(torch.float)\n self.prev_actions[step + 1, env_id,\n :] *= self.masks[step, env_id]\n self.prev_rewards[step + 1, env_id] = reward * \\\n self.masks[step, env_id]\n\n ob2 = torch.from_numpy(ob2).to(device)\n s2 = torch.from_numpy(s2).to(device)\n self.observations[self.n_steps, env_id] = ob2\n self.states[self.n_steps, env_id] = s2\n\n flat_observations = self.observations.reshape(\n -1, env.ob_channels, height, width)\n flat_states = self.states.reshape(-1, env.state_dims)\n flat_hx = self.hx.reshape(-1, len_lstm)\n flat_cx = self.cx.reshape(-1, len_lstm)\n flat_prev_actions = self.prev_actions.reshape(-1, env.n_keys)\n flat_prev_r = self.prev_rewards.reshape(-1, 1)\n values, probs, log_probs, _ = model(\n (flat_observations, flat_states, (flat_hx, flat_cx), flat_prev_actions, flat_prev_r))\n values = values.view(self.n_steps + 1, self.n_batches, -1)\n # probs = F.softmax(logits, dim=1).view(args.n_steps + 1, args.n_batches, -1)\n # log_probs = F.log_softmax(logits, dim=1).view(args.n_steps + 1, args.n_batches, -1)\n probs = probs.view(self.n_steps + 1, self.n_batches, -1)\n log_probs = log_probs.view(self.n_steps + 1, self.n_batches, -1)\n action_log_probs = log_probs[:-1].gather(2, self.actions.unsqueeze(2))\n\n # v-trace\n entropies = -(log_probs * probs).sum(-1).unsqueeze(2)\n value_loss = torch.zeros(self.n_batches, 1).to(device)\n policy_loss = torch.zeros(self.n_batches, 1).to(device)\n gae = torch.zeros(self.n_batches, 1).to(device)\n returns = torch.zeros(self.n_batches, 1).to(device)\n returns.copy_(values[-1].data)\n rhos = torch.exp(action_log_probs.data -\n self.action_log_probs_mu.unsqueeze(2)).data.clamp(max=1.0)\n cs = torch.ones(self.n_batches, 1).to(device)\n\n for step in reversed(range(self.n_steps)):\n delta = rhos[step] * (self.rewards[step] + args.gamma *\n values[step + 1].data - values[step].data)\n cs = (cs * rhos[step]).clamp(max=1.0)\n returns = values[step].data + delta + args.lmbd * args.gamma * \\\n cs * (returns - values[step + 1].data) * self.masks[step - 1]\n # returns = values[step].data + delta + args.lmbd * args.gamma * cs * (returns - values[step + 1].data) * self.masks[step]\n value_loss = value_loss + 0.5 * \\\n (values[step] - returns.data).pow(2)\n # value_loss = value_loss + 0.5 * (self.rewards[step] + args.gamma * values[step + 1].data - values[step]).pow(2)\n if args.no_gae:\n advantages = returns - values[step].data\n policy_loss = policy_loss - \\\n action_log_probs[step] * advantages - \\\n args.ent_coef * entropies[step].data\n else:\n gae = args.gamma * args.tau * gae + delta\n policy_loss = policy_loss - \\\n action_log_probs[step] * gae - \\\n args.ent_coef * entropies[step].data\n\n self.value_dict['value'] = values.mean().item()\n\n value_loss = value_loss.mean()\n policy_loss = policy_loss.mean()\n entropy_loss = entropies.mean()\n self._loss_variable = args.value_coef * \\\n value_loss + args.policy_coef * policy_loss\n self.loss_dict['value_loss'] = value_loss.item()\n self.loss_dict['policy_loss'] = policy_loss.item()\n self.loss_dict['entropy_loss'] = entropy_loss.item()\n\n loss_dict = dict(\n total_loss=self._loss_variable.item(),\n value_loss=args.value_coef * value_loss.item(),\n policy_loss=args.policy_coef * policy_loss.item(),\n entropy_loss=args.ent_coef * entropy_loss.item())\n\n for aux_task in self.aux_tasks:\n if aux_task.ready():\n name = aux_task.loss_name\n coef = aux_task.get_coef(args)\n aux_task_loss = aux_task.step(device, args, env, model)\n aux_task_loss = aux_task_loss.mean()\n self._loss_variable += coef * aux_task_loss\n self.loss_dict[name] = aux_task_loss.item()\n loss_dict[name] = coef * aux_task_loss.item()\n\n self.loss_dict['total_loss'] = self._loss_variable.item()\n\n grad_dict = self._optimize(args, model)\n\n # self.hx[0].copy_(self.hx[-1])\n # self.cx[0].copy_(self.cx[-1])\n\n self.hx[0] = self.hx[-1]\n self.cx[0] = self.cx[-1]\n\n return loss_dict, grad_dict\n\n def _optimize(self, args, model):\n\n self.optimizer.zero_grad()\n self._loss_variable.backward()\n torch.nn.utils.clip_grad_norm_(model.parameters(), args.max_grad_norm)\n grad_dict = OrderedDict()\n for name, param in model.named_parameters():\n if param.grad is not None:\n grad_dict[name] = param.grad.data.mean().item()\n else:\n grad_dict[name] = 0.0\n self.optimizer.step()\n\n return grad_dict\n\n @property\n def buffer_size(self):\n buffer_size = OrderedDict()\n\n for bid, episode in enumerate(self.episodes):\n buffer_size[f'b{bid:03d}'] = len(episode)\n\n for aux_task in self.aux_tasks:\n buffer_size.update(aux_task.buffer_size)\n\n return buffer_size\n\n\nclass UnrealRewardPredction(BasePolicyOptimizer):\n\n loss_name = 'reward_pred_loss'\n\n def __init__(self, rank, device, args, env, model, mailbox):\n super(UnrealRewardPredction, self).__init__(\n rank, device, args, env, model, mailbox)\n self.device = device\n self.n_batches = args.n_batches\n self.base_reward = env.base_reward\n self.zero_reward_memory = deque(maxlen=args.replay_memory_size)\n self.nonzero_reward_memory = deque(maxlen=args.replay_memory_size)\n self._new_data_added = False\n\n @staticmethod\n def get_coef(args):\n return getattr(args, 'reward_pred_coef', 0.0)\n\n def put(self, episode):\n buff = deque(maxlen=4)\n pa, pr = None, 0\n\n for sample in episode:\n ob1 = sample.get('ob1')\n s1 = sample.get('state1')\n action = sample.get('action')\n reward = sample.get('reward')\n ob2 = sample.get('ob2')\n\n buff.append(ob1)\n if len(buff) == 4:\n ob1 = np.expand_dims(buff[0], 0)\n ob2 = np.expand_dims(buff[1], 0)\n ob3 = np.expand_dims(buff[2], 0)\n if reward == self.base_reward:\n self.zero_reward_memory.append([ob1, ob2, ob3, s1, reward])\n else:\n self.nonzero_reward_memory.append(\n [ob1, ob2, ob3, s1, reward])\n pa, pr = action, reward\n\n self._new_data_added = True # 새로운 데이터 추가: ready() -> True\n\n def ready(self):\n if self._new_data_added:\n # 데이터가 새로 추가되어야만 학습할 수 있음\n if len(self.zero_reward_memory) > self.n_batches // 2 and \\\n len(self.nonzero_reward_memory) > self.n_batches // 2:\n return True\n return False\n\n def step(self, device, args, env, model):\n\n zero_samples = random.sample(\n self.zero_reward_memory, self.n_batches // 2)\n nonzero_samples = random.sample(\n self.nonzero_reward_memory, self.n_batches // 2)\n samples = zero_samples + nonzero_samples\n\n ob1, ob2, ob3, st, rw = zip(*samples)\n ob1 = torch.from_numpy(np.vstack(ob1)).to(self.device)\n ob2 = torch.from_numpy(np.vstack(ob2)).to(self.device)\n ob3 = torch.from_numpy(np.vstack(ob3)).to(self.device)\n st = torch.from_numpy(np.vstack(st)).to(self.device)\n rwc = torch.zeros(len(rw), 3).to(self.device)\n for idx, r in enumerate(rw):\n if r > self.base_reward:\n rwc[idx, 0] = 1\n elif r < self.base_reward:\n rwc[idx, 1] = 1\n else:\n rwc[idx, 2] = 1\n\n logit = model((ob1, ob2, ob3, st), mode=Mode.REWARD_PREDICTION)\n log_prob = F.log_softmax(logit, dim=1)\n loss = -(rwc * log_prob).mean()\n\n self._new_data_added = False # 새로운 데이터 추가 필요: ready() -> False\n return loss\n\n def optimize(self):\n pass\n\n @property\n def buffer_size(self):\n buffer_size = OrderedDict()\n value = 100 * len(self.zero_reward_memory) / \\\n self.zero_reward_memory.maxlen\n buffer_size['zero'] = value\n value = 100 * len(self.nonzero_reward_memory) / \\\n self.nonzero_reward_memory.maxlen\n buffer_size['non-zero'] = value\n return buffer_size\n","sub_path":"toolbox/policy_optimizer/meta_optimizer.py","file_name":"meta_optimizer.py","file_ext":"py","file_size_in_byte":22945,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"227148223","text":"#!/usr/bin/env python\n\nimport logging\nimport pathlib\nimport platform\nimport shutil\nimport shlex\nimport subprocess\nimport warnings\n\n\ndef get_script_path() -> pathlib.Path:\n return pathlib.Path.cwd() / __file__\n\n\ndef eval_os_cmd(cmd: str) -> (int, str):\n proc = subprocess.Popen(shlex.split(cmd),\n stdout=subprocess.PIPE,\n stderr=subprocess.PIPE)\n stdout, stderr = proc.communicate()\n\n if proc.returncode:\n return proc.returncode, f'Evaluation of {cmd} raised the error {stderr}'\n else:\n return proc.returncode, stdout\n\n\ndef install_brew():\n if platform.system() == 'Darwin':\n if shutil.which('brew') is None:\n eval_os_cmd(\n '/bin/bash -c \"$(curl -fsSL https://raw.githubusercontent.com/Homebrew/install/master/install.sh)\"'\n )\n else:\n raise ValueError(f'{platform.system()} is not supported')\n\n\ndef install_git():\n if platform.system() == 'Darwin':\n if shutil.which('git') is None:\n eval_os_cmd('brew install git')\n elif platform.system() == 'Linux':\n if shutil.which('git') is None:\n eval_os_cmd('sudo apt install -y git')\n else:\n raise ValueError(f'{platform.system()} is not supported')\n\n\ndef install_curl():\n if platform.system() == 'Darwin':\n if shutil.which('curl') is None:\n eval_os_cmd('brew install curl')\n elif platform.system() == 'Linux':\n if shutil.which('curl') is None:\n eval_os_cmd('sudo apt install -y curl')\n else:\n raise ValueError(f'{platform.system()} is not supported')\n\n\ndef copy_tmux_config(dot_folder: pathlib.Path):\n logging.debug('Installing tmux config ...')\n tmux_conf_src = dot_folder / 'tmux.conf'\n tmux_conf_dst = pathlib.Path.home() / '.tmux.conf'\n\n if tmux_conf_dst.exists():\n if tmux_conf_dst.is_symlink() and tmux_conf_dst.resolve() == tmux_conf_src:\n logging.info(f'{tmux_conf_dst} is alread symlink to the {tmux_conf_src}')\n else:\n raise ValueError(f'{tmux_conf_dst} exists.')\n else:\n tmux_conf_dst.symlink_to(tmux_conf_src)\n\n\ndef copy_neovim_config(dot_folder: pathlib.Path):\n logging.debug('Installing neovim config ...')\n\n neovim_conf_src = dot_folder / 'nvim'\n neovim_conf_dst = pathlib.Path.home() / '.config/nvim'\n\n if neovim_conf_dst.exists():\n if neovim_conf_dst.is_symlink() and neovim_conf_dst.resolve(\n ) == neovim_conf_src:\n logging.info(\n f'{neovim_conf_dst} is alread symlink to the {neovim_conf_src}')\n else:\n raise ValueError(f'{neovim_conf_dst} exists.')\n else:\n neovim_conf_dst.symlink_to(neovim_conf_src)\n\n\ndef copy_kitty_config(dot_folder: pathlib.Path):\n logging.debug(\"Installing kitty config ...\")\n if platform.system() == 'Linux':\n\n kitty_conf_src = dot_folder / 'kitty'\n kitty_conf_dst = pathlib.Path.home() / '.config/kitty'\n\n if kitty_conf_dst.exists():\n if kitty_conf_dst.is_symlink() and kitty_conf_dst.resolve(\n ) == kitty_conf_src:\n logging.info(\n f'{kitty_conf_dst} is alread symlink to the {kitty_conf_src}')\n else:\n raise ValueError(f'{kitty_conf_dst} exists.')\n else:\n kitty_conf_dst.symlink_to(kitty_conf_src)\n else:\n logging.info(f'Platform {platform.platform()} is not supported')\n\n\ndef copy_zsh_config(dot_folder: pathlib.Path):\n logging.debug(\"Installing zsh config ...\")\n zsh_conf_src = dot_folder / 'zshrc'\n zsh_conf_dst = pathlib.Path.home() / '.zshrc'\n\n if zsh_conf_dst.exists():\n if zsh_conf_dst.is_symlink() and zsh_conf_dst.resolve() == zsh_conf_src:\n logging.info(f'{zsh_conf_dst} is alread symlink to the {zsh_conf_src}')\n else:\n raise ValueError(f'{zsh_conf_dst} exists.')\n else:\n zsh_conf_dst.symlink_to(zsh_conf_src)\n\n\ndef install_fira_code():\n logging.debug(\"Installing fira code...\")\n if platform.system() == 'Linux':\n rcode, msg = eval_os_cmd(\"apt list --installed fonts-firacode\")\n if not rcode and len(msg.decode('utf8').rstrip().split('\\n')) == 1:\n rcode, msg = eval_os_cmd(\"sudo apt install -y fonts-firacode\")\n if rcode:\n logging.critical(msg)\n elif platform.system() == 'Darwin':\n install_brew()\n rcode, _ = eval_os_cmd(\"brew list --cask font-fira-code\")\n if rcode:\n warnings.warn(f\"list firacode returns {rcode}\")\n eval_os_cmd(\n \"brew tap homebrew/cask-fonts && brew cask install font-fira-code\")\n\n\ndef install_dircolors():\n logging.debug(\"Installing dircolors...\")\n if platform.system() == 'Darwin':\n rcode, _ = eval_os_cmd(\"brew list coreutils\")\n if rcode:\n rcode, msg = eval_os_cmd(\"brew install coreutils\")\n if rcode:\n logging.critical(msg)\n elif platform.system() == 'Linux':\n logging.info(\"Linux has dircolors by default\")\n else:\n raise ValueError(f'{platform.system()} platform is not supported yet')\n\n\ndef install_zsh():\n logging.debug(\"Installing zsh...\")\n if platform.system() == 'Linux':\n rcode, msg = eval_os_cmd(\"apt list --installed zsh\")\n if not rcode and len(msg.decode('utf8').rstrip().split('\\n')) == 1:\n rcode, msg = eval_os_cmd(\"sudo apt install -y zsh\")\n if rcode:\n logging.critical(msg)\n rcode, msg = eval_os_cmd(\"chsh -s $(which zsh) dbihbka\")\n if rcode:\n logging.critical(msg)\n elif platform.system() == 'Darwin':\n logging.info(\"MacOS X uses zsh by default\")\n else:\n raise ValueError(f'{platform.system()} platform is not supported yet')\n\n\ndef install_kitty():\n logging.debug(\"Installing kitty...\")\n if platform.system() == 'Linux':\n rcode, msg = eval_os_cmd(\"apt list --installed kitty\")\n if not rcode and len(msg.decode('utf8').rstrip().split('\\n')) == 1:\n rcode, msg = eval_os_cmd(\"sudo apt install -y kitty\")\n if rcode:\n logging.critical(msg)\n elif platform.system() == 'Darwin':\n logging.info(\"with use iTerm2 in MacOS X usually\")\n else:\n raise ValueError(f'{platform.system()} platform is not supported yet')\n\n\ndef install_vim_plug():\n logging.info(\"Install vim plug...\")\n config_folder = pathlib.Path.home() / \".local/share\"\n logging.critical(f'curl -fLo {config_folder}/nvim/site/autoload/plug.vim --create-dirs https://raw.githubusercontent.com/junegunn/vim-plug/master/plug.vim')\n\n\ndef install_node():\n logging.info(\"Installing node...\")\n if shutil.which('node') is None:\n rcode, msg = eval_os_cmd('curl -sL https://deb.nodesource.com/setup_14.x | sudo -E bash - && sudo apt-get install -y nodejs')\n if rcode:\n logging.critical(msg)\n\ndef install_neovim():\n if platform.system() == 'Darwin':\n if shutil.which('nvim') is None:\n eval_os_cmd('brew install neovim')\n elif platform.system() == 'Linux':\n if shutil.which('nvim') is None:\n eval_os_cmd('sudo apt install -y neovim')\n else:\n raise ValueError(f'{platform.system()} is not supported')\n\ndef install_tmux():\n if platform.system() == 'Darwin':\n if shutil.which('tmux') is None:\n eval_os_cmd('brew install tmux')\n elif platform.system() == 'Linux':\n if shutil.which('tmux') is None:\n eval_os_cmd('sudo apt install -y tmux')\n else:\n raise ValueError(f'{platform.system()} is not supported')\n\n\n\ndef install_zsh_plugins():\n install_git()\n zsh_folder = pathlib.Path.home() / \".zsh\"\n if not zsh_folder.exists():\n zsh_folder.mkdir()\n rcode, msg = eval_os_cmd(\n f\"git clone https://github.com/zsh-users/zsh-autosuggestions {zsh_folder}/zsh-autosuggestions\"\n )\n if rcode:\n warnings.warn(msg)\n\n rcode, msg = eval_os_cmd(\n f\"git clone --recursive https://github.com/joel-porquet/zsh-dircolors-solarized {zsh_folder}/zsh-dircolors-solarized\"\n )\n if rcode:\n logging.critical(msg)\n\n\ndef install_starship():\n logging.debug(\"Installing starship...\")\n if platform.system() == 'Darwin':\n rcode, _ = eval_os_cmd('brew list starship')\n if rcode:\n eval_os_cmd('brew install starship')\n elif platform.system() == 'Linux':\n if shutil.which('starship') is None:\n rcode, msg = eval_os_cmd(\n 'curl -fsSL https://starship.rs/install.sh | bash')\n if rcode:\n raise RuntimeError(msg)\n else:\n raise ValueError(f'{platform.system()} platform isnot supported yet')\n\n\ndef install_exa():\n logging.debug('Installing exa ...')\n if platform.system() == 'Darwin':\n install_brew()\n rcode, _ = eval_os_cmd('brew list exa')\n if rcode:\n rcode, msg = eval_os_cmd('brew install exa')\n if rcode:\n logging.warn(msg)\n elif platform.system() == 'Linux':\n rcode, msg = eval_os_cmd(\"apt list --installed exa\")\n if not rcode and len(msg.decode('utf8').rstrip().split('\\n')) == 1:\n rcode, msg = eval_os_cmd('sudo apt install -y exa')\n if rcode:\n logging.critical(msg)\n\n\ndef install_fd():\n logging.debug('Installing fd ...')\n if platform.system() == 'Darwin':\n install_brew()\n rcode, _ = eval_os_cmd('brew list fd')\n if rcode:\n rcode, msg = eval_os_cmd('brew install fd')\n if rcode:\n logging.warn(msg)\n elif platform.system() == 'Linux':\n rcode, msg = eval_os_cmd(\"apt list --installed fd-find\")\n if not rcode and len(msg.decode('utf8').rstrip().split('\\n')) == 1:\n rcode, msg = eval_os_cmd(\"sudo apt install -y fd-find\")\n if rcode:\n logging.critical(msg)\n\n\ndef install_bat():\n logging.debug('Installing bat ...')\n if platform.system() == 'Darwin':\n install_brew()\n rcode, _ = eval_os_cmd('brew list bat')\n if rcode:\n rcode, msg = eval_os_cmd('brew install bat')\n if rcode:\n logging.warn(msg)\n elif platform.system() == 'Linux':\n rcode, msg = eval_os_cmd(\"apt list --installed bat\")\n if not rcode and len(msg.decode('utf8').rstrip().split('\\n')) == 1:\n rcode, msg = eval_os_cmd('sudo apt install -y bat')\n if rcode:\n logging.critical(msg)\n\n\ndef generate_alaises(dot_folder):\n logging.debug(\"Generating aliases ...\")\n alias_conf_src = dot_folder / 'aliases'\n alias_conf_dst = pathlib.Path.home() / '.aliases'\n\n with open(alias_conf_src, 'w') as src:\n if platform.system() == 'Darwin':\n dircolors_cmd = 'dircolors=gdircolors'\n cat_cmd = 'cat=bat'\n elif platform.system() == 'Linux':\n dircolors_cmd = 'dircolors=dircolors'\n cat_cmd = 'cat=batcat'\n ls_cmd = 'ls=exa'\n find_cmd = 'find=fdfind'\n src.write('\\n'.join(\n [f'alias {cmd}' for cmd in [ls_cmd, dircolors_cmd, find_cmd, cat_cmd]]))\n\n if alias_conf_dst.exists():\n if alias_conf_dst.is_symlink() and alias_conf_dst.resolve(\n ) == alias_conf_src:\n logging.info(\n f'{alias_conf_dst} is alread symlink to the {alias_conf_src}')\n else:\n raise ValueError(f'{alias_conf_dst} exists.')\n else:\n alias_conf_dst.symlink_to(alias_conf_src)\n\n\ndef main():\n dot_folder = get_script_path().parent\n\n install_git()\n install_curl()\n install_fira_code()\n install_zsh()\n install_zsh_plugins()\n install_starship()\n install_exa()\n install_dircolors()\n install_fd()\n install_bat()\n install_kitty()\n install_node()\n install_vim_plug()\n install_neovim()\n install_tmux()\n\n generate_alaises(dot_folder)\n\n copy_zsh_config(dot_folder)\n copy_tmux_config(dot_folder)\n copy_neovim_config(dot_folder)\n\n if platform.system() == 'Linux':\n copy_kitty_config(dot_folder)\n\n\nif __name__ == \"__main__\":\n import sys\n logging.basicConfig(stream=sys.stderr, level=logging.DEBUG)\n main()\n","sub_path":"run.py","file_name":"run.py","file_ext":"py","file_size_in_byte":11270,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"356987545","text":"#!/usr/bin/env micropython\n\n\nfrom ev3dev2.sensor.lego import InfraredSensor\nfrom ev3dev2.sensor import INPUT_4\nfrom ev3dev2.sound import Sound\n\n\nIR_SENSOR = InfraredSensor(address=INPUT_4)\nSPEAKER = Sound()\n\n\ndef detect_object(\n distance: float = 30,\n sound_to_play: str = 'Hi'):\n if IR_SENSOR.proximity <= distance:\n SPEAKER.play_file(\n wav_file='/home/robot/sound/{}.wav'.format(sound_to_play),\n volume=100,\n play_type=Sound.PLAY_WAIT_FOR_COMPLETE)\n\n \nwhile True:\n detect_object(\n distance=30,\n sound_to_play='Magic wand')\n","sub_path":"Computing-Platforms/EV3/Education-Edition/Lessons/Animal-Rescue/1-Explore-Idea-1/4a-Make-Your-Own-with-IR-Sensor.EV3Dev2.py","file_name":"4a-Make-Your-Own-with-IR-Sensor.EV3Dev2.py","file_ext":"py","file_size_in_byte":599,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"540776651","text":"\"\"\"\nTrain a model on TACRED.\n\"\"\"\n\nimport os\nimport sys\nfrom datetime import datetime\nimport time\nimport numpy as np\nimport random\nimport argparse\nfrom shutil import copyfile\nimport torch\nimport pickle\nimport torch.nn as nn\nimport torch.optim as optim\nfrom torch.autograd import Variable\nfrom collections import defaultdict\n\nfrom data.semeval_loader import DataLoader\nfrom model.trainer import GCNTrainer\nfrom utils import torch_utils, scorer, constant_semeval as constant, helper\nfrom utils.vocab import Vocab\n\ncwd = os.getcwd()\non_server = 'Desktop' not in cwd\n\n\ndef str2bool(v):\n return v.lower() in ('true')\n\n# Local paths\nif not on_server:\n data_dir = '/Volumes/External HDD/dataset/semeval/data/json'\n vocab_dir = '/Volumes/External HDD/dataset/semeval/data/vocab'\n model_save_dir = '/Volumes/External HDD/dataset/semeval/saved_models'\n test_save_dir = os.path.join(cwd, '{dataset}_test_performances')\n os.makedirs(test_save_dir, exist_ok=True)\n# Server paths\nelse:\n data_dir = '/usr0/home/gis/data/semeval/data/json'\n vocab_dir = '/usr0/home/gis/data/semeval/data/vocab'\n model_save_dir = '/usr0/home/gis/research/tacred-exploration/saved_models'\n test_save_dir = '/usr0/home/gis/research/tacred-exploration/semeval_test_performances'\n os.makedirs(test_save_dir, exist_ok=True)\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--data_dir', type=str, default=data_dir)\nparser.add_argument('--vocab_dir', type=str, default=vocab_dir)\nparser.add_argument('--model_save_dir', type=str, default=model_save_dir)\nparser.add_argument('--emb_dim', type=int, default=300, help='Word embedding dimension.')\nparser.add_argument('--ner_dim', type=int, default=30, help='NER embedding dimension.')\nparser.add_argument('--pos_dim', type=int, default=30, help='POS embedding dimension.')\nparser.add_argument('--hidden_dim', type=int, default=200, help='RNN hidden state size.')\nparser.add_argument('--num_layers', type=int, default=2, help='Num of RNN layers.')\nparser.add_argument('--input_dropout', type=float, default=0.5, help='Input dropout rate.')\nparser.add_argument('--gcn_dropout', type=float, default=0.5, help='GCN layer dropout rate.')\nparser.add_argument('--word_dropout', type=float, default=0.04, help='The rate at which randomly set a word to UNK.')\nparser.add_argument('--topn', type=int, default=1e10, help='Only finetune top N word embeddings.')\nparser.add_argument('--lower', dest='lower', action='store_true', help='Lowercase all words.')\nparser.add_argument('--no-lower', dest='lower', action='store_false')\nparser.add_argument('--test_save_dir', default=test_save_dir, type=str)\nparser.set_defaults(lower=False)\n\nparser.add_argument('--prune_k', default=-1, type=int,\n help='Prune the dependency tree to <= K distance off the dependency path; set to -1 for no pruning.')\nparser.add_argument('--conv_l2', type=float, default=0, help='L2-weight decay on conv layers only.')\nparser.add_argument('--pooling', choices=['max', 'avg', 'sum'], default='max',\n help='Pooling function type. Default max.')\nparser.add_argument('--pooling_l2', type=float, default=0, help='L2-penalty for all pooling output.')\nparser.add_argument('--mlp_layers', type=int, default=2, help='Number of output mlp layers.')\nparser.add_argument('--no_adj', dest='no_adj', action='store_true', help=\"Zero out adjacency matrix for ablation.\")\n\nparser.add_argument('--no-rnn', dest='rnn', action='store_false', help='Do not use RNN layer.')\nparser.add_argument('--rnn_hidden', type=int, default=200, help='RNN hidden state size.')\nparser.add_argument('--rnn_layers', type=int, default=1, help='Number of RNN layers.')\nparser.add_argument('--rnn_dropout', type=float, default=0.5, help='RNN dropout rate.')\n\nparser.add_argument('--lr', type=float, default=1.0, help='Applies to sgd and adagrad.')\nparser.add_argument('--lr_decay', type=float, default=0.9, help='Learning rate decay rate.')\nparser.add_argument('--decay_epoch', type=int, default=5, help='Decay learning rate after this epoch.')\nparser.add_argument('--optim', choices=['sgd', 'adagrad', 'adam', 'adamax'], default='sgd',\n help='Optimizer: sgd, adagrad, adam or adamax.')\nparser.add_argument('--num_epoch', type=int, default=100, help='Number of total training epochs.')\nparser.add_argument('--batch_size', type=int, default=50, help='Training batch size.')\nparser.add_argument('--max_grad_norm', type=float, default=5.0, help='Gradient clipping.')\nparser.add_argument('--log_step', type=int, default=20, help='Print log every k steps.')\nparser.add_argument('--log', type=str, default='logs.txt', help='Write training log to file.')\nparser.add_argument('--save_epoch', type=int, default=100, help='Save model checkpoints every k epochs.')\nparser.add_argument('--id', type=str, default='00', help='Model ID under which to save models.')\nparser.add_argument('--info', type=str, default='', help='Optional info for the experiment.')\nparser.add_argument('--test_confusion_save_file', default='')\nparser.add_argument('--seed', type=int, default=1234)\nparser.add_argument('--cuda', type=bool, default=torch.cuda.is_available())\nparser.add_argument('--cpu', action='store_true', help='Ignore CUDA.')\n\nparser.add_argument('--load', dest='load', action='store_true', help='Load pretrained model.')\nparser.add_argument('--model_file', type=str, help='Filename of the pretrained model.')\n\nparser.add_argument('--adj_type', type=str, default='regular')\nparser.add_argument('--deprel_emb_dim', type=int, default=200)\nparser.add_argument('--deprel_dropout', type=float, default=.5)\n\nparser.add_argument('--use_bert_embeddings', type=str2bool, default=False)\nparser.add_argument('--emb_dropout', type=float, default=.0)\nparser.add_argument('--dataset', type=str, default='semeval')\nargs = parser.parse_args()\n\ntorch.manual_seed(args.seed)\nnp.random.seed(args.seed)\nrandom.seed(1234)\nif args.cpu:\n args.cuda = False\nelif args.cuda:\n torch.cuda.manual_seed(args.seed)\ninit_time = time.time()\n\n# make opt\nopt = vars(args)\nlabel2id = constant.LABEL_TO_ID\nopt['num_class'] = len(label2id)\n\n# load vocab`\nvocab_file = opt['vocab_dir'] + '/vocab.pkl'\nvocab = Vocab(vocab_file, load=True)\nopt['vocab_size'] = vocab.size\nemb_file = opt['vocab_dir'] + '/embedding.npy'\n\n# Change embedding size for BERT\nif opt['use_bert_embeddings']:\n opt['emb_dim'] = 1024\n emb_matrix = None\nelse:\n emb_matrix = np.load(emb_file)\n assert emb_matrix.shape[0] == vocab.size\n assert emb_matrix.shape[1] == opt['emb_dim']\n\nif opt['use_bert_embeddings']:\n print('Loading BERT Embeddings...')\n embeddings_file = '/usr0/home/gis/data/bert_saves/id2embeddings.pkl'\n id2embeddings = pickle.load(open(embeddings_file, 'rb'))\n print('Embeddings Loaded')\nelse:\n id2embeddings = None\nprint(\"Loading data from {} with batch size {}...\".format(opt['data_dir'], opt['batch_size']))\ntrain_batch = DataLoader(opt['data_dir'] + '/train.json', opt['batch_size'], opt,\n vocab, evaluation=False, bert_embeddings=id2embeddings)\ntest_batch = DataLoader(opt['data_dir'] + '/test.json', opt['batch_size'], opt,\n vocab, evaluation=True, bert_embeddings=id2embeddings)\n\nmodel_id = opt['id'] if len(opt['id']) > 1 else '0' + opt['id']\nmodel_save_dir = opt['model_save_dir'] + '/' + model_id\nopt['model_save_dir'] = model_save_dir\nhelper.ensure_dir(model_save_dir, verbose=True)\n\n# save config\nhelper.save_config(opt, model_save_dir + '/config.json', verbose=True)\nvocab.save(model_save_dir + '/vocab.pkl')\nfile_logger = helper.FileLogger(model_save_dir + '/' + opt['log'],\n header=\"# epoch\\ttrain_loss\\ttest_loss\\ttrain_score\\tbest_train_score\")\n\ntest_save_dir = os.path.join(opt['test_save_dir'], opt['id'])\nos.makedirs(test_save_dir, exist_ok=True)\ntest_save_file = os.path.join(test_save_dir, 'test_records.pkl')\ntest_confusion_save_file = os.path.join(test_save_dir, 'test_confusion_matrix.pkl')\ntrain_confusion_save_file = os.path.join(test_save_dir, 'train_confusion_matrix.pkl')\ndeprel_save_file = os.path.join(test_save_dir, 'deprel_embs.pkl')\n# print model info\nhelper.print_config(opt)\n\n# model\nif not opt['load']:\n trainer = GCNTrainer(opt, emb_matrix=emb_matrix)\nelse:\n # load pretrained model\n model_file = opt['model_file']\n print(\"Loading model from {}\".format(model_file))\n model_opt = torch_utils.load_config(model_file)\n model_opt['optim'] = opt['optim']\n trainer = GCNTrainer(model_opt)\n trainer.load(model_file)\n\nid2label = dict([(v, k) for k, v in label2id.items()])\ntrain_score_history = []\ncurrent_lr = opt['lr']\n\nglobal_step = 0\nglobal_start_time = time.time()\nformat_str = '{}: step {}/{} (epoch {}/{}), loss = {:.6f} ({:.3f} sec/batch), lr: {:.6f}'\nmax_steps = len(train_batch) * opt['num_epoch']\nbest_train_metrics = defaultdict(lambda: -np.inf)\ntest_metrics_at_best_train = defaultdict(lambda: -np.inf)\n\n# start training\nupdate_gap = max(int(50 / opt['batch_size']), 1)\nfor epoch in range(1, opt['num_epoch'] + 1):\n train_loss = 0\n # Training in-case of mini-batches\n trainer.model.train()\n trainer.optimizer.zero_grad()\n\n for i, batch in enumerate(train_batch):\n start_time = time.time()\n global_step += 1\n loss = trainer.update(batch)\n loss.backward()\n loss_val = loss.item()\n step_num = i + 1\n if step_num % update_gap == 0:\n torch.nn.utils.clip_grad_norm_(trainer.model.parameters(), trainer.opt['max_grad_norm'])\n trainer.optimizer.step()\n trainer.optimizer.zero_grad()\n\n train_loss += loss_val\n if global_step % opt['log_step'] == 0:\n duration = time.time() - start_time\n print(format_str.format(datetime.now(), global_step, max_steps, epoch,\n opt['num_epoch'], loss, duration, current_lr))\n # Update grads if needed\n torch.nn.utils.clip_grad_norm_(trainer.model.parameters(), trainer.opt['max_grad_norm'])\n trainer.optimizer.step()\n trainer.optimizer.zero_grad()\n\n print(\"Saving Deprel Embeddings...\")\n with open(deprel_save_file, 'wb') as handle:\n pickle.dump(trainer.get_deprel_emb(), handle)\n\n # eval on train\n print(\"Evaluating on train set...\")\n train_predictions = []\n train_eval_loss = 0\n for i, batch in enumerate(train_batch):\n preds, _, loss = trainer.predict(batch)\n train_predictions += preds\n train_eval_loss += loss\n train_predictions = [id2label[p] for p in train_predictions]\n train_eval_loss = train_eval_loss / train_batch.num_examples * opt['batch_size']\n\n train_p, train_r, train_f1 = scorer.score(train_batch.gold(), train_predictions)\n print(\"epoch {}: train_loss = {:.6f}, train_eval_loss = {:.6f}, train_f1 = {:.4f}\".format(\n epoch, train_loss, train_eval_loss, train_f1))\n train_score = train_f1\n file_logger.log(\"{}\\t{:.6f}\\t{:.6f}\\t{:.4f}\\t{:.4f}\".format(epoch, train_loss, train_eval_loss, train_f1,\n max([train_score] + train_score_history)))\n current_train_metrics = {'f1': train_f1, 'precision': train_p, 'recall': train_r}\n # eval on test\n test_predictions = []\n for i, batch in enumerate(test_batch):\n preds, _, loss = trainer.predict(batch)\n test_predictions += preds\n test_predictions = [id2label[p] for p in test_predictions]\n\n test_p, test_r, test_f1 = scorer.score(test_batch.gold(), test_predictions)\n current_test_metrics = {'f1': test_f1, 'precision': test_p, 'recall': test_r}\n\n if best_train_metrics['f1'] < current_train_metrics['f1']:\n best_train_metrics = current_train_metrics\n test_metrics_at_best_train = current_test_metrics\n trainer.save(os.path.join(model_save_dir, 'best_model.pt'), epoch)\n print(\"New best model saved\")\n file_logger.log(\"new best model saved at epoch {}: {:.2f}\\t{:.2f}\\t{:.2f}\" \\\n .format(epoch, test_p * 100, test_r * 100, test_f1 * 100))\n\n # Compute Confusion Matrices over triples excluded in Training\n test_preds = np.array(test_predictions)\n test_gold = np.array(test_batch.gold())\n train_preds = np.array(train_predictions)\n train_gold = np.array(train_batch.gold())\n test_confusion_matrix = scorer.compute_confusion_matrices(ground_truth=test_gold,\n predictions=test_preds)\n dev_confusion_matrix = scorer.compute_confusion_matrices(ground_truth=train_gold,\n predictions=train_preds)\n print(\"Saving Excluded Triple Confusion Matrices...\")\n with open(test_confusion_save_file, 'wb') as handle:\n pickle.dump(test_confusion_matrix, handle)\n\n print(\"Best Train Metrics | F1: {} | Precision: {} | Recall: {}\".format(\n best_train_metrics['f1'], best_train_metrics['precision'], best_train_metrics['recall']\n ))\n print(\"Test Metrics at Best Train | F1: {} | Precision: {} | Recall: {}\".format(\n test_metrics_at_best_train['f1'], test_metrics_at_best_train['precision'], test_metrics_at_best_train['recall']\n ))\n\n # save\n model_file = model_save_dir + '/checkpoint_epoch_{}.pt'.format(epoch)\n trainer.save(model_file, epoch)\n if epoch == 1 or train_score > max(train_score_history):\n copyfile(model_file, model_save_dir + '/best_model.pt')\n print(\"new best model saved.\")\n file_logger.log(\"new best model saved at epoch {}: {:.2f}\\t{:.2f}\\t{:.2f}\" \\\n .format(epoch, train_p * 100, train_r * 100, train_score * 100))\n if epoch % opt['save_epoch'] != 0:\n os.remove(model_file)\n\n # lr schedule\n if len(train_score_history) > opt['decay_epoch'] and train_score <= train_score_history[-1] and \\\n opt['optim'] in ['sgd', 'adagrad', 'adadelta']:\n current_lr *= opt['lr_decay']\n trainer.update_lr(current_lr)\n\n train_score_history += [train_score]\n print(\"\")\n\nprint(\"Training ended with {} epochs.\".format(epoch))\n\n","sub_path":"train_semeval.py","file_name":"train_semeval.py","file_ext":"py","file_size_in_byte":14163,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"}
+{"seq_id":"23756264","text":"#!/usr/bin/python\n#-*-coding:utf8-*-\n\nimport sys\nimport os\nimport cPickle as pickle\nfrom random import randint\n\ntm_lambda=0.2\nlm_lambda=0.8\n\ndef get_uppers(token_list):\n uppers=[]\n for token in token_list:\n uppers.append([])\n for index,char in enumerate(token):\n if char.isupper():\n uppers[-1].append(index)\n return uppers\n\ndef apply_uppers(uppers,token_list):\n for token_index,indices in enumerate(uppers):\n token=token_list[token_index]\n for index in indices:\n if index